As part of our ongoing efforts to simplify the Google Search results page, we will be phasing out support for a few structured data features in Search. We regularly evaluate the usefulness of Search features, both for users and website owners.
Category Archives: Google Webmaster Central Blog
Adding markup support for loyalty programs
Member benefits, such as lower prices and earning loyalty points are a major factor considered by shoppers when buying products online. Today we're adding support for defining loyalty programs under Organization structured data combined with loyalty benefits under Product structured data.
Source: Google Search Central Blog
Top ways to ensure your content performs well in Google’s AI experiences on Search
As a site owner, publisher or creator, you may be wondering how to best succeed in our AI search experiences, such as AI Overviews and our new AI Mode. The underpinnings of what Google has long advised carries across to these new experiences. Focus on your visitors and provide them with unique, satisfying content. Then you should be well positioned as Google Search evolves, as our core goal remains the same: to help people find outstanding, original content that adds unique value. With that in mind, here are some things to consider for success in Google Search all around, including our AI experiences.
Source: Google Search Central Blog
App deep links: connecting your website and app
Since 2013, Search has recognized the importance of app deep links in a mobile-centric world. In this post, we'll review the current state of app deep links — take a look at what they are, the benefits of using them, and how to implement them effectively.
Source: Google Search Central Blog
Register now for Search Central Live Deep Dive 2025
We're ready to open registrations for the first ever Search Central Live Deep Dive, a 3-day event that will be held in Bangkok, Thailand this year on July 23-25!
Source: Google Search Central Blog
The Search Analytics API now supports hourly data
A few months ago, we announced an improved way to view recent performance data in Search Console. The "24 hours" view includes data from the last available 24 hours and appears with a delay of only a few hours. This view can help you find information about which pages and queries are performing in this recent timeframe and how content you recently published is picking up.
Source: Google Search Central Blog
Robots Refresher: Future-proof Robots Exclusion Protocol
In the previous posts about the Robots Exclusion Protocol (REP) we explored what's already possible to do with its various components — namely robots.txt and the URI level controls. In this post we will explore how the REP can play a supporting role in the ever-evolving relation between automatic clients and the human web.
Source: Google Search Central Blog
Get Ready for Search Central Live Asia Pacific 2025
Hello 2025! (Yeah, we know, time flies!) We've had some exciting plans in the works for Search Central Live (SCL) Asia Pacific this year, and we're super excited to let you in on what we've been up to. We've been listening closely to your feedback, and we're cooking up something different than what we usually do–something bigger, deeper, and more tailored to you!
Source: Google Search Central Blog
Robots Refresher: page-level granularity
With the robots.txt file, site owners
have a simple way to control which parts of a website are accessible by crawlers.
To help site owners further express how search engines and web
crawlers can use their pages, the web standards group came
up with robots meta
tags in 1996, just a few months after meta
tags
were proposed for HTML (and anecdotally, also before Google
was founded). Later, X-Robots-Tag
HTTP response headers were added.
These instructions are sent together with a URL, so crawlers can only take them into account
if they're not disallowed from crawling the URL through the robots.txt file. Together, they
form the Robots Exclusion Protocol (REP).
Source: Google Search Central Blog
Robots Refresher: robots.txt — a flexible way to control how machines explore your website
A long-standing tool for website owners, robots.txt has been in active use for over 30 years and is broadly supported by crawler operators (such as tools for site owners, services, and search engines). In this edition of the robots refresher series, we'll take a closer look at robots.txt as a flexible way to tell robots what you want them to do (or not do) on your website.