Sitemaps

A sitemap is a list of all of the live URLs which exist on a site and is used to inform search engine crawlers of the most important pages and therefore which ones should be crawled and indexed. There are several things to consider when creating sitemaps, as well as understanding how search engines view them. We cover a range of these topics within our Hangout Notes, along with best practice recommendations and advice from Google.

Submit Old URLs in Sitemaps when Moving Domains

November 29, 2016 Source

When moving domains, you can submit old redirecting URLs in a Sitemap with a changed date, on either domain, to help get them re-crawled more quickly.


Sitemaps on Separate Domains Require Both Sites to be Verified in Search Console

September 23, 2016 Source

You can submit Sitemaps with URLs on a different domain if both of the site are verified in the Search Console account.


Sitemaps Hosted in Subdirectories are Only Valid for URLs in the Same Directory

September 20, 2016 Source

Sitemaps located in a subdirectory only be valid for URLs in that sub-directory.


Search Console Indexed URLs Counts Report Exact URLs in Sitemaps

September 20, 2016 Source

Search Console indexed pages in Sitemaps uses exact URLs, so variations including www/non-www, trailing slash variations, etc, won’t be reported as indexed.


Google Measures Sitemap Trust

September 9, 2016 Source

Google has a trust rating per Sitemap, based on incorrect use of last modified data. Google can learn that the last modified date is providing useful information which increases the trust. Otherwise it will start to ignore the last modified dates.


Crawl Errors Priority Metric includes Mixture of Signals

September 9, 2016 Source

The priority metric for crawl errors in search console is a mixture of pages being returned in search results, included in Sitemaps, and if it has internal links. The higher the priority are the ones Google thinks might have content which Google wants to index.


Add Last Modified to Redirects in Sitemaps

August 9, 2016 Source

When Redirecting URLs include them in a Sitemap with a last modified date set after the redirect was put in place, it will encourage them to be crawled more quickly


Last Modified In Sitemaps Aids Crawling

July 8, 2016 Source

Google thinks the Last Modified date in an XML Sitemap can be very useful to help them recrawl URLs, and they also support RSS and Atom feeds.


Google Ignores Irrelevant Sitemap Content

July 8, 2016 Source

Google will ignore any information in Sitemaps which it doesn’t recognise, so you can include additional information for other purposes.


Related Topics

Crawling Indexing Crawl Budget Crawl Errors Crawl Rate Disallow Last Modified Nofollow Noindex RSS Canonicalization Fetch and Render