A sitemap is a list of all of the live URLs which exist on a site and is used to inform search engine crawlers of the most important pages and therefore which ones should be crawled and indexed. There are several things to consider when creating sitemaps, as well as understanding how search engines view them. We cover a range of these topics within our Hangout Notes, along with best practice recommendations and advice from Google.

A Sitemap File Won’t Replace Normal Crawling

February 5, 2019 Source

A sitemap will help Google crawl a website but it won’t replace normal crawling, such as URL discovery from internal linking. Sitemaps are more useful for letting Google know about changes to the pages within them.

Google Doesn’t Mind How Sitemaps Are Split up

January 11, 2019 Source

Google combines separate sitemaps together so that they can be processed. This means it is up to webmasters to decide how they want to split up sitemaps.

Google Treats XML Sitemaps Differently From HTML Pages

January 11, 2019 Source

Google treats XML sitemaps differently from HTML pages, as they are a machine-readable file and not meant to be indexed by search engines.

Image Sitemaps Help Google Understand Which Images You Want to Be Indexed

December 21, 2018 Source

Google can find images to index in the source code, but Sitemaps can help them to confirm which images you want to be indexed.

Google News Sitemap is Fastest Way to Get Pages Crawled for Publishers

December 11, 2018 Source

Submitting a Google News sitemap is the fastest way to get Google to crawl pages for publishing sites.

Nested Sitemap Index Files Aren’t Supported by Google

December 11, 2018 Source

Google doesn’t support nested sitemap index files, where one sitemap references another one. Instead set up separate sitemap index files and submit them separately.

You can use non-English Language Locations Names in Image Sitemaps Geolocation Tags

November 30, 2018 Source

You can use non-English languge names for locations in Image Sitemaps geolocation tags. You can test them by searching for the location in Google Maps and see if Google can figure out where the location is. John thinks that Google’s algorithms may not use the information.

It’s Normal for Google to Index XML Sitemap Files

November 27, 2018 Source

If you see an XML sitemap file showing in the search results when you search for a specific URL on your website, this is normal and won’t cause any issues. If you don’t want XML sitemaps to be indexed, then add an x-robots tag in the HTTP header.

Only Use Sitemap Files Temporarily for Serving Removed URLs to be Deindexed

November 16, 2018 Source

Sitemap files are a good temporary solution for getting Google to crawl and deindex lists of removed URLs quickly. However, make sure these sitemaps aren’t being served to Google for too long.

Related Topics

Crawling Indexing Crawl Budget Crawl Errors Crawl Rate Disallow Last Modified Nofollow Noindex RSS Canonicalization Fetch and Render