Deepcrawl is now Lumar. Read more.
DeepcrawlはLumarになりました。 詳細はこちら

Sitemaps

A sitemap is a list of all of the live URLs which exist on a site and is used to inform search engine crawlers of the most important pages and therefore which ones should be crawled and indexed.

There are several things to consider when creating sitemaps, as well as understanding how search engines view them. We cover a range of these topics within our SEO Office Hours Notes below, along with best practice recommendations and Google’s advice on sitemaps.

For more on sitemaps and SEO, check out our article:  How to Improve Website Crawlability with Sitemaps.

Image Sitemaps Help Google Understand Which Images You Want to Be Indexed

Google can find images to index in the source code, but Sitemaps can help them to confirm which images you want to be indexed.

21 Dec 2018

Google News Sitemap is Fastest Way to Get Pages Crawled for Publishers

Submitting a Google News sitemap is the fastest way to get Google to crawl pages for publishing sites.

11 Dec 2018

Nested Sitemap Index Files Aren’t Supported by Google

Google doesn’t support nested sitemap index files, where one sitemap references another one. Instead set up separate sitemap index files and submit them separately.

11 Dec 2018

You can use non-English Language Locations Names in Image Sitemaps Geolocation Tags

You can use non-English languge names for locations in Image Sitemaps geolocation tags. You can test them by searching for the location in Google Maps and see if Google can figure out where the location is. John thinks that Google’s algorithms may not use the information.

30 Nov 2018

It’s Normal for Google to Index XML Sitemap Files

If you see an XML sitemap file showing in the search results when you search for a specific URL on your website, this is normal and won’t cause any issues. If you don’t want XML sitemaps to be indexed, then add an x-robots tag in the HTTP header.

27 Nov 2018

Only Use Sitemap Files Temporarily for Serving Removed URLs to be Deindexed

Sitemap files are a good temporary solution for getting Google to crawl and deindex lists of removed URLs quickly. However, make sure these sitemaps aren’t being served to Google for too long.

16 Nov 2018

Use X-Robots-Tag HTTP Header to Noindex Indexed Sitemap Files

If sitemap files are indexed for normal search queries, then you can use the X-Robots-Tag HTTP header to noindex all pages ending in .xml or .gz.

19 Oct 2018

Compressing Sitemaps Saves Bandwidth But Doesn’t Reduce Processing Time

Compressing sitemap files using Gzip can save bandwidth but doesn’t impact the speed that Googlebot processes these files.

19 Oct 2018

Check URLs Match Exactly When GSC Reports a URL is Not in a Sitemap

If the URL Inspection tool says a page is indexed but not submitted in the XML sitemap, John recommends checking that the exact URL seen in Search Console is present in the sitemap. For example, you should check that there are no differences with trailing slashes or the case used.

19 Oct 2018

Update Last Modified Date in Sitemap & Use Validate Fix in GSC to Get Pages Crawled Sooner

If technical issues cause pages to show incorrectly (e.g. serving a blank page), you can get Googlebot to crawl these sooner by submitting sitemap files with the last modification date set to when the affected pages were restored. You can also click ‘validate fix’ on pages with errors in Search Console to get Googlebot to recrawl these pages faster.

5 Oct 2018

Back 3/10 Next