Deepcrawl is now Lumar. Read more.
DeepcrawlはLumarになりました。 詳細はこちら

URL Parameters

URL parameters are additional query strings that follow the main URL for a page, typically contained after a question mark. Multiple query string parameters can be used within a single URL, which can impact a website’s SEO and organic search performance in several ways, including causing duplicate content and crawling issues.  
 
This collection of SEO Office Hours notes addresses common problems seen with URL parameter handling, with recommendations on how to fix these issues with best practices suggested by Google Search Central.

Google Learns Which URL Parameters Return Irrelevant Pages

Google learns which parameters are returning irrelevant pages partly based on canonicalised URLs.

27 Jan 2017

Excessive URL Parameters and Rewrites Can Causing Problems

Google can have problems crawling your site if your URL structure has an excessive number of URL parameters and rewrites which redirect to a few pages.

7 Oct 2016

Increased Crawl Rates Can Be Caused by Authority, Server Performance or Duplicate Content

If you experience a temporarily high crawl rate, it might be caused because of an increase in authority, or Google thinks the server can handle an increased load, or they might be finding a lot of duplicate content caused by things like URL parameters.

8 Jul 2016

Robots.txt Overrides Parameter Settings

URL Parameter settings in Search Console are a hint for Google, and they will validate them periodically. The Robots.txt disallow overrides the parameter removal, so it’s better to use the parameter tool to consolidate duplicate pages instead of disallow.

17 May 2016

Use GL Parameter to See Google SERPs for a Different Country

If you want to see SERPs from a different geo location then change the ���gl�۪ parameter in the URL to the country code.

29 Jan 2016

Googlebot Can Be Redirected to Canonical URLs

John says that althought it’s technically cloaking, it’s actually OK to redirect Googlebot from URLs with tracking parameters to canonical URLs, but allow users not to be redirected, so they can be tracked in analytics.

1 Dec 2015

URLs with Hashbangs Won’t be Indexed

URLs with hashbangs won’t be indexed, so if they are required to produce unique pages, you’ll need to migrate to a traditional URL structure without the hashbang.

16 Oct 2015

Use Disallow to Improve Crawling Efficiency

John recommends against robots.txt, because it prevents Google consolidating authority signals, but then says there are occassions when crawling efficiency is more important.

10 Oct 2014

Back 4/4 Next