You can restrict the overall size and depth of a crawl before you start or during your crawl.
This is useful to prevent a lot of URL credits being used unintentionally, or to run a discovery crawl, when you first start to crawl a website and don't yet know the optimal settings.
Crawl Depth & Max URL Limits
On step 3 of the crawl setup, you will see a set of options to limit the crawl.
The first option is crawl depth. The allows you to only crawls pages reached within the number of clicks or levels from the starting page.
The starting page of a website is level 1. All the pages linked from the start page are level 2. All the pages linked from level 2 are level 3 etc.
The drop-down will be set to the default setting showing ‘1000 Levels’. If you click on this you will see that it drops down to give a number of preset options to get you started. These settings are fantastic if you’re setting up a crawl quickly and just want to select an approximate number of URLs.
Another option, which gives more control and versatility, is to choose the bottom option from the drop-down list, which is Custom. The Custom setting allow you to choose a specific number of levels, or clicks from the home page.
The second option is the URLs Limit. This allows you to specify the maximum amount of URLs to be crawled.
This also offers some preset options along with a custom option where you can specify the exact maximum amount of URLs you want to crawl.
The crawl will stop after reaching the number of URLs you specify or the depth limit and either finalise or pause depending on your settings.
The max URLs per crawl setting in your account can be adjusted, to prevent regular crawls from using excessive credits. Contact email@example.com to change it.
Pause & Finalize
Select ‘Finish anyway’ to generate your reports automatically when the crawl limits have been reached.
Select ‘If this limit was not enough, notify me’ if you want the option to adjust your crawl settings, resume the crawl or to finalize it when the crawl limits are reached.
It's also worth noting that you can change these settings whilst a crawl is running.
It’s not possible to restrict a List or Sitemap crawl based on max pages or max level restrictions. All pages in the list or Sitemaps will be crawled although you can manually pause and finalise at any point.
It’s possible to limit the number of pages crawled for any page type using the Page Grouping function in Advanced Settings.
This is useful if you have a large number of pages with a certain URL type that you know you want to leave out of the crawl e.g. user profile pages.
Simply add a regular expression to match the URLs you want to limit and add a % value.
- The ".*" regular expression is being appended by default to the end of Page URL Match so /search will match /search/anything, /search/anything2 etc.
- The actual number of URLs crawled may vary slightly from the limits
- Forward slashes do not need to be escaped with a backslash
- You can match the full URL including http://