Published on 15th March 2022 by Sharon McClintic

Webinar On-Demand: Advanced SEO Crawling Techniques – How to Use Custom Extractions

 

Our latest Deepcrawl webinar dives deep into how you can use custom extractions to get even better insights from your crawl data. 

Custom Extractions are an advanced crawling technique that offer new ways to understand your own websites — and those of your competitors. 

This session’s guest, Reuben Yau, has been working in SEO since 2000 and now serves as Vice President of SEO at Vizion Interactive. In conversation with Deepcrawl’s Chris Spann, the two talk through what custom extractions are, how to use them to build custom data feeds, how to use them for competitor analysis, and suggest a few challenges to watch out for along the way. 

It’s a jam-packed presentation. Read on for our key takeaways for the session, but, as always, be sure to check out the embedded video above for the full talk (as well as the fascinating Q&A!).

 
 

What are custom extractions?

If you are familiar with Deepcrawl’s Analytics Hub crawler already, you will no doubt recognize the many default extractions included in the platform. These are some of the most crucial extractions for SEO, such as title tags, meta descriptions, canonical tags, etc.

But Deepcawl’s platform also allows for custom extractions. This gives SEOs the option to pull out virtually any other type of page content or data available – including information that isn’t overtly search-centric but can help with competitor analysis and make our overall SEO and website health reports more robust.  

As Yau points out, custom extractions are really only limited to the kind of content contained on the site in question. For example, on an eCommerce website, perhaps you want to pull out SKUs or prices. For content-led sites or blogs, perhaps you want to extract the dates on which posts were published or author names.

Custom extractions allow you to do all that and more.

What are custom extractions? Some examples of what you can find.

 
 

Custom extractions for competitor analysis

Yau talks us through how custom extractions can be used to compare the pricing on our own sites with those of our competitors. Here’s how to do it within the Deepcrawl platform:

First, we want to create a new project, then go to ‘Advanced Settings’ in step 4.

How to use custom extractions in Deepcrawl - step 1

In Advanced Settings, click on ‘New Extraction Rule’.  [You can create up to 30.]

Step 2 - set up new extraction rule

Viewing the source HTML in our example, we see the price is contained in a <span> tag:

Step 3 - examine the HTML in your target site

In the ‘New Extraction Rule’ tab we can add a label to make it easier to use later on (as opposed to ‘custom extraction 1’ etc.). We also need to add the regex pattern — i.e. the code used to help identify the price. [Tip: use Rubular to test your regex – it uses Ruby, like Deepcrawl does, so it will provide the best consistency.] 

Step 4- Add label in New Extraction Rule tab

This tab also gives us the option to pull out just the first match (in our case, there is only a single price on each page crawled). And we can also choose to strip out any additional code/text from the crawl if we don’t want it to be included in our results.

Once you’ve added all the custom extractions you want, simply run the crawl. When it’s finished, go to the ‘Extraction’ menu, then ‘Overview’, and ‘Extract Report’.

Step 5 - Extract Report

Click on the extractions, select the fields/columns you want to show and export as a CSV.

Step 6 - Export as CSV

From there, we can import the CSV into an Excel file and do the same for the competitor crawls across separate tabs. 

Yau notes how we can then use the VLOOKUP function on SKUs to pull pricing for all competitors into a single view. We can use different calculations and conditional formatting to highlight price differences.

Step 7 - use VLOOKUP in Excel to view custom crawl data

 

 
 

Building custom data feeds

Yau also shows us how to build a data feed to submit to a 3rd party (e.g. Google Merchant Center) for a new website. 

Traditionally, in order to do this, we might have to contact our developers and request a database report. But with Deepcrawl, we can do it ourselves. Here’s how:

  1. Identify the required fields you need to capture, based on the feed’s requirements.
  2. Build custom extraction rules into a new crawl project.
  3. Extract the necessary fields.
  4. Export to a CSV.
  5. Transform and add extra fields as necessary.
  6. Upload to the data feed provider.

 
 

Boost SEO reporting with custom extractions

Custom extractions are also great for boosting our SEO reports. As an example, Yau shows us how to analyze content performance to get meaningful insights for a new content-focused site/blog by augmenting Google Analytics data with our own customized Deepcrawl data.

Google Analytics is a powerful tool in its own right, but how do we best decide what kind of content to continue producing? Alternatively, how might we justify ceasing the production of certain types of content?

Yau notes that in order to work this out, we need to ask more in-depth questions (and get more detailed data!). For example…

Sometimes we can get some of this information on Google Analytics, within the URLs (e.g. dates, categories, etc.), but other times a method like custom extractions will be needed to get the information we’re really looking for. 

Example of URL data available in Google Analytics

There are also other details we can extract that might help shape our future content campaigns, such as tags, or how many comments certain types of posts are getting.

As Yau details, the process is similar to how we used custom extractions in our competitor analysis example. It goes something like this:

  1. Set up project.
  2. Inspect the HTML to see how its structured.
  3. Build custom extraction rules for the information you’re looking for.
  4. Run the crawl.
  5. Export to CSV.

 

In this example, we want to pull out traffic numbers by blog post and by month in a custom Google Analytics report. We can then export that report to CSV and combine the GA and Deepcrawl custom extraction in one Excel file. Here we can create a pivot table, use VLOOKUP to combine the two data sets, add sparklines, and filter/sort as needed.

Example of content analysis with custom extractions and pivot tables

Then we can identify the posts that are performing the best, which ones may be declining, and observe commonalities by author, category, etc.

 
 

Challenges and pitfalls in custom extractions

Yau highlights the following crawl issues as potential issues that SEOs need to be aware of when performing custom extractions:

He also notes the following potential HTML issues:

But if SEOs are aware of these potential issues, custom extractions are a powerful way to get the most from your crawls and ensure your sites are performing as well as they can be—for both users and search engines.

— 

Interested in working with Reuben Yau? Consider booking a coffee break (bonus: free coffee!) with his team.

Ready to get started with your own custom extractions in Deepcrawl Analytics Hub? Schedule a demo anytime.  

 

Choose a better way to grow

With tools that will help you realize your website’s true potential, and support to help you get there, growing your enterprise business online has never been so simple.

Book a Demo