How to save crawl budget while keeping internal linking for SEO?

benperry
Visitor
3 0 1

Hello,

 

A well-known problem on Shopify, the

?pr_prod_strat=collection_fallback&pr_rec_id

parameters create a bunch of URLs that are uninteresting for Google.

 

In addition to consuming crawl budget for nothing (thanks Shopify) this can cause serious problems:
- Google not choosing the same canonical URL as declared
- Google indexing the same two pages with different URL parameters, creating duplicate content...

 

This article got a lot of attention: https://ed.codes/blog/weird-shopify-seo-issue-millions-of-indexed-pages
Which recommends setting all URLs containing parameters to disallow for crawling.


However, since these URLs (with parameters) are the ones in the recommendatios etc sections, they are the ones responsible for internal linking.

recommended-product-links.e4839b68_Z1G5iJp.pngSo if I disallow them, Google will no longer take internal linking into account for all the pages on my site?
Or will it return the internal linking juice to the declared canonical URLs?

 

A priori, if it's in disallow, it doesn't crawl the page, so it doesn't even see the declared canonical URL...
At that point, what's the solution to avoid consuming unnecessary crawl budget (millions of parameter combinations) and also avoid indexing the same two pages?

 

I've seen this solution passed around the Shopify community: https://community.shopify.com/c/shopify-discussions/removing-url-parameters-recommended-products-sec...
It would solve the problem vis-a-vis crawlers... but it would prevent Shopify tracking apparently?

 

Very simply, what I'm looking for is:
1) disallow URLs with parameters (to save crawl budget)
2) while keeping internal linking for crawlers

Replies 2 (2)

benperry
Visitor
3 0 1

N/A

SEOProStaff
Shopify Partner
89 0 4

Hi @benperry ,

 

The issue of URL parameters like ?pr_prod_strat=collection_fallback&pr_rec_id can indeed create numerous URLs that could dilute your SEO efforts by wasting crawl budget and risking duplicate content issues. One strategy to address this is to use the Disallow directive in your robots.txt file for the specific parameters. However, this needs to be done with caution to avoid disrupting valuable internal linking.

 

While disallowing these URLs, search engines like Google should ideally consolidate the 'link juice' to the canonical URLs, provided that the canonical tags are implemented correctly on your product pages. It's essential to ensure that your canonical URLs are consistent and point to the preferred version of each page.

 

SEOPro could help in this context by automating and simplifying the process of SEO optimization. It runs SEO audits to identify issues and provides recommendations for improvement, including insights on meta titles and descriptions. SEOPro's tools can also assist in bulk updates, potentially easing the task of ensuring that your canonical URLs are correctly set across your site, without losing track of the internal links critical for SEO.

 

Full disclosure: We are the developers of this app.

Phil
SeoPro Developer
For any question, feel free to reach out to us: support@seo-pro.app