|
Hopefully, Google will still crawl all the unique, distinct content eventually, but it may take longer. As new content is published, it will take longer to perform. Several other scenarios can cause this same phenomenon. For example, different filtering combinations on a site with faceted navigation may result in exponentially expanding volumes of parameter URLs spawning on a website. A non-filtered category page may end up with ten or even 100 parameter variations as different filtering is applied. We can just put canonical tags on the highly-duplicate pages that we don’t want the content duplication issues, right? While that is true.
Google still has to crawl and visit the non-canonical DB to Data to see that they are non-canonical (to read their embedded canonical tags). Canonical tags only help to alleviate content duplication, but they don’t really help much with crawl efficiency and content discovery. You could argue that this is where you deploy complex wildcard robots.txt rules. Still, you must be careful in that area, as you can easily unintentionally cut off chunks of organic search traffic. The best practice is to implement correct URL architecture and supporting redirects. If you have those in place, not much can go wrong. Often, canonical tags are deployed as a band-aid solution after issues have already arisen.
But they’re quite a messy patch to a more fundamental problem. Links and content: Quality vs. quantity On the surface, this seems like a no-brainer. Google has constantly stated that quality content and links matter more than mass-manufactured spam. SEOs and digital PR specialists can often spend weeks attempting to create great content and ascertain a single high-value placement to knock the competition off their ranking pedestal. No SEO worth their salt would argue that mass-spun content and spam links are an effective tool. These tactics are ineffective if you expect to maintain a long-term online brand, a foundation of business that you can build upon over time.
|
|