Rel="canonical" questions?
-
On our site we have some similar pages for example in our parts page we have the link to all the electrical parts you can see here http://www.rockymountainatvmc.com/c/43/53/160/Electrical and we have a very similar page going from our accessories page to electrical here http://www.rockymountainatvmc.com/c/43/72/221/Electrical We are thinking about putting rel="canonical" from the accessories electrical page to the parts one. We would do this for several pages not just this one.
Thoughts???
-
We put rel=canonical on it. Now there are several other reasons that we have way to many pages indexed. Big project coming soon, I hope!
-
Oh, ouch - yeah that's definitely has potential to spin out of control. I think rel=canonical would actually be great there, because the product page really is 100% duplicated.
-
just read your article
http://www.seomoz.org/blog/duplicate-content-in-a-post-panda-world
The Indexation “Cap”
Similarly, there’s no set “cap” to how many pages of a site Google will index. There does seem to be adynamic limit, though, and that limit is relative to the authority of the site. If you fill up your index with useless, duplicate pages, you may push out more important, deeper pages. For example, if you load up on 1000s of internal search results, Google may not index all of your product pages. Many people make the mistake of thinking that more indexed pages is better. I’ve seen too many situations where the opposite was true. All else being equal, bloated indexes dilute your ranking ability.
I liked it thanks!
-
Just started here about 1 month ago. We are thinking that it is the search causing this issue. We should have it fixed in the next few weeks. For example if you go to http://www.rockymountainatvmc.com/search?term=dirt+bikes and then click on the first result. http://www.rockymountainatvmc.com/p/43/53/210/1157/-/25928/Dirt-Tricks-KTM-Timing-Chain-Tensioner/dirt+bikes it has the search result as part of the url for tracking. I am thinking this could be causing hundreds of thousands of urls. We will turn this into a parametersso they can still track it but google will not index it. Any thoughts on this, or maybe it is to far off topic. Thanks for all your input.
-
I was about to say that 85 sounds significant, but then I noticed that Google has indexed 1.13M pages on your site. For your link profile/authority (which is respectable but still pretty medium-sized) that's a massive index. I strongly suspect you have a ton of duplicate content that's completely unrelated to this. You've probably got bigger fish to fry, honestly.
One thing you might want to look at is your search pagination, as that can easily spin out thousands of pages. Unfortunately, really sorting this out takes a pretty thorough audit.
-
of the first pair (ones like these ttp://www.rockymountainatvmc.com/c/43/53/160/Electrical http://www.rockymountainatvmc.com/c/43/72/221/Electrical) around around 15 in each Classification and there are 5 classification so 75.
of the second pair (like this
http://www.rockymountainatvmc.com/m/43/53/2/ATV-Parts-KAWASAKI
http://www.rockymountainatvmc.com/m/43/72/2/ATV-Accessories-KAWASAKI
) about 17 per classification so 85.
Third level like (http://www.rockymountainatvmc.com/v/43/72/10864/ATV-KAWASAKI/BRUTE-FORCE-300-2013 http://www.rockymountainatvmc.com/v/43/53/10864/ATV-KAWASAKI/BRUTE-FORCE-300-2013) there are about 1100-1300
-
How many pairs of these duplicates are there across the whole site?
-
I agree one URL would be best; however, they said how the system is set up not really possible right now.
There are only two page that are the same and they are the same here on the next leave down too.
http://www.rockymountainatvmc.com/m/43/53/2/ATV-Parts-KAWASAKI
http://www.rockymountainatvmc.com/m/43/72/2/ATV-Accessories-KAWASAKI
So you think if there are just two page that are the same it is nothing to worry about? I know adding one or two paragraph of good content not just seo copy to the http://www.rockymountainatvmc.com/c/43/53/160/Electrical http://www.rockymountainatvmc.com/c/43/72/221/Electrical Having one focus on parts and the other one accessories. I know that is not my ideal situation but a possibility. I would rather just have one URL like you suggested.
Thoughts?
-
Yeah, I think Matthew's right about the duplication - it looks like these are basically landing on the same results. It's not a disaster, but it's certainly going to look like thin content. I guess I'd want to step back and ask if these two paths are really necessary - could they converge on one URL if you're really serving up exactly the same products? The best canonicalization solution is ultimately not to have multiple URLs.
The canonical tag is slightly odd here, just because the pages aren't 100% duplicates, but it's probably fine. Honestly, a lot of it boils down to how many search pages you've got across the site and how many are duplicated. If you're talking about half-a-dozen, it's probably not worth worrying about. If you're talking about hundreds of pages, then cleaning up your architecture and removing these duplicates could have real dividends for SEO (and possibly make life easier for your visitors, too).
-
Your welcome.
Have a great day.
-
Thanks for your input.
-
There are a few things here that I think you should be aware of.
1 if you leave it like that, you will have duplicate content issues, because both of those pages are the same, the pictures are just in different places.
2. The page http://www.rockymountainatvmc.com/c/43/72/221/Electrical is a pr3 and the other is a pr1. I would personally keep the pr3 page and 301 redirect the other page to that page so you no longer have those duplicate content issues on your site. That way if you have anyone still visit that page through old links they will still end up on that pr3 page.
3. The url structure, is a bit messy for your site. I generally advise others to have something a long these lines. http://www.rockymountainatvmc.com/electrical this is an SEO friendly url. I am not saying the ones you have now won't work, it's just a bit messy that's all. Personally though if your getting good traffic to it, I wouldn't mess with it at all. Now if you were just starting out I would say change them asap.
In the future when making new pages, I would set it up to make a proper url structure.
Have a great day.
Matthew Boley
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Internal Clicks and CTR. Is REL=canonical better than Noindex in this case?
I currently have a search facility in a website that noindexes the search results which is ok. But when you click one of the results it takes you to a product which is noindexes as it has URL params. e.g. https://www.visitliverpool.com/accommodation/albion-guest-house-p305431?bookurl=%2Fbook-online%3Fstage%3Dunitsel%26isostartdate%3D2017-10-31%26nights%3D1%26roomReq_1_adults%3D1%26NumRoomReqs%3D1%26fuzzy%3D0%26product%3D305431 The product also exists as this which is indexed : - https://www.visitliverpool.com/accommodation/albion-guest-house-p305431 Should I canonicalise is this instance instead of no index? Does CTR apply to internal links? i.e. Does search console consider internal clicks? Are internal clicks a ranking factor?
Intermediate & Advanced SEO | | Andrew-SEO0 -
Best Way To Go About Fixing "HTML Improvements"
So I have a site and I was creating dynamic pages for a while, what happened was some of them accidentally had lots of similar meta tags and titles. I then changed up my site but left those duplicate tags for a while, not knowing what had happened. Recently I began my SEO campaign once again and noticed that these errors were there. So i did the following. Removed the pages. Removed directories that had these dynamic pages with the remove tool in google webmasters. Blocked google from scanning those pages with the robots.txt. I have verified that the robots.txt works, the pages are longer in google search...however it still shows up in in the html improvements section after a week. (It has updated a few times). So I decided to remove the robots.txt file and now add 301 redirects. Does anyone have any experience with this and am I going about this the right away? Any additional info is greatly appreciated thanks.
Intermediate & Advanced SEO | | tarafaraz0 -
What is better? No canonical or two canonicals to different pages?
I have a blogger site that is adding parameters and causing duplicate content. For example: www.mysite.com/?spref=bl
Intermediate & Advanced SEO | | TMI.com
www.mysite.com/?commentPage=1 www.mysite.com/?m=1 www.mysite.com/?m=0 I decided to implement a canonical tag on these pages pointing to the correct version of the page. However, for the parameter ?m=0, the canonical keeps pointing to itself. Ex: www.mysite.com/?m=0 The canonical = www.mysite.com/?m=0 So now I have two canonicals for the same page. My question is if I should leave it, and let Google decide, or completely remove the canonicals from all pages?0 -
Using unique content from "rel=canonical"ized page
Hey everyone, I have a question about the following scenario: Page 1: Text A, Text B, Text C Page 2 (rel=canonical to Page 1): Text A, Text B, Text C, Text D Much of the content on page 2 is "rel=canonical"ized to page 1 to signalize duplicate content. However, Page 2 also contains some unique text not found in Page 1. How safe is it to use the unique content from Page 2 on a new page (Page 3) if the intention is to rank Page 3? Does that make any sense? 🙂
Intermediate & Advanced SEO | | ipancake0 -
Fetch as GoogleBot "Unreachable Page"
Hi, We are suddenly having an error "Unreachable Page" when any page of our site is accessed as Googlebot from webmaster tools. There are no DNS errors shown in "Crawl Errors". We have two web servers named web1 and web2 which are controlled by a software load balancer HAProxy. The same network configuration has been working for over a year now and never had any GoogleBot errors before 21st of this month. We tried to check if there could be any error in sitemap, .htaccess or robots.txt by excluding the loadbalancer and pointing DNS to web1 and web2 directly and googlebot was able to access the pages properly and there was no error. But when loadbalancer was made active again by pointing the DNS to it, the "unreachable page" started appearing again. This very same configuration has been working properly for over a year till 21st of this month. Website is properly accessible from browser and there are no DNS errors either as shown by "Crawl Errors". Can you guide me about how to diagnose the issue. I've tried all sorts of combinations, even removed the firewall but no success. Is there any way to get more details about error instead of just "Unreachable Page" error ? Regards, shaz
Intermediate & Advanced SEO | | shaz_lhr0 -
How Rel=Prev & Rel=Next work for me?
I have implemented Rel=Prev & Rel=Next tag on my website. I would like to give example URL to know more about it. http://www.vistapatioumbrellas.com/market-umbrellas?limit=40&p=3 http://www.vistapatioumbrellas.com/market-umbrellas?limit=40&p=4 http://www.vistapatioumbrellas.com/market-umbrellas?limit=40&p=5 Right now, I have blocked paginated pages by Robots.txt by following query. Disallow: /*?p= I have removed disallow syntax from Robots.txt for paginated pages. But, I have confusion with duplicate page title. If you will check all 3 pages so you will find out duplicate page title across all pages. I know that, duplicate page title is harmful for SEO. Will Google crawl + index all paginated pages? If yes so which page will get maximum benefits in organic ranking? Is there any specific way which may help me to solve this issue?
Intermediate & Advanced SEO | | CommercePundit0 -
"Original Content" Dynamic Hurting SEO? -- Strategies for Differentiating Template Websites for a Nationwide Local Business Segment?
The Problem I have a stable of clients spread around the U.S. in the maid service/cleaning industry -- each client is a franchisee, however their business is truly 'local' with a local service area, local phone/address, unique business name, and virtually complete control over their web presence (URL, site design, content; apart from a few branding guidelines). Over time I've developed a website template with a high lead conversion rate, and I've rolled this website out to 3 or 4 dozen clients. Each client has exclusivity in their region/metro area. Lately my white hat back linking strategies have not been yielding the results they were one year ago, including legitimate directories, customer blogging (as compelling as maid service/cleaning blogs can really be!), and some article writing. This is expected, or at least reflected in articles on SEO trends and directory/article strategies. I am writing this question because I see sites with seemingly much weaker back link profiles outranking my clients (using SEOMoz toolbar and Site Explorer stats, and factoring in general quality vs. quantity dynamics). Questions Assuming general on-page optimization and linking factors are equal: Might my clients be suffering because they're using my oft-repeated template website (albeit with some unique 'content' variables)? If I choose to differentiate each client's website, how much differentiation makes sense? Specifically: Even if primary content (copy, essentially) is differentiated, will Google still interpret the matching code structure as 'the same website'? Are images as important as copy in differentiating content? From an 'machine' or algorithm perspective evaluating unique content, I wonder if strategies will be effective such as saving the images in a different format, or altering them slightly in Photoshop, or using unique CSS selectors or slightly different table structures for each site (differentiating the code)? Considerations My understanding of Google's "duplicate content " dynamics is that they mainly apply to de-duping search results at a query specific level, and choosing which result to show from a pool of duplicate results. My clients' search terms most often contain client-specific city and state names. Despite the "original content" mantra, I believe my clients being local businesses who have opted to use a template website (an economical choice), still represent legitimate and relevant matches for their target user searches -- it is in this spirit I ask these questions, not to 'game' Google with malicious intent. In an ideal world my clients would all have their own unique website developed, but these are Main St business owners balancing solutions with economics and I'm trying to provide them with scalable solutions. Thank You! I am new to this community, thank you for any thoughts, discussion and comments!
Intermediate & Advanced SEO | | localizedseo0 -
Use rel=canonical to save otherwise squandered link juice?
Oftentimes my site has content which I'm not really interested in having included in search engine results. Examples might be a "view cart" or "checkout" page, or old products in the catalog that are no longer available in our system. In the past, I'd blocked those pages from being indexed by using robots.txt or nofollowed links. However, it seems like there is potential link juice that's being lost by removing these from search engine indexes. What if, instead of keeping these pages out of the index completely, I use to reference the home page (http://www.mydomain.com) of the business? That way, even if the pages I don't care about accumulate a few links around the Internet, I'll be capturing the link juice behind the scenes without impacting the customer experience as they browse our site. Is there any downside of doing this, or am I missing any potential reasons why this wouldn't work as expected?
Intermediate & Advanced SEO | | cadenzajon1