Magento and Duplicate content
-
I have been working with Magento over the last few weeks and I am becoming increasingly frustrated with the way it is setup. If you go to a product page and remove the sub folders one by one you can reach the same product pages causing duplicate content. All magento sites seem to have this weakness. So use this site as an example because I know it is built on magento,
http://www.gio-goi.com/men/clothing/tees/throve-t-short.html?cid=756
As you remove the tees then the clothing and men sub folders you can still reach the product page. My first querstion is how big an issue is this and two does anyone have any ideas of how to solve it?
Also I was wondering how does google treat question marks in urls? Should you try and avoid them unless you are filtering?
Thanks
-
Gregster,
I assume that you have found an answer to your question by now. However, I wanted to offer up what looks to be an extremely in depth and comprehensive walkthrough on Magento SEO from yoast.com. They have several sections on duplicate content, as well as a canonical plugin you may find useful.
http://yoast.com/articles/magento-seo/
Best of Luck!
-
"I recommend you nofollow the login, search, and cart pages through XML layout. That will cross off another 500 pages or so." Not nofollow. Don't use nofollow . This is for untrusted links - so should not be used for internal links.
It's Noindex. And then use the canonical tag if 301 Redirects are not an option. To make life more complicated, you need to be careful not to do use noindex and canonical tag simultaneously.
-
Hi Kevin,
I would be interested to talk more with you about this issue. What does your custom extension do that others don't?
Thanks again.
-
Hi Gregster. I feel your pain. Having worked on Magento for the past three years, I've come across a lot of "issues" you'd expect a top-tier e-commerce solution provider to have under control.
I've written about getting canonical URLs in CMS pages here, something that many Magento SEO extensions don't do. I also had a custom SEO extension created and would be happy to share with you. No cost. Just use it.
I don't know if you have multiple languages, but that alone will create an exponential amount of duplicate content from dynamic parameters. Go into your WMT and set those parameters to be ignored. If you aren't sure how to do that, it's well documented here and on Google, Yahoo, and Bing webmaster sites.
I recommend you nofollow the login, search, and cart pages through XML layout. That will cross off another 500 pages or so.
One last mention is that RocketTheme has created a pretty neat extension that will get rid of the p parameter altogether by using JS to switch from grid and list views. Or you could just select in your admin to only allow either grid or list instead of both.
Any more questions just ask.
-
Hi,
Magento is surely a "beast"... the way to solve your problem, as far as I understood it, is to use the rel="canonical", in order to show to the Search Engines what URL they have to consider in case of duplicated content.
The solutions?
- or you have very good devs skills (or a developer very fond of Magento);
- or you have to rely to the many extensions existing.
Very well know is the Yoast extension, but it seems it can give serious problem on the lastest version of Magento.
Another SEO extension is SEO Suite Pro Magento Extension (which exists also in a Ultimate version), Very good extension, but not for free.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What is the best way to handle these duplicate page content errors?
MOZ reports these as duplicate page content errors and I'm not sure the best way to handle it. Home
Technical SEO | | ElykInnovation
http://myhjhome.com/
http://myhjhome.com/index.php Blog
http://myhjhome.com/blog/
http://myhjhome.com/blog/?author=1 Should I just create 301 redirects for these? 301 http://myhjhome.com/index.php to http://myhjhome.com/ ? 301 http://myhjhome.com/blog/?author=1 to http://myhjhome.com/ ? Or is there a better way to handle this type of duplicate page content errors? and0 -
Duplicated content in moz report due to Magento urls in a multiple language store.
Hi guys, Moz crawl is reporting as duplicated content the following urls in our store: http://footdistrict.com and http://footdistrict.com?___store=footdistrict_es The chain: ___store=footdistrict_es is added as you switch the language of the site. Both pages have the http://footdistrict.com" /> , but this was introduced some time after going live. I was wondering the best action to take considering the SEO side effects. For example: Permanent redirect from http://footdistrict.com?___store=footdistrict_es to http://footdistrict.com. -> Problem: If I'm surfing through english version and I switch to spanish, apache will realize that http://footdistrict.com?___store=footdistrict_es is going to be loaded and automatically it will redirect you to http:/footdistrict.com. So you will stay in spanish version for ever. Deleting the URLS with the store code from Google Web Admin tools. Problem: What about the juice? Adding those URL's to robots.txt. Problem: What about the juice? more options? Basically I'm trying to understand the best option to avoid these pages being indexed. Could you help here? Thanks a lot.
Technical SEO | | footd0 -
SEOMOZ and non-duplicate duplicate content
Hi all, Looking through the lovely SEOMOZ report, by far its biggest complaint is that of perceived duplicate content. Its hard to avoid given the nature of eCommerce sites that oestensibly list products in a consistent framework. Most advice about duplicate content is about canonicalisation, but thats not really relevant when you have two different products being perceived as the same. Thing is, I might have ignored it but google ignores about 40% of our site map for I suspect the same reason. Basically I dont want us to appear "Spammy". Actually we do go to a lot of time to photograph and put a little flavour text for each product (in progress). I guess my question is, that given over 700 products, why 300ish of them would be considered duplicates and the remaning not? Here is a URL and one of its "duplicates" according to the SEOMOZ report: http://www.1010direct.com/DGV-DD1165-970-53/details.aspx
Technical SEO | | fretts
http://www.1010direct.com/TDV-019-GOLD-50/details.aspx Thanks for any help people0 -
Duplicate video content question
This is really two questions in one. 1. If we put a video on YouTube and on our site via Wistia, how would that affect our rankings/authority/credibility? Would we get punished for duplicate video content? 2. If we put a Wistia hosted video on our website twice, on two different pages, we would get hit for having duplicate content? Any other suggestions regarding hosting on Wistia and YouTube versus just Wistia for product videos would be much appreciated. Thank you!
Technical SEO | | ShawnHerrick1 -
Does Google count the same article in different languages as duplicate content?
If the "content" is the same, but is used in separate articles written in entirely different languages, will Google see the articles as duplicate content? How do international companies handle information that needs to be distributed in different languages?
Technical SEO | | BlueLinkERP0 -
Is Noindex Enough To Solve My Duplicate Content Issue?
Hello SEO Gurus! I have a client who runs 7 web properties. 6 of them are satellite websites, and 7th is his company's main website. For a long while, my company has, among other things, blogged on a hosted blog at www.hismainwebsite.com/blog, and when we were optimizing for one of the other satellite websites, we would simply link to it in the article. Now, however, the client has gone ahead and set up separate blogs on every one of the satellite websites as well, and he has a nifty plug-in set up on the main website's blog that pipes in articles that we write to their corresponding satellite blog as well. My concern is duplicate content. In a sense, this is like autoblogging -- the only thing that doesn't make it heinous is that the client is autoblogging himself. He thinks that it will be a great feature for giving users to his satellite websites some great fresh content to read -- which I agree, as I think the combination of publishing and e-commerce is a thing of the future -- but I really want to avoid the duplicate content issue and a possible SEO/SERP hit. I am thinking that a noindexing of each of the satellite websites' blog pages might suffice. But I'd like to hear from all of you if you think that even this may not be a foolproof solution. Thanks in advance! Kind Regards, Mike
Technical SEO | | RCNOnlineMarketing0 -
Problem with duplicate content
Hi, My problem is this: SEOmoz tells me I have duplicate content because it is picking up my index page in three different ways: http://www.web-writer-articles.co.uk http://www.web-writer-articles.co.uk/ and http://www.web-writer-articles.co.uk/index.php Can someone give me some advice as to how I can deal with this issue? thank you for your time, louandel15
Technical SEO | | louandel150 -
Forget Duplicate Content, What to do With Very Similar Content?
All, I operate a Wordpress blog site that focuses on one specific area of the law. Our contributors are attorneys from across the country who write about our niche topic. I've done away with syndicated posts, but we still have numerous articles addressing many of the same issues/topics. In some cases 15 posts might address the same issue. The content isn't duplicate but it is very similar, outlining the same rules of law etc. I've had an SEO I trust tell me I should 301 some of the similar posts to one authoritative post on the subject. Is this a good idea? Would I be better served implementing canonical tags pointing to the "best of breed" on each subject? Or would I be better off being grateful that I receive original content on my niche topic and not doing anything? Would really appreciate some feedback. John
Technical SEO | | JSOC0