Is My Boilerplate Product Description Causing Duplicate Content Issues?
-
I have an e-commerce store with 20,000+ one-of-a-kind products. We only have one of each product, and once a product is sold we will never restock it. So I really have no intention to have these product pages showing up in SERPs. Each product has a boilerplate description that the product's unique attributes (style, color, size) are plugged into. But a few sentences of the description are exactly the same across all products.
Google Webmaster Tools doesn't report any duplicate content. My Moz Crawl Report show 29 of these products as having duplicate content. But a Google search using the site operator and some text from the boilerplate description turns up 16,400 product pages from my site.
Could this duplicate content be hurting my SERPs for other pages on the site that I am trying to rank? As I said, I'm not concerned about ranking for these products pages. Should I make them "rel=canonical" to their respective product categories? Or use "noindex, follow" on every product? Or should I not worry about it?
-
My SERPs for a competitive term I felt I underperforming for dropped about 10 spots overnight after I added "noindex,follow" to the product pages. From the 3rd page to the 4th page, so it's not like I had a lot to lose. My SERPs for less competitive long tail keywords, which is where I'm getting most of my traffic, have dropped slightly or stayed the same.
Should I cross my fingers and hope for a recovery? Revert the product pages back to "index, follow"? Any thoughts?
-
Hi Tom,
Thanks so much for the thorough response.
Based on several comparative metrics with sites that are outranking me significantly, I do feel the site is underperforming. Because our traffic is ridiculously seasonal the Panguin Tool doesn't provide any clues.
I just added to all my products using Yoast's Wordpress SEO plugin. We'll see what happens.
Thanks,
Zach -
Hi Zachary
I really can't be sure if it's having an adverse affect, but I wouldn't be surprised if it was.
Having looked at just 3 of the product pages, there is a problem with content being repeated, but I think it is being compounded by there being no other content on the page either to make it look unique.
Both are the hallmarks to a potential Panda penalty, which could affect the pages performance themselves and/or the whole domain. So, if you're seeing subpar performance (and even if you're performing well it's worth reading on) I would look at the following solution.
For every product that you do not intend to restock or reuse, I would either add a tag, add a 301 redirect or simply remove the page and serve a 404. If we're talking tens of thousands, then having that many redirects might bloat out your .htaccess file (making it larger and longer to load/process) and having an instant drop of 20k URLs and 404 errors might look a bit odd to Google as well. However, adding 20k tags is a bit of a nightmare as well.
You might want to try a combination of all 3 - a few 404 errors is nothing to worry about - but the logic is that you will be removing a number of pages that have this duplicate content on it, thus improving the quality of the domain. For your remaining 'live' pages, I'd highly recommend taking the time to add 200+ words of unique content about the product in order to avoid this happening again.
An alternative solution would be to block the bots from accessing the /shop/ subfolder in your robots.txt file - and then setting up the shop and the currently active product listings on a different subdomain. You'd lose the ability to use the /shop/ folder, but it would be quicker than manually adding tags or 301 redirects.
That's the method I would use if I wanted to address the issue. However, this **may not be necessary **if your site is not performing badly. You can check this to a degree with the Panguin Tool - this overlays your organic traffic in analytics with Google updates - if a drop in traffic coincides with a Panda penalty, you may be under the affect of one - in which case you should take action ASAP.
Hope this helps.
-
Hi Zachery
29 pages showing up in your Moz crawl report out of 16,400 indexed pages on your site is such a small percentage (0.18% to be accurate) it is not worth worrying about. Also, if GWT is not reporting any issues I think you should be fine.
Don't worry, be happy!
Peter
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is using part of a meta description already on your site for another product considered duplicate?
I'm writing meta descriptions for this site, trying to keep them different, however, for two product types, I want to add the same info I added in the other likeminded product's meta descriptions. Is this ok as long as it's not the whole sentence or am I really to rewrite the same info another way, which is hard for " quick shipping available for x amount of colors ". Any Advice?
Technical SEO | | Deacyde0 -
Another Duplicate Content - eCommerce Question!
We are manufacturers of about 15 products and our website provides information about the products. We also offer them for sale on the site. Recently we partnered with a large eCommerce site that sells many of these types of products. They lifted descriptions from our site for theirs and are now selling our products. They have higher DA than us. Will this cause a ranking problem for us? Should we write unique descriptions for them? Thanks!
Technical SEO | | Chris6610 -
Duplicate Page Content but where?
Hi All Moz is telling me I have duplicate page content and sure enough the PA MR mT are all 0 but it doesnt give me a link to this content! This is the page: http://www.orsgroup.com/index.php?page=Scanning-services But I cant find where the duplicate content is other than on our own youtube page which I will get removed here: http://www.youtube.com/watch?v=Pnjh9jkAWuA Can anyone help please? Andy
Technical SEO | | ORS-Group0 -
Duplicate content in product listing
We have "duplicate content" warning in our moz report which mostly revolve around our product listing (eCommerce site) where various filters return 0 results (and hence show the same content on the page). Do you think those need to be addressed, and if so how would you prevent product listing filters that appearing as duplicate content pages? should we use rel=canonical or actually change the content on the page?
Technical SEO | | erangalp0 -
Localized domains and duplicate content
Hey guys, In my company we are launching a new website and there's an issue it's been bothering me for a while. I'm sure you guys can help me out. I already have a website, let's say ABC.com I'm preparing a localized version of that website for the uk so we'll launch ABC.co.uk Basically the websites are going to be exactly the same with the difference of the homepage. They have a slightly different proposition. Using GeoIP I will redirect the UK traffic to ABC.co.uk and the rest of the traffic will still visit .com website. May google penalize this? The site itself it will be almost the same but the homepage. This may count as duplicate content even if I'm geo-targeting different regions so they will never overlap. Thanks in advance for you advice
Technical SEO | | fabrizzio0 -
Duplicate Content in Wordpress.com
Hi Mozers! I have a client with a blog on wordpress.com. http://newsfromtshirts.wordpress.com/ It just had a ranking drop because of a new Panda Update, and I know it's a Dupe Content problem. There are 3900 duplicate pages, basically because there is no use of noindex or canonical tag, so archives, categories pages are totally indexed by Google. If I could install my usual SEO plugin, that would be a piece of cake, but since Wordpress.com is a closed environment I can't. How can I put a noindex into all category, archive and author peges in wordpress.com? I think this could be done by writing a nice robot.txt, but I am not sure about the syntax I shoud use to achieve that. Thank you very much, DoMiSol Rossini
Technical SEO | | DoMiSoL0 -
Filter Tag Duplicate Content E-Commerce Issue
Hello, I just launched a new site for a client but am seeing some duplicate content issues in the campaign crawl. It has to do with the drill-down, filter "tags" that helps users find the product they are looking for. You can see them in the sidebar here: http://www.ssmd.com/shop/ In my crawl report this is what is showing up as duplicate content (attached image). How do I keep these widgets from generating duplicate content on the site? Also, not sure if it's important or not, but I am using Wordpress, WooCommerce and Yoast's SEO Tool. Any suggestions are appreciated! Screen%20Shot%202012-10-23%20at%202.56.00%20PM.png
Technical SEO | | kylehungate0 -
An odd duplicate content issue...
Hi all, my developers have just assured me that nothing has changed form last week but in the today's crawl I see all the website duplicated: and the difference on the url is the '/' so basically the duplicated urls are: htts://blabla.bla/crop htts://blabla.bla/crop/ Any help in understanding why is much appreciated. thanks
Technical SEO | | LeadGenerator0