Is there a way to no index no follow sections on a page to avoid duplicative text issues?
-
I'm working on an event-related site where every blog post starts with an introductory header about the event and then a Call To Action at the end which gives info about the Registration Deadline. I'm wondering if there is something we can and should do to avoid duplicative content penalties. Should these go in a widget or is there some way to No Index, No Follow a section of text? Thanks!
-
Hi Karl,
Google doesn't mind a bit of content that is the same page on page - they have said so in the page. If the page requires that there is a bit of the content that is duplicated, as long as it serves a purpose, you will be absolutely fine.
Look at the pages and see if what is there would be considered beneficial to the page.
The issues for duplication tend to arise when you are copying huge swathes of text from one page to another, plagiarising others work or even spinning.
By what you have said, it sounds like you will be absolutely fine, but feel free to post an example here or PM it to me if you wish and I will happily take a look
-Andy
-
This might help. https://www.hochmanconsultants.com/how-to-block-part-of-a-page-from-being-indexed-by-google-or-other-search-engines/ let me know if it doesn't work.
-
Hi,
Short answer: No you can’t hide a selection on a page from Google. If Google reaches your page it will crawl all of it. I personally wouldn’t worry that much about duplicate content here if it is only a small part of the page.One solution could be to make an image for the returning content if this is possible. This way Google doesn’t crawl the text, but make sure you make the content visual. It is not userfriendly to put a lot of text in images.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do we avoid duplicate/thin content on +150,000 product pages?
Hey guys! We got a rather large product range (books) on our eCommerce site (+150,000 titles). We get book descriptions as meta data from our publishers, which we display on the product pages. This obviously is not unique, as many other sites display the same piece of description of the book. It is important for us to rank on those book titles, so my question to You is: How would you go about it? I mean, it seems like a rather unrealistic task to paraphrase +150,000 (and growing) book descriptions. As I see it, there are these options: 1. Don't display the descriptions on the product pages (however then those pages will get even thinner!)
Intermediate & Advanced SEO | | Jacob_Holm
2. Display the (duplicate) descriptions, but put no-index on those product pages in order not to punish the rest of the site (not really an option, though).
3. Hire student workers to produce unique product descriptions for all 150,000 products (seems like a huge and expensive task) But how would You solve such a challenge?
Thanks a lot! Cheers, Tommy.0 -
Blank Cart Pages Showing as Duplicate, HELP
Hi Everyone, I'm seeing a bunch of URLs that look something like this [ domain.com/cart?add&id_product=42&token=776d4a08721f3d8c920e287248797547] showing as duplicate content in my Moz crawls. I think these are just blank pages for the most part. Is there anything to be concerned with here? Is there a way to clean this up? Thanks! Ricky
Intermediate & Advanced SEO | | RickyShockley0 -
Duplicated Meta Descriptions on Dynamic Paginated Pages
If the title didn't put you off please read on! 🙂 According to our latest Moz Crawl Report we have circa 700 instances of duplicate Meta Description on pages that are both dynamically created and also paginated, however, I believe that number to be greater! We are unable to manual make changes to these pages (because they are dynamic) and so we need ask our web devs to create a change in how the Meta is created... If I am not making myself clear (and there is a good chance that I'm not!) then here is an example of what I mean; http://www.bolsovercruiseclub.com/cruise-deals/silversea-cruise-deals/ There are 92 pages of cruise deals for this particular operator with the results of each page having the option to sort by 4 categories; Recommended Cruise Price Sail Date Best Value 4 x 92 = 368 instances just for this one operator! The current Meta Desc is; A selection of the best Silversea cruise deals taking in over 800 destinations across all 7 continents. ...which isn't great I know! The problem is how to make each page (in each category) unique If any of you have incurred anything similar and have any kind of solution or recommendation then please respond - I would be most grateful! Andy
Intermediate & Advanced SEO | | TomKing0 -
Why are these pages considered duplicate content?
I have a duplicate content warning in our PRO account (well several really) but I can't figure out WHY these pages are considered duplicate content. They have different H1 headers, different sidebar links, and while a couple are relatively scant as far as content (so I might believe those could be seen as duplicate), the others seem to have a substantial amount of content that is different. It is a little perplexing. Can anyone help me figure this out? Here are some of the pages that are showing as duplicate: http://www.downpour.com/catalogsearch/advanced/byNarrator/narrator/Seth+Green/?bioid=5554 http://www.downpour.com/catalogsearch/advanced/byAuthor/author/Solomon+Northup/?bioid=11758 http://www.downpour.com/catalogsearch/advanced/byNarrator/?mediatype=audio+books&bioid=3665 http://www.downpour.com/catalogsearch/advanced/byAuthor/author/Marcus+Rediker/?bioid=10145 http://www.downpour.com/catalogsearch/advanced/byNarrator/narrator/Robin+Miles/?bioid=2075
Intermediate & Advanced SEO | | DownPour0 -
Thousands of 404 Pages Indexed - Recommendations?
Background: I have a newly acquired client who has had a lot of issues over the past few months. What happened is he had a major issue with broken dynamic URL's where they would start infinite loops due to redirects and relative links. His previous SEO didn't pay attention to the sitemaps created by a backend generator, and it caused hundreds of thousands of pages to be indexed. Useless pages. These useless pages were all bringing up a 404 page that didn't have a 404 server response (it had a 200 response) which created a ton of duplicate content and bad links (relative linking). Now here I am, cleaning up this mess. I've fixed the 404 page so it creates a 404 server response. Google webmaster tools is now returning thousands of "not found" errors, great start. I fixed all site errors that cause infinite redirects. Cleaned up the sitemap and submitted it. When I search site:www.(domainname).com I am still getting an insane amount of pages that no longer exist. My question: How does Google handle all of these 404's? My client wants all the bad pages removed now but I don't have as much control over that. It's a slow process getting Google to remove these pages that are returning a 404. He is continuously dropping in rankings still. Is there a way of speeding up the process? It's not reasonable to enter tens of thousands of pages into the URL Removal Tool. I want to clean house and have Google just index the pages in the sitemap.
Intermediate & Advanced SEO | | BeTheBoss0 -
Should we deindex duplicate pages?
I work on an education website. We offer programs that are offered up to 6 times per year. At the moment, we have a webpage for each instance of the program, but that's causing duplicate content issues. We're reworking the pages so the majority of the content will be on one page, but we'll still have to keep the application details as separate pages. 90% of the time, application details are going to be nearly identical, so I'm worried that these pages will still be seen as duplicate content. My question is, should we deindex these pages? We don't particularly want people landing on our application page without seeing the other details of the program anyway. But, is there problem with deindexing such a large chunk of your site that I'm not thinking of? Thanks, everyone!
Intermediate & Advanced SEO | | UWPCE0 -
Mobile site version - Is it a duplication issue?
There is a blog www.blogname.com and someone creates 2 mobile versions: iphone.blogname.com mobile.blogname.com they are the perfect copy of www.blogname.com (articles, tags, links, etc etc) How Google will manage them? Right now, my article gets backlink by three sites www.blogname.com iphone.blogname.com mobile.blogname.com
Intermediate & Advanced SEO | | Greenman0 -
Dynamically generated page issues
Hello All! Our site uses dynamically generated pages. I was about to begin the process of optimising our product category pages www.pitchcare.com/shop I was going to use internal anchor text from some high ranking pages within our site but each of the product category pages already have 1745 links! Am I correct in saying that internal anchor text links works to a certain point? (maybe 10 or so links) So any new internal anchor text links will count for nothing? Thanks Todd
Intermediate & Advanced SEO | | toddyC0