Multiple Instances of the Same Article
-
Hi, I'm having a problem I cannot solve about duplicate article postings.
As you will see from the attached images, I have a page with multiple variants of the same URL in google index and as well as duplicate title tag in the search console of webmasters tools. Its been several months I have been using canonical meta tags to resolve the issue, aka declare all variants to point to a single URL, however the problem remains. Its not just old articles that stay like that, even new articles show the same behaviour right when they are published even thought they are presented correctly with canonical links and sitemap as you will see from the example bellow.
Example URLs of the attached Image
-
All URLs belonging to the same article ID, have the same canonical link inside the html head.
-
Also because I have a separate mobile site, I also include in every desktop URL an "alternate" link to the mobile site.
-
At the Mobile Version of the Site, I have another canonical link, pointing back to the original Desktop URL. So the mobile site article version also has
-
Now, when it comes to the xml sitemap, I pass only the canonical URL and none of the other possible variants (to avoid multiple indexing), and I also point to the mobile version of the article.
<url><loc>http://www.neakriti.gr/?page=newsdetail&DocID=1300357</loc>
<xhtml:link rel="alternate" media="only screen and (max-width: 640px)" href="http://mobile.neakriti.gr/fullarticle.php?docid=1300357"><lastmod>2016-02-20T21:44:05Z</lastmod>
<priority>0.6</priority>
<changefreq>monthly</changefreq>
image:imageimage:lochttp://www.neakriti.gr/NewsASSET/neakriti-news-image.aspx?Doc=1300297</image:loc>
image:titleΟΦΗ</image:title></image:image></xhtml:link></url>
The above Sitemap snippet Source: http://www.neakriti.gr/WebServices/sitemap.aspx?&year=2016&month=2
The main sitemap of the website: http://www.neakriti.gr/WebServices/sitemap-index.aspxDespite my efforts you see that webmasters tools reports three variants for the desktop URL, and google search reports 4 URLs (3 different desktop variant urls and the mobile url).
I get this when I type the article code to see if what is indexed in google search: site:neakriti.gr 1300297
So far I believe I have done all I could in order to resolve the issue by addressing canonical links and alternate links, as well as correct sitemap.xml entry. I don't know what else to do... This was done several months ago and there is absolutelly no improvement.
Here is a more recent example of an article added 5 days ago (10-April-2016), just type
site:neakriti.gr 1300357
at google search and you will see the variants of the same article in google cache. Open the google cached page, and you will see the cached pages contain canonical link, but google doesn't obey the direction given there.Please help!
-
-
Hi all,
sorry for the delay, I am away on a business trip, this is why I stopped communicating the past few days.
I can confirm that the latest entries (those after March) come as a single instance.
However there are some minor exceptions like the one hereExample of a recent article indexed in both desktop (even though desktop url is not the canonical) and mobile URL
https://www.google.gr/search?q=site:neakriti.gr&biw=1527&bih=899&source=lnms&sa=X&ved=0ahUKEwiIxODGt5_MAhUsKpoKHdcUAkYQ_AUIBigA&dpr=1.1#q=site:neakriti.gr+1315539&tbs=qdr:w&filter=0Also I noticed that with the "alternate" and "canonical" links the mobile version of the site doesn't get indexed anymore (with minor exceptions like the one above).
-
Hi Ioannis!
How's this going? We'd love an update.
-
Hmm, interestingly, when I followed your link, I only saw the canonical version of the article. Is this what you're seeing now?
Also, in response to your earlier question, yes, you can disallow parameters with robots.txt. If these canonical issues continue, that may be the best next step.
-
Thank you for your response, I will take a look at this.
However I have two questions regarding your suggestion
- Since I have canonical links at the loading page, doesn't that resolve the issue?
- the printerfriendly variation has a noindex meta at the head, shouldn't that be taken into account?
- Can I put regular expressions in my robots.txt? How can I block url params? Because printerfriendly and newsdetailsports are values of the "page" GET param
Infact the printerfriendly contains canonical link and noindex meta to inform search engines not to index content, and let them know where the original content exists
-
Hi there
The printer friendly URL is coming from the print this article button (attached) and the /default.aspx URL is coming from the ^ TOP button (attached).
What you could do is use your robots.txt to ignore these URLs. You can all tell Google what URL parameters to ignore, but please be EXTREMELY careful doing this. It's not a fine comb tool, not a hatchet.
Let me know if you have any questions or comments, good luck!
Patrick
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best-practice URL structures with multiple filter combinations
Hello, We're putting together a large piece of content that will have some interactive filtering elements. There are two types of filters, topics and object types. The architecture under the hood constrains us so that everything needs to be in URL parameters. If someone selects a single filter, this can look pretty clean: www.domain.com/project?topic=firstTopic
Intermediate & Advanced SEO | | digitalcrc
or
www.domain.com/project?object=typeOne The problems arise when people select multiple topics, potentially across two different filter types: www.domain.com/project?topic=firstTopic-secondTopic-thirdTopic&object=typeOne-typeTwo I've raised concerns around the structure in general, but it seems to be too late at this point so now I'm scratching my head thinking of how best to get these indexed. I have two main concerns: A ton of near-duplicate content and hundreds of URLs being created and indexed with various filter combinations added Over-reacting to the first point above and over-canonicalizing/no-indexing combination pages to the detriment of the content as a whole Would the best approach be to index each single topic filter individually, and canonicalize any combinations to the 'view all' page? I don't have much experience with e-commerce SEO (which this problem seems to have the most in common with) so any advice is greatly appreciated. Thanks!0 -
Multiple Keywords for a site
I have a client that is OBSESSED with KWP ranking (don't go there...I know) This client offers multiple services, dog boarding, dog grooming, dog training, dog daycare and dog walking. Essentially these are our focus. She ranks on page one for all of these words (locally of course) BUT she wants to rank in positions 1 and 2 for all of these words. Here's my rub, with her limited budget, we focus on 1 word (and associated long tails like "dog boarding in the south loop) and it takes a couple of months to zoom up to positions 1 or 2 (not counting map pack....she wants ORGANIC) While we're focusing on this 1 word, the others maintain their ranking or slip a few spots (like from 6 to 😎 Conversions average about about 1 a day, organic traffic is roughly 1000 hits a month. In your opinion is it better to split this focus between the 5 target words every month, more slowly building ranking, but maintaining it for longer periods of time. Or do it the way we have been chase dog boarding, then chase training, and so on. It just seems like we are CONSTANTLY chasing something while something else falls. Thanks Tracy
Intermediate & Advanced SEO | | lkilera0 -
Multiple, Partial Redirecting URLs from old SEO company
Received quite a surprise when I gained access to the Google webmaster account and saw 4 domains that are link to my clients domain and the number of links for each domain range between 10,000 and 90,000. Come to find out this was a result of their former agency. The business is very local central. I will use the example of a burger place. They main site is burgers.com and burger places are listed by city and state. Their former agency bought several domains like californiaburgers.com and duplicated the listings for that state on this domain. You can view certain pages of the second domain, but the home page is redirected as are most of the city pages with 301s to the main burgers.com domain. However, there are pages on the additional domains that do not redirect, as they are not duplicated on the main domain so nowhere to redirect. Google has only found four of them but looks like there could be at least 50. Pages that are not redirected are indexed by the engines - but not ranking (at least not well). There is a duplicate content issue, although "limited" in the sense that it really is just the name of the business, address and phone number - there is not much to these listings. What is the best approach to overcome? Right now GWT is showing over 300,000 links, however at least 150,000 to 200,000 of that is from these domains.
Intermediate & Advanced SEO | | LeverSEO0 -
Multiple Authors Google + Authorship
Hello, I took a look through past questions but can't seem to find a definitive answer on setting up Google + Authorship credit (for multiple authors) using a Wordpress blog. Has anyone had experience setting this up? Or could you recommend solid reading/research? I took a look at a couple of Wordpress plug in's but just found them very confusing (so did our IT contact who will ultimately be setting up code for this.) Any direction or advice is appreciated.
Intermediate & Advanced SEO | | SEOSponge0 -
Multiple domain level redirects to unique sub-folder on one domain...
Hi, I have a restaurant menu directory listing website (for example www.menus.com). Restaurant can have there menu listed on this site along with other details such as opening hours, photos ect. An example of a restaurant url might be www.menus.com/london/bobs-pizza. A feature i would like to offer is the ability for Bob's pizza to use the menus.com website listing as his own website (let assume he has no website currently). I would like to purchase www.bobspizza.com and 301 redirect to www.menus.com/london/bobs-pizza Why?
Intermediate & Advanced SEO | | blackrails
So bob can then list bobspizza.com on his advertising material (business cards etc, rather than www.menus.com/london/bobs-pizza). I was considering using a 301 redirect for this though have been told that too many domain level redirects to one single domain can be flagged as spam by Google. Is there any other way to achieve this outcome without being penalised? Rel canonical url, url masking? Other things to note: It is fine if www.bobspizza.com is NOT listed in search results. I would ideally like any link juice pointing to www.bobspizza.com to pass onto www.menus.com though this is a nice to have. If it comes at the cost of being penalised i can live without the link juice from this. Thanks0 -
Multiple sites in the same niche
Hi All A question regarding multiple sites in the same niche... If I have say 10 sites all targetting the same niche yet all on different C-class IPs with different hosts, registrars, whois data and ages can I use the same template, or will Google discern a pattern? Basically I have developed a WordPress template which I want to use on the sites albeit with different logos / brand colours. NB/ All of the 10 sites will have unique, original content and they will NOT be interlinked
Intermediate & Advanced SEO | | danielparry1 -
Duplicating an article I wrote on an external blog
Hi, I wrote a blog article on another site. I would like to add the article to my site as well and would like to know the best way to do it. If I duplicate the article that I wrote would I then risk getting a penalty for duplicate content? If so, then what is the best way for me to include the article on my site for the benefit of my readers, but not lead to the duplicate content problem? Would it be better to use a canonical tag? Or to noindex the page? If I use the canonical tag, am I helping to make the article on the external blog stronger? Where is I use the noindex tag I am not helping my site nor that article I think, is that right? Last question, if I offer the copy of the article on my site and use the canonical or noindex tag then my site does not receive any direct benefit from the article for SEO. In other words the article wont appear in the search index with a link to my site. What about the comments that people write on the article on my site? That is unique content which may have great questions or points. I want to ensure those can be indexed properly. If I noindex the page I lose out. If I canonicalize (is that a word?) the page then I don't know if will send search results based on those comments to the external blog where that information (the comments from my site) does not exist. Thank you for any help to better understand this part of seo.
Intermediate & Advanced SEO | | NikkiGaul0 -
How many articles should I write per day & how many backlinks should I get per day to be natural!
hey.. I"m working in review blog one day per 1 or 2 weeks and I post up to 6 articles one time; is it unnatural for SEO ? how many articles should I post in blog per day? another question..how many backlinks should I get to just one post? I'm using Magic Submitter software to get help but I don't get more than 50 backlinks one time..what's real number of backlinks should I get and for how much time to be 100% natural for Google? any helpful info about backlinks techniques worth to hear..thnx
Intermediate & Advanced SEO | | akitmane0