Duplicate content when changing a site's URL due to algorithm penalty
-
Greetings
A client was hit by penguin 2.1, my guess is that this was due to linkbuilding using directories. Google webmaster tools has detected about 117 links to the site and they are all from directories. Furthermore, the anchor texts are a bit too "perfect" to be natural, so I guess this two factors have earned the client's site an algorithm penalty (no manual penalty warning has been received in GWT).
I have started to clean some of the backlinks, on Oct the 11th. Some of the webmasters I asked complied with my request to eliminate backlinks, some didn´t, I disavowed the links from the later.
I saw some improvements on mid october for the most important KW (see graph) but ever since then the rankings have been falling steadily.
I'm thinking about giving up on the domain name and just migrating the site to a new URL. So FINALLY MY QUESTION IS: if I migrate this 6-page site to a new URL, should I change the content completely ? I mean, if I just copy paste the content of the curent site into a new URL I will incur in dpolicate content, correct?.
Is there some of the content I can copy ? or should I just start from scratch?
Cheers
-
Hey Masoko -
In the past, I've had luck with 410ing the previous site and putting a link from it saying that we've moved. This way, you keep any direct traffic by referring them, but you also don't redirect your pages via 301.
Penalties pass through redirects. You don't want to keep both sites and duplicate content. I'd kill off the old site (it's only 6 pages, so that's pretty easy) and take the chance to, as has been said, refresh the content. Also, think about adding more pages to the site so you can rank for more longtail terms.
Good luck.
-
Thanks everyone for answering my question!!!
-
As long as you 410 (delete) the old pages, they are no longer indexed and will not cause a duplicate content issue.
-
You can safely move to a new domain, move the content over (upgrade it a little) and there should be no duplicate content issues. The duplicate content issues were designed for things like just scraping content from news feeds and posting them on your own site, and not having any unique or original. Or selling products as a reseller and not doing anything to the manufacturers text etc.
If you move the site to a new domain - I would just 410 the pages on the old site and not do any redirects. You were probably only ranking for a short period of time because of the unnatural back links. If you redirect them you will pass the negative link values over to the new site (those that were not fixed or disavowed anyway) and there is probably not much for good link metrics to warrant a redirect. You will lose any traffic from people who are trying to visit the old site, so maybe you can put up a message on the old site's homepage that it has moved to a new domain, but not link to it.
-
Masoko-T,
If you're sure that the penalty is from link building, you should have no problem. As mentioned above a refresh of the content, might be a good idea though.
-
Hi Tuzzel
Thanks for your reply. Are you sure there are no duplicate content risks?, I thought that, since google had already indexed the original content, finding the same content in a different (newer) site will cause the later to be considered "duplicate".
I hadn't thought about the 302 redirects, that's not a bad idea :).
-
If you're moving a site, Google's recommendations are to move the content and redirect. However, it sounds like you're looking for a fresh start.
Are you sure it's the links? Are you also concerned about EMD penalty or just hoping for a fresh start?
-
You should be ok just to replicate it, but by all means use the opportunity to refresh the content, 6 pages shouldn’t take too long. If you want to be extra safe then you can of course just rewrite from scratch. The Penalty will be at the domain level so you should be ok to redirect the existing pages to the New URLs, this will signal to Search engines that the pages have been moved and not to count the redirected pages as unique content, avoiding Dupe content issues. You can also use a cross domain Canonical tag.
If you don’t want to do any redirects to totally severe your links to the old domain profile then remove the original pages from Google’s index in your webmaster tools account and ensure you return 410 status codes to individuals that request the page. If you do still want the users to redirect however 302 the page to the new location as this won’t pass link equity.
Hope this proves useful.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Partial Match or RegEx in Search Console's URL Parameters Tool?
So I currently have approximately 1000 of these URLs indexed, when I only want roughly 100 of them. Let's say the URL is www.example.com/page.php?par1=ABC123=&par2=DEF456=&par3=GHI789= All the indexed URLs follow that same kinda format, but I only want to index the URLs that have a par1 of ABC (but that could be ABC123 or ABC456 or whatever). Using URL Parameters tool in Search Console, I can ask Googlebot to only crawl URLs with a specific value. But is there any way to get a partial match, using regex maybe? Am I wasting my time with Search Console, and should I just disallow any page.php without par1=ABC in robots.txt?
Intermediate & Advanced SEO | | Ria_0 -
Why is this url redirecting to our site?
I was doing an audit on our site and searching for duplicate content using some different terms from each of our pages. I came across the following result: www.sswug.org/url/32639 redirects to our website. Is that normal? There are hundreds of these url's in google all with the exact same description. I thought it was odd. Any ideas and what is the consequence of this?
Intermediate & Advanced SEO | | Sika220 -
Don't affiliate programs have an unfair impact on a company's ability to compete with bigger businesses?
So many coupon sites and other websites these days will only link to your website if you have a relationship with Commission Junction or one of the other large affiliate networks. It seems to me that links on these sites are really unfair as they allow businesses with deep pockets to acquire links unequitably. To me it seems like these are "paid links", as the average website cannot afford the cost of running an affiliate program. Even worse, the only reason why these businesses are earning a link is because they have an affiliate program; that to me should violate some sort of Google rule about types and values of links. The existence of an affiliate program as the only reason for earning a link is preposterous. It's just as bad as paid link directories that have no editorial standards. I realize the affiliate links are wrapped in CJ's code, so that mush diminish the value of the link, but there is still tons of good value in having the brand linked to from these high authority sites.
Intermediate & Advanced SEO | | williamelward0 -
How do I best handle Duplicate Content on an IIS site using 301 redirects?
The crawl report for a site indicates the existence of both www and non-www content, which I am aware is duplicate. However, only the www pages are indexed**, which is throwing me off. There are not any 'no-index' tags on the non-www pages and nothing in robots.txt and I can't find a sitemap. I believe a 301 redirect from the non-www pages is what is in order. Is this accurate? I believe the site is built using asp.net on IIS as the pages end in .asp. (not very familiar to me) There are multiple versions of the homepage, including 'index.html' and 'default.asp.' Meta refresh tags are being used to point to 'default.asp'. What has been done: 1. I set the preferred domain to 'www' in Google's Webmaster Tools, as most links already point to www. 2. The Wordpress blog which sits in a /blog subdirectory has been set with rel="canonical" to point to the www version. What I have asked the programmer to do: 1. Add 301 redirects from the non-www pages to the www pages. 2. Set all versions of the homepage to redirect to www.site.org using 301 redirects as opposed to meta refresh tags. Have all bases been covered correctly? One more concern: I notice the canonical tags in the source code of the blog use a trailing slash - will this create a problem of inconsistency? (And why is rel="canonical" the standard for Wordpress SEO plugins while 301 redirects are preferred for SEO?) Thanks a million! **To clarify regarding the indexation of non-www pages: A search for 'site:site.org -inurl:www' returns only 7 pages without www which are all blog pages without content (Code 200, not 404 - maybe deleted or moved - which is perhaps another 301 redirect issue).
Intermediate & Advanced SEO | | kimmiedawn0 -
What do I do about sites that copy my content?
I've noticed that there are a number of websites that are copying my content. They are putting the full article on their site, mentioning that it was reposted from my site, but contains no links to me. How should I approach this? What are my rights and should I ask them to remove it or add a link? Will the duplicate content affect me?
Intermediate & Advanced SEO | | JohnPeters0 -
Pagination Question: Google's 'rel=prev & rel=next' vs Javascript Re-fresh
We currently have all content on one URL and use # and Javascript refresh to paginate pages, and we are wondering if we transition to the Google's recommended pagination if we will see an improvement in traffic. Has anyone gone though a similar transition? What was the result? Did you see an improvement in traffic?
Intermediate & Advanced SEO | | nicole.healthline0 -
Can you see the 'indexing rules' that are in place for your own site?
By 'index rules' I mean the stipulations that constitute whether or not a given page will be indexed. If you can see them - how?
Intermediate & Advanced SEO | | Visually0 -
Duplicate content
I have just read http://www.seomoz.org/blog/duplicate-content-in-a-post-panda-world and I would like to know which option is the best fit for my case. I have the website http://www.hotelelgreco.gr and every image in image library http://www.hotelelgreco.gr/image-library.aspx has a different url but is considered duplicate with others of the library. Please suggest me what should i do.
Intermediate & Advanced SEO | | socrateskirtsios0