How Best to Handle Inherited 404s on Purchased Domain
-
We purchased a domain from another company and migrated our site over to it very successfully. However, we have one artifact of the original domain in that there was a page that was exploited by other sites on the web. This page allowed you to pass any URL to it and redirect to that URL (e.g. http://example.com/go/to/offsite_link.asp?GoURL=http://badactor.com/explicit_content).
This page does not exist on our site so the results always go to a 404 on our site. However, we find that crawlers are still attempting to access these invalid pages.
We have disavowed as many of the explicit sites as we can, but still some crawlers come looking for those links. We are considering blocking the redirect page in our robots.txt but we are concerned that the links will remain indexed but uncrawlable.
What's the best way to pull these pages from search engines and never have them crawled again?
UPDATE: Clarifying that what we're trying to do it get search engines to just never try to get to these pages. We feel the fact they're even wasting their time on getting a 404 is what we're trying to avoid. Is there any reason we shouldn't just block these in our robots.txt?
-
@gastonriera calm down mate. We have actually tested this at not seen any negative effect on any site we have done it on. It is the "easiest" option, but it won't cause the death and destruction your comment implies. Good day sir.
-
Hi there,
I'm considering that you have over 500k URLs, to be worrying about crawl efficiency. If you have less than that, please don't worry.
Having 404s is completely fine, and google will eventually lower their crawl frequency to those pages.
Blocking them in robots.txt will cause to google stop crawling them, but never to never remove them from the index.
My advice here: don't block them in robots.txtAs Rajesh pointed out, you could force those 404s into 410 to tell Google that they are gone forever. Yet, Google said that they treat 404s and 410s as the same.
John Mueller said over a year ago that 4xx status codes don't incur in crawl wastage. You can check it our in these Webmasters hangout notes - DeepcrawlHope it helps,
Best luck.
Gaston -
FOR THE LOVE OF GOD DONT REDIRECT 404s TO THE HOME!
This is terrible advice. Doing that you'll turn those 404s into soft 404s, making them more problematic than ever.
-
I would actually recommend redirecting it to the homepage. If you have a Wordpress website and a bunch of 404 pages, you can install a free plugin called "All 404 to Homepage" and this will solve the problem. I would, however, recommend that if you have replacement pages or pages covering similar content, that you redirect those to the corresponding replacement page.
-
You need to do one thing with those 404 pages. Move them as 410 status code. Redirection is not good practice for the same.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Which is the best option for these pages?
Hi Guys, We have product pages on our site which have duplicate content, the search volume for people searching for these products is very, very small. Also if we add unique content, we could face keyword cannibalisation issues with category/sub-category pages. Now based on proper SEO best practice we should add rel canonical tags from these product pages to the next relevant page. Pros Can rank for product oriented keywords but search volume is very small. Any link equity to these pages passed due to the rel canonical tag would be very small, as these pages barely get any links. Cons Time and effort involved in adding rel canonical tags. Even if we do add rel canonical tags, if Google doesn't deem them relevant then they might ignore causing duplicate content issues. Time and effort involved in making all the content unique - not really worth it - again very minimal searchers. Plus if we do make it unique, then we face keyword cannibalisation issues. -- What do you think would be the optimal solution to this? I'm thinking just implementing a: Across all these product based pages. Keen to hear thoughts? Cheers.
Intermediate & Advanced SEO | | seowork2140 -
How to switch brand domain and address previous use of domain
We recently acquired a new domain to replace existing as it better fits our brand. We have little/no organic value on existing domain so switching is not an issue. However the newly acquired domain was previously used in a different industry and has inbound links with significant spam scores. How can we let Google know that these links are not valid for our business and start rebuilding reputation of the domain? Disavow tool?
Intermediate & Advanced SEO | | Marlette0 -
What's the best way to deal with deleted .php files showing as 404s in WMT?
Disclaimer: I am not a developer During a recent site migration I have seen a bit of an increase in WMT of 404 errors on pages ending .php. Click on the link in WMT and it just shows as File Not Found - no 404 page. There are about 20 in total showing in webmaster tools and I want to advise the IT department what to do. What is the best way to deal with this for on-page best practice? Thanks
Intermediate & Advanced SEO | | Blaze-Communication0 -
No Domain Authority after 6 weeks
Hi, We have just launched a new e-commerce site and 301'd a lot of the products and categories to the new site. I have also added a link or 2 from other domains. We launched it on 10th of June - but still using site explorer, there is no domain authority showing - just a 1. Why is that? Any ideas?
Intermediate & Advanced SEO | | bjs20100 -
Duplicate Sub-domains Being Indexed
Hi all, I have this site that has a sub-domain that is meant to be a "support" for clients. Some sort of FAQ pages, if you will. A lot of them are dynamic URLs, hence, the title and most of the content are duplicated. Crawl Diagnostics found 52 duplicate content, 138 duplicate title and a lot other errors. My question is, what would be the best practice to fix this issue? Should I noindex and nofollow all of its subdomains? Thanks in advance.
Intermediate & Advanced SEO | | EdwardDennis0 -
Purchase of domain name from a different industry.
Hi I am thinking of acquiring a domain name, although it is currently being used in a completely different industry to the one I hoping to use it for. The site only has 46 links and was registered in 2009. It has a DA of 25, Home PA of 37 and PR of 2 I was just wondering how easy or hard it would be to optimise the website for a completely different industry, i.e. lets say it was initially bought to sell hair-care products and I want to use it to sell electronics. Would I leave the existing links in? Could I use that new disavow tool in webmaster tools to wipe the slate clean and start again? Really haven't come across this before, does anyone have any ideas? Thanks for your time, Steve
Intermediate & Advanced SEO | | Townpages0 -
Affects of vanity domains?
Hi there! My client's domain name is his name, let's say (www,myname.com) because he is well-known in his industry. He is a physician with a very specific specialty and organic competition is fierce for the most relevant keyword to his specialty. A domain has just become available that includes the keyword. If we bought the domain, how could we use it to our advantage? I'm confused about redirecting, etc., with this type of situation. Am i making any sense here? Help! Thanks. 🙂
Intermediate & Advanced SEO | | Mills0 -
For multi language sites, what is best - two domains or one with both languages?
We are assisting a client in setting up English and Spanish sites in Texas. They want to be able to find customers who are Spanish speaking predominantly or totally along with the customers they now get who are English speakers. We are building them a new site and I have researched to find answers all over the board or less than clear. Should the structure be such that we have one site with a set of English and Spanish pages all with Spanish links to Spanish pages and English links to English pages. Should we instead just have an English site for those people who utilize English and a different site for those who utilize Spanish? Thanks
Intermediate & Advanced SEO | | RobertFisher0