Site architecture change - +30,000 404's in GWT
-
So recently we decided to change the URL structure of our online e-commerce catalogue - to make it easier to maintain in the future.
But since the change, we have (partially expected) +30K 404's in GWT - when we did the change, I was doing 301 redirects from our Apache server logs but it's just escalated.
Should I be concerned of "plugging" these 404's, by either removing them via URL removal tool or carry on doing 301 redirections? It's quite labour intensive - no incoming links to most of these URL's, so is there any point?
Thanks,
Ben
-
Hi Ben,
The answer to your question boils down to usability and link equity:
- Usability: Did the old URLs get lots of Direct and Referring traffic? E.g., do people have them bookmarked, type them directly into the address bar, or follow links from other sites? If so, there's an argument to be made for 301 redirecting the old URLs to their equivalent, new URLs. That makes for a much more seamless user experience, and increases the odds that visitors from these traffic sources will become customers, continue to be customers, etc.
- Link equity: When you look at a Top Pages report (in Google Webmaster Tools, Open Site Explorer, or ahrefs), how many of those most-linked and / or best-ranking pages are old product URLs? If product URLs are showing up in these reports, they definitely require a 301 redirect to an equivalent, new URL so that link equity isn't lost.
However, if (as is common with a large number of ecommerce sites), your old product URLs got virtually zero Direct or Referring traffic, and had virtually zero deep links, then letting the URLs go 404 is just fine. I think I remember a link churn report in the early days of LinkScape when they reported that something on the order of 80% of the URLs they had discovered would be 404 within a year. URL churn is a part of the web.
If you decide not to 301 those old URLs, then you simply want to serve a really consistent signal to engines that they're gone, and not coming back. Recently, JohnMu from Google suggested recently that there's a tiny difference in how Google treats 404 versus 410 response codes - 404s are often re-crawled (which leads to those 404 error reports in GWT), whereas 410 is treated as a more "permanent" indicator that the URL is gone for good, so 410s are removed from the index a tiny bit faster. Read more: http://www.seroundtable.com/google-content-removal-16851.html
Hope that helps!
-
Hi,
Are you sure these old urls are not being linked from somewhere (probably internally)? Maybe the sitemap.xml was forgotten and is pointing to all the old urls still? I think that for 404's to show in GWT there needs to be a link to them from somewhere, so in the first instance in GWT go to the 404s and have a look at where they are linked from (you can do this with moz reports also). If it is an internal page like a sitemap, or some forgotten menu/footer feature or similar that is still linking to old pages then yes you certainly want to clear this up! If this is the case, once you have fixed the internal linking issues you should have significantly reduced list of 404s and can then concentrate on these on a more case by case basis (assuming they are being triggered by external links).
Hope that helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Site traffic halved not sure why
Hi guys, not sure if anyone can help, but we had a client's google organic traffic literally halve from the week at the end of August to September (29 Aug 2016 to be precise) and it hasn't recovered since (here's a screenshot from GA http://puu.sh/sAmd3/b071dd1e57.png) I've been doing a lot of digging around on Moz and elsewhere about any Google updates that may have gone through around that time and there doesn't seem to be anything that I would think would affect it. I thought it might be to do with Penguin, but that doesn't seem to be the case. A while ago before then we did have some domains and pages 301 redirected to the main site when multiple other sites were rolled into the one, but I wouldn't have thought that should affect it. After that I've also gone and removed all those sites and redirects too (couple of weeks ago) but that doesn't seem to have fixed it. There's no black hat SEO done on the site so very odd to have this happen. I'm rather out of ideas what it could be that has impacted things so suddenly and that we couldn't get it recovered from. Any ideas would be much appreciated.
White Hat / Black Hat SEO | | BrisbaneSEOWorks0 -
Should I submit a sitemap for a site with dynamic pages?
I have a coupon website (http://couponeasy.com)
White Hat / Black Hat SEO | | shopperlocal_DM
Being a coupon website, my content is always keeps changing (as new coupons are added and expired deals are removed) automatically. I wish to create a sitemap but I realised that there is not much point in creating a sitemap for all pages as they will be removed sooner or later and/or are canonical. I have about 8-9 pages which are static and hence I can include them in sitemap. Now the question is.... If I create the sitemap for these 9 pages and submit it to google webmaster, will the google crawlers stop indexing other pages? NOTE: I need to create the sitemap for getting expanded sitelinks. http://couponeasy.com/0 -
How to make second site in same niche and do white hat SEO
Hello, As much as we would like, there's a possibility that our site will never recover from it's Google penalties. Our team has decided to launch a new site in the same niche. What do we need to do so that Google will not mind us having 2 sites in the same niche? (Menu differences, coding differences, content differences, etc.) We won't have duplicate content, but it's hard to make the sites not similar. Thanks
White Hat / Black Hat SEO | | BobGW0 -
International web site - duplicate content?
I am looking at a site offering different language options via a javascript drop down chooser. Will google flag this as duplicate content? Should I recommend the purchase of individual domains for each country? i.e. .uk
White Hat / Black Hat SEO | | bakergraphix_yahoo.com1 -
Seeking Top Notch Marketing Company with experience in growing sites post manual penalty
Does anyone know of a company who has direct experience with growing websites AFTER a manual link penalty has been lifted? Any referrals would be great!
White Hat / Black Hat SEO | | WebServiceConsulting.com0 -
Are directory listings still appropriate in 2013? Aren't they old-style SEO and Penguin-worthy?
We have been reviewing our off-page SEO strategy for clients and as part of that process, we are looking at a number of superb info-graphics on the subject. I see that some of current ones still list "Directories" as being part of their off-page strategy. Aren't these directories mainly there for link-building purposes and provide Users no real benefit? I don't think I've ever seen a directory that I would use, apart for SEO research. Surely Google's Penguin algorithm would see directories in the same way and give them less value, or even penalise websites that use them to try to boost page rank? If I were to list my websites on directories it wouldn't be to share my lovely content with people that use directories to find great sites, it would be to sneakily build page rank. Am I missing the point? Thanks
White Hat / Black Hat SEO | | Crumpled_Dog
Scott0 -
Can anyone tell me why this site ranks so well?
Site in question: cellphoneshop.net From what I can tell from their link profile, the links they garner don't appear to be particularly high value but they dominate organic listings for my vertical (cell phone accessories), esp. in the last 2-3 months when Google was supposedly increasing the quality of their search results. Can anyone tell me why in particular this site ranks so well for competitive short and long tail terms?
White Hat / Black Hat SEO | | eugeneku0 -
It Shows as "google results" but it's an incoming links, is it spaming me...?
Hello everyone I have 2 issues to share: 1) We have a site (personal-loans.org), In the past few weeks we notice that there are sites that have links to our site and we get traffic from them but...! when you go online to these sites they show you that all they do is provide "google search" results, because we where in first page on the results we had hits there as well what leads me to think that this is the reason we are at page 7 now after yesterday the ranking was at page 4. these are some of these sites so you can see it: internetpayadvances.com fastlivecashadvance.com assistancemoney.com scoutcashnow.com officialpayday.net Does anyone else got to see anything like that...??? I have many more links like that, these are only 5 out of 9 that had hits yesterday only, site traffic went from 250-300 to 63 a day... For the same site - it was on google search results 1st page and ranked 4-7, even after the big penguin changes. What we did notice is that A LOT of non related sites like surfing (yes ocean surfing) and sites that had no content AT ALL - all the text was inside of an image and ranked 3! 3rd on payday loans search result. (and the rest was and still just looks the same with different content...) Google say they want quality but does not do homework for the 2nd largest search for keywords such as loans and payday loans market, same goes for the cash advance. Please help, need your advice.... Thanks
White Hat / Black Hat SEO | | Yonnir0