404's in WMT are old pages and referrer links no longer linking to them.
-
Within the last 6 days, Google Webmaster Tools has shown a jump in 404's - around 7000. The 404 pages are from our old browse from an old platform, we no longer use them or link to them.
I don't know how Google is finding these pages, when I check the referrer links, they are either 404's themselves or the page exists but the link to the 404 in question is not on the page or in the source code. The sitemap is also often referenced as a referrer but these links are definitely not in our sitemap and haven't been for some time. So it looks to me like the referrer data is outdated. Is that possible?
But somehow these pages are still being found, any ideas on how I can diagnose the problem and find out how google is finding them?
-
How long ago did you switch platforms? It can take months for Google to come back around to a page that linked to your site. Page on your site will stay in the cache until a few passes.
When you switch, did you do any 301 redirects? Examine the back links to your domain - any that come from good pages should be redirected to the new URL. If not, they will be scooped up by active SEOs. (finding 404 links is a popular link building technique).
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=93633
If you know the links will be dead forever, try using a 410 response as it is supposed to make search engines drop the page faster.
http://www.seroundtable.com/404-410-google-15225.html (bottom)
Have you requested Google remove old directories/pages? If the content is gone and has no back links, try a removal request.
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=1663427
-
Having a similar problem with a new site that was created by copying an old site in its entirety. Went through the trouble of cleaning everything up, having pages that were no longer relevant removed, fixed the sitemaps, etc. and now months later WMT showed me a spike of 404s for the old pages with the referrers as the XML sitemap and sitemap page... but they are definitely not be linked from there. I'm assuming there was some sort of hiccup with Google using an older, cached version of the sitemap to find these links.
I wound up just clearing the errors out of WMT and waiting to see if it will recrawl the error pages again. If Google continues to crawl them even though they aren't being linked to, then our next course of action was going to be 301ing them all just in case.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
#Page Jump link sharing
Hi I'm managing an in-house link building campaign in order to help in our key search term 'Location Holidays'. We were historically number 1 for this term until a recent re-design in May where our web design agency butchered our SEO. All of the main issued fixed, we're now fluctuating between 3rd & 4th on a daily basis. I'm putting together a social share comp to promote through the press in order to boost our backlink profile. We're nesting the competition within the body of the page we want to improve the rankings for. I will be including a #page jump link to quickly access it as it will be further down the page. My question is that if we get press to link to http://holidaycompany.com/destination/#comp will http://holidaycompany.com/destination/ receive the link juice or will http://holidaycompany.com/destination/#comp be looked upon as a whole new page? Thanks in advance!
Technical SEO | | MattHolidays0 -
How to find all crawlable links on a particular page?
Hi! This might sound like a newbie question, but I'm trying to find all crawlable links (that google bot sees), on a particular page of my website. I'm trying to use screaming frog, but that gives me all the links on that particular page, AND all subsequent pages in the given sub-directory. What I want is ONLY the crawlable links pointing away from a particular page. What is the best way to go about this? Thanks in advance.
Technical SEO | | AB_Newbie0 -
How to Remove Old Comment Page Query String URLs
I used to use a comments program on my website that created comment pages in the form of http://www.example.com/web-page.htm?comm_page=2. When I switched to a new comments program, I worried that these old comment URLs would be considered duplicate content. I created a 301 redirect that, for example, would redirect http://www.example.com/web-page.htm?comm_page=2 to http://www.example.com/web-page.htm and disallowed them in robots.txt, which I later learned was not the thing to do.. I have removed the URLs from being disallowed in robots.txt. However, many months later, these comment page URLs keep appearing in Google's index from time to time. I use the "Remove URLs" tool in Google Webmaster Tools to remove the URLs from Google's index, but more URLs appear a few days later. How can I get rid of these URLs for good? Thanks!
Technical SEO | | MrFrost0 -
New domain's Sitemap.xml file loaded to old domain - how does this effect SEO?
I have a client who recently changed their domain when they redesigned their site. The client wanted the old site to remain live for existing customers with links to the new domain. I guess as a workaround, the developer loaded the new domain's sitemap.xml file to the old domain. What SEO ramifications would this have if any on the primary (new) domain?
Technical SEO | | julesae0 -
What is the best approach to specifying a page's language?
I have read about a number of different tags that can accomplish this so it is very confusing. For example, should I be using: OR
Technical SEO | | BlueLinkERP0 -
Should i redirect my lost links to my home page
Hi, as some of you maybe aware, i had a major problem last year that has caused me nothing but trouble. in short, my hosting company lost me over 10,000 pages from my site and i had to rebuild the site from stratch which is still on going. I lost thousands of links to my site and i have been over the past week pointing the pages not found to the sections that is best suited to them. But i am just wondering if it would harm my site if i also point some of those links to my home page. I was a page rank four before disaster happened to my site and now i am a page rank two and i want to build this up. so i am just wondering if i should point some of those good links to my home page i am redirecting the pages using 301 in my htaccess file any advice would be great
Technical SEO | | ClaireH-1848860 -
Blocking URL's with specific parameters from Googlebot
Hi, I've discovered that Googlebot's are voting on products listed on our website and as a result are creating negative ratings by placing votes from 1 to 5 for every product. The voting function is handled using Javascript, as shown below, and the script prevents multiple votes so most products end up with a vote of 1, which translates to "poor". How do I go about using robots.txt to block a URL with specific parameters only? I'm worried that I might end up blocking the whole product listing, which would result in de-listing from Google and the loss of many highly ranked pages. DON'T want to block: http://www.mysite.com/product.php?productid=1234 WANT to block: http://www.mysite.com/product.php?mode=vote&productid=1234&vote=2 Javacript button code: onclick="javascript: document.voteform.submit();" Thanks in advance for any advice given. Regards,
Technical SEO | | aethereal
Asim0 -
Access To Client's Google Webmaster Tools
Hi, What's the best/easiest way for a client to grant access to his Google Webmaster Tools to me? Thanks! Best...Michael
Technical SEO | | 945010