Severe Health issue on my site through Webmaster tools
-
I use Go Daddy Website Tonight. I keep getting a severe health message in Google Webmaster tools stating that my robots.txt file is blocking some important page. When I try to get more details the blocked file will not open. When I asked the Go Daddy peeps they told me that it was just image and backup files that do not need to be crawled. But if Google spiders keep thinking an important page is blocked will this hurt my SERPS?
-
I would just like to add: If you're considering signing up for something (SEV), you may as well get a real hosting package.
-
Thanks for letting us know, and glad you found a work-around. A 0-second META REFRESH sometimes acts like a 301 - it's not ideal, as you said, but it's something.
-
For anyone else with Website Tonight, I have finally found a work around if not a fix. Being that Website Tonight will not allow you to do a 301 Redirect of an old page, I have figured out that if you re-create the deleted page (just the URL not the content) and use the Meta Tag to do a REFRESH to the new page, anytime the old page is clicked on it will bring them to the new page. Not ideal, of course, for SEO purposes but at least they are no longer going to a 404 and HOPEFULLY your old link juice will pass on.
-
While creating the copy of the home-page isn't ideal, if Google hasn't indexed it, it's very likely not creating duplicate content problems. Either they're filtering it out or haven't indexed it at all (since it probably has no links/paths).
I don't think that this alone is the cause of your ranking drop, but you've got a few things going on, so it's tough to say. Unfortunately, most of the ideal solutions seem to be impossible in the Godaddy system, and that's going to continue to cause you some problems.
-
No I have not made any changes yet. Google has never preferred the /shakeology.html page. I don't think it's ever been indexed. My only problem is that since I tried to CHANGE the root url not CREAT ANOTHER VERSION my serps have seemed to tank and I am trying to avoid the duplicate content issues that I believe the /shakeology.html is causing.
-
No I have not made any changes yet. Google has never preferred the /shakeology.html page. I don't think it's ever been indexed. My only problem is that since I tried to CHANGE the root url not CREAT ANOTHER VERSION my serps have seemed to tank and I am trying to avoid the duplicate content issues that I believe the /shakeology.html is causing.
-
It's a bit dangerous to simply block "shakeology.html", if Google has preferred it for a reason - you could end up getting your root page back in the rankings, or you could end up just falling out completely. I think you'd be better off leaving it and having the "wrong" page rank, if that's the only viable option.
I'm actually still showing your root home-page ranking, though, and now the "shakeology.html" page isn't even appearing in the index. Did you already make a change?
-
My original wanted homepage is www .hompage .com
My duplicate is www. hompage .com/shakeology.html
Would it be possible and/or advisable to use a Parameter in Web Master tools to ignore the /shakeology.html?
-
Unfortunately, there just comes a point where sometimes these very narrow CMS systems hit their limits, and it can start to harm you. I don't know Website Tonight well enough to help on that (hopefully someone else does), but there may come a point where you want to consider to a more advanced platform. These days, there are a lot of options that aren't budget-breakers, although switching is always a bit tough.
-
Unfortunately no. The edit page section of Website Tonight only shows the newer homepage.com/shakeology.html version and not the original homepage.com
I am afraid to delete the homepage.com/shakeology.html in fear that I will be left with neither one.
You probably aren't seeing the preview of the .com/shakeology.html because it is not the indexed homepage. It shows for the homepage.com version. The canonical tag was me trying to redirect search engines from the new (unwanted) homepage to the original because Website tonight won't allow me to 301 it.
-
Sounds like SEO Executive has got you covered on the Godaddy front - just wanted to point out a couple of things:
(1) I'm not seeing a preview for your home-page, and I had trouble connecting to it the first time. It seems to be cahced, so this could be a fluke.
(2) Not sure if this is part of the Godaddy code, but there's a really weird tag on the home-page:
name="canonical tag" content=""/>
That might just be a reference, but it doesn't do anything. If it's supposed to actually be a canonical, then something is broken.
-
Yes I saw that but unfortunately the organize your site page on website tonight only shows the new page. I'm afraid to delete it and lose both.
-
I found some great info here that I believe explains it: http://support.godaddy.com/help/2986/organizing-your-website-using-the-organize-site-page
-
I really do appreciate all of your help. Here's the issue I'm having with this though... After I renamed the homepage file to add /shakeology.html to it (because I thought it would be beneficial to have a main keyword in the url) Website Tonight only shows me the homepage.com/shakeology.html and not the homepage.com. I'm afraid that if I delete /shakeology.html I will show neither one and in essence, according to Website Tonight just be deleting my homepage. I'm not sure how to properly accomplish what I'm looking to do without screwing myself any further??
-
Personally, that's what I would do is delete it. Unless, there is a reason you need that page.
-
Since I can't 301 it, would it be bad to delete the dupe page?
-
Yes thanks. I foolishly renamed my home page and caused a duplicate page. Website tonight will not allow me to do a 301 redirect. I put a canonical tag on the /shakeology. Should this do the trick?
-
Your welcome! I'm also sending the other side of the story not to confuse you but to allow you to make a decision based on both sides: http://groups.google.com/a/googleproductforums.com/forum/#!category-topic/webmasters/crawling-indexing--ranking/8nyxCtv9RHM
-
Oh OK great. Thanks so much for your help. I just got nervous because Google puts up the Severe Health Issue warning everytime I get crawled.
-
This is a javascript file and I don't see it being an issue unless Google thinks your hiding it to be spammy . Also, there are some that say it's a benefit to block js files from the search for SEO purposes. Here is an example of that situation: http://www.seomofo.com/advanced/do-not-let-google-crawl-javascript.html I think since this is out of your control and goes by the standard of how Godaddy sets up there sites, then it shouldn't be an issue.
-
It was just crawled. And it was after robots.txt was uploaded. This is the page it lists: siteUtil.js
-
Also, the following are duplicates: http://www.shakes4life.com/shakeology.html & http://www.shakes4life.com
-
When did Google last index your site? You can check this through webmaster tools. When did you instal the robots.txt file. The reason I ask: If Google's last crawl was before you uploaded your robots file then that could be the issue. Please look at these statistics and verify this before we move further.
-
Is Google webmaster tools giving you the specific name of the files that are being blocked?
-
Is there something on it that would be detrimental to my SERPS?
-
Yes. When I type www.shakes4life.com/robots.txt the same list shows.
-
Can you place the following in your browser and replace website with your domain name and www or non www in front. website.com/robots.txt
Let me know if you see the same stuff you sent me in your last response
-
Below is the robots.txt Website Tonight Creates when I tell it to allow all pages:
User-agent: *
Allow: /
User-agent: *
Disallow: /cache/
Disallow: /_backup/
Disallow: /_mygallery/
Disallow: /_temp/
Disallow: /_tempalbums/
Disallow: /_tmpfileop/
Disallow: /dbboon/
Disallow: /Flash/
Disallow: /images/
Disallow: /plugins/
Disallow: /scripts/
Disallow: /stats/
Disallow: /statshistory/
Disallow: /WstxSearchResults.html
Disallow: /WstxSearchResults.php
Disallow: /QSC/ -
Yes you are correct. I forgot to mention (sorry) that I do use S.E.V. It allows you to create a robots.txt and lets you choose pages to block. However, even when you choose allow all, by default it blocks certain files. Go Daddy tells me they are only system files but Google tells me an important page is blocked.
-
From what I know, Godaddy Website Tonight does not offer you the opportunity to create a custom robots.txt. I believe you have to sign up for there Search Engine Visibility services. Here is some more information: http://support.godaddy.com/help/article/5321
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Internal Linking issue
So i am working with a review company and I am having a hard time with something. We have created a category which lists and categorizes every one of our properties. For example a specific property in the category "restaurant" would be as seen below: /restaurant/mcdonalds /restaurant/panda-express And so on and so on. What I am noticing however is that our more obscure properties are not being linked to by any page. If I were to visit the page myurl.com/restaurant I would see 100+ pages of properties, however it seems like only the properties on the first few pages are being counted as having links. So far the only way I have been able to work around this issue is by creating a page and hiding it in our footer called "all restaurants". This page lists and links to every one of our properties. However it isn't exactly user friendly and I would prefer scrapers not to be able to scrape all properties at once! Anyway, any suggestions would be greatly appreciated.
Technical SEO | | HashtagHustler0 -
Website indexing issues
My website is being indexed with both https - https with www. and no leader at all. example. https//www.example.com and https//example.com and example.com 3 different versions are being indexed. How would I begin resolving this? Hosting?
Technical SEO | | DigitalRipples0 -
Bulk URL Removal in Webmaster Tools
One of Wordpress sites was hacked (for about 10 hours), and Google picked up 4000+ urls in the index. The site is fixed, but I'm stuck with all those urls in the index. All the urls of of the form: walkerorthodontics.com/index.php?online-payday-cash-loan.htmloncewe The only bulk removal option I could find was to remove an entire folder, but I can't do that, as it would only leave the homepage and kill off everything else. For some crazy reason, the removal tools doesn't support wildcards, so that obvious solution is right out. So, how do it get rid of 4000 results? And no, waiting around for them to 404 out of the index isn't an option.
Technical SEO | | MichaelGregory0 -
Site hacked in Jan. Redeveloped new site. Still not ranking. Should we change domain?
Our top ranking site in the UK was hacked at the end of 2014. http://www.ultimatefloorsanding.co.uk/ The site was the subject of a manual spam action from Google. After several unsuccessful attempts to clean it up, using Securi.net and reinstating old versions of the site, changing passwords etc. we took the decision to redevelop the site. We also changed hosting provider as we had received absolutely no support from them whatsoever in resolving the issue. So far we have: Removed the old website files off the server Developed a new website having implemented 301's for all the old URL's (except the spam ones) Submitted a reconsideration request for the manual spam action, which was accepted. Disavowed all the spammy inbound links through Webmaster Tools Implemented custom URL parameters through Google to not index the SPAM URLs ( which were using parameters) Our organic traffic is down by 63% compared to last year, and we are not ranking for most of our target keywords any longer. Is there anything that I am missing in the actions I have taken so far? We were advised that at this stage changing domain and starting again might be the way to go. However the current domain has been used by us since 2007, so it would be a big call. Any advice is appreciated, thanks. Sue - http://www.ultimatefloorsanding.co.uk/
Technical SEO | | galwaygirl0 -
Switching site from http to https. Should I do entire site?
Good morning, As many of you have read, Google seems to have confirmed that they will give a small boost to sites with SSL certificates this morning. So my question is, does that mean we have to switch our entire site to https? Even simple information pages and blog posts? Or will we get credit for the https boost as long as the sensitive parts of our site have it? Anybody know? Thanks in advance.
Technical SEO | | rayvensoft1 -
Pagination and Canonocal Issue
Hi, I have a site which have city wise pages and in a given city we have categories. The listed products can be listed in different categories which have separate URL. The site have different URL, meta, title for each category. We want to Rank these pages based on category also... What is best way to avoid duplicate and canonical issue.. Thanks,
Technical SEO | | dsingh1079
Darshan..0 -
Bogus Crawl Errors in Webmaster Tools?
I am suddenly seeing a ton of crawl errors in webmaster tools. Almost all of them are URL links coming from scraper sites.that I do not own. Do you see these in your Webmaster Tools account? Do you mark them as "fixed" if they are on a scraper site? There are waaaay too many of these to make redirects. Thanks!
Technical SEO | | EGOL0 -
MSNbot Issues
We found msnbot is doing lots of request at same time to one URL, even considering we have caching, it triggers many requests at same time so caching does not help at the moment: For sure we can use mutex to make sure URL waits for cache to generate, but we are looking for solution for MSN boot. 123.253.27.53 [11/Dec/2012:14:15:10 -0600] "GET //Fun-Stuff HTTP/1.1" 200 0 "-" "msnbot/2.0b (+http://search.msn.com/msnbot.htm)" 1.253.27.53 [11/Dec/2012:14:15:10 -0600] "GET //Type-of-Resource/Fun-Stuff HTTP/1.1" 200 0 "-" "msnbot/2.0b (+http://search.msn.com/msnbot.htm)" 1.253.27.53 [11/Dec/2012:14:15:10 -0600] "GET /Browse//Fun-Stuff HTTP/1.1" 200 6708 "-" "msnbot/2.0b (+http://search.msn.com/msnbot.htm)" We found the following solution: http://www.bing.com/community/site_blogs/b/webmaster/archive/2009/08/10/crawl-delay-and-the-bing-crawler-msnbot.aspx Bing offers webmasters the ability to slow down the crawl rate to accommodate web server load issues. User-Agent: * Crawl-Delay: 10 Need to know if it’s safe to apply that. OR any other advices. PS: MSNBot gets so bad at times that it could trigger a DOS attack – alone! (http://www.semwisdom.com/blog/msnbot-stupid-plain-evil#axzz2EqmJM3er).
Technical SEO | | tpt.com0