Crawl error robots.txt
-
Hello, when trying to access the site crawl to be able to analyze our page, the following error appears:
**Moz was unable to crawl your site on Nov 15, 2017. **Our crawler was banned by a page on your site, either through your robots.txt, the X-Robots-Tag HTTP header, or the meta robots tag. Update these tags to allow your page and the rest of your site to be crawled. If this error is found on any page on your site, it prevents our crawler (and some search engines) from crawling the rest of your site. Typically errors like this should be investigated and fixed by the site webmaster.
Can help us?
Thanks!
-
@Linda-Vassily yes
-
The page is: https://frizzant.com/ And don't have noindex
-
Thanks Lind and Tawny! i 'll check it
-
Hey there!
This is a tricky one — the answer to these questions is almost always specific to the site and the Campaign. For this Campaign, it looks like your robots.txt file returned a 403 forbidden response to our crawler: https://www.screencast.com/t/f42TiSKp
Do you use any kind of DDOS protection software? That can give our tools trouble and cause us to be unable to access the robots.txt file for your site.
I'd recommend checking with your web developer to make sure that your robots.txt file is accessible to our user-agent, rogerbot, and returning a 200 OK status for that user-agent. If you're still having trouble, it'll be easier to assist you if you contact us through [email protected], where we can take a closer look at your account and Campaign directly.
-
I just popped that into ScreamingFrog and I don't see a noindex on that page, but I do see it on some other pages. (Though that shouldn't stop other pages from being crawled.)
Maybe it was just a glitch that happened to occur at the time of the crawl. You could try doing another crawl and see if you get the same error.
-
The page is: http://www.yogaenmandiram.com/ And don't have noindex
-
Hmm. How about on the page itself? Is there a noindex?
-
Yes, our robots.txt it's very simple:
User-agent: *
Disallow: /wp-admin/
Allow: /wp-admin/admin-ajax.php -
That just says that you are blocking the Moz crawler. Take a look at your robots.txt file and see if you have any exclusions in there that might cause that page not to be crawled. (Try going to yoursite.com/robots.txt or you can learn more about this topic here.)
-
Sorry, the image don't appear
Try now -
It looks like the error you are referring to did not come through in your question. Could you try editing it?
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Site Crawl Status code 430
Hello, In the site crawl report we have a few pages that are status 430 - but that's not a valid HTTP status code. What does this mean / refer to?
Product Support | | ianatkins
https://en.wikipedia.org/wiki/List_of_HTTP_status_codes#4xx_Client_errors If I visit the URL from the report I get a 404 response code, is this a bug in the site crawl report? Thanks, Ian.0 -
Crawl test
I used to use the crawl test tool to crawl websites and it presented the information in a really useful hierarchy of pages. The new on-demand crawl test doesn't seem to do this. Is there another tool I should be using to get the data?
Product Support | | Karen_Dauncey0 -
How do I fix the 500 error when trying to use the page optimization tool?
I keep getting an error when using the page optimization tool - Moz staff replied when I used the chatbot and said that they're receiving a 500 error from my server and to whitelist pagella however my server is not blocking anything. I don't know how to fix this issue any ideas? I've attached a picture of the error message I'm receiving for reference. zzwUlt0
Product Support | | GogoBusinessAviation1 -
My site crawl has been in progress since last week
Hi there, I've been waiting on my site crawl to complete since Friday (it's Tuesday now), but it still has the 'in progress' notification at the top. Is it normal for it to take over 3 days? Or is there something holding it up?
Product Support | | VAPartners0 -
Why can I not crawl this site
I wanted to add this site as new campaign: new.kbc.be But it won't accept it. Why?
Product Support | | KBC0 -
Receiving incapsula error codes when trying to respond to questions
I have been receiving multiple error codes from incapsula when trying to edit answers or add an additional answer. I want to save the code I have seen is 112 or 12 I will take a screenshot the next time I see it. However, it has happened to me quite a few times and I have to refresh my browser every time I choose to post, edit basically do anything on the site. Using both Safari, Chrome and Firefox have yielded the same results. Tom PS I have tried to replicate the results when posting this no luck.
Product Support | | BlueprintMarketing0 -
503 error on Moz
Is any one else getting a 503 error (I'll copy it below) whilst trying to get on Moz, I can get to the home page but going to internal links I get the error, I managed to get onto Q&A via an external link. I've been having some troubles the last few weeks with logging on etc. Most importantly Is Roger ok ? Error 503 Service Unavailable Service Unavailable Guru Meditation: XID: 1257248781 Varnish cache server
Product Support | | GPainter0 -
Showing 302 redirection error instead of 404
Moz is showing 302 error instead of 404 in the Crawl Diagnostics Summary report. But there is no such page uploaded ever to the site. Please let me know why this is happening. eg: http://www.zco.com/images/2d-game-development.aspx
Product Support | | zco_seo0