Get a list of robots.txt blocked URL and tell Google to crawl and index it.
-
Some of my key pages got blocked by robots.txt file and I have made required changes in robots.txt file but how can I get the blocked URL's list.
My webmaster page Health>blocked URL's shows only number not the blocked URL's.My first question is from where can I fetch these blocked URL's and how can I get them back in searches,
One other interesting point I see is that blocked pages are still showing up in searches.Title is appearing fine but Description shows blocked by robots.txt file.
I need urgent recommendation as I do not want to see drop in my traffic any more.
-
"changing the lastmod of those pages to today".
How can I make these changes?
Right now the news is that Resubmitted the Sitemap and no warnings this time.
-
I imagine that since you've got a robots txt error you'll probably ended closing a whole directory to bots which you wanted to be indexed. You can easily spot the directory and resubmit a sitemap to google changing the lastmod of those pages to today and the priority to 1 but only of those pages.
If you still receive warnings it may be due to errors in your sitemap. You're probably including some directory you don't want. You can try it in GWT putting in the box at the bottom the url you want to maintain in the index and then trying to see if some urls are being blocked by your robots.
If you want you can post here your robots and the URIs you want to be indexed without knowing the domain so that won't be public. Hope this may help you
-
Ok Resubmitted it.but even with updated file it gives a lot of errors.I think it takes some time.20,016 warnings
I have not added no index attribute in my header region.It was all messy stuff with robots.txt file.It means that with site showing up in SERP the rank will probably be the same or it was deranked?
-
Go into GWMT and resubmit sitemap.xml files (with the URLs you want indexed) for recrawling and Google will digest the sitemaps again, instead of waiting for Googlebot to come around on their own, you are requesting it to come around, also include those new sitemap files in your robots.txt file.
-
In Google Webmaster Tools, go to Health -> Fetch As Google. Then add the previously blocked URL and click Fetch. Once you've done that, refresh the page and click "Submit to index". That should get Google indexing those pages again.
Getting external links to your pages also helps get pages crawled & indexed, so it may be worth submitting your pages to social bookmarking sites, or get other types of backlinks to your previously blocked pages if possible.
-
Since you fixed your robots.txt file you should be good to. It will probably take a few days for Google to recrawl your site and update the index with the URLs they are now allow to crawl.
Blocked URLs can still show up in SERPs if you haven't defined the no-index attribute in your section.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Was Google Analytics and Adsense Down Today?
For the last 4 hours of so we were registering zero users and Adsense reporting has not changed. We checked the site, and there were no problems. It seems for some reason there was no reporting. Just now it came back up and we are showing live traffic. Trying to figure out if this was a problem specific to us or if it is on Google's end. Thanks,
Reporting & Analytics | | akin670 -
Are RSS fees tracked by Google Analytics?
In both the new GA and former version, how are RSS feeds handled by Google Analytics. If not, is there an easy means by which to account for this traffic, in order to have a better picture of traffic. Thanks. Alan
Reporting & Analytics | | ahw0 -
Does GWT "Fetch as Google Bot" feature affect crawl rate?
Hello Mozians, I have noticed many people saying using GWT fetch as GoogleBot can affect your crawl rate in future, if used regularly. Though, i am not very sure if this is true or just another stale SEO myth. As currently GWT provides a limit of 500 URLs to fetch every month. I hope my doubts will be cleared by the Moz community experts. Thanks!
Reporting & Analytics | | pushkar630 -
PDFs and indexing
Hello and good morning. I work for a paint manufacturing company in the UK on their seo campaigns across a couple of websites, this is my question. as paint and chemicals require data and tech sheets by law, available to be downloadable for said product, should these be included in the sitemap, we auto generate our sitemaps which they include these files, with low priorities and never change in terms of name etc. they basically have a name of say 092847.pdf for example which cannot be changed, but from an seo view this doesn't mean a thing? so theres my question should they be included and would they carry any value?
Reporting & Analytics | | TeamacPaints0 -
Implications of Google discontinuing Website Optimizer
Hi Guys, As most of you probably know, Google is discontinuing Website Optimizer and introducing Experiments within Google Analytics. However, doesn't this mean that now, every site you want to run an experiment for has to be using Google Analytics? This is possibly one of the motivations for them making the change I guess? I also find it inconvenient that every 'experiment' now has to be based on improving a pre-defined goal in Google Analytics. This means that for a lot of situations we'll be creating goals just for the experiment and the clients actual goal conversions will appear quite inflated. I guess we'll just have to filter the new 'goals' out from the actual goals.
Reporting & Analytics | | David_ODonnell1 -
Strange 404 Error URL
Can anyone help determine how a URL like "www.mycompany.com/lago_www.bad-nsfw-content.com" would appear on the "not found" crawl error list in Google Webmaster Tools? The "www.bad-nsfw-content" site has nothing to do with our company and I don't how it would get associated with our site.
Reporting & Analytics | | pbhatt0 -
Google parameter
All Welcome! In the old interface, Google Analytics, you can find rare queries (organic traffic) in such a setting: ^ ([^] +) {5.50} [^] + $ (where 5.50 - spaces between words). As an option to use the new interface Google Analytics? This option does not work there?
Reporting & Analytics | | meteorr0