Any idea why this is reporting a 404 in MozTools?
-
I did away with a vague category and 301 redirected the category url to the home page. However the link is reporting as a 404 in Moz Tools when it scans my site. Here's the link, and as you can see it redirects to the home page. Just curious if I did something wrong. Thanks.
-
Totally agree with Ryan, you should redirect https to http
Trivial maybe, but can see no reason not to do it as it takes no effort and makes your site even more accessible to users who mis-type, key your address directly into the url bar while on a https site etc
*This was in response to Ryans post which started "Perhaps i am being too picky..." I accidentally clicked the wrong reply button!!
-
Hi Rick,
While I know that Ryan decided to drop the https:// protocol problem, I just wanted to explain why it could be an issue for you.
It may not be a concern for a lot of personal sites, but for those where the site may be serving a strongly recognized brand it definitely should be a concern. That recognition could be coming both from your online visibility and/or from your involvement in offline communities or activities. Basically, if you have a reputation or following, people who know of your site will be much more likely to type your URL straight into a browser to go there.
I, for example, have come to know and love noahsdad.com through your involvement here in SEOmoz Q&A. I've visited the site, love your work and from time to time it crosses my mind to drop in and see what Noah has been up to lately. Since I know the site's domain, when that happens, I click inside the field at the top of my browser and replace everything after the www of the site that is open with noahsdad.com.
Now, in the event that the page I had open in my browser when I did that happened to be using the https:// protocol and I didn't realize that (which often happens), I would actually be asking my browser to go to https://www.noahsdad.com...and I think now you see why this could be an issue for you.
Hope that helps
Sha
Thumbs up for the catch too Ryan!
-
It could be that MozTools is looking at a cached version . You could try fetching it in Google Webmasters Tool to see how Google sees it ( which is what you should really worry about )
But as far as I can see you have 301 in place for that category to your home page : www.webconfs.com/http-header-check.php?submit=submit&url=http://noahsdad.com/mom-md
-
I would need to see the full record from the crawl report in order to respond. Perhaps you can upload the record to a web server and share the link?
-
No worries.
My real question is still why the redirect is showing up as a 404 when the site is crawled. I'd still be interested in figuring that out if you have any thoughts.
Thanks.
-
I am going to let this issue drop since it is such a small item for a personal site. A few last thoughts:
**Wordpress automatically redirects to the non http version. **
No, it does not. On the SEOmoz site, you are seeing the proper redirect. For example, if you take this Q&A post and prefix it with https:// you will wind up on this exact Q&A post in the http:// protocol. If you look at the Mozbar, you will see the redirect.
On your site, you are taken to another page from your hosting company.
**I dont know many personal sites that pay for a https cert **
There is no need for you to purchase a SSL certificate. That is not what I was suggesting.
I apologize for bringing this trivial matter up. Please disregard.
-
Ryan,
Thanks for your feedback. Wordpress automatically redirects to the non http version.
Regarding https sites I dont know many personal sites that pay for a https cert (or even why there would be a need.) Also I don't link to anything with https. It doesn't seem really seem necessary (unless I"m missing something.)
Heck, I just tried to go to https://seomoz.com and their site doesn't even go there.
-
Perhaps I am being too picky, especially for a non-business site.
I would share that even though your site is noahsdad.com, you took the effort to redirect www.noahsdad.com to noahsdad.com, right? Why did you take that extra step?
Whatever the response, the same concept would apply to the redirect from https protocol to http. The issue may never come up, but then again it only requires minimal effort to close this gap.
-
Yah, but my site is http://noahsdad.com/ - should I expect https to also work?
-
Try going to the following address: https://noahsdad.com
-
Yelp, the crawl just happened today.
Also you'll have to excuse my ignorance but I'm not sure what you mean by the last half of your comment. "my site does not handle https protocol well?"
Can you explain that to me. (I'm still learning.)
Thanks for taking the time to help by the way.
-
Are you certain you crawled the site after the redirect was in place?
If so, can you share the full record?
By the way, while looking at this is issue I noticed your site does not handle https protocol well. Try using it on your home page and see what happens. If your site does not use https, I would suggest redirecting all https requests to their http equivalent.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google Search Console Showing 404 errors for product pages not in sitemap?
We have some products with url changes over the past several months. Google is showing these as having 404 errors even though they are not in sitemap (sitemap shows the correct NEW url). Is this expected? Will these errors eventually go away/stop being monitored by Google?
Technical SEO | | woshea0 -
404 Errors for Form Generated Pages - No index, no follow or 301 redirect
Hi there I wonder if someone can help me out and provide the best solution for a problem with form generated pages. I have blocked the search results pages from being indexed by using the 'no index' tag, and I wondered if I should take this approach for the following pages. I have seen a huge increase in 404 errors since the new site structure and forms being filled in. This is because every time a form is filled in, this generates a new page, which only Google Search Console is reporting as a 404. Whilst some 404's can be explained and resolved, I wondered what is best to prevent Google from crawling these pages, like this: mydomain.com/webapp/wcs/stores/servlet/TopCategoriesDisplay?langId=-1&storeId=90&catalogId=1008&homePage=Y Implement 301 redirect using rules, which will mean that all these pages will redirect to the homepage. Whilst in theory this will protect any linked to pages, it does not resolve this issue of why GSC is recording as 404's in the first place. Also could come across to Google as 100,000+ redirected links, which might look spammy. Place No index tag on these pages too, so they will not get picked up, in the same way the search result pages are not being indexed. Block in robots - this will prevent any 'result' pages being crawled, which will improve the crawl time currently being taken up. However, I'm not entirely sure if the block will be possible? I would need to block anything after the domain/webapp/wcs/stores/servlet/TopCategoriesDisplay?. Hopefully this is possible? The no index tag will take time to set up, as needs to be scheduled in with development team, but the robots.txt will be an quicker fix as this can be done in GSC. I really appreciate any feedback on this one. Many thanks
Technical SEO | | Ric_McHale0 -
Why is Google Webmaster Tools showing 404 Page Not Found Errors for web pages that don't have anything to do with my site?
I am currently working on a small site with approx 50 web pages. In the crawl error section in WMT Google has highlighted over 10,000 page not found errors for pages that have nothing to do with my site. Anyone come across this before?
Technical SEO | | Pete40 -
GWT False Reporting or GoogleBot has weird crawling ability?
Hi I hope someone can help me. I have launched a new website and trying hard to make everything perfect. I have been using Google Webmaster Tools (GWT) to ensure everything is as it should be but the crawl errors being reported do not match my site. I mark them as fixed and then check again the next day and it reports the same or similar errors again the next day. Example: http://www.mydomain.com/category/article/ (this would be a correct structure for the site). GWT reports: http://www.mydomain.com/category/article/category/article/ 404 (It does not exist, never has and never will) I have been to the pages listed to be linking to this page and it does not have the links in this manner. I have checked the page source code and all links from the given pages are correct structure and it is impossible to replicate this type of crawl. This happens accross most of the site, I have a few hundred pages all ending in a trailing slash and most pages of the site are reported in this manner making it look like I have close to 1000, 404 errors when I am not able to replicate this crawl using many different methods. The site is using a htacess file with redirects and a rewrite condition. Rewrite Condition: Need to redirect when no trailing slash RewriteCond %{REQUEST_FILENAME} !-f
Technical SEO | | baldnut
RewriteCond %{REQUEST_FILENAME} !.(html|shtml)$
RewriteCond %{REQUEST_URI} !(.)/$
RewriteRule ^(.)$ /$1/ [L,R=301] The above condition forces the trailing slash on folders. Then we are using redirects in this manner: Redirect 301 /article.html http://www.domain.com/article/ In addition to the above we had a development site whilst I was building the new site which was http://dev.slimandsave.co.uk now this had been spidered without my knowledge until it was too late. So when I put the site live I left the development domain in place (http://dev.domain.com) and redirected it like so: <ifmodule mod_rewrite.c="">RewriteEngine on
RewriteRule ^ - [E=protossl]
RewriteCond %{HTTPS} on
RewriteRule ^ - [E=protossl:s] RewriteRule ^ http%{ENV:protossl}://www.domain.com%{REQUEST_URI} [L,R=301]</ifmodule> Is there anything that I have done that would cause this type of redirect 'loop' ? Any help greatly appreciated.\0 -
Site was infected with spam webmaster tools still reporting it
I have recently been working with a site that was hacked. It suffered from a pharma injection into Joomla. The site has been cleaned for several months, but WMT is still reporting "pharmacy" as occuring 421 times. The url it gives reports a 500 error. I also removed it in Google. Can this still be hurting the site? How can I clean this up?
Technical SEO | | smcmark0 -
Page not Accesible for crawler in on-page report
Hi All, We started using SEOMoz this week and ran into an issue regarding the crawler access in the on-page report module. The attached screen shot shows that the HTTP status is 200 but SEOMoz still says that the page is not accessible for crawlers. What could this be? Page in question
Technical SEO | | TiasNimbas
http://www.tiasnimbas.edu/Executive_MBA/pgeId=307 Regards, Coen SEOMoz.png0 -
Any good ideas on how to quickly rank a new site for a new product laucnh?
Hi, I'm thinking of examples like the wii u - when it was announced all wii u sites were new. what ways could one get an edge quickly when competing against other sites without a great deal of authority on this 'new' keyword? obvious ones are social bookmarking, what else could easily be used to quickly (<24hrs) at least help a bit in non competitive kws?
Technical SEO | | dmn020