Best free tool to check internal broken links
-
Question says it all I guess. What would your recommend as the best free tool to check internal broken links?
-
[Spammy comment removed by forum moderator.]
-
How awesome is that Screaming Frog tool?
-
Hey Kevin,
If you download the CSV report, we should list the referring page in the CSV. I know it'd be better to have it on the on-screen report, and it is on our wish list, but you should be able to see the info in an export.
-
Hi Kevin
Thanks - yes, although I am a Moz Associate, out of full transparency I agree with that you are saying. It is not a tool I specifically use with Moz, and do prefer Screaming Frog for the exact functions you are referring to. Screaming Frog will tell you what pages the bad links are on.
You can definitely pass this on to the product team as a feature request on this page. Personally I love tools like their analytics, keyword difficulty, rank tracker, open site explorer, but I agree from this aspect the crawler is not the strongest, and I would suggest supplementing with Screaming Frog (as I do).
-Dan
-
Dan, although I love me some MOZ the crawl diagnostics kinda suck... all they do is report errors but they don't give any insight on where the bad link is originated... it merely shows the page which is 404ed, which is a BIG FAT WASTE OF TIME.
You guys should know that we need to know WHICH PAGE THE BAD LINK IS ON, and furthermore WHAT LINE OF CODE HAS THE BAD LINK. Who cares about the broken page!?!?!?!?!?
-
Yes when you set up a campaign with Moz Analytics, Moz will crawl your website and return a whole report of suggestions. It will start a sample crawl which gets returned to you pretty quickly, and a full crawl a little bit thereafter. It will report things to you like 400 errors etc.
You can also use the Moz Crawl Test (PRO only as well) which will return to you the HTTP status code of each URL crawled.
Optionally, you can also use Screaming Frog SEO Spider (free up to 500 pages to crawl, paid for bigger sites).
-Dan
-
Does anyone have any clue on how I can check my internal broken links using MOZ tools?
-
Hello,
If you have a full Sitemap, transform it to a .txt by putting only URL (Excel can do that easily) try Screaming Frog SEO on list mode, which is free regardless of the number of links.
Even if you search a free tool, I strongly recommand to invest 99£/year on this tool, it's worth it.
-
Ok, first, does the url with /errors/error_404 (without the query string) exist?
Also - are you using something to create your 404 pages for tracking purposes?
The query string on the url would appear to be stating the referring url anyway - hence ?q=-i3 basically equates to the page /i3
At a guess - it could be there is a script running to create logs on specific 404's, or create a new log each time one occurs - if this folder is visable to the crawlers, it would get spidered and subsequently the problem would arise.
-
Yeah I tried that but the URL in the URL coloumn shows for example:
/errors/error_404?q=-i3
and the Referrer shows:
/-i3
Neither page exists and so I do not know where the page which contains this broken link is...
-
It does, when you get the report, filter column D marked "4XX (Client Error)" and you will see your 404's there - further along in the report you will also be able to see referring url which will show you which page is linking to it.
-
Checked out the tool I just linked to on here and it doesn't include broken link data
-
Is the the custom crawl tool on here you are referring too? http://pro.seomoz.org/tools/crawl-test
-
Custom crawl right here on SEOmoz, or you could use Xenu
-
Hi Dan,
Thanks for that tool. It works great and especially like that it works in Chrome.
Do you however know of a tool which would do this site wide rather than per page?
-
This tool is a great extension for Chrome
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
how to find broken outbound links ?
hey guy, can anyone help me in finding broken outbound link on my website by using moz ? does Moz has this function ?
Technical SEO | | rogerdam0 -
Brushing up on my SEO skills - how do I check my website to see if Javascript is blocking search engines from crawling the links within a javascript-enabled drop down menu?
I set my user agent in my Chrome browser to Googlebot and I disable javascript within my Chrome settings, but then what?
Technical SEO | | MagnitudeSEO0 -
Check my website loading time
Kindly check my website loading time for the home page and deep pages. Do I need to make it fast or it is Okey? Website - brandstenmedia.com.au
Technical SEO | | Green.landon0 -
Will multiple internal links with the same anchor text hurt a site's ranking?
Hello, I just watched this video from the Google Webmasters channel at YouTube: http://www.youtube.com/watch?v=6ybpXU0ckKQ My question: If a site is built up on subdomains, will linking the different subdomains with exact anchor text hurt the site's ranking? Thanks
Technical SEO | | arnoldwender0 -
Sitemap international websites
Hey Mozzers,Here is the case that I would appreciate your reply for: I will build a sitemap for .com domain which has multiple domains for other countries (like Italy, Germany etc.). The question is can I put the hreflang annotations in sitemap1 only and have a sitemap 2 with all URLs for EN/default version of the website .COM. Then put 2 sitemaps in a sitemap index. The issue is that there are pages that go away quickly (like in 1-2 days), they are localised, but I prefer not to give annotations for them, I want to keep clear lang annotations in sitemap 1. In this way, I will replace only sitemap 2 and keep sitemap 1 intact. Would it work? Or I better put everything in one sitemap?The second question is whether you recommend to do the same exercise for all subdomains and other domains? I have read much on the topic, but not sure whether it worth the effort.The third question is if I have www.example.it and it.example.com, should I include both in my sitemap with hreflang annotations (the sitemap on www.example.com) and put there it for subdomain and it-it for the .it domain (to specify lang and lang + country).Thanks a lot for your time and have a great day,Ani
Technical SEO | | SBTech0 -
Internal Links
In OSE, it is reporting that i don't have any internal links to my homepage. In the header on every page is my logo in the top left hand corner which links back to my homepage. Shouldn't this mean then that every page should link to the home page? Similarly, internal pages which link from my main nav aren't showing up as having any internal links in OSE. Any ideas?
Technical SEO | | Santaur0 -
Is this seen as a Link Exchange
If i give a self serve banner ad to someone on my blog or a image with a link and they give me a text link ad is that in googles eyes a link exchange or a one way link.
Technical SEO | | DavidKonigsberg0 -
Which version of pages should I build links to?
I'm working on the site www.qualityauditor.co.uk which is built in Moonfruit. Moonfruit renders pages in Flash. Not ideal, I know, but it also automatically produces an HTML version of every page for those without Flash, Javascript and search engines. This HTML version is fairly well optimised for search engines, but sits on different URLs. For example, the page you're likely to see if browsing the site is at http://www.qualityauditor.co.uk/#/iso-9001-lead-auditor-course/4528742734 However, if you turn Javascript off you can see the HTML version of the page here <cite>http://www.qualityauditor.co.uk/page/4528742734</cite> Mostly, it's the last version of the URL which appears in the Google search results for a relevant query. But not always. Plus, in Google Webmaster Tools fetching as Googlebot only shows page content for the first version of the URL. For the second version it returns HTTP status code and a 302 redirect to the first version. I have two questions, really: Will these two versions of the page cause my duplicate content issues? I suspect not as the first version renders only in Flash. But will Google think the 302 redirect for people is cloaking? Which version of the URL should I be pointing new links to (bearing in mind the 302 redirect which doesn't pass link juice). The URL's which I see in my browser and which Google likes the look at when I 'fetch as Googlebot'. Or those Google shows in the search results? Thanks folks, much appreciated! Eamon
Technical SEO | | driftnetmedia0