Help with Roger finding phantom links
-
It Monday and Roger has done another crawl and now I have a couple of issues:
- I have two pages showing 404->302 or 500 because these links do not exist. I have to fix the 500 but the 404 is trapped correctly.
http://www.oznappies.com/nappies.faq & http://www.oznappies.com/store/value-packs/\
The issue is when I do a site scan there is no anchor text that contains these links. So, what I would like to find out is where is Roger finding them. I cannot see any where in the Crawl Report that tells me where the origin of these links is.
- I also created a blog on Tumblr and now every tag and rss feed entry is producing a duplicate content error in the crawl stats. I cannot see anywhere in Tumblr to fix this issue.
Any Ideas?
-
Thanks again Ryan, you have been very helpful answering al lot of my questions.
-
Someone else asked the same question regarding tag pages yesterday. I would suggest asking a separate Q&A on that topic.
Tag pages & forum category pages are both often used as containers. They don't have any content except links to articles. I would ask for feedback as to the best practice. I suspect noindex, following those pages would be best, but I don't have the experience to feel comfortable offering that advice.
-
I have been looking at the data that Roger is reporting for the duplicate content and in ALL cases there is either a 301 or a NoIndex. So now I do not know why Roger is reporting them as a duplicate, robots should not see the second entry.
-
I did not think of looking at the csv report. I see it now thanks Ryan. There should be a soft 404 handler in place to process the bad urls, I will have to see why it is not working.
With tumblr, I was looking for an easy way to add a blog to the site.
The RSS is coming from tumblr as is all the content.
When we specify Tags in tumblr it creates urls e.g. mypage.com/article/tag1 mypage.com/article/tag2 mypage.com/article/tag3 which all contain the content of mypage.com/article with out a canonical to the original. It is a really strange non-seo friendly approach, and so I wondered if anyone had similar problems.
-
The crawl report offers a "referrer" field. That field offers where Roger found the offending link. In my experience that field has always been accurate.
When I try to access www.oznappies.com/faq I receive a 302 redirect and a 500 error. I would recommend adjusting non-existant pages to a soft 404 page. Still provide a 404 response to browsers, but offer users a friendly way to find information (i.e. links / search) and stay on your site.
A great example of a soft 404 page is http://www.orangecoat.com/a-404-page.html
For the Tumblr issue, I am not clear on the problem. Are you writing content and publishing on both the oznappies.com site and your tumblr site? Then this content is being published again on your site via a RSS import?
-
I removed the links and just left the text so these will cut and paste now. It confuses me where Roger found the links.
Thanks for running the Xenu scan. I have tried other site scanner and come up blank.
-
That second link is anchored to the wrong place.
Regardless I also cannot find the .faq page. I just ran Xenu over it to see what it could find, but no broken links showed up.
Afraid I don't use Tumblr either, so eh, pretty useless post. Sorry.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How important are Internal Equity-Passing Links in rankings?
Hi,I ran opensite explorer of a customer site: www.enviosadomicilio.com and compared it with the most ranked competence sites. In almost all the factor enviosadomicilio.com has the best scores, except in Internal Equity-Passing Links. The competence has aprox 20K Equity links and another 2K meanwhile enviosadomicilio.com only have 296. Do you think that factor could be the reason that enviosadomicilio.com are in ranking 15th while the competence in 1st and 3rd?. How could we get Internal-Passing Links?, we don´t have a lot of products or categories in the site. Thanks a lot.
Moz Pro | | ramirez_salvador0 -
How bad are poor link profiles for a website?
A client has recently switched to us for their SEO and is hugely concerned about their previous company's link building efforts.They say that they have recently removed over 300 links but OSE still lists 160 links.So three questions: How long will it take for OSE to recognise if the links have been removed? How great could the effects of this be on the site's rankings? If we continue to struggle with getting the links removed would Google's disavow tool make much of a difference or are there negatives that sould be considered first?
Moz Pro | | SoundinTheory0 -
Find ClickBank Affiliates
Hey everyone, just a quick question. I am interested in using the Open Site Explorer tool to find ClickBank affiliates (advertisers not merchants) for current ClickBank products. Is there a way to do this? What I have been doing: 1.) Run an inbound link search for current products 2.) Manually go through all the back links and pull links that could be from affiliate sites (and delete those that look like links from directories). Thoughts? Thanks!
Moz Pro | | goproworkouts0 -
Number of available links limited?
OK, I've been making use of the free LinkScape API (on behalf of a client of mine) and trying to get links (and info on those links) to a specific domain/page/etc. NOTE : I've been using it without any issue in the past, however we are currently facing some weird issues. Let's take this simple query as an example : http://lsapi.seomoz.com/linkscape/links/wikipedia.org?SourceCols=4&TargetCols=4&Sort=page_authority&Scope=page_to_domain What this one supposedly does is to get links to "wikipedia.org", right? I'm reading : The Page_to_* scopes will by default return 25 links per source domain if no limit is specified, so you can see domain diversity. Due to space limitations in our API, a general link query for a given page will return at most 25 pages for every unique domain linking to that page. And I'm saying OK, that's fine. The thing is that (instead of the 1000 links I had been getting before), I'm now getting just 25 links. NOT per... "source domain"... but obviously per "target domain" (= wikipedia.org) - or am I missing something? (well, probably wikipedia suddenly has just about 25 links pointed to it... makes sense! 🙂 ) Please, let me know what's going on with the above, simply because getting just 25 links is close to worthless... Thanks a lot, in advance!
Moz Pro | | drkameleon0 -
SEOMoz Link Analysis Not Updating?
Hi there, I am wondering why the Link Analysis in SEOMoz takes so long to recognise new back-links. I have had the same figures showing for months when I know that I have added a lot more backlinks that are showing in my Google WebMaster account and I have also proven exist by searching directly in Google's search engine. How often does the Link Analysis section update itself? Are these figures worth following at all or are they useless? Thanks!
Moz Pro | | onlineexpression
Karl0 -
How can i find competitor title and descriptions in SEO moz PRO
Hi I can't seem to find competitor title and descriptions in SEO moz PRO. I am missing something?
Moz Pro | | smashseo0 -
Help with duplicate title tags?
I was looking in Google webmaster tools and it says I have 95 duplicate title tags for my site Noah's Dad. When I look through the list it appears the pages with duplicate title tags are some of my category pages, archive pages, and some author pages... Not sure if you guys can use some of the tools to see what is actually showing up duplicate or not, and if you need more info just let me know. But I wanted to see if this is something I should be concerned with? Should WMT also say 0 in duplicate content? It seems like when I started my blog I was told no to be conceded with this sort of stuff in gwmt. Anyways...I just wanted to see what you guys think. (By the way, is there any way to tell what this duplicate content is having (or has had) on my SERP results? Thanks.
Moz Pro | | NoahsDad0 -
Best way to count number of inbound links to a subfolder?
What's the best way to count and track the number of inbound links to a subfolder like http://www.domain.com/folder ? I've tried using open site explorer but I can't seem to get it to show just inbound links for things under a subfolder. i.e. links to domain.com/folder domain.com/folder/subfolder etc. Thanks!
Moz Pro | | twobitoperation0