Help with Roger finding phantom links
-
It Monday and Roger has done another crawl and now I have a couple of issues:
- I have two pages showing 404->302 or 500 because these links do not exist. I have to fix the 500 but the 404 is trapped correctly.
http://www.oznappies.com/nappies.faq & http://www.oznappies.com/store/value-packs/\
The issue is when I do a site scan there is no anchor text that contains these links. So, what I would like to find out is where is Roger finding them. I cannot see any where in the Crawl Report that tells me where the origin of these links is.
- I also created a blog on Tumblr and now every tag and rss feed entry is producing a duplicate content error in the crawl stats. I cannot see anywhere in Tumblr to fix this issue.
Any Ideas?
-
Thanks again Ryan, you have been very helpful answering al lot of my questions.
-
Someone else asked the same question regarding tag pages yesterday. I would suggest asking a separate Q&A on that topic.
Tag pages & forum category pages are both often used as containers. They don't have any content except links to articles. I would ask for feedback as to the best practice. I suspect noindex, following those pages would be best, but I don't have the experience to feel comfortable offering that advice.
-
I have been looking at the data that Roger is reporting for the duplicate content and in ALL cases there is either a 301 or a NoIndex. So now I do not know why Roger is reporting them as a duplicate, robots should not see the second entry.
-
I did not think of looking at the csv report. I see it now thanks Ryan. There should be a soft 404 handler in place to process the bad urls, I will have to see why it is not working.
With tumblr, I was looking for an easy way to add a blog to the site.
The RSS is coming from tumblr as is all the content.
When we specify Tags in tumblr it creates urls e.g. mypage.com/article/tag1 mypage.com/article/tag2 mypage.com/article/tag3 which all contain the content of mypage.com/article with out a canonical to the original. It is a really strange non-seo friendly approach, and so I wondered if anyone had similar problems.
-
The crawl report offers a "referrer" field. That field offers where Roger found the offending link. In my experience that field has always been accurate.
When I try to access www.oznappies.com/faq I receive a 302 redirect and a 500 error. I would recommend adjusting non-existant pages to a soft 404 page. Still provide a 404 response to browsers, but offer users a friendly way to find information (i.e. links / search) and stay on your site.
A great example of a soft 404 page is http://www.orangecoat.com/a-404-page.html
For the Tumblr issue, I am not clear on the problem. Are you writing content and publishing on both the oznappies.com site and your tumblr site? Then this content is being published again on your site via a RSS import?
-
I removed the links and just left the text so these will cut and paste now. It confuses me where Roger found the links.
Thanks for running the Xenu scan. I have tried other site scanner and come up blank.
-
That second link is anchored to the wrong place.
Regardless I also cannot find the .faq page. I just ran Xenu over it to see what it could find, but no broken links showed up.
Afraid I don't use Tumblr either, so eh, pretty useless post. Sorry.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do you find popular blog topics (methodology) to rank in Google (organically)
Hi guys Do you guys have any advice and experience into how to find very popular blog topics? The purpose is to be able to rank them organically in Google. If you could share your methodology I would be really thankful as well if possible. Cheers John
Moz Pro | | igniterman750 -
What impact has external links reported as http503
Hello I have 72 external broken links who are reported with a http 503 status. When I ask the owner of that site, he confirmed he has removed it from GA by using this status. My question is. Does this have an impact for the quality of my site? The site is available, there is no delay when you open the link. Thank you for your support. br
Moz Pro | | kjerstibakke
Kjersti Bakke0 -
Inbound external links
Hi I have a question about inbound external links The good news is that I have 135 inbound links showing up in open site explorer, where my main competitor has 4 The bad news is that these are not the links that I have been building myself. The links are sometimes in a foreign language, rarely in an entertainment category, which my website is, and there seems no way to add a link on the page. I did buy some links when I first started out, but I was advised not to do it again But these are clearly working despite their apparent lack of value Then the question is where are my web directory listings, social bookmarks, social media content and web 2.0 articles, and the links I found my competitor had, that I was able to add my own website on, I have manually grafted for these links, have I been wasting my time with all of these?.
Moz Pro | | singingtelegramsuk0 -
Finding affiliate links with opensiteexplorer
Hey guys, I'm trying to develop lists of our competitors affiliates using opensiteexplorer using the following method When logged in to seomoz, go to http://www.opensiteexplorer.org/ Search for<rival company="" domain=""></rival> Select Show ALL links from ALL pages to PAGES ON THIS ROOT DOMAIN and SHOW LINKS UNGROUPED Click 'Filter' Click_ 'Download csv'_ Open the csv in excel and filter the Target URL column for any known part of the affiliate link (example ?a_aid) - this leaves you with a list of the affiliate links Filter out duplicates Is this the best process? I have a feeling I'm missing some as OSE truncates to 25 links per domain. It would be nice also if OSE could do the filtering for me. Any better ways to do this?
Moz Pro | | pilkster2 -
How do you check the outbound links of a site?
There are great tools like http://www.opensiteexplorer.org that will tell you all about the inbound links. What about the more basic and easier question: What outgoing links does this site have?
Moz Pro | | SkinLaboratory2 -
Does Google have a direct link with facebook and twitter?
Google monitor social media. What I'm wondering is do Google use the same tools we have on Facebook's API, Twitter's API etc to use in their SERPs Or do Facebook grant Google more detailed access to see who has liked links etc. I think it's quite an interesting point as surely I can push up my own count by repeatedly sharing my own links, which wouldn't be genuine. If Google had better access they could then determine what's been faked etc.
Moz Pro | | PhotoGazza0 -
Inbound Links Report Problem
While looking over my competitors inbound link report, they have a reported 5K inbound links. The first 20 listed inbound domains are things like UPS, and other sites that when I look at them, and the source file has no link to their domain at all??? So are they using some kind of Black Hat technique, or is SEOMoz not reporting properly? Help please.
Moz Pro | | WBConsulting0 -
Can I get a list of all links on a given domain?
Sorry, this is actually kind of a tripartite question: I was looking at the Competitive Link Analysis on one my clients' campaigns. Sometime between June and September their total links went up by about 120,000. We have no idea where those links came from (although the numbers would indicate that they're mostly internal). Question 1: In none of the other tools can I figure out how to list these links on a domain level. Is there a way to get a list of all links for our given domain? I've been playing around with the page-by-page and even that doesn't show me everything. For example, I'm looking at OSE for their homepage and it lists 45 links for a page that it claims has 151 total. Question 2: How did it pick those 45 to display out of the 151 possible? If these are only external links, why do half of them come from one of our subdomains? Also... Question 3: If our client hasn't made any major changes recently, why has the number of internal links gone up so dramatically? Thanks.
Moz Pro | | MackenzieFogelson1