Sitmap Page - HTML and XML
-
Hi there
I have a domain which has a sitemap in html for regular users and a sitemap in xml for the spiders.
I have a warning via seomoz saying that i have too many links on the html version.
What do i do here?
regards
Stef
-
Sorry for late reply guys. Great advice by both of you.
@ Alan, great display on how Page Rank flows. Great illustration which i totally could never explain to clients
-
220 links on a page is absolutely not too many on any level. Many of the highest ranked sites on the internet present more then 220 links.
The particular page is question is simply a sitemap, and the page is being offered to help users navigate the site. The VerizonWireless.com sitemap I shared has 370+ links on it.
The SEOmoz "warning" is a simple feature which will be set off on any internet page with 100+ links. The SEOmoz tool does not care how well those links are presented, whether they are footer links, whether they are on a content page, what the PA of the page is nor any other SEO factor. It is simply a >100 or not warning. As such, it offers very little value.
I am in the process of compiling a list of suggested features for the tool which will help improve it's usefulness. One of the feature recommendations I am proposing is to allow users to adjust the 100 count to any number they want. Each SEO can then choose to use the default 100 number, or use a number more suited to the particular site.
The link Alan shared is a nice explanation of PR flow. It is a nice page for learning PR, but with respect to this topic it over-complicates an otherwise very simple and straight-forward question. The simple point is, the more links on a page the less link juice will flow to each link.
The goals for any web page links should be as follows:
1. Ensure all links are useful for your site. For example, you probably want PR flowing to your most profitable product/service, and to your latest additions.
2. Ensure your links are actually used. Check analytics.
3. If a link is not used or not useful, remove it.
4. Along the lines above, your links should be presented in a very user-friendly manner. You don't want a page to look like a list of nothing but links as users will have a difficult time choosing what they want. An exception would be a sitemap.
With the above in mind, keep as many links as you see fit on the page. If it is 40, that is fine. If there are 250 links on the page, that is fine as well. When you start down a path of chasing numbers such as forcing your content into "500 words" or forcing your links into "100 maximum" you fall into a pit of SEO fallacies. You are not providing the best experience for your users nor SEO.
TL;DR - Provide your links in a manner which is visually appealing, non-spammy and helpful to users. Keep in mind your need to flow PR to important pages such as your money pages. Otherwise remove unnecessary links. Whatever that number of links is, so be it. Don't try to fit your links into a "I must be under 100" or any other number mindset.
-
too many according to google. make of it what you will, does not look like it is for any technial reason anyymore, but obviously there is a limit to how much of page they will crawl.
http://www.mattcutts.com/blog/how-many-links-per-page/You see how page rank flows, having a lot of links on your home page works to your advantage. Using numbers from Googles original algo,
Assuming every page starts with 1PR, a page passes %85 of its link juice, so if you have 100 links that’s 0.0085 each. To 100 internal pages, making them 1. 0085each , now they all pass back 85% that’s 0.857225 each, x 100 = 85.7225 back to your home page, now we do the sums all over again and again till they numbers stay static. Now this calculation relies on the internal pages having no other links, so you are unlikely to get figures as good as this, but you get the idea.
See link for better explanation.
http://www.webworkshop.net/pagerank.html check out calculator
Remember don’t stuff up your linking stuckture for the users just for the sake of page rank.I see it as like a golf swing after a lesson, if you try to do what you just learnt too much, you will get all stiff and un-natural, it’s better to swing naturally with what you have learnt in the back of your head.
-
Yes, ignore the warning.
It is possible to present 220 links in a neat, categorized manner. It is also possible to present 100 links as a jumble which is not user friendly.
You shared your presentation is similar to the example I shared which means it is user friendly so ignoring the warning is fine.
-
Nice, i really like that example that you gave. My one is similar and categorized too. Question still remains, do i ignore this warning for this page?
-
I have about 220 links
-
Wel how many do you have.
A quick way of checking is with IE, press F12, go to view menu, then link report
-
Your HTML sitemap is for users. It should present your links in such a manner as to be useful for users who are looking for a page on your site.
An example sitemap for a large site: http://www.verizonwireless.com/b2c/sitemap.jsp
It does not contain a link to every last page. It is more of a helpful directory. I would suggest you adjust your HTML sitemap in a similar manner. Treat is as a page of links for users.
-
So do you think that i should ignore this warning for the sitemap html page?
-
Well have a look if you can move a few out, it is good to link to as many pages as you can from the home page for the sake of PR flow. but not go over the limit, Some say the limit is 100, some say 150
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Which pages should I index or have in my XML sitemap?
Hi there, my website is ConcertHotels.com - a site which helps users find hotels close to concert venues. I have a hotel listing page for every concert venue on my site - about 12,000 of them I think (and the same for nearby restaurants). e.g. https://www.concerthotels.com/venue-hotels/madison-square-garden-hotels/304484 Each of these pages list the nearby hotels to that concert venue. Users clicking on the individual hotel are brought through to a hotel (product) page e.g. https://www.concerthotels.com/hotel/the-new-yorker-a-wyndham-hotel/136818 I made a decision years ago to noindex all of the /hotel/ pages since they don't have a huge amount of unique content and aren't the pages I'd like my users to land on . The primary pages on my site are the /venue-hotels/ listing pages. I have similar pages for nearby restaurants, so there are approximately 12,000 venue-restaurants pages, again, one listing page for each concert venue. However, while all of these pages are potentially money-earners, in reality, the vast majority of subsequent hotel bookings have come from a fraction of the 12,000 venues. I would say 2000 venues are key money earning pages, a further 6000 have generated income of a low level, and 4000 are yet to generate income. I have a few related questions: Although there is potential for any of these pages to generate revenue, should I be brutal and simply delete a venue if it hasn't generated revenue within a time period, and just accept that, while it "could" be useful, it hasn't proven to be and isn't worth the link equity. Or should I noindex these "poorly performing pages"? Should all 12,000 pages be listed in my XML sitemap? Or simply the ones that are generating revenue, or perhaps just the ones that have generated significant revenue in the past and have proved to be most important to my business? Thanks Mike
Technical SEO | | mjk260 -
3,511 Pages Indexed and 3,331 Pages Blocked by Robots
Morning, So I checked our site's index status on WMT, and I'm being told that Google is indexing 3,511 pages and the robots are blocking 3,331. This seems slightly odd as we're only disallowing 24 pages on the robots.txt file. In light of this, I have the following queries: Do these figures mean that Google is indexing 3,511 pages and blocking 3,331 other pages? Or does it mean that it's blocking 3,331 pages of the 3,511 indexed? As there are only 24 URLs being disallowed on robots.text, why are 3,331 pages being blocked? Will these be variations of the URLs we've submitted? Currently, we don't have a sitemap. I know, I know, it's pretty unforgivable but the old one didn't really work and the developers are working on the new one. Once submitted, will this help? I think I know the answer to this, but is there any way to ascertain which pages are being blocked? Thanks in advance! Lewis
Technical SEO | | PeaSoupDigital0 -
Crawl Test Report only shows home page and no inner site pages?
Hi, My site is [removed] When I first tried to set up a new campaign for the site, I received the error: Roger has detected a problem: We have detected that the root domain [removed] does not respond to web requests. Using this domain, we will be unable to crawl your site or present accurate SERP information. I then ran a Crawl Test per the FAQ. The SEOmoz crawl report only shows my home page URL and does not have any inner site pages. This is a Joomla site. What is the problem? Thanks! Dave
Technical SEO | | crave810 -
Page Content
Our site is a home to home moving listing portal. Consumers who wants to move his home fills a form so that moving companies can cote prices. We were generating listing page URL’s by using the title submitted by customer. Unfortunately we have understood by now that many customers have entered exactly same title for their listings which has caused us having hundreds of similar page title. We have corrected all the pages which had similar meta tag and duplicate page title tags. We have also inserted controls to our software to prevent generating duplicate page title tags or meta tags. But also the page content quality not very good because page content added by customer.(example: http://www.enakliyat.com.tr/detaylar/evden-eve--6001) What should I do. Please help me.
Technical SEO | | iskq0 -
Drupal duplicate pages
Anyone else encountered massive numbers of duplicate pages being reported on SEO Moz crawls for Drupal based sites? I assumed it was b/c there was no redirect on the print format pages, so I fixed that with a cannonical tag. But still seeing 2 or 3 duplicate pages reported for many pages. Any experience fixing this would be awesome to hear about. Thanks, Kevin
Technical SEO | | kevgrand0 -
Duplicate page titles
Hi, I have a Joomla 2.5 site and I use categoryblogs. So I have a page with "reviews". All the reviews are shown on this page and there are about 15 pages of it. In my SEOMoz crawl result I get 71 errors ! about "duplicate titles". How can I diminish this? I don't know how to show all the reviews in a proper way other than what I have accomplished with categoryblog. Patrick
Technical SEO | | paddydaddy0 -
Google News not indexing .index.html pages
Hi all, we've been asked by a blog to help them better indexing and ranking on Google News (with the site being already included in Google News with poor results) The blog had a chronicle URL duplication problem with each post existing with 3 different URLs: #1) www.domain.com/post.html (currently in noindex for editorial choices as showing all the comments) #2) www.domain.com/post/index.html (currently indexed showing only top comments) #3) www.domain.com/post/ (very same as #2) We've chosen URL #2 (/index.html) as canonical URL, and included a rel=canonical tag on URL #3 (/) linking to URL #2.
Technical SEO | | H-FARM
Also we've submitted yesterday a Google News sitemap including consistently the list of URLs #2 from the last 48h . The sitemap has been properly "digested" by Google and shows that all URLs have been sent and indexed. However if we use the site:domain.com command on Google News we see something completely different: Google News has indexed actually only some news and more specifically only the URLs #3 type (ending with the trailing slash instead of /index.html). Why ? What's wrong ? a) Does Google News bot have problems indexing URLs ending with .index.html ? While figuring out what's wrong we've found out that http://news.google.it/news/search?aq=f&pz=1&cf=all&ned=us&hl=en&q=inurl%3Aindex.html gives no results...it seems that Google News index overall does not include any URLs ending with /index.html b) Does Google News bot recognise rel=canonical tag ? c) Is it just a matter of time and then Google News will pick up the right URLs (/index.html) and/or shall we communicate Google News team any changes ? d) Any suggestions ? OR Shall we do the other way around. meaning make URL #3 the canonical one ? While Google News is showing these problems, Google Web search has actually well received the changes, so we don't know what to do. Thanks for your help, Matteo0 -
Duplicate Page Title
The crawl of my website http://www.aboutaburningfire.com revealed an error showing a duplicate page title. Can someone please explain to me how to fix this? I'm not sure what it means or how to fix it. | House Church Chicago, Organic Church, Illinois http://www.aboutaburningfire.com/ 1 Pending Pending House Church Chicago, Organic Church, Illinois http://www.aboutaburningfire.com/index.html |
Technical SEO | | severity0