Which pages should I index or have in my XML sitemap?
-
Hi there,
my website is ConcertHotels.com - a site which helps users find hotels close to concert venues. I have a hotel listing page for every concert venue on my site - about 12,000 of them I think (and the same for nearby restaurants).
e.g.
https://www.concerthotels.com/venue-hotels/madison-square-garden-hotels/304484
Each of these pages list the nearby hotels to that concert venue. Users clicking on the individual hotel are brought through to a hotel (product) page e.g.
https://www.concerthotels.com/hotel/the-new-yorker-a-wyndham-hotel/136818
I made a decision years ago to noindex all of the /hotel/ pages since they don't have a huge amount of unique content and aren't the pages I'd like my users to land on . The primary pages on my site are the /venue-hotels/ listing pages.
I have similar pages for nearby restaurants, so there are approximately 12,000 venue-restaurants pages, again, one listing page for each concert venue.
However, while all of these pages are potentially money-earners, in reality, the vast majority of subsequent hotel bookings have come from a fraction of the 12,000 venues. I would say 2000 venues are key money earning pages, a further 6000 have generated income of a low level, and 4000 are yet to generate income.
I have a few related questions:
-
Although there is potential for any of these pages to generate revenue, should I be brutal and simply delete a venue if it hasn't generated revenue within a time period, and just accept that, while it "could" be useful, it hasn't proven to be and isn't worth the link equity. Or should I noindex these "poorly performing pages"?
-
Should all 12,000 pages be listed in my XML sitemap? Or simply the ones that are generating revenue, or perhaps just the ones that have generated significant revenue in the past and have proved to be most important to my business?
Thanks
Mike
-
-
Hi Chris,
thank you very much for your help and suggestions, it is much appreciated. I'll de-noindex a handful of my biggest artist pages and see if they attract much interest from users.
As for the /venues/ pages, these have been fairly neglected to date, so perhaps I need to really focus my attention on them, as you say, and bring in some cross referencing.
I have also wondered whether allowing companies to create pages dedicated to their events would be a good route to take - it could be done with ease, so perhaps I should investigate further.
Again, thanks very much, and hopefully I can report back with good news at some point.
Best wishes
Mike
-
I think they should be indexed, but keyword research should shed light on this topic for you. It will let you know if your audience is searching for those things and in what numbers. Even as they are, though, they might make sufficient landing pages for google. You could de-noindex a group of those pages at a time, starting with the ones most likely to be popular and see how google treats them. I think I'd go that route rather than release them into the wild all at once.
To me, the pages with the most interesting potential are the /venues/ pages like /venues/md-concert-venues/a, for example. I think the potential lies in populating them with venue grouping, upcoming artists grouping, and state. How hard would it be to populate an area above the black line with all/some of the upcoming artists playing near the hotels that show on that page. That 3-way cross referencing would make those pages fairly unique on the web and unique on your site and would give google a number of good reasons to send traffic there. They'd probably be good pages to publish advertising on, too.
Also wondering if there is a thing such as "licensing" dedicated pages out to companies/hotels that are putting on non-musical events like conferences, etc, so they can link to a kind of pre-fab hotels-close-by page up for their attendees?
-
Thanks Chris,
appreciate your comments. Google in indexing a high percentage of the key pages, and does not have any noindex pages indexed. Pages are loading at a decent speed. And only indexed pages are in the sitemap. So perhaps the non-performing pages are not something I should be particularly concerned about, especially they don't necessarily take up much of my time. I guess if I start to run into issues with overall site speed then perhaps then is is the time to consider whether they should continue to be listed. So perhaps, you're right, it's more of a business decision, rather than an SEO one.
I have a further question if you don't mind, which is related but I think is an SEO one. I have a large number of /artist/ pages - these are pages that list which venues a particular artist is performing at, and allows a user to then check hotel availability for the specific venue and date they will be attending. At the minute the pages are fairly light on content - they just list venues and dates, although I'm planning to start introducing more content in the near future. An example page can be seen here:
https://www.concerthotels.com/artist/hotels-near-guns-n-roses-events/1227
At the minute, I've noindexed every artist page on the site, because I was worried Google would see them as thin pages. But I actually think they are potentially very useful to users, and a powerful landing page for quickly taking a user to the correct venue page with the correct dates for the concert. I also think that not all users will search for "Hotels near Metlife Stadium" - they might instead search for "Hotels for Guns n roses in NJ..." etc etc. so perhaps I can pick up some long tail searches with these additional landing pages.
The question is, should I index these pages?
If the answer to that question is yes..... obviously, artists/bands do a tour and then generally disappear into a recording studio for a year or two - as a result, there will be many /artist/ pages that, for a while, have lots of useful event dates/venues listed, but at the end of the tour, the pages will simply be empty, and no longer useful, at least until the next tour. Would you recommend that such pages are indexed when there are events, but when no future events are listed, I set them to noindex?
Many thanks
Mike
-
Mike,
I'm wondering...is that an SEO question? It sounds like a business decision to me. From what you've said, I don't see any reason for Google to ding you on anything. My only questions would be--Is google indexing all the pages you want it to and does not have your noindex pages indexed? Any bad links coming in? Pages are loading at a decent speed? Oh, and I don't see a reason to have your noindex pages in the the sitemap.
Other than that, if those non-performing page are taking up time that you could be spending on more productive pages or on exploring more productive opportunities, then, again, it's time to put on your CEO cap.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
.xml sitemap showing in SERP
Our sitemap is showing in Google's SERP. While it's only for very specific queries that don't seem to have much value (it's a healthcare website and when a doctor who isn't with us is search with the brand name so 'John Smith Brand,' it shows if there's a first or last name that matches the query), is there a way to not make the sitemap indexed so it's not showing in the SERP. I've seen the "x-robots-tag: noindex" as a possible option, but before taking any action wanted to see if this was still true and if it would work.
Technical SEO | | Kyleroe950 -
Japanese URL-structured sitemap (pages) not being indexed by Bing Webmaster Tools
Hello everyone, I am facing an issue with the sitemap submission feature in Bing Webmaster Tools for a Japanese language subdirectory domain project. Just to outline the key points: The website is based on a subdirectory URL ( example.com/ja/ ) The Japanese URLs (when pages are published in WordPress) are not being encoded. They are entered in pure Kanji. Google Webmaster Tools, for instance, has no issues reading and indexing the page's URLs in its sitemap submission area (all pages are being indexed). When it comes to Bing Webmaster Tools it's a different story, though. Basically, after the sitemap has been submitted ( example.com/ja/sitemap.xml ), it does report an error that it failed to download this part of the sitemap: "page-sitemap.xml" (basically the sitemap featuring all the sites pages). That means that no URLs have been submitted to Bing either. My apprehension is that Bing Webmaster Tools does not understand the Japanese URLs (or the Kanji for that matter). Therefore, I generally wonder what the correct way is to go on about this. When viewing the sitemap ( example.com/ja/page-sitemap.xml ) in a web browser, though, the Japanese URL's characters are already displayed as encoded. I am not sure if submitting the Kanji style URLs separately is a solution. In Bing Webmaster Tools this can only be done on the root domain level ( example.com ). However, surely there must be a way to make Bing's sitemap submission understand Japanese style sitemaps? Many thanks everyone for any advice!
Technical SEO | | Hermski0 -
Over 500 thin URLs indexed from dynamically created pages (for lightboxes)
I have a client who has a resources section. This section is primarily devoted to definitions of terms in the industry. These definitions appear in colored boxes that, when you click on them, turn into a lightbox with their own unique URL. Example URL: /resources/?resource=dlna The information for these lightboxes is pulled from a standard page: /resources/dlna. Both are indexed, resulting in over 500 indexed pages that are either a simple lightbox or a full page with very minimal content. My question is this: Should they be de-indexed? Another option I'm knocking around is working with the client to create Skyscraper pages, but this is obviously a massive undertaking given how many they have. Would appreciate your thoughts. Thanks.
Technical SEO | | Alces0 -
Page missing from Google index
Hi all, One of our most important pages seems to be missing from the Google index. A number of our collections pages (e.g., http://perfectlinens.com/collections/size-king) are thin, so we've included a canonical reference in all of them to the main collection page (http://perfectlinens.com/collections/all). However, I don't see the main collection page in any Google search result. When I search using "info:http://perfectlinens.com/collections/all", the page displayed is our homepage. Why is this happening? The main collection page has a rel=canonical reference to itself (auto-generated by Shopify so I can't control that). Thanks! WUKeBVB
Technical SEO | | leo920 -
3,511 Pages Indexed and 3,331 Pages Blocked by Robots
Morning, So I checked our site's index status on WMT, and I'm being told that Google is indexing 3,511 pages and the robots are blocking 3,331. This seems slightly odd as we're only disallowing 24 pages on the robots.txt file. In light of this, I have the following queries: Do these figures mean that Google is indexing 3,511 pages and blocking 3,331 other pages? Or does it mean that it's blocking 3,331 pages of the 3,511 indexed? As there are only 24 URLs being disallowed on robots.text, why are 3,331 pages being blocked? Will these be variations of the URLs we've submitted? Currently, we don't have a sitemap. I know, I know, it's pretty unforgivable but the old one didn't really work and the developers are working on the new one. Once submitted, will this help? I think I know the answer to this, but is there any way to ascertain which pages are being blocked? Thanks in advance! Lewis
Technical SEO | | PeaSoupDigital0 -
Post Site Migration - thousands of indexed pages, 4 months after
Hi all, Believe me. I think I've already tried and googled for every possible question that I have. This one is very frustrating – I have the following old domain – fancydiamonds dot net. We built a new site – Leibish dot com and done everything by the book: Individual 301 redirects for all the pages. Change of address via the GWT. Trying to maintain and improve the old optimization and hierarchy. 4 months after the site migration – we still have to gain back more than 50% of our original organic traffic (17,000 vs. 35,500-50,000 The thing that strikes me the most that you can still find 2400 indexed pages on Google (they all have 301 redirects). And more than this – if you'll search for the old domain name on Google – fancydiamonds dot net you'll find the old domain! Something is not right here, but I have no explanation why these pages still exist. Any help will be highly appreciated. Thanks!
Technical SEO | | skifr0 -
50,000 pages or a page with parameters
I have a site with about 12k pages on a topic... each of these pages could use another several pages to go into deeper detail about the topic. So, I am wondering, for SEO purposes would it be better to have something like 50,000 new pages for each sub topic or have one page that I would pass parameters to and the page would be built on the fly in code behind. The drawback to the one page with parameters is that the URL would be static but the effort to implement would be minimal. I am also not sure how google would index a single page with parameters. The drawback to the 50k pages model is the dev effort and possibly committed some faux pas by unleashing so many links to my internal pages. I might also have to mix aspx with html because my project can't be that large. Anyone here ever have this sort of choice to make? Is there a third way I am not considering?
Technical SEO | | Banknotes0 -
Page MozRank and MozTrust 0 for Home Page, Makes No Sense?
Hey Mozzers! I'm a bit confused by a site that is showing a 0 for home page MozRank and MozTrust, while its subdomain and root domain metrics look decent (relatively). I am posting images of the page metrics and subdomain metrics to show the disparity: http://i.imgur.com/3i0jq.png http://i.imgur.com/ydfme.png Is it normal to see this type of disparity? The home page has very little inbound links, but the big goose egg has me wondering if there is something else going on. Has anyone else experienced this? Or, does anyone have speculation as to why a home page would have a 0 MozRank while the subdomain metrics look much better? Thanks!
Technical SEO | | ClarityVentures0