Large number of thin content pages indexed, affect overall site performance?
-
Hello Community,
Question on negative impact of many virtually identical calendar pages indexed.
We have a site that is a b2b software product. There are about 150 product-related pages, and another 1,200 or so short articles on industry related topics. In addition, we recently (~4 months ago) had Google index a large number of calendar pages used for webinar schedules. This boosted the indexed pages number shown in Webmaster tools to about 54,000.
Since then, we "no-followed" the links on the calendar pages that allow you to view future months, and added "no-index" meta tags to all future month pages (beyond 6 months out). Our number of pages indexed value seems to be dropping, and is now down to 26,000.
When you look at Google's report showing pages appearing in response to search queries, a more normal 890 pages appear. Very few calendar pages show up in this report.
So, the question that has been raised is: Does a large number of pages in a search index with very thin content (basically blank calendar months) hurt the overall site? One person at the company said that because Panda/Penguin targeted thin-content sites that these pages would cause the performance of this site to drop as well.
Thanks for your feedback.
Chris
-
Unless a page can give value to a searcher (not just an existing customer) it shouldn't be in Google's search index.
Sometimes I like to go back to the basics. Remember that search engines exist to help people find information that they WANT to find. Realistically, people are not going to want to find every page on your websites in SERPS.I suggest you ask yourself this question; does this page offer information that someone would actually want to search for, and make your decision accordingly.
p.s. Having said all of that, I'll answer your question. The answer is yes, having thin pages on your site can hurt your domain. If your pages offer value to searchers, I suggest you improve them instead of remove them, but if they don't offer value to searchers don't waste your time, and just no-index them.
-
So, the question that has been raised is: Does a large number of pages in a search index with very thin content (basically blank calendar months) hurt the overall site?
Yes.
We had a site with some image content pages that had not a lot of text. They ranked great for years. Then, BAM, rankings across the site dropped on a Panda update.
We added noindex/follow to these pages, redirected some that were obsolete and our rankings came back with the next update.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
About porn sites and ranking
Hello, I'm thinking to extend my website into porn. At the moment there is no pornography on it, although we do talk about sex related topics and products (from dating to tutorials, to toys etc.) Would it be dangerous to keep the porn section on the same domain as the rest? Would this negatively affect my non-porn content as Googlebot would "flag" my website as being pornographic (although only a few pages would be)? Or simply Googlebot would leave the current non-porn pages ranking as they are now, just fine, and plus it would rank the porn pages if they "deserve" to? I hope my question is clear. I don't want to create a subdomain.
Algorithm Updates | | fabx0 -
Google & Site Architecture
Hi I've been reading the following article about Google's quality signals here: https://searchenginewatch.com/2016/10/10/guide-to-google-ranking-signals-part-6-trust-authority-and-expertise/?utm_source=Search+Engine+Watch&utm_campaign=464594db7c-11_10_2016_NL&utm_medium=email&utm_term=0_e118661359-464594db7c-17828341 They mention - 3) All your categories should be accessible from the main menu. All your web pages should be labelled with the relevant categories. Is this every category? We have some say 3 levels deep, and they aren't all in the menu. I'd like them to be, so would be good to make a case for it. Thank you
Algorithm Updates | | BeckyKey1 -
Home Page not ranking?
Hey guys, I'm working on a relatively new client and have noticed that all the initial rankings that we have point towards a sub page rather than the home page. Only a branded search appears to bring up the home page which seems really strange. Any ideas? Home page; www.geraldmiller.com Sub page (ranking); www.geraldmiller.com/dwi-dui-defense
Algorithm Updates | | Webrevolve0 -
Duplicate content on a sub domain
I have two domains www.hairremoval.com and a sub domain www.us.hairromoval.com both sites have virtual the same content apart from around 8 pages and the sub domain is more focused to US customers so the spelling are different, it is also hosted in the states. Would this be classed as duplicate content ? (The url’s are made up for the question but the format is correct)
Algorithm Updates | | Nettitude0 -
Links from high Domain authority sites
I have a relatively uncompetitive niche ranking around number 6 for my keywords. Would getting a few links from some Moz DA 80-90 and DA 90-100 sites help my rankings a lot? Some of the pages linking to me from these sites might be deep in the site pretty far away from the home page with pagerank of "unranked" or a grayed out bar and these pages linking to me might not have many links at all other than from the internal links of the site itself and would have a Moz PA of 10 or 20. Would these pass much pagerank or authority to my site or would they not be worth going after? These links to my site would be in context on a blog. Thanks mozzers!
Algorithm Updates | | Ron100 -
Why is a sub page ranking over home page?
Hey guys! I was wondering whether any of you Mozzers out there could shed some light on this query for me. Currently, one of our clients is ranking (on the second page, at least) for one of their target keywords. However, it's not the home page that is ranking - it is a sub page. I guess you could say both are targeted to rank for the keyword in question but the home page has a considerable more PA (+10) and has a lot more incoming links so it's a little bit baffling as to why the sub page has been given an advantage. Does anyone know why this may be? Also, on a secondary note, should I continue to build links to the home page or target this particular sub page to have a better chance of ranking higher for the keyword? Any advice on this welcome! Cheers!
Algorithm Updates | | Webrevolve0 -
Sudden drop in rankings and indexed pages!
Over the past few days I have noticed some apparent major changes. Before I explain, let me say this: Checking my analytics and WMT: There is an increase in traffic (even via google organic) There is no drop in impressions or clicks There is no drop in indexed pages in GWT Having said that; When I check my indexed pages using site:www.mywebsite.com, I see only 30 results as opposed to the 120K that I was seeing before (it was steadily climbing). The indexed pages have increase 3 fold in the past year, because of the increase in pages, updates, and products on the site. I see a sudden drop in rankings for major keywords that had been steadily rising. For example, I had some major keywords that were on page 7-8, not they are on page 20+ or not at all. Also, the page that used to show in the rankings has changed. I have only done white-hat guest blogging in the past year for link building, on a small scale (maybe 20-30 links in a year). They only other change recently, is that we are: Posting products on Houzz and Pinterest daily adding our site to all local directories (white pages, Yelp, citysearch, etc.) My site got hit by Penguin more than a year ago, but we have done everything right since, and our traffic via organic results has more than doubled since the Penguin release. What the hell is going on? Should I be concerned?
Algorithm Updates | | inhouseseo0 -
Stop google indexing CDN pages
Just when I thought I'd seen it all, google hits me with another nasty surprise! I have a CDN to deliver images, js and css to visitors around the world. I have no links to static HTML pages on the site, as far as I can tell, but someone else may have - perhaps a scraper site? Google has decided the static pages they were able to access through the CDN have more value than my real pages, and they seem to be slowly replacing my pages in the index with the static pages. Anyone got an idea on how to stop that? Obviously, I have no access to the static area, because it is in the CDN, so there is no way I know of that I can have a robots file there. It could be that I have to trash the CDN and change it to only allow the image directory, and maybe set up a separate CDN subdomain for content that only contains the JS and CSS? Have you seen this problem and beat it? (Of course the next thing is Roger might look at google results and start crawling them too, LOL) P.S. The reason I am not asking this question in the google forums is that others have asked this question many times and nobody at google has bothered to answer, over the past 5 months, and nobody who did try, gave an answer that was remotely useful. So I'm not really hopeful of anyone here having a solution either, but I expect this is my best bet because you guys are always willing to try.
Algorithm Updates | | loopyal0