How To Implement Pagination Properly? Important and Urgent!
-
I have seen many instructions but I am still uncertain. Here is the situation
We will be implementing rel prev rel next on our paginted pages.
The question is:
- Do we implement self referencing canonical URL on the main page and each paginated page?
- Do we implement noindex/follow meta robots tag on each paginated page?
- Do we include the canonical URL for each paginated page in the sitemap if we do not add the meta robots tag?
- We have a view all but will not be using it due to page load capabilities...what do we do with the viewl all URL? Do we add meta robots to it?
- For website search results pages containing pagination should we just put a noindex/follow meta robots tag on them?
- We have seperate mobile URL's that also contain pagination. Do we need to consider these pages as a seperate pagination project? We already canonical all the mobile URL's to the main page of the desktop URL.
Thanks!
-
Hello SEO32,
I apologize for the delayed response. There are several good questions here. They're also complicated questions, which don't really always have a single "correct" answer. So much revolves around the specific situation, and without seeing your website it is difficult to say what is best for you. Also, much of what we think we know about this kind of stuff is either based on what Google tells us (which isn't always the truth) and what we've observed and deduced from our own experiences (which aren't always the same). True "testing" of this stuff one variable at a time isn't always possible so we rely on best practices and our own experience.
That said, I will attempt to answer your questions with what I would probably do in most situations, including links to more information when possible.
Do we implement self referencing canonical URL on the main page and each paginated page?
Here's what Rand says, and he's probably seen way more data than I have:
"Whatever you do, DO NOT:
- Put a rel=canonical directive on paginated results pointing back to the top page in an attempt to flow link juice to that URL. You'll either misdirect the engines into thinking you have only a single page of results or convince them that your directives aren't worth following (as they find clearly unique content on those pages).
- Add nofollow to the paginated links on the results pages. This tells the engines not to flow link juice/votes/authority down into the results pages that desperately need those votes to help them get indexed and pass value to the deeper pages.
- Create a conditional redirect so that when search engines request paginated results, they 301 redirect or meta refresh back to the top page of results."
Keep in mind that post is from 2010, and I think before Google said a "View All" canonical <a>was their preference</a>.
I have seen plenty of sites do well ranking the canonical category page, and with indexing most of the product pages, while all paginated pages had a rel canonical that referenced the first page in the series (i.e. .com/category/ or .com/categry1/category2/). It probably helps that they had good XML sitemaps for product pages, and plenty of internal linking, unique content on category pages, etc.
I have also seen sites do well using rel next/prev without rel canonical, or rel next/prev with self-referencing canonicals on paginated category pages.
I think where you run into problems is when you also allow the facet/filter/sort versions to have self-referencing rel canonical tags.
Here is what I advise in most cases:
Use rel next/prev (not because I think it works, but because Google says to and I don't think it hurts) along with self-referencing rel canonical tags, and "follow,noindex" robots meta tags on paginated pages.
Always include a followable link to the first page in the series from every subsequent page. For example:
<previous>first...1...25...26...27...last...</previous>
I recommend always having a first and last page link. The first is obvious because it means pagerank is going to flow into it from every other page in the set, giving it the most internal links of all. The last is more of a crawlability and usability thing. For users it helps us figure out how much further we have to go. It does the same thing for search engines. Instead of blindly following a path that may or may not have an end, a message is sent that tells a spider how much further it has to go. I don't know if Google takes advantage of that signal or not, but it just makes sense to include it. If you want to get fancy you can try making the 'last' link flash or javascript or something so it doesn't pass (as much?) page rank.
The category root pages usually have links from site-wide navigation, unlike the paginated versions, which further establishes it as the page that should be ranked highest.
Make sure the first page in each series is indexable, and has content that does not appear on the paginated versions. Also, make sure that ?p=1 doesn't have a self-referencing canonical tag, but references the root page for that series (e.g. /category1/category2/).
All subsequent variations (e.g. color, size) should rel canonical back to their root page. For example:
/category1/category2/?page=2&size=s&color=blue would have the following URL in the rel canonical tag:
/category1/category2/?page=2
Which happens to be followable, but not-indexable, and has a self-referencing rel canonical tag.In this way you give search engines a strong signal about which URL in the whole set is the strongest (i.e. /category1/category2/) because it is indexable, has its own content, has the most internal and external links, is the simplest version of this URL pattern, and is at the root of the directory. You're telling search engines which page is next in the series, and that this page is first in the series. You're telling search engines which page is last in the series, as well. Google usually does an awesome job figuring it out from there. There are always exceptions.
Do we implement noindex/follow meta robots tag on each paginated page?
I would. Consider this from Google's perspective, or from that of a searcher. Someone types "Blue Flower Dress" into Google. Is the best page to return a deep category page full of blue dresses, one of which happens to have flowers? Or would it be the Blue Flower Dress product page? I can't think of any reason why I would want to land on page 3, where what I'm looking for is listed among dozens of other things, when I could just go straight to the thing I'm looking for.
Likewise, if someone searches for "Blue Dresses" is the best page /dresses/blue/?page=3 (paginated page in the Blue Dresses category), OR /dresses/blue/ (the very first page of the Blue Dresses category), which also has useful content about blue dresses?
Long story short, when it comes to transactional eCommerce queries, they're usually either looking for a product page or the first page of a specific category or sub-category. Or sometimes the home page. Therefore, I don't see any reason for allowing paginated URLs to be indexable in most cases. Non-transactional eCommerce content is different (e.g. buying guides, comparison charts, reviews...) but I still wouldn't allow paginated pages to be indexed in most cases.
Slightly Off Topic - Filters/Facets/Sorts
Or perhaps the category is "casual dresses" and "blue" is specified in the "color" attribute. In this case, would the best page be /dresses/casual/?color=blue , /dresses/casual/ or /dresses/casual/?color=blue&page=4 for someone who Googled "blue dresses"? I've bolded the one I'd prefer as a searcher.
Here again, as with the internal search results, there is an opportunity to use real data to inform your decision. Pay attention to the facet/filter/sort URLs most accessed by shoppers and consider turning those into category or collections pages with their own URL pattern (e.g. /dresses/casual/blue/). One example I come across all the time is when "Brand" is a filter instead of its own limb in the category structure. If people are shopping by brand, as they do with most consumer products, then you should have a brand subcategory under each major top-level category. If I search for Levi Jeans Google doesn't want to send me to a "pants" page where I have to set a filter to see only Levis. I should go to pants/brand/levi/ . If I Google Chefmate Pots I want to see cookware/pots/brands/chefmate so I don't have to set a filter after I get there.
This doesn't mean all filter pages should be turned into category pages either. Use your best judgement based on the pages most of your users are accessing from the navigation and filters.
Do we include the canonical URL for each paginated page in the sitemap if we do not add the meta robots tag?
I would add the robots meta tag. Please let me know if I've misunderstood the question.
We have a view all but will not be using it due to page load capabilities...what do we do with the view-all URL? Do we add meta robots to it?
I would add a meta robots "index,nofollow" tag, and would also use the canonical page's URL (e.g. /category1/category2) in the rel canonical tag.
For website search results pages containing pagination should we just put a noindex/follow meta robots tag on them?
This is one of those situations involving crawl budget potentially being eaten up by an infinite amount of pages. I would consider blocking the internal search result URLs in the robots.txt file. They are of no use to Google, as they consider a search engine returning search results with links to more search results somewhere else a bad user experience. This is also what Google recommends in their Webmaster Guidelines:
"Use robots.txt to prevent crawling of search results pages or other auto-generated pages that don't add much value for users coming from search engines."
However, I would also make use of those pages internally. Rather than relying on a search result page for things people often look for, track what is being searched for and create static, indexable pages. For example, try "Collections" pages on eCommerce sites, as well as FAQ pages, or "Industries" or "Use Case"-type pages on lead generation sites. This is a much better user experience for someone arriving on that page from a search engine.
We have separate mobile URL's that also contain pagination. Do we need to consider these pages as a separate pagination project? We already canonical all the mobile URL's to the main page of the desktop URL.
I think you should if that's the way you're handling it. Here is a post I did on mobile best practices. It covers some other options. I would also add a rel=”alternate” tag in the HTML header of the desktop page, which alerts search engines to the corresponding mobile URL and helps define the relationship between the two pages.
The bottom line for me is to always think about what would be the best experience for someone searching from Google for something, and to try and use all of the various technical options to ensure that is the page I'm telling Google they should rank for that query, or those types of queries. The 'best practice' changes, depending on the situation.
I hope others will join the discussion with their own experiences and findings.
-
Hi,
Please check this moz article on this @ SEO Guide to Google Webmaster Recommendations for Pagination
Hope this helps you.
Thanks
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Importance (or lack of) Meta keywords tags and Tags in Drupal
I'm wondering should I put any effort in making Meta Keywords tags for my pages or normal Tags (they're separate in Drupal), since apparently first are not considered by most of search engines, while not sure about normal tags. Obviously SERPS has to determine partial valu of the page by content, thus consider keywords / tags to some extend. What's your opinion on that. Thank you.
Intermediate & Advanced SEO | | Optimal_Strategies1 -
Google Pagination Changes
What with Google recently coming out and saying they're basically ignoring paginated pages, I'm considering the link structure of our new, sooner to launch ecommerce site (moving from an old site to a new one with identical URL structure less a few 404s). Currently our new site shows 20 products per page but with this change by Google it means that any products on pages 2, 3 and so on will suffer because google treats it like an entirely separate page as opposed to an extension of the first. The way I see it I have one option: Show every product in each category on page 1. I have Lazy Load installed on our new website so it will only load the screen a user can see and as they scroll down it loads more products, but how will google interpret this? Will Google simply see all 50-300 products per category and give the site a bad page load score because it doesn't know the Lazy Load is in place? Or will it know and account for it? Is there anything I'm missing?
Intermediate & Advanced SEO | | moon-boots0 -
How To Implement Breadcrumbs
Hi, I'm looking to implement breadcrumbs for e-commerce store so they will appear in the SERP results like the attached image. In terms of implementing to a site, would you simply add HTML to each page like this Google example? Which looks like this: Books › Science Fiction Award Winners Then is there anything you need to do, to get this showing in the SERPs results e.g. doing something in search console. Or do you just wait into google has crawled and hopefully starts showing in the SERPs results? Cheers. wn3ybMMOQFW98fNQkxtJkA.png [SERP results with bread crumbs](SERP results with bread crumbs)
Intermediate & Advanced SEO | | jaynamarino0 -
Most important things for seo a travel website
Hello everyone a website working on travel field with this address : https://goo.gl/4gaoAn Let me know what do you think about it and please give me some advises about it get improve on google rankings. If you be able to take time and give me some advises based on what you see on the website, would be great for me. Also what would work best for me to have a great link building strategy after penguin 4.0 update? and what does my site lack right now? Thanks and waiting to hear from you asap.
Intermediate & Advanced SEO | | BahadorGh0 -
DeIndexing pagination
I have a custom made blog with boat loads of undesirable URLs in Google's index like this:
Intermediate & Advanced SEO | | Rich_Coffman
.com/resources?start=150
.com/resources?start=160
.com/resources?start=170 I've identified this is a source of duplicate title tags and had my programmer put a no index tag to automatically go on all of these undesirable URLs like this: However doing a site: search in google shows the URLs to still be indexed even though I've put the tag up a few weeks ago. How do I get google to remove these URLs from the index? I'm aware that the Search Console has an answer here https://support.google.com/webmasters/topic/4598466?authuser=1&authuser=1&rd=1 but it says that blocking with meta tags should work. Do I just get google to crawl the URL again so it sees the tag and then deindexes the URLs? Or is there another way I'm missing.0 -
Proper 301 in Place but Old Site Still Indexed In Google
So i have stumbled across an interesting issue with a new SEO client. They just recently launched a new website and implemented a proper 301 redirect strategy at the page level for the new website domain. What is interesting is that the new website is now indexed in Google BUT the old website domain is also still indexed in Google? I even checked the Google Cached date and it shows the new website with a cache date of today. The redirect strategy has been in place for about 30 days. Any thoughts or suggestions on how to get the old domain un-indexed in Google and get all authority passed to the new website?
Intermediate & Advanced SEO | | kchandler0 -
Dilemma: Should we use pagination or 'Load More' Function
In the interest of pleasing Google with their recent updates and clamping down on duplicate content and giving a higher preference to pages with rich data, we had a tiny dilemma that might help others too. We have a directory like site, very similar to Tripadvisor or Yelp, would it be best to: A) have paginated content with almost 40 pages deep of data < OR > B) display 20 results per page and at the bottom have "Load More" function which would feed more data only once its clicked. The problem we are having now is that deep pages are getting indexed and its doing us no good, most of the juice and page value is on the 1st one, not the inner pages. Wondering what are the schools of thought on this one. Thanks
Intermediate & Advanced SEO | | danialniazi0 -
Need some urgent Panda advice. Open discussion about recovering from the Panda algorithm.
I have a site that has been affected by Panda, and I think I have finally found the problem. When I created this site in the year 2006, I bought content without checking it. Recently, when I went through the site I found out that this content had many duplicates around the web. Not 100% exact, but close to. The first thing I did is ask my best writer to rewrite these topics, as they are a must on my site. This is a very experienced writer, and she will make the categories and subpages outstanding. Second thing I did was putting a NOINDEX, FOLLOW robots meta in place for the pages I determined being bad. They haven't been de-indexed yet. Another thing I recently did is separate other languages and move these over to other domains (with 301's redirecting the old locations to the new.) This means that the site now has a /en/ directory in the URL which is no longer used. With this in mind I was thinking to relocate the NEW content, and 301 the old (to preserve the juice for a while.) For example: http://www.mysite.com/en/this-is-a-pandalized-page/ 301 to http://www.mysite.com/this-is-the-rewritten-page/ The benefits of doing this are: decreasing the amounts of directories in the URL getting rid of pages that are possibly causing trouble getting fresh pages added to the site Now, the advice I am looking for is basically this: Do you agree with the above? Or don't you agree? If you don't, please be so kind to include a reason with your answer. If you do, and have any additional information, or would like to discuss, please go ahead 🙂 Thanks, Giorgio PS: Is it proven that Panda is now a running update? Or is it still periodically executed?
Intermediate & Advanced SEO | | VisualSense1