Duplicate Page title - PHP Experts!
-
After running a crawl diagnostics i was surprised to see 336 duplicate page titles. But i am wondering if it is true or not. Most of them are not a page at all but a .php variation.
for example: The following are all the same page, but a different limit on viewing listings. Limiting your view to 5, 10, 15, 20, 25 as you choose.
.com/?lang=en&limit=5
.com/?lang=en&limit=5&limitstart=10
.com/?lang=en&limit=5&limitstart=15
.com/?lang=en&limit=5&limitstart=20
.com/?lang=en&limit=5&limitstart=25Same type of things are going on all over the site causing 228 duplicate content errors and the already mentioned 336 duplicate pages. But is "crawl diagnostic telling the truth" or is it just some php thing? I am not a php expert so any input would be much appreciated.
What should i do?
-
Both Ryan and Marie provided good answers, but let me elaborate further.
This is not a php thing. You can verify this yourself by visiting these pages.
Does the page content change when you visit /?lang=en&limit=5&limitstart=20 and switch to /?lang=en&limit=5&limitstart=20? If these look like the same page, then you have a duplicate content/title issue on your hands.
Google's take on the matter is simple...
"Provide one version of a URL to reach a document
To prevent users from linking to one version of a URL and others linking to a different version (this could split the reputation of that content between the URLs), focus on using and referring to one URL in the structure and internal linking of your pages. If you do find that people are accessing the same content through multiple URLs, setting up a 301 redirect from non-preferred URLs to the dominant URL is a good solution for this. You may also use canonical URL or use the rel="canonical" link element if you cannot redirect"
You have several options to deal with this, depending on the content_._
- Use the canonical tag to point the different versions of the page towards a single, definitive URL. Best to use this option only if the pages are actually duplicate.
- If the pages are not duplicate in content, but duplicate in title, then make sure your CMS writes unique title tags for every unique page of content, as Marie describes.
- If these extra pages are producing lists of only slightly varying entries, like a list of blog archives, you may consider adding a Meta Robots tag to NOINDEX, FOLLOW those superfluous pages.
- If dealing with pagination, here's an excellent article from Adam Audette to help.
This is a tricky area, and you must be careful when dealing with duplicate content, because it's easy to make a mistake and have Google de-index your content. That said, by correcting these errors, you just might see an improvement in your indexation and traffic stats.
-
I'm not sure if this will answer your question or just confuse things, but here's my take on the situation.
Are the pages all duplicate content, or are the titles just duplicated? If this is all the exact same page but simply with a different url then the solution is to use a rel canonical tag. (See http://www.google.com/support/webmasters/bin/answer.py?answer=139394). This tells Google what the original version of the page is and to not index the other pages.
However, if these are all different pages and you are trying to create a unique title for each then there are some ways to do this. Like Ryan mentioned, if you're using a CMS (i.e. Wordpress) there's likely a plugin that will do this for you. But, if you have a static site then here is what I do. I have a static site and I have a page called, let's say, www.example.com/information?id=123. I have thousands of these pages but they all have different content. The content is generated from a database.
So, what I do is I put php in the head of the page, before the title. The php reads the database and grabs some information about the page and stores that information in a variable called $title. Then, in the title tag I do this:
<title><?php echo $title; ?></title>
Hopefully that makes sense!
-
Let me preface by saying I am not a PHP expert. I do have some familiarity with PHP and the issue you mention.
The bottom line, you do not wish to use the same title for multiple URLs. Using the URLs you shared, some possible titles would be:
Best Restaurants in Miami
Best Restaurants in Miami - Page 3
Best Restaurants in Miami - Page 4
You need to adjust your site code to add a page number to the existing URLs. Most popular software is set up to do this automatically.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Possible duplicate content issue with my Blog and archive pages . Any help greatly appreciated
Dear Mozzers, I have been looking at my news section on my eCommerce site and I think I may have a duplicate content issue and wanted some advice on whether I do and if so , how best I handle this. http://www.website.co.uk/news
Technical SEO | | PeteC12
http://www.website.co.uk/news/page:1
http://www.website.co.uk/news/page:2
http://www.website.co.uk/news/page:3
http://www.website.co.uk/news/limit:9999 (This is show all) I also have the ability of showing articles by month : http://www.website.co.uk/news/archive/2015/04 (April)
http://www.website.co.uk/news/archive/2015/03 (March)
http://www.website.co.uk/news/archive/2015/02 (Feb)
http://www.website.co.uk/news/archive/2015/01 (Jan) I am wondering if there's a duplicate issue here or not given that I also articles by month as well and if so how best I handle this.? I already do pagination on my news pages (page 1 , page 2) by using rel=next and rel=Prev but I don't have an canconical or anything as yet. I enclose a couple of links if this would help and would appreciate if someone could take a browse. I have a View All link on my homepage for for all news items - http://goo.gl/JPPIvQ I which have a different urls - March 2015 Articles - http://goo.gl/0O1wYD and April 2015 articles - http://goo.gl/GdW2oK On another note, These articles are also linked to from the relevant category landing pages on my website to help with SEO. I have not used H tags on the article links in my landing pages , just displaying the weblink back to the news article.I've done this to try and improve the PR and rankings of my landing pages. Just wondered if anyone has any comments as to whether thats a good or bad idea and whether I could improve it in any way - An example is here (scroll down the page to the pressure washing guides) - http://goo.gl/nnRE49 Thanks Pete0 -
Duplicate Pages on GWT when redesigning website
Hi, we recently redesigned our online shop. We have done the 301 redirects for all product pages to the new URL (and went live about 1.5 week ago), but GWT indicated that the old product URL and the new product URL are 2 different pages with the same meta title tags (duplication) - when in fact, the old URL is 301 redirecting to the new URL when visited. I found this article on google forum: https://productforums.google.com/forum/#!topic/webmasters/CvCjeNOxOUw
Technical SEO | | Essentia
It says we either just wait for Google to re-crawl, of use the fetch URL function for the OLD URLs. Question is, after i fetch the OLD URL to tell Google that it's being redirected, should i click the button 'submit to index' or not? (See screengrab - please note that it was the OLD URL that was being fetched, not the NEW URL). I mean, if i click this button, is it telling Google that: a. 'This old URL has been redirected, therefore please index the new URL'? or
b. 'Please keep this old URL in your index'? What's your view on this? Thanks1 -
Diagnostics say I'm missing Page titles... but I am not?
I've been running a crawl of one of our new site builds for a couple of weeks. The Diagnostics picked up a couple of issues, which was great, but it's saying we're missing Page Titles and Descriptions on pages that we have Page Titles and Descriptions. Anyone come across this before?
Technical SEO | | niamhomahony0 -
Anyone know how to fix duplicate content and titles with news section?
We use django for out site and it's working really well, but we're having an issue with duplicate titles and content via the news section. The news is basically stories sourced from other sites and we link to them via our news section. I'm not sure how to fix the duplicate title issue in this case. I noticed people recommend archiving or using a canonical, but because the news section is set up how it is I don't think that would work. Does anyone have a way around this?A
Technical SEO | | KateGMaker0 -
How can I prevent duplicate content between www.page.com/ and www.page.com
SEOMoz's recent crawl showed me that I had an error for duplicate content and duplicate page titles. This is a problem because it found the same page twice because of a '/' on the end of one url. e.g. www.page.com/ vs. www.page.com My question is do I need to be concerned about this. And is there anything I should put in my htaccess file to prevent this happening. Thanks!
Technical SEO | | onlineexpression
Karl0 -
Once duplicate content found, worth changing page or forget it?
Hi, the SEOmoz crawler has found over 1000 duplicate pages on my site. The majority are based on location and unfortunately I didn't have time to add much location based info. My question is, if Google has already discovered these, determined they are duplicate, and chosen the main ones to show on the SERPS, is it worth me updating all of them with localalized information so Google accepts the changes and maybe considers them different pages? Or do you think they'll always be considered duplicates now?
Technical SEO | | SpecialCase0 -
Does duplicate content on word press work against the site rank? (not page rank)
I noticed in the crawl that there seems to be some duplicate content with my word press blog. I installed a seo plugin, Yoast's wordpress seo plugin, and set it to keep from crawling the archives. This might solve the problem but my main question is can the blog drag my site down?
Technical SEO | | tommr10 -
Discrepency between # of pages and # of pages indexed
Here is some background: The site in question has approximately 10,000 pages and Google Webmaster shows that 10,000 urls(pages were submitted) 2) Only 5,500 pages appear in the Google index 3) Webmaster shows that approximately 200 pages could not be crawled for various reasons 4) SEOMOZ shows about 1,000 pages that have long URL's or Page Titles (which we are correcting) 5) No other errors are being reported in either Webmaster or SEO MOZ 6) This is a new site launched six weeks ago. Within two weeks of launching, Google had indexed all 10,000 pages and showed 9,800 in the index but over the last few weeks, the number of pages in the index kept dropping until it reached 5,500 where it has been stable for two weeks. Any ideas of what the issue might be? Also, is there a way to download all of the pages that are being included in that index as this might help troubleshoot?
Technical SEO | | Mont0