How to prevent duplicate content at a calendar page
-
Hi,
I've a calender page which changes every day.
The main url is
/calendarFor every day, there is another url:
/calendar/2012/09/12
/calendar/2012/09/13
/calendar/2012/09/14So, if the 13th september arrives, the content of the page
/calendar/2012/09/13
will be shown at
/calendarSo, it's duplicate content.
What to do in this situation?
a) Redirect from /calendar to /calendar/2012/09/13 with 301? (but the redirect changes the day after to /calendar/2012/09/14)
b) Redirect from /calendar to /calendar/2012/09/13 with 302 (but I will loose the link juice of /calendar?)
c) Add a canonical tag at /calendar (which leads to /calendar/2012/09/13) - but I will loose the power of /calendar (?) - and it will change every day...
Any ideas or other suggestions?
Best wishes,
Georg.
-
Ah... yeah, that's tricky. There's no magic solution, I'm afraid. You've really got three options:
(1) Leave it alone
(2) Re-organize your site architecture to push individual date pages down a level or two, so that they get less internal link-juice.
(3) Re-organize such that you focus search engines on chunks of time or maybe date/aspect combinations, but then de-index the individual date combos. This would take a much better understanding of your site structure than I currently have. The goal would be to focus your index on some smaller combination of pages that still covers 80% of your search traffic.
The big problem is just that this is a lot potential dilution, and I suspect that many of these pages look very similar to Google. I'm also certain that not all pages have the same value, either for SEO or users, so there's some hybrid approach where you could prune back but not lose everything. Long-term, I think that's worth the time and trouble to sort out, but it's not an emergency or something I'd rush into.
-
Hi Peter,
thanks for your answer!
Well, it's even more complicated!
It's an astrology calendar with planet aspect data for each day starting from 1900-01-01 to 2099-12-31, so there are around 73,000 pages, it's a big database.
People are searching for a date and the planet aspects. So I need the "old pages" and the future pages in the index.
People are also searching their birthday and want to know their zodiac. My calendar is providing this info.
This is an example:
http://www.schicksal.com/horoskop/tageshoroskop/1951/09/10The best thing is to do nothing at the moment I think. The alternativ is to cut the content of the current day from the main page and let the user click a button which redirects to the current day page. But this is not user friendly and I will do nothing at them moment.
Any other idea would be great
Best wishes,
Georg.
-
Sadly, the short answer is that you can't have it all. Either you index the separate calendar pages, get more pages/content out there and risk some "thinning" of your index, or you focus on one page, maximize the SEO value, but then lose the individual pages.
I would not 301 or 302 to the individual calendar URLs - that kind of daily URL shifting is going to look suspicious, Google will not re-cache consistently, and you're going to end up with a long-term mess, I strongly suspect.
I actually tend to agree with Muhammed and Paragon that a viable option would be to let the individual days have their own content, but then canonical to the main calendar page to focus the search results. That way, users can still cycle through each individual day, but Google will focus on the core content. In a way, that's how a blog home-page works - the content changes daily, but you're still keeping the bots focused on one URL.
Think of it in terms of usability, too. How valuable is old/outdated content to search users? They might find something relevant on an old page, but they still probably want to see the main calendar and view recent content.
Where are the links to the individual days, if "/calendar" always has today's content? I'm wondering if there's a hybrid approach, like letting the most recent 30 days all have their own URLs, but then redirecting or using rel-canonical to point to the main page after 30 days.
-
What about adding to all of the other pages i.e not to /calendar/ the links will be followed but not indexed by Google.
-
Hi Georg,
Setting up a redirect or canonicalization for the the calendar page in the ways you describe might make it harder to build up any kind of authority for your calendar.
You could consider adding canonicalization for all the individual day pages that points to the main calendar page. ie. Each page /calendar/YYYY/MM/DD would have rel canonlical=/calendar/. Not sure this is the best idea though.
I don't know how your calendar is setup but you could also look at differentiating the pages by doing just a listing of events on the main page and including summaries or detail on the current day page. Or maybe including some additional information about your calendar on the main page like what type of events are included and how to submit events and not including that information on the individual day pages.
I've always taken the approach of minimizing duplicate content as much as possible but not getting excessive with it. I think in a case like this you could do more harm than good. The calendar page is an ever changing page, it's not like you have the exact same static content on two pages.
Hope this helps!
Zach
-
Hi Muhammed,
because the content is different. This would devaluate all calendar pages.
Best wishes,
Georg. -
Hi Georg What about adding canonical tag(s) from each days (/calendar/2012/09/13) calender pages to the main page (/calendar)
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Page Indexing without content
Hello. I have a problem of page indexing without content. I have website in 3 different languages and 2 of the pages are indexing just fine, but one language page (the most important one) is indexing without content. When searching using site: page comes up, but when searching unique keywords for which I should rank 100% nothing comes up. This page was indexing just fine and the problem arose couple of days ago after google update finished. Looking further, the problem is language related and every page in the given language that is newly indexed has this problem, while pages that were last crawled around one week ago are just fine. Has anyone ran into this type of problem?
Technical SEO | | AtuliSulava1 -
Despite canonical duplicate content in WMT
Hi, 2 weeks ago we've made big changes in title and meta descriptions. To solve the missing title and descriptions. Also set the right canonical. Now i see that in WMT despite the canonical it shows duplicates in meta descriptions and titles. i've setup the canonical like this:
Technical SEO | | Leonie-Kramer
1. url: www.domainname.com/category/listing-family/productname
2. url: www.domainname.com/category/listing-family/productname-more-info The canonical on both pages is like this: I'm aware of creating duplicate titles and descriptions, caused by the cms we use and also caused by wrong structure of category/products (we'll solve that nest year) that's why i wanted the canonical, but now it's not going any better, did i do something wrong with the canonical?0 -
Cloud Hosting and Duplicate content
Hi I have an ecommerce client who has all their images cloud hosted (amazon CDN) to speed up site. Somehow it seems maybe because the pinned the images on pinterest but the CDN got indexed and there now seems to be about 50% of the site duplicated (about 2500 pages eg: http://d2rf6flfy1l.cloudfront.net..) Is this a problem with duplicate content? How come Moz doesnt show it up as crawl errors? Why is thisnot a problem that loads of people have?I only found a couple of mentions of such a prob when I googled it.. any suggestion will be grateful!
Technical SEO | | henya0 -
Duplicate page errors from pages don't even exist
Hi, I am having this issue within SEOmoz's Crawl Diagnosis report. There are a lot of crawl errors happening with pages don't even exist. My website has around 40-50 pages but SEO report shows that 375 pages have been crawled. My guess is that the errors have something to do with my recent htaccess configuration. I recently configured my htaccess to add trailing slash at the end of URLs. There is no internal linking issue such as infinite loop when navigating the website but the looping is reported in the SEOmoz's report. Here is an example of a reported link: http://www.mywebsite.com/Door/Doors/GlassNow-Services/GlassNow-Services/Glass-Compliance-Audit/GlassNow-Services/GlassNow-Services/Glass-Compliance-Audit/ btw there is no issue such as crawl error in my Google webmaster tool. Any help appreciated
Technical SEO | | mmoezzi0 -
How damaging is duplicate content in a forum?
Hey all; I hunted around for this in previous questions in the Q&A and didn't see anything. I'm just coming back to SEO after a few years out of the field and am preparing recommendations for our web dev team. We use a custom-coded software for our forums, and it creates a giant swathe of duplicate content, as each post has its own link. For example: domain.com/forum/post_topic domain.com/forum/post_topic/post1 domain.com/forum/post_topic/post2 ...and so on. However, since every page of the forum defaults to showing 20 posts, that means that every single forum thread that's 20 posts long has 21 different pages with identical content. Now, our forum is all user-generated content and is not generally a source of much inbound traffic--with occasional exceptions--but I was curious if having a mess of duplicate content in our forums could damage our ability to rate well in a different directory of the site. I've heard that Panda is really cracking down on duplicate content, and last time I was current on SEO trends, rel="canonical" was the hot new thing that everyone was talking about, so I've got a lot of catching up to do. Any guidance from the community would be much appreciated.
Technical SEO | | TheEnigmaticT0 -
Duplicate Page Content and Titles
A few weeks ago my error count went up for Duplicate Page Content and Titles. 4 errors in all. A week later the errors were gone... But now they are back. I made changes to the Webconfig over a month ago but nothing since. SEOmoz is telling me the duplicate content is this http://www.antiquebanknotes.com/ and http://www.antiquebanknotes.com Thanks for any advise! This is the relevant web.config. <rewrite><rules><rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^www.antiquebanknotes.com$" negate="true"></add></conditions>
Technical SEO | | Banknotes
<action type="Redirect" url="<a href=" http:="" www.antiquebanknotes.com="" {r:1"="">http://www.antiquebanknotes.com/{R:1}" />
</action></match></rule>
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule></rules></rewrite>0 -
Magento and Duplicate content
I have been working with Magento over the last few weeks and I am becoming increasingly frustrated with the way it is setup. If you go to a product page and remove the sub folders one by one you can reach the same product pages causing duplicate content. All magento sites seem to have this weakness. So use this site as an example because I know it is built on magento, http://www.gio-goi.com/men/clothing/tees/throve-t-short.html?cid=756 As you remove the tees then the clothing and men sub folders you can still reach the product page. My first querstion is how big an issue is this and two does anyone have any ideas of how to solve it? Also I was wondering how does google treat question marks in urls? Should you try and avoid them unless you are filtering? Thanks
Technical SEO | | gregster10001 -
Complex duplicate content question
We run a network of three local web sites covering three places in close proximity. Each sitehas a lot of unique content (mainly news) but there is a business directory that is shared across all three sites. My plan is that the search engines only index the business in the directory that are actually located in the place the each site is focused on. i.e. Listing pages for business in Alderley Edge are only indexed on alderleyedge.com and businesses in Prestbury only get indexed on prestbury.com - but all business have a listing page on each site. What would be the most effective way to do this? I have been using rel canonical but Google does not always seem to honour this. Will using meta noindex tags where appropriate be the way to go? or would be changing the urls structure to have the place name in and using robots.txt be a better option. As an aside my current url structure is along the lines of: http://dev.alderleyedge.com/directory/listing/138/the-grill-on-the-edge Would changing this have any SEO benefit? Thanks Martin
Technical SEO | | mreeves0