Is publishing a large quantity of content at once a bad idea?
-
If you plan on doubling the size of your site with original, unique content, is it better to publish it all at once or over a period of time? Is there any penalty for publishing it all at once?
-
Hey Michelleh,
Matt Cutts gives his advice about it in this video about publishing 100 pages at once....
-
If I have a lot of content I publish it as soon as I can. Content in the index is as good as money. Why wait? Cash in!
-
There is no penalty to publishing all at once. There is no advantage to holding back.
You could try an approach where you published an article each day in an attempt to build hype and focus attention. From a pure SEO standpoint, I would recommend publishing content as soon as it becomes available. A quality article can easily take a day or more to produce. It usually doesn't make sense to hold back.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is an iframe redirect on the same Domain bad for SEO
Good morning. We have a vendor that has created a landing page with content that we want to use. Because of the way we built the site, the only way to use the content is to create an i-frame. The i-frame is re-directingon the same Domain. Would we benefit from the SEO Content?
Intermediate & Advanced SEO | | jdenbo_edf0 -
Best Sitemap for Large Website
i have more than 3500 pages on my website. Please let me know the best sitemap plugin for my website.
Intermediate & Advanced SEO | | Michael.Leonard1 -
Internal Duplicate Content Question...
We are looking for an internal duplicate content checker that is capable of crawling a site that has over 300,000 pages. We have looked over Moz's duplicate content tool and it seems like it is somewhat limited in how deep it crawls. Are there any suggestions on the best "internal" duplicate content checker that crawls deep in a site?
Intermediate & Advanced SEO | | tdawson091 -
Duplicate content on subdomains
Hi All, The structure of the main website goes by http://abc.com/state/city/publication - We have a partnership with public libraries to give local users access to the publication content for free. We have over 100 subdomains (each for an specific library) that have duplicate content issues with the root domain, Most subdomains have very high page authority (the main public library and other local .gov websites have links to this subdomains).Currently this subdomains are not index due to the robots text file excluding bots from crawling. I am in the process of setting canonical tags on each subdomain and open the robots text file. Should I set the canonical tag on each subdomain (homepage) to the root domain version or to the specific city within the root domain? Example 1:
Intermediate & Advanced SEO | | NewspaperArchive
Option 1: http://covina.abc.com/ = Canonical Tag = http://abc.com/us/california/covina/
Option 2: http://covina.abc.com/ = Canonical Tag = http://abc.com/ Example 2:
Option 1: http://galveston.abc.com/ = Canonical Tag = http://abc.com/us/texas/galveston/
Option 2: http://galveston.abc.com = Canonical Tag = http://abc.com/ Example 3:
Option 1: http://hutchnews.abc.com/ = Canonical Tag = http://abc.com/us/kansas/hutchinson/
Option 2: http://hutchnews.abc.com/ = Canonical Tag = http://abc.com/ I believe it makes more sense to set the canonical tag to the corresponding city (option 1), but wondering if setting the canonical tag to the root domain will pass "some link juice" to the root domain and it will be more beneficial. Thanks!0 -
Quantity or quality in Google+ authorship?
Hi folks, here goes a (hopefully) easy one for the local authorship gurus. For our blog content strategy we currently have two inhouse contributors. Both have decent Google+ profiles and one is in the process of really establishing authorship/influence by submitting guest posts to several industry sites, sharing content in Google+, engaging in conversations in twitter, etc. Posts by this latter contributor already rank page 1 for the main keywords. We now have a new content contributor who is a retired employee from the company and a good friend. He has written excellent content that will be published in our blog in the coming few months. He does not have a Google+ profile but he can have one if we ask him to, but he is not going to use it for anything other than writting on our blog. He does not mind having his content published under any of our current Google+ profiles. Question: should we include this new content under our current profiles or should we create a new Google+ profile for this new contributor knowing that it will be an 'empty' profile? Thanks in advance!
Intermediate & Advanced SEO | | TIBA0 -
International SEO - cannibalisation and duplicate content
Hello all, I look after (in house) 3 domains for one niche travel business across three TLDs: .com .com.au and co.uk and a fourth domain on a co.nz TLD which was recently removed from Googles index. Symptoms: For the past 12 months we have been experiencing canibalisation in the SERPs (namely .com.au being rendered in .com) and Panda related ranking devaluations between our .com site and com.au site. Around 12 months ago the .com TLD was hit hard (80% drop in target KWs) by Panda (probably) and we began to action the below changes. Around 6 weeks ago our .com TLD saw big overnight increases in rankings (to date a 70% averaged increase). However, almost to the same percentage we saw in the .com TLD we suffered significant drops in our .com.au rankings. Basically Google seemed to switch its attention from .com TLD to the .com.au TLD. Note: Each TLD is over 6 years old, we've never proactively gone after links (Penguin) and have always aimed for quality in an often spammy industry. **Have done: ** Adding HREF LANG markup to all pages on all domain Each TLD uses local vernacular e.g for the .com site is American Each TLD has pricing in the regional currency Each TLD has details of the respective local offices, the copy references the lacation, we have significant press coverage in each country like The Guardian for our .co.uk site and Sydney Morning Herlad for our Australia site Targeting each site to its respective market in WMT Each TLDs core-pages (within 3 clicks of the primary nav) are 100% unique We're continuing to re-write and publish unique content to each TLD on a weekly basis As the .co.nz site drove such little traffic re-wrting we added no-idex and the TLD has almost compelte dissapread (16% of pages remain) from the SERPs. XML sitemaps Google + profile for each TLD **Have not done: ** Hosted each TLD on a local server Around 600 pages per TLD are duplicated across all TLDs (roughly 50% of all content). These are way down the IA but still duplicated. Images/video sources from local servers Added address and contact details using SCHEMA markup Any help, advice or just validation on this subject would be appreciated! Kian
Intermediate & Advanced SEO | | team_tic1 -
What to do with bad webpage
Hello everyone, I have a page in my website that has a terrible link profile (95% exact match keyword links.) What is the best thing to do with this page? It provides no value, and if anything it is hurting me. Should I just delete the page, 301 it to an obscure page or something else? Thanks!
Intermediate & Advanced SEO | | Mjstout0 -
What do I do about sites that copy my content?
I've noticed that there are a number of websites that are copying my content. They are putting the full article on their site, mentioning that it was reposted from my site, but contains no links to me. How should I approach this? What are my rights and should I ask them to remove it or add a link? Will the duplicate content affect me?
Intermediate & Advanced SEO | | JohnPeters0