Title tag solution for a med sized site
-
Its the same old story, we all know it well. I have a client that has a site with 20k+ pages (not too big) and traffic levels around 450k/month.
Now we have identified 15 pages with various conversion points/great backlink metrics etc. that we are going to explicitly target in the first round of recs. However, we are looking at about 18,000 dup title tags that I'd like to clean up.
The site is not on a CMS and in the past I've had the dev team write a script to adopt the h1 tag or the name of the page etc as the title tag. This can cause a problem when some of these pages that are being found in long tail search lose their positions etc. I'm more hesitant than ever to make this move with this current client because they get a ton of long tail traffic spread over a ton of original content they wrote.
How does everyone else usually handle this? Thoughts?
Thanks in advance Mozzers!
-
How about having your developers script something, that scrapes all 18.000 h1, h2, h3 for each article and store them in a database. Finding dupes then would be a piece of cake, even for a less experienced developer You could easily export all your duplicates to csv and then manually rename them based on their content.
Dev time: about 1 day max. (Developed a lot of software myself and IMHO a good developer should get this up and running within 4 hours)
If you don't have toooooo many duplicate tags, correcting those in question shouldn't be taking too long aswell.
If you have done your chores you could reimport your corrected title-tags to the database. Your developer could write a script in the meantime, that sets the title-tag of a page according to the title-tag you stated in your database.
Hope that helped If you have further questions on this, just go ahead. Had a similar problem with 25k+ pages for a major health insurance and we figured out, that the best way to prevent problems was to do most of the work manually than with a script. Helped us a lot to stay within the budget and given timeframe.
-
This is sound advice. Test out a percentage of pages before rolling out the change site-wide.
I also agree that 18K duplicate titles isn't helping the site.
One thing I would do is review analytics and define the top X % pages and hand optimize those. The balance can be optimized via rules utilizing the system you outlined. As to whether to use the H1 or the file name or some other element, I'd probably lean towards the h1 as it would likely accurately describe the content and not be truncated or contain stop words.
-
-
Can you implement on a section or on a % of your pages first? then you can test the effect without risking your whole catalogue
-
Ryan - excellent points! The benefits of adding a CMS to this site would be quite good, at the very least for providing some sort of grounds for moving forward on a unified platform.
-
The 18k is a hard piece of the puzzle to wrap your mind around...I'd like to give more details there but can't...currently. Hopefully when this campaign starts to show results they will let me write a case for it...I'll be sure to share.
There is a "templating system" for various sections. However, as I mentioned, many developers have had their hands in it and didn't follow a standardized system.
I am considering EGOLs comment
Thanks!
-
Is there any form of standardization? I can't imagine 18k pages which were independently developed.
There should be a templating system or some logic which controls code common to all pages. Most pages should share the same header, footer and sidebar, along with standards for things like a canonicalization tag, title and meta description.
If that is not the case, the EGOL's comment should be considered. It is not reasonable to maintain a site which lacks standards.
-
It's possible that putting a reasonably intelligent human on the job for a couple of months could pay back big time. I'll bet a good title tag job would pull in thousands of dollars worth of sales every month.
-
Hey Ryan,
Thanks for the response!
There was 18k title tag duplicates but top content that I can tell is being found in search is about 1,500 pages. Its not a forum site or a site with UGC. Its a very successful tech hardware company that has put out a lot of great unique content over time.
Determining the logic is the tough part because there isn't a lot of consistency throughout the site...different developers have had their hands in it over time.
-
What kind of site is it?
With 18k+ pages I will take a guess that it is a forum site. Definitely check with your forum software provider. There should be some form of "page container" which is used as a template for all the site's pages. If you can determine the logic you want to use, such as go with the post title or H1 tag, then you can modify the template according to your logic and take care of your entire site quickly and easily.
-
Thanks for the response! I should rephrase my question...
I'm either looking for tricks/tips others use in this situation or messages like yours that will give me the confidence to go for it haha.
I think we've all experienced the fear of doing what we know is technically correct and risk being at the mercy of the algo. I've gone this route a lot in the past but I've never done it on a site that gets traffic so deep into so many pages from search.
Have you ever gone the script route? If so, what did you have it pull to use as a title tag? Like I mentioned above, I usually have used h1's in the past..
-
If a lot of traffic is coming in through 18,000 pages that have duplicate title tags I am willing to bet that there will be a huge increase in the amount of traffic that those pages pull when unique and relevant title tags are put in place.
So, although there is a small chance that traffic will go down, I think that there is a much higher chance that traffic will immedately shoot up spectacularly - and the quality of that traffic might also improve.
I would archive the site, run a script to replace the title tags, see what happens. You can always put the old title tags back up if this doesn't work - but I bet it works great.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Htaccess maximum size?
Hello all, The company that develops our website recently contacted and asked me if we could remove a large amount of URL rewrites. I've described a few factors and my main questions below. Some information: One year ago we did a large migration. We went from 27 websites to one main website. We have got about 2000 rewrites in the htaccess file. And the file is 208kb. A lot of links from our old domains still have incoming traffic which are handled by the rewrite rules mentioned above. Questions:
Intermediate & Advanced SEO | | DPA
The company that develops our website said that the htaccess file is too large and is causing or could be causing us website performance issues. They have asked us to remove URL rewrites.
My question is:
a) How many rewrites is too much?
b) Is the filesize of the htaccess of any importance or is it just the amount of rewrites in the file?
c) Could we solve any potential server/website performance issues due to a large htaccess file in any other way? Increasing some values like 'post_max_size' or by any other solutions handled serverside? I do not have a lot of knowledge of htaccess rules but I've seen websites that handled over a million of rewrite rules. This is why I'm having doubts on whether removing URL rewrites is the only solution and possibly not the best solution for us. Hopefully you can help me any further and with the best way to proceed without losing traffic or causing 404 pages. Thanks in advance!
Iordache Voicu0 -
Indexed Pages Different when I perform a "site:Google.com" site search - why?
My client has an ecommerce website with approx. 300,000 URLs (a lot of these are parameters blocked by the spiders thru meta robots tag). There are 9,000 "true" URLs being submitted to Google Search Console, Google says they are indexing 8,000 of them. Here's the weird part - When I do a "site:website" function search in Google, it says Google is indexing 2.2 million pages on the URL, but I am unable to view past page 14 of the SERPs. It just stops showing results and I don't even get a "the next results are duplicate results" message." What is happening? Why does Google say they are indexing 2.2 million URLs, but then won't show me more than 140 pages they are indexing? Thank you so much for your help, I tried looking for the answer and I know this is the best place to ask!
Intermediate & Advanced SEO | | accpar0 -
Original Source Tag or Canonical Tag for News Publishers?
I have been sourcing content from a news publisher who is my partner for publishing content online. My website deals with sourcing content from a couple of websites. I did use a canonical tag pointing towards the respective syndicated source but I have not seen traffic for those articles. I did some research and found out that Google does have a tag for news publishers which is the "original-source" tag which helps news publishers to give proper credit for their work. Here's a link to the official word by Google" https://news.googleblog.com/2010/11/credit-where-credit-is-due.html Although Google has officially stated that the "syndication-source" tag has been replaced by the "canonical" tag. However, there is no mention about the "original-source" tag.
Intermediate & Advanced SEO | | Starcom_Search
Can I still use the "original-source" tag to syndicate content from my partner site instead of the "canonical" tag? P.S.: The reason why I am not convinced with the use of the canonical tag is because:
1. As per what Google says, duplicate content won't harm my website unless it is spam. (And since we are rightfully content from our partner'website and showcasing it to a larger audience by hosting it on our website as well, we are thereby not indulging in any unethical practices) 2. The canonical tag could possibly hamper my crawl bandwidth issues as it would essentially need the crawler to crawl the whole page to figure out that the canonical is present, post which any possible valuation that my site could have garnered gets lost.3. Moreover, since I am from the news, media and publication industry, content republication is a widely accepted practice and in such cases simply including a link to the original source of the article or using the original source tag should suffice, That being mentioned, I do not want to go ahead without taking a second opinion about this. Kindly help me to resolve this issue.0 -
WordPress posts Title field inserts title into blog posts like a headline but doesn't ad H1 tag how to change?
I have a Wordpress website which is just using the Default theme, when I post in the blog, whatever I put in the "Title" field at the top of the editor is automatically is placed within the body of the blog post, like a headline, but it doesn't include any H1 tags that I can see. If I add my own headline within in the blog editor, it still inserts the Title like a headline. I am using the Yoast SEO Plugin and also write the meta title there, should I just leave the Wordpress title field blank so it doesn't insert into the blog post? Or is that inserted Title being recognized as an H1 even though I don't see h1 tags anywhere? Hope this isn't too confusing.
Intermediate & Advanced SEO | | SEO4leagalPA1 -
SEO Site Analysis
I am looking for a company doing a SEO analysis on our website www.interelectronix.com and write a optimization proposal incl. a budgetary quote for performing those optimizations.
Intermediate & Advanced SEO | | interelectronix0 -
Weird title tag in SERps (see attachment)
Hi Mozzers Does anyone know why my clients title tag appears like it does in the image attached? It seems as though Google is pulling the parent page url and putting that at the front. All other title tags are normal. Anyone any ideas and is it anything to be worried about? Thanks Anthony @Anthony_Mac85 9KhTgk5
Intermediate & Advanced SEO | | Tone_Agency0 -
Site #2 beats site #1 in every aspect?
Hey guys, loving SEOMoz so far and will definitely continue my subscription after the free trial. I have a question however, which I am really confused about. When researching my primary keyword, I have found that the second ranked site beats the top site in every single aspect, apart from domain age, which is almost 6 years for the top one and 6 months for the second. When I say every single aspect, I mean everything. More authority for the page and domain, more links, more anchor text links, more authoritive links, more social signals, more relevant links, better domain (although second ranked site is a .net), better MozRank, better MozTrust etc.... I have noticed though, that in the UK SERPs, those sites are switched, so #2 is actually #1. Could it be that the US SERPs just haven't updated yet, or am I missing something completely different.
Intermediate & Advanced SEO | | darrenspeed1 -
Help with a Sticky Site
Hey Everyone - I work for a company that is just getting into SEO. We have had some successes, but one project lately has got us stumped. We have been working hard, but have been unable to make an impact in Google rankings with the following site: http://stoneycreekinn.com/locations/index.cfm/DesMoines We are trying to optimize for the keyword phrase, "des moines hotel" This hotel is a branch location of a hotel chain in the Midwest. *Note we've already moved up some other branch locations for this hotel chain successfully. We've used several tools including the SEOmoz tool and seem to have higher marks than those sites that rank above us in Google surprisingly. Any idea what we're missing? Thanks!
Intermediate & Advanced SEO | | markhope0