Title tag solution for a med sized site
-
Its the same old story, we all know it well. I have a client that has a site with 20k+ pages (not too big) and traffic levels around 450k/month.
Now we have identified 15 pages with various conversion points/great backlink metrics etc. that we are going to explicitly target in the first round of recs. However, we are looking at about 18,000 dup title tags that I'd like to clean up.
The site is not on a CMS and in the past I've had the dev team write a script to adopt the h1 tag or the name of the page etc as the title tag. This can cause a problem when some of these pages that are being found in long tail search lose their positions etc. I'm more hesitant than ever to make this move with this current client because they get a ton of long tail traffic spread over a ton of original content they wrote.
How does everyone else usually handle this? Thoughts?
Thanks in advance Mozzers!
-
How about having your developers script something, that scrapes all 18.000 h1, h2, h3 for each article and store them in a database. Finding dupes then would be a piece of cake, even for a less experienced developer You could easily export all your duplicates to csv and then manually rename them based on their content.
Dev time: about 1 day max. (Developed a lot of software myself and IMHO a good developer should get this up and running within 4 hours)
If you don't have toooooo many duplicate tags, correcting those in question shouldn't be taking too long aswell.
If you have done your chores you could reimport your corrected title-tags to the database. Your developer could write a script in the meantime, that sets the title-tag of a page according to the title-tag you stated in your database.
Hope that helped If you have further questions on this, just go ahead. Had a similar problem with 25k+ pages for a major health insurance and we figured out, that the best way to prevent problems was to do most of the work manually than with a script. Helped us a lot to stay within the budget and given timeframe.
-
This is sound advice. Test out a percentage of pages before rolling out the change site-wide.
I also agree that 18K duplicate titles isn't helping the site.
One thing I would do is review analytics and define the top X % pages and hand optimize those. The balance can be optimized via rules utilizing the system you outlined. As to whether to use the H1 or the file name or some other element, I'd probably lean towards the h1 as it would likely accurately describe the content and not be truncated or contain stop words.
-
-
Can you implement on a section or on a % of your pages first? then you can test the effect without risking your whole catalogue
-
Ryan - excellent points! The benefits of adding a CMS to this site would be quite good, at the very least for providing some sort of grounds for moving forward on a unified platform.
-
The 18k is a hard piece of the puzzle to wrap your mind around...I'd like to give more details there but can't...currently. Hopefully when this campaign starts to show results they will let me write a case for it...I'll be sure to share.
There is a "templating system" for various sections. However, as I mentioned, many developers have had their hands in it and didn't follow a standardized system.
I am considering EGOLs comment
Thanks!
-
Is there any form of standardization? I can't imagine 18k pages which were independently developed.
There should be a templating system or some logic which controls code common to all pages. Most pages should share the same header, footer and sidebar, along with standards for things like a canonicalization tag, title and meta description.
If that is not the case, the EGOL's comment should be considered. It is not reasonable to maintain a site which lacks standards.
-
It's possible that putting a reasonably intelligent human on the job for a couple of months could pay back big time. I'll bet a good title tag job would pull in thousands of dollars worth of sales every month.
-
Hey Ryan,
Thanks for the response!
There was 18k title tag duplicates but top content that I can tell is being found in search is about 1,500 pages. Its not a forum site or a site with UGC. Its a very successful tech hardware company that has put out a lot of great unique content over time.
Determining the logic is the tough part because there isn't a lot of consistency throughout the site...different developers have had their hands in it over time.
-
What kind of site is it?
With 18k+ pages I will take a guess that it is a forum site. Definitely check with your forum software provider. There should be some form of "page container" which is used as a template for all the site's pages. If you can determine the logic you want to use, such as go with the post title or H1 tag, then you can modify the template according to your logic and take care of your entire site quickly and easily.
-
Thanks for the response! I should rephrase my question...
I'm either looking for tricks/tips others use in this situation or messages like yours that will give me the confidence to go for it haha.
I think we've all experienced the fear of doing what we know is technically correct and risk being at the mercy of the algo. I've gone this route a lot in the past but I've never done it on a site that gets traffic so deep into so many pages from search.
Have you ever gone the script route? If so, what did you have it pull to use as a title tag? Like I mentioned above, I usually have used h1's in the past..
-
If a lot of traffic is coming in through 18,000 pages that have duplicate title tags I am willing to bet that there will be a huge increase in the amount of traffic that those pages pull when unique and relevant title tags are put in place.
So, although there is a small chance that traffic will go down, I think that there is a much higher chance that traffic will immedately shoot up spectacularly - and the quality of that traffic might also improve.
I would archive the site, run a script to replace the title tags, see what happens. You can always put the old title tags back up if this doesn't work - but I bet it works great.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Htaccess maximum size?
Hello all, The company that develops our website recently contacted and asked me if we could remove a large amount of URL rewrites. I've described a few factors and my main questions below. Some information: One year ago we did a large migration. We went from 27 websites to one main website. We have got about 2000 rewrites in the htaccess file. And the file is 208kb. A lot of links from our old domains still have incoming traffic which are handled by the rewrite rules mentioned above. Questions:
Intermediate & Advanced SEO | | DPA
The company that develops our website said that the htaccess file is too large and is causing or could be causing us website performance issues. They have asked us to remove URL rewrites.
My question is:
a) How many rewrites is too much?
b) Is the filesize of the htaccess of any importance or is it just the amount of rewrites in the file?
c) Could we solve any potential server/website performance issues due to a large htaccess file in any other way? Increasing some values like 'post_max_size' or by any other solutions handled serverside? I do not have a lot of knowledge of htaccess rules but I've seen websites that handled over a million of rewrite rules. This is why I'm having doubts on whether removing URL rewrites is the only solution and possibly not the best solution for us. Hopefully you can help me any further and with the best way to proceed without losing traffic or causing 404 pages. Thanks in advance!
Iordache Voicu0 -
Adult Toys Sites
Does anyone know of any changes SEOwise when running an adult toy site versus a normal eCommerce site? Is there any tips or suggestions that are worth knowing to achieve rankings faster? Thanks,
Intermediate & Advanced SEO | | the-gate-films0 -
My site shows 503 error to Google bot, but can see the site fine. Not indexing in Google. Help
Hi, This site is not indexed on Google at all. http://www.thethreehorseshoespub.co.uk Looking into it, it seems to be giving a 503 error to the google bot. I can see the site I have checked source code Checked robots Did have a sitemap param. but removed it for testing GWMT is showing 'unreachable' if I submit a site map or fetch Any ideas on how to remove this error? Many thanks in advance
Intermediate & Advanced SEO | | SolveWebMedia0 -
Site: inurl: Search
I have a site that allows for multiple filter options and some of these URL's have these have been indexed. I am in the process of adding the noindex, nofollow meta tag to these pages but I want to have an idea of how many of these URL's have been indexed so I can monitor when these have been re crawled and dropped. The structure for these URL's is: http://www.example.co.uk/category/women/shopby/brand1--brand2.html The unique identifier for the multiple filtered URL's is --, however I've tried using site:example.co.uk inurl:-- but this doesn't seem to work. I have also tried using regex but still no success. I was wondering if there is a way around this so I can get a rough idea of how many of these URL's have been indexed? Thanks
Intermediate & Advanced SEO | | GrappleAgency0 -
Regional and Global Site
We have numerous versions of what is basically the same site, that targets different countries, such as United States, United Kingdom, South Africa. These websites use Tlds to designate the region, for example, co.uk, co.za I believe this is sufficient (with a little help from Google Webmastertools) to convince the search engines what site is for what region. My question is how do we tell the search engines to send traffic from other regions besides the above to our global site, which would have a .com TLD. For example, we don't have a Brazilian site, how do we drive traffic from Brazil to our global .com site? Many thanks, Jason
Intermediate & Advanced SEO | | Clickmetrics0 -
Merging Sites: Will redirecting the old homepage to an internal page on the new site cause issues?
I've ended up with two sites which have similar content (but not duplicate) and target similar keywords, rather than trying to maintain two sites I would like to merge the sites together. The old site is more of a traditional niche site and targets a particular set of keywords on its homepage, the new site is more of an authority site with a magazine type homepage and targets the same set of keywords from an internal page. My question is: Should I redirect the old site's homepage to the relevant internal page on the new website...
Intermediate & Advanced SEO | | lara_dar
...or should I redirect the old site's homepage to the new site's homepage? (the old site's homepage backlinks are a mixture of partial match keyword anchor text, naked URLs and branded anchor text) I am in two minds (a & b!) (a) Redirecting to the internal page would be great for ranking as there are some decent backlinks and the content is similar (b) But usually when you do a 301 redirect the homepage usually directs to the new homepage and some of the old site's links are related to the domain rather than the keyword (e.g. http://www.site.com) and some people will be looking for the site's homepage. What do you think? Your help is much appreciated (and hope this makes sense...!)0 -
This site got hit but why..?
I am currently looking at taking on a small project website which was recently hit but we are really at a loss as to why so I wanted to open this up to the floor and see if anyone else had some thoughts or theories to add. The site is Howtotradecommodities.co.uk and the site appeared to be hit by Penguin because sure enough it drops from several hundred visitors a day to less than 50. Nothing was changed about the website, and looking at the Analytics it bumbled along at a less than 50 visitors a day. On June 25th when Panda 3.8 hit, the site saw traffic increase to between 80-100 visitors a day and steadily increases almost to pre-penguin levels. On August 9th/10th, traffic drops off the face of the planet once again. This site has some amazing links http://techcrunch.com/2012/02/04/algorithmsdata-vs-analystsreports-fight/
Intermediate & Advanced SEO | | JamesAgate
http://as.exeter.ac.uk/library/using/help/business/researchingfinance/stockmarket/ That were earned entirely naturally/editorially. I know these aren't "get out of jail free cards" but the rest of the profile isn't that bad either. Normally you can look at a link profile and say "Yep, this link and that link are a bit questionable" but beyond some slightly off-topic guest blogging done a while back before I was looking to get involved in the project there really isn't anything all that fruity about the links in my opinion. I know that the site design needs some work but the content is of a high standard and it covers its topic (commodities) in a very comprehensive and authoritative way. In my opinion, (I'm not biased yet because it isn't my site) this site genuinely deserves to rank. As far as I know, this site has received no unnatural link warnings. I am hoping this is just a case of us having looked at this for too long and it will be a couple of obvious/glaring fixes to someone with a fresh pair of eyes. Does anyone have any insights into what the solution might be? [UPDATE] after responses from a few folks I decided to update the thread with progress I made on investigating the situation. After plugging the domain into Open Site Explorer I can see quite a few links that didn't show up in Link Research Tools (which is odd as I thought LRT was powered by mozscape but anyway... shows the need for multiple tools). It does seem like someone in the past has been a little trigger happy with building links to some of the inner pages.0 -
Scapers and Other Sites Outranking
Post panda, there is definitely more talk about scrapers or other (more authoritative) sites outranking the original content creators in the SERPS. The most common way this problem is addressed (from what I've seen) is by rewriting the content and try your hardest to be the first one to be indexed or just ignoring it from an on page standpoint and do more link dev. Does anyone have any advice on the best way to address? Should site owners be looking deeper into their analytics and diagnostics before doing the rewrites?
Intermediate & Advanced SEO | | Troyville0