Is there any benefit in using a subdomain redirected to a single page?
-
For example if we have a domain www.bobshardware.com.au and we setup a subdomain sydneysupplies.bobshardware.com.au and then brisbanescrewdrivers.bobshardware.com.au and used those in ad campaigns. Each subdomain being redirected back to a single page such as bobshardware.com.au/brisbane-screw-drivers etc.
Is there a benefit ?
Cheers
-
Thanks Rick. When you say unless links are involved what do you mean?
-
There will be only a single benefit, which is tracking. Separate subdomains will allow you track visitors properly. No positive or negative result - unless links are involved.
-
Having looked at that white board Friday I did find it helpful.
I did just go look at wotif.com.au and lastminute.com.au one of which I do recall using subdomains to divide their sites with. Neither appear to be using it any more. Which would be another indication that subdomains are in fact bad.
Seems to be subdomains are not really the way to go which from my point of view is a shame. It makes more sense to work that way.
-
Hi David,
Rand covered this very topic in a white board friday. Perhaps you may find it helpful and provide insight on what can happen and why he thinks the way he does.
Hope it helps,
Don
-
The main reasoning behind wishing to use a subdomain is more organisational.
Simply looking at having the subdomain house information on a particular topic or item, for instance screwdrivers in Brisbane. Any deals, latest arrivals etc could be found on that particular subdomain. And further to that thinking being able to redirect to a different page for 2 weeks and then bring the original page back with out changing or adding a new url on which it can be found.
Possibly just me and the way I like things organisationally but the idea appealed and I was wondering if there were any benefits or for that matter negatives to running a particular section that way.
-
Hi David. The benefits associated with 301 redirection come from either relocating your site, combining sites, cleaning up 404 pages, aligning page names within your site architecture, things of that nature. If you have links or visits to those third level pages and want to house all pages on your root domain instead of third levels, then 301 redirection would be the way to go. Cheers!
-
There would not be a direct SEO benefit for doing this. There maybe however a benefit in tracking. If you only used that sub-domain for X ad campaign than you would know all traffic from referral sub-domain would be coming from that ad campaign.
There may be some slight non-optimization for doing it this way. Sub-domains are treated as their own domains to a degree, so you are in affect giving the ad-campaign's link to juice to a new domain entirely. Then forwarding that to a specific page. Opposed to just directly giving the link juice an ad campaign can generate to the actual page.
A couple things here depending on the type of ad campaign there may not be any link juice to worry about, like Google's ad words don't pass link juice. However, if you purchased direct advertisement on certain sites you may get some link juice from those ads running.
The second thing is actually a question. What is the purpose of creating a sub-domain to point to a sub directory? Is it just for tracking? Or were you wondering if you could benefit from a sub-domain being treated as a new domain linking to you? If for tracking; I would think there are other tracking methods that could handle referring traffic. If it were in hopes of gaining a new backlink from a different domain than I would say it isn't helpful this way. First because it is simply forwarding to the sub-directory and secondly even it weren't forwarding the link would be considered from the same server and not very helpful anyway.
So in short, no benefit other than a potential way to help with tracking.
Hope that makes sense and helps,
Don
edit some grammar
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
301 Domain Redirect And Old Domain to a New one including pages
Hi, I need to 301 an old domain to a new one (new website) I need to 301 the domain to a new page not the new domain direct for example www.olddomain.co.uk to www.newdomain.co.uk/pagenew Then I need to also 301 all the other pages on the old domain to the new one for example... www.oldmain.co.uk/oldpage to www.newdomain.co.uk/newpage Issue is I can do one or the other not both, I can get the other pages to redirect but then the main domain wont redirect to the correct new page. Or I can get the old domain to redirect but not the internal pages. Thanks
Technical SEO | | David-Sharpe0 -
Does using a reverse proxy to make a subdomain appear as a subdirectory affect SEO?
Using a reverse proxy only makes it appear that a subdomain is really a subfolder. However, links in the end remain the same. Does this have any negative (or positive) impact on SEO? Does it make it difficult from the blog's (subdomain's) sitemap or robots.txt file to be properly read by search engines?
Technical SEO | | rodelmo41 -
Are image pages considered 'thin' content pages?
I am currently doing a site audit. The total number of pages on the website are around 400... 187 of them are image pages and coming up as 'zero' word count in Screaming Frog report. I needed to know if they will be considered 'thin' content by search engines? Should I include them as an issue? An answer would be most appreciated.
Technical SEO | | MTalhaImtiaz0 -
3,511 Pages Indexed and 3,331 Pages Blocked by Robots
Morning, So I checked our site's index status on WMT, and I'm being told that Google is indexing 3,511 pages and the robots are blocking 3,331. This seems slightly odd as we're only disallowing 24 pages on the robots.txt file. In light of this, I have the following queries: Do these figures mean that Google is indexing 3,511 pages and blocking 3,331 other pages? Or does it mean that it's blocking 3,331 pages of the 3,511 indexed? As there are only 24 URLs being disallowed on robots.text, why are 3,331 pages being blocked? Will these be variations of the URLs we've submitted? Currently, we don't have a sitemap. I know, I know, it's pretty unforgivable but the old one didn't really work and the developers are working on the new one. Once submitted, will this help? I think I know the answer to this, but is there any way to ascertain which pages are being blocked? Thanks in advance! Lewis
Technical SEO | | PeaSoupDigital0 -
Duplicate page errors from pages don't even exist
Hi, I am having this issue within SEOmoz's Crawl Diagnosis report. There are a lot of crawl errors happening with pages don't even exist. My website has around 40-50 pages but SEO report shows that 375 pages have been crawled. My guess is that the errors have something to do with my recent htaccess configuration. I recently configured my htaccess to add trailing slash at the end of URLs. There is no internal linking issue such as infinite loop when navigating the website but the looping is reported in the SEOmoz's report. Here is an example of a reported link: http://www.mywebsite.com/Door/Doors/GlassNow-Services/GlassNow-Services/Glass-Compliance-Audit/GlassNow-Services/GlassNow-Services/Glass-Compliance-Audit/ btw there is no issue such as crawl error in my Google webmaster tool. Any help appreciated
Technical SEO | | mmoezzi0 -
301 redirects
Hello. Our site was recently rebuilt, and we switched from using index.php in all the urls to not using it at all. We also changed the names of many of our pages. So the urls have been renamed from "example.com/index.php/old_page_name/" to "example.com/new-page-name/". While we were at it, we changed from "_" to "-" as our word separators in the urls. In the .htaccess file, we have a small block of code that strips out "index.php/" from all requests. This code redirects a request for "example.com/index.php/old_page_name/" to "example.com/old_page_name/" For your information, the code that strips out "index.php/" is: RewriteCond %{THE_REQUEST} ^GET.index.php [NC]
Technical SEO | | nyc-seo
RewriteCond %{THE_REQUEST} !/uSZWTLna/.
RewriteRule (.?)index.php/(.*) /$1$2 [R=301,L] Then we have 301 redirects from "example.com/old_page_name/" to "example.com/new-page-name/" QUESTION 1: Is this two-step redirect approach okay, or would it be better to skip the separate index.php stripping code and simply have 301 redirects that include "index.php" in the urls? QUESTION 2: Will we lose some of the benefit of the links that have to pass through a 301 redirect? QUESTION 3: We have 50 or so redirects. Will this affect performance of the site? How many redirects does it take to start affecting performance? Thank you!0 -
What factors matters the most when using a 301 permanent redirect?
Hi SEOmozzers, I have a client that has couple of duplicates but I am debating if i should just kill those pages or use 301 Permanent redirects. I know SEO moz provides 2 important factors to look at which are PA and link root domain. 1.Which one matters the most? or which one should I look at first to make a decision? 2. I have empty pages creating duplicate content with a PA of 14 and 1 linking root domain. my thought is to kill the page by inserting a meta NO INDEX. If you don't agree and think I should 301 to an existing page that needs link juice, let me know. Thank you mozzers 🙂
Technical SEO | | Ideas-Money-Art0 -
SEO Benefit from Redirecting New Exact Match Domains?
Hi, All! This is a question asked in the old Q & A section, but the answer was a little ambiguous and it was about 3 years ago, so I decided to repost and let the knowledgeable SEO public answer... From David LaFerney: It’s clear that it’s much easier to get high rankings for a term if your domain is an exact match for the query. If you own several such domains that are very related such as – investmentrealestate.com, positivecashflow.com, and rentalproperty.com – would you be able to benefit from those by 301ing them to a single site, or would you have to maintain separate sites to help capture those targeted phrases? In a nutshell – SEO wise, is it worth owning multiple domains to exactly match valuable search phrases? Or do you lose the exact match benefit when you redirect?>> To clarify: redirecting an old domain with lots of history and links to a new exact match domain seems to contain SEO benefit. (You get links+exact match domain, approximately.) But the other way around? Redirecting a new exact match domain to an older domain with links? Does that do anything for the ranking of the old domain for the exact match keyword? Or absolutely nothing? (My impression has been that it's nothing, but the question came up for a client and I just wanted to make sure I wasn't missing something.) Thanks in advance!
Technical SEO | | debi_zyx0