Www vs non-www which is better?
-
Is it better to have all your pages point to the www version or non www version.
-
I am needing help with this same thing. Did you ever find a solution to redirecting with yahoo web hosting? TIA
-
Joel, i prefer www version cause i think from a technical perspective, there are several benefits to including the WWW.
- Ability to restrict cookies when using multiple subdomains. Cookies of a main domain (i.e. example.com) are sent to all subdomains: If you are going to have subdomains for other purposes (blog for instance), you may want to differentiate the sites and have a www prefix for the regular site.
- WWW actually MEANS something. As mentioned above, WWW is a hostname, and the hostname names the specific service being used a computer network; WWW names the web service for a domain.
- Using the WWW hostname allows for easy segregation in the file structure of your website. Everything in the “www” folder (and at the www.example.com domain) is directly related to serving the site to the public. This allows for simple root-level site organization, eg you could also have a dev folder and have a subdomain dev.example.com for your development site, etc.
- More flexibility with DNS. Your domain’s “Zone” file controls where traffic to your domain is directed and using the non-WWW version of your domain can complicate things.
you may still want to use the WWW simply because it’s conventional to do so. On a business card, the WWW clearly conveys, This is our address on the World Wide Web. People are used to looking for, and seeing, the WWW and that’s sufficient reason for many to stick to the convention
-
Personally, I'd dump yahoo hosting and have my stuff hosted elsewhere. For less than $40/mo you can get hosting and have access to edit the .htaccess file to your heart's content.
-
I spoke with Yahoo, apparently they only offer the 301 redirect for the higher cost hosting plans that run about $40. Any ideas?
-
-
Ok, does anyone know how to do a proper 301 redirect in yahoo web hosting?
-
As long as your consistent, but it just comes down to which have the higest ranksing if on an existing site.
I tend to prefer non-www for new sites as its less typing and un-necessary.
There is a moment for non-www http://no-www.org/
-
There is no better method they do not affect rankings, it is purely personal preference. However you must implement proper redirect rules to resolve http://mysite.com to http://www.mysite.com or vice versa which ever one you choose.
I tend to always go for www. as it just looks better to me.
-
I prefer www, because folks will generally tend to use that version when they link to you. It's reflex.
But you can check this. Run Open Site Explorer for both versions of your domain.
If more people link to you using 'www' than non-www, use www and 301 redirect the non-www to www.
If more people use non-www, do the reverse.
-
If you do choose to keep the www, make sure you have redirects in place so when a user doesn't enter the www, he or she will get to your home page. Just FYI, www.domain.com is a subdomain of domain.com, so if your site can be access through both, search engines view these as two different pages and possibly split rankings.
-
Neither one is better, but whichever one you choose, make sure you remain consistent for your entire site.
As for me, I use the www because that's what google uses.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google tries to index non existing language URLs. Why?
Hi, I am working for a SAAS client. He uses two different language versions by using two different subdomains.
Technical SEO | | TheHecksler
de.domain.com/company for german and en.domain.com for english. Many thousands URLs has been indexed correctly. But Google Search Console tries to index URLs which were never existing before and are still not existing. de.domain.com**/en/company
en.domain.com/de/**company ... and an thousand more using the /en/ or /de/ in between. We never use this variant and calling these URLs will throw up a 404 Page correctly (but with wrong respond code - we`re fixing that 😉 ). But Google tries to index these kind of URLs again and again. And, I couldnt find any source of these URLs. No Website is using this as an out going link, etc.
We do see in our logfiles, that a Screaming Frog Installation and moz.com w opensiteexplorer were trying to access this earlier. My Question: How does Google comes up with that? From where did they get these URLs, that (to our knowledge) never existed? Any ideas? Thanks 🙂0 -
Discrepancy in actual indexed pages vs search console
Hi support, I checked my search console. It said that 8344 pages from www.printcious.com/au/sitemap.xml are indexed by google. however, if i search for site:www.printcious.com/au it only returned me 79 results. See http://imgur.com/a/FUOY2 https://www.google.com/search?num=100&safe=off&biw=1366&bih=638&q=site%3Awww.printcious.com%2Fau&oq=site%3Awww.printcious.com%2Fau&gs_l=serp.3...109843.110225.0.110430.4.4.0.0.0.0.102.275.1j2.3.0....0...1c.1.64.serp..1.0.0.htlbSGrS8p8 Could you please advise why there is discrepancy? Thanks.
Technical SEO | | Printcious0 -
Redirect non slash to slash
Hello SEO gurus We have an issue here ( www.xyz.com.au) is having 200 responses www.xyz.com.au and www.xyz.com.au/ ( when i ran the crawl test i found this ) We have been advised to do a 301 from non slash to slash ( as our other pages are showing up with slash ) for the consistency we decided to go with this but our devs just couldnt do it. Error is - redirect loop and this site is a wordpress one Can anyone help us with this issue? Help is much appreciated.
Technical SEO | | Pack0 -
Google Search Console Site Map Anomalies (HTTP vs HTTPS)
Hi I've just done my usual Monday morning review of clients Google Search Console (previously Webmaster Tools) dashboard and disturbed to see that for 1 client the Site Map section is reporting 95 pages submitted yet only 2 indexed (last time i looked last week it was reporting an expected level of indexed pages) here. It says the sitemap was submitted on the 10th March and processed yesterday. However in the 'Index Status' its showing a graph of growing indexed pages up to & including yesterday where they numbered 112 (so looks like all pages are indexed after all). Also the 'Crawl Stats' section is showing 186 pages crawled on the 26th. Then its listing sub site-maps all of which are non HTTPS (http) which seems very strange since the site is HTTPS and has been for a few months now and the main sitemap index url is an HTTPS: https://www.domain.com/sitemap_index.xml The sub sitemaps are:http://www.domain.com/marketing-sitemap.xmlhttp://www.domain.com/page-sitemap.xmlhttp://www.domain.com/post-sitemap.xmlThere are no 'Sitemap Errors' reported but there are 'Index Error' warnings for the above post-sitemap, copied below:_"When we tested a sample of the URLs from your Sitemap, we found that some of the URLs were unreachable. Please check your webserver for possible misconfiguration, as these errors may be caused by a server error (such as a 5xx error) or a network error between Googlebot and your server. All reachable URLs will still be submitted." _
Technical SEO | | Dan-Lawrence
Also for the below site map URL's: "Some URLs listed in this Sitemap have a high response time. This may indicate a problem with your server or with the content of the page" for:http://domain.com/en/post-sitemap.xmlANDhttps://www.domain.com/page-sitemap.xmlAND https://www.domain.com/post-sitemap.xmlI take it from all the above that the HTTPS sitemap is mainly fine and despite the reported 0 pages indexed in GSC sitemap section that they are in fact indexed as per the main 'Index Status' graph and that somehow some HTTP sitemap elements have been accidentally attached to the main HTTPS sitemap and the are causing these problems.What's best way forward to clean up this mess ? Resubmitting the HTTPS site map sounds like right option but seeing as the master url indexed is an https url cant see it making any difference until the http aspects are deleted/removed but how do you do that or even check that's what's needed ? Or should Google just sort this out eventually ? I see the graph in 'Crawl > Sitemaps > WebPages' is showing a consistent blue line of submitted pages but the red line of indexed pages drops to 0 for 3 - 5 days every 5 days or so. So fully indexed pages being reported for 5 day stretches then zero for a few days then indexed for another 5 days and so on ! ? Many ThanksDan0 -
Robots.txt on http vs. https
We recently changed our domain from http to https. When a user enters any URL on http, there is an global 301 redirect to the same page on https. I cannot find instructions about what to do with robots.txt. Now that https is the canonical version, should I block the http-Version with robots.txt? Strangely, I cannot find a single ressource about this...
Technical SEO | | zeepartner0 -
My www. domain has less page authroity than my non www.
I have a webmaster working on the site and I don't really remember how it got this way but it now appears that the non www have more PA than the www. I am redirecting it so I guess it's a good thing? should I move the site so that it's the root and www is the redirect? any help is greatly appreciated!
Technical SEO | | ecb090 -
2 sites versus a subdomain: Which is better?
I have a client that sponsors a couple of events during the year. They currently have pages within a single website for these events but are interested in creating a separate website so they can brand the events differently. I'm not sure this is the most effective way to do it for fear of losing the "google juice" already there for these pages.Here's what I'm thinking is a better strategy: 1) Host the content both on the main domain and the sub-domain2) Make sure there is a tag on each page of the sub-domain version that points to the main version.That will give them the branding they are seeking while pushing all juice across to the main domain.What are your thoughts?
Technical SEO | | Britewave0 -
Capitals URLs to Non Capitals...
Hi, I am working on a website which has capital urls and non capital urls which will be generating duplicate content, and I know it is better to use all lower case. The problem is that the page authority is better for the capital versions and I was wondering will it negatively impact the SEO of we 301 redirect the uppercase urls to the lowercase counterparts? Thanks.
Technical SEO | | J_Sinclair0