Link Juice passing through a redirect of a disallowed URL
-
Hey guys!
Suppose I disallow search bots from indexing anything on my secure server in my robots.txt, and 301 redirect all of my secure server traffic to my non-secure site. Will the search bots see the redirect before they realize that they're disallowed from accessing that page? Or will they see that page is disallowed and not follow the redirect?
Should I change my robots.txt to allow search bots to crawl my secure site so they can find the redirects?
-
Thanks Nick!
-
If you think there are any links out there that point to your secure pages and want to preserve any SEO related value from those links, you should remove the disallow from robots.txt and use 301 redirects. However, if you redirect your secure pages to non-secure, what is the point of having the secure pages? They would never be used since users would be 301 redirected to the non-secure address.
If you are trying to keep the secure pages out of the index to avoid duplicate content, and want users to be able to have access to the secure pages, disallow in robots.txt and remove the 301 redirects.
In the case of pages that don't really need to be secure, but are available both as http and https, you could also use the canonical tag to show either the secure or non-secure as the original. Individually redirecting the potentially duplicate pages rather than a sitewide https to http redirect would preserve any "juice" you may be getting from links to either version.
If it is more complicated, take a look at this thread on Googles support forum: http://www.google.com/support/forum/p/Webmasters/thread?tid=6a7fac6f81489bc2&hl=en
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Redirects
Hello, i would like to ask which is the best redirect to use in a website now i have all my links to go to https://www.anavasis.gr Which is the best between: https://www.anavasis.gr https://anavasis.gr https://www.anavasis.gr/index.php https://anavasis.gr/index.php thanks in advance
Technical SEO | | anavasis0 -
Canonical sitemap URL different to website URL architecture
Hi, This may or may not be be an issue, but would like some SEO advice from someone who has a deeper understanding. I'm currently working on a clients site that has a bespoke CMS built by another development agency. The website currently has a sitemap with one link - EG: www.example.com/category/page. This is obviously the page that is indexed in search engines. However the website structure uses www.example.com/page, this isn't indexed in search engines as the links are canonical. The client is also using the second URL structure in all it's off and online advertising, internal links and it's also been picked up by referral sites. I suspect this is not good practice... however I'd like to understand whether there are any negative SEO effectives from this structure? Does Google look at both pages with regard to visits, pageviews, bounce rate, etc. and combine the data OR just use the indexed version? www.example.com/category/page - 63.5% of total pageviews
Technical SEO | | MikeSutcliffe
www.example.com/page - 34.31% of total pageviews Thanks
Mike0 -
Canconical tag on site with multiple URL links but only one set of pages
We have a site www.mezfloor.com which has a number of Url's pointing at one site. As the url's have been in use for many years there are links from many sources include good old fashioned hard copy advertising. We have now decided that it would be better to try to start porting all sources to the .co.uk version and get that listing as the prime/master site. A couple of days ago I went through and used canonical tags on all the pages thinking that would set the priority and that would also strengthen the page in terms of trust due to the reduced duplication. However when I went to scan the site in MOZ the warning that the page redirects came up and I am beginning to think that I need to remove all these canonical tags so that search engines do not get into a confused spiral where we loose the little page rank we have. Is there a way that I can redirect everything except the target URL without setting up a separate master site just for all the other pages to point at.
Technical SEO | | Eff-Commerce0 -
What to do with a 302 redirect after a while
Hi guys, A client of ours has a website with a very bad linkprofile. We adressed this issue and we migrated the website to another domain. We redirected the bad website (cornelisbedding.be) to the new domain (cornelisbedding.com) with a 302 redirect. We didn't want to pass the bad link juice. The problem we are having now is that we can't afford to lose the redirect on cornelisbedding.be. We would lose to much traffic because the old domain still has alot of links that generate good quality traffic. I have read that Google will treat 302 redirects as 301's in the long run. We really want to avoid this.
Technical SEO | | Jacobe
We were thinking of using a meta refresh with a delay on, but in Google's eyes that would be considered spammy. Are their any other suggestions on how to handle this? Thanks you!0 -
How can I see the SEO of a URL? I need to know the progress of a specific landing-page of my web. Not a keyword, an url please. Thanks.
I need to know the evolution on SEO of a specific landing-page (an URL) of my web. Not a keyword, a url. Thanks. (Necesito saber si es posible averiguar el progreso de una URL específica en el posicionamiento de Google. Es decir, lo que hace SEOmoz con las palabras clave pero al revés. Yo tengo una url concreta que quiero posicionar en las primeras posiciones de Google pero quiero ver cómo va progresando en función a los cambios que le voy aplicando. Muchas gracias)
Technical SEO | | online_admiral0 -
Spider Indexed Disallowed URLs
Hi there, In order to reduce the huge amount of duplicate content and titles for a cliënt, we have disallowed all spiders for some areas of the site in August via the robots.txt-file. This was followed by a huge decrease in errors in our SEOmoz crawl report, which, of course, made us satisfied. In the meanwhile, we haven't changed anything in the back-end, robots.txt-file, FTP, website or anything. But our crawl report came in this November and all of a sudden all the errors where back. We've checked the errors and noticed URLs that are definitly disallowed. The disallowment of these URLs is also verified by our Google Webmaster Tools, other robots.txt-checkers and when we search for a disallowed URL in Google, it says that it's blocked for spiders. Where did these errors came from? Was it the SEOmoz spider that broke our disallowment or something? You can see the drop and the increase in errors in the attached image. Thanks in advance. [
Technical SEO | | ooseoo](<a href=)" target="_blank">a> [
](<a href=)" target="_blank">a> LAAFj.jpg
0 -
Can I reduce link count by no following links?
Hi, A large number of my pages contain over 100 links. This is due to a large drop down navigation which is on every page. To reduce my link count could I just no follow these navigation links or would I have to remove the navigation completely?
Technical SEO | | moesian0 -
Linking from other language websites passes juice or not?
If i get links from websites with different language than english - has the same sort of field (business type) ... will that pass juice or not? Is it worth linking or not?
Technical SEO | | mosaicpro0