Moving from http to https: image duplicate issue?
-
Hello everyone,
We have recently moved our entire website virtualsheetmusic.com from http:// to https:// and now we are facing a question about images.
Here is the deal: All webpages URLs are properly redirected to their corresponding https if they are called from former http links. Whereas, due to compatibility issues, all images URLs can be called either via http or https, so that any of the following URLs work without any redirect:
http://www.virtualsheetmusic.com/images/icons/ResponsiveLogo.png
https://www.virtualsheetmusic.com/images/icons/ResponsiveLogo.png
Please note though that all internal links are relative and not absolute.
So, my question is: Can that be a problem from the SEO stand point? In particular: We have thousands of images indexed on Google, mostly images related to our digital sheet music preview image files, and many of them are ranking pretty well in the image pack search results. Could this change be detrimental in some way? Or doesn't make any difference in the eyes of Google? As I wrote above, all internal links are relative, so an image tag like this one:
Hasn't changed at all, it is just loaded in a https context.
I'll wait for your thoughts on this. Thank you in advance!
-
No problem
-
Great! Glad to know that. Thank you Dimitrii, I appreciated your help very much!
-
Oh, I see. Yeah, there shouldn't be any problems, if someone else links to your images with http. And yes, your assumption is correct
-
Thank you Dimitrii to clarifying, actually all our webpages now load images only via the https://, but since many external websites are hard-linking to many of our images via the regular http:// protocol, I was thinking to allow linking to them the "insecure" way if requested. Do you see my point? So... to better clarify my initial question, let's say Google is spidering one of those external affiliates and finds an image tag like this:
Will Google consider the image found at:
http://www.virtualsheetmusic.com/image.jpg
a duplicate of:
https://www.virtualsheetmusic.com/image.jpg
?? This was my original question...
In any case, I have made some testings today, and I have been able to redirect all images via .htaccess permanently (301) to https:// and looks like even if an image is requested with the http:// from the browser, it shows up correctly because the web browser handles redirects for images in the same way it handles them for the web page itself.
So... my concern should be solved this way. But in case, for any reason, I need to be able to serve the same image from both protocols (http or https) it is my understand that that shouldn't be an issue anyway. Is my assumption correct?
Thanks again.
-
I did quick search, and there are lots of good articles about why images are not duplicate content: http://bfy.tw/9Qy4
-
So, the reason I recommend having images loading only through one resource is the "insecurity" of https connection, if any resources are loaded not over https. You might have seen that sometimes instead of green lock in a browser bar, it can show yellow exclamation mark - that's one of the reasons. And also it's just cleaner, if everything is loaded the same way.
Here is a link to resource about mixed content: https://developers.google.com/web/fundamentals/security/prevent-mixed-content/fixing-mixed-content
-
Thank you Dimitrii for your reply.
Well, your two statements above contradicts each other, in my opinion. You see, what really concerns me is your last suggestion:
"it's better to make sure that images (and all the other resources) available only through one protocol - http or https."
And hence my original concern. Why should we make sure that images are available only through one protocol if you say first that there isn't such thing as duplicate content for images? Why should we concern about that then?
Sorry for my further request for clarification. I really appreciated your help!
-
Howdy.
As far as I understand, there is no such thing as duplicate content just for images. Duplicate content is more for the page as a whole. Especially, since you guys redirected all the links, you shouldn't have any problems, since google will simply "realize" the change.
Now, it's better to make sure that images (and all the other resources) available only through one protocol - http or https.
Hope this helps
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Same content, different languages. Duplicate content issue? | international SEO
Hi, If the "content" is the same, but is written in different languages, will Google see the articles as duplicate content?
Intermediate & Advanced SEO | | chalet
If google won't see it as duplicate content. What is the profit of implementing the alternate lang tag?Kind regards,Jeroen0 -
Browser Cacheing - HTTPS redirects to HTTP
Howdy lovely Moz people. A webmaster redirected https protocol links to http a number of years ago in order to try and capture as many links as possible on a site we now manage. We have recently tried to implement https and realised that because of this existing redirect rule, they are now causing infinite loops when trying to test an http redirect. http redirecting to https redirecting back to http, etc. The https version works by itself weirdly enough. We believe that this is due to the permanent browser caching. So unless users clear their cache, they will get this infinite loop. Does anyone have any advice on how we can get round this? a) index both sites and specify in GSC that the https is the canonical version of the site and hope that Google sees that and removes the http version for the https version b) stick with http as infinite loops will kill the site c) ??????????? Thanks all.
Intermediate & Advanced SEO | | HenryFrance0 -
Duplicate Content: Is a product feed/page rolled out across subdomains deemed duplicate content?
A company has a TLD (top-level-domain) which every single product: company.com/product/name.html The company also has subdomains (tailored to a range of products) which lists a choosen selection of the products from the TLD - sort of like a feed: subdomain.company.com/product/name.html The content on the TLD & subdomain product page are exactly the same and cannot be changed - CSS and HTML is slightly differant but the content (text and images) is exactly the same! My concern (and rightly so) is that Google will deem this to be duplicate content, therfore I'm going to have to add a rel cannonical tag into the header of all subdomain pages, pointing to the original product page on the TLD. Does this sound like the correct thing to do? Or is there a better solution? Moving on, not only are products fed onto subdomain, there are a handfull of other domains which list the products - again, the content (text and images) is exactly the same: other.com/product/name.html Would I be best placed to add a rel cannonical tag into the header of the product pages on other domains, pointing to the original product page on the actual TLD? Does rel cannonical work across domains? Would the product pages with a rel cannonical tag in the header still rank? Let me know if there is a better solution all-round!
Intermediate & Advanced SEO | | iam-sold0 -
How to resolve duplicate content issues when using Geo-targeted Subfolders to seperate US and CAN
A client of mine is about to launch into the USA market (currently only operating in Canada) and they are trying to find the best way to geo-target. We recommended they go with the geo-targeted subfolder approach (___.com and ___.com/ca). I'm looking for any ways to assist in not getting these pages flagged for duplicate content. Your help is greatly appreciated. Thanks!
Intermediate & Advanced SEO | | jyoung2220 -
Duplicate page content query
Hi forum, For some reason I have recently received a large increase in my Duplicate Page Content issues. Currently it says I have over 7,000 duplicate page content errors! For example it says: Sample URLs with this Duplicate Page Content http://dikelli.com.au/accessories/gowns/news.html http://dikelli.com.au/accessories/news.html
Intermediate & Advanced SEO | | sterls
http://dikelli.com.au/gallery/dikelli/gowns/gowns/sale_gowns.html However there are no physical links to any of these page on my site and even when I look at my FTP files (I am using Dreamweaver) these directories and files do not exist. Can anyone please tell me why the SEOMOZ crawl is coming up with these errors and how to solve them?0 -
Fixing Duplicate Content Errors
SEOMOZ Pro is showing some duplicate content errors and wondered the best way to fix them other than re-writing the content. Should I just remove the pages found or should I set up permanent re-directs through to the home page in case there is any link value or visitors on these duplicate pages? Thanks.
Intermediate & Advanced SEO | | benners0 -
Duplicate content for area listings
Hi, I was slightly affected by the panda update on the 14th oct generaly dropping by about 5-8 spots in the serps for my main keywords, since then I've been giving my site a good looking over. On a site I've got city listings urls for certain widget companys, the thing is many areas and thus urls will have the same company listed. What would be the best way of solving this duplicate content as google may be seeing it? I was thinking of one page per company and prominenly listing the areas they operate so still hopefully get ranked for area searches. But i'd be losing the city names in the url as I've got them now for example: mywidgetsite.com/findmagicwidgets/new-york.html mywidgetsite.com/findmagicwidgets/atlanta.html Any ideas on how best to proceed? Cheers!
Intermediate & Advanced SEO | | NetGeek0 -
Page URL Issue
Hey Friend, I am having sort of a problem. I currently have a subpage with the url of: /musclecars/ I also have a subpage at /muscle-cars/muscle-car-restoration.html Obviously my main url is not listed here. My problem is I am trying to rank for the term Muscle Cars but the first URL does not have the keywords seperated so I rank no where. If I type MuscleCars into google I rank though (but nobody types the keyword in like that). So my question is can I create muscle-cars.mydomainname.com and rank well with that? Or is it better to just use mydomainname.com/muscle-cars/ even though that second term I am ranking for already has that in its url?
Intermediate & Advanced SEO | | shandaman0