This might be a silly question...
-
I have 14,000 pages on my website, but when I do a site:domain.com search on google, it shows around 55,000.
I first thought.."hmm, maybe it is including subdomains". So I tried site:www.domain.com and now it shows 35,000. That still is more than double the pages I have.
Any ideas why? When you filter a google search using "site", isn't it meant to pick up just that site's pages?
*P.S I tried using the SEOquake add-on to download search results as a CSV file to review, but the add-on only downloads the first 100 search results
-
Thanks, I'll look at manually specifying these parameters and see if they make an impact.
-
Thank you streamline,
That's interesting, I have provided 'searchType', 'searchTerm', 'search', 'cat', 'filter2name', 'filter1name' as URL Parameters
- Are URL Parameters case sensitive?
- Should these be not set as CRAWL - 'Let Googlebot decide' and instead manually given as best practise? It looks like Google is still indexing from what you guys have found.
-
Easy way to be sure is to do a quick search on Google to see if they are ranking. If you know for sure the Parameters make no difference its usually better to specifically signal that through the WMT console. While Google tend to be pretty smart at these kind of things they can always make mistakes so may as well give as much info as possible.
-
Hi there,
I am doing a crawl on the site listed in your profile (www.abdserotec.com) using Screaming Frog SEO Spider using Googlebot as the User Agent, and I am seeing many more URLs than the 14,000 pages you have. The bulk majority of these excess pages are the Search Results pages (such as http://www.abdserotec.com/search.html?searchType=BASIC&searchTerm=STEM CELL FACTOR&cat=&Filter2Name=GO&Filter2Value=germ-cell development&filterCount=2&type=&filter1name=Spec&filter1value=STEM CELL FACTOR). While these URLs are not showing up in the Google Index when you try searching your site with the site: command, Google is still definitely accessing them and crawling them. As Tuzzell just suggested, I also highly recommend configuring the parameters within GWT.
-
We have 49 Parameters listed and given 'Let Googlebot decide'. I thought adding the parameters here would avoid google from indexing those URLs? I believe our setup already does this?
-
What do you mean by "multiple ways"? We have a search page which isn't indexed and internal links from pages but that wouldn't count would it? It's not like the URL string changes from a search page or internal hyperlink?
-
Have you discounted URL parameters through Google Webmaster tools? This would be particularly prevalent for an ecommerce site as if you have not Google could be looking at /page, /page?p=x, /page?p=y etc and counting these as unique pages. This creates obvious dupe content issues and is easily fixed in WMT by going to:
Crawl>URL Parameters
Hope that helps.
-
what about multiple ways of getting to the same product?
-
There are no blog posts, it's an ecommerce site and every product page and article page has the URL www.domain.com/.
I even looked at my GA and it reports 14,000 pages
If there was a tool to export all the search results, I could've manually looked into why the big count.
-
Hi Cyto,
Does that include your blog pages? If you have a blog, such as Wordpress, then it may be picking up the different URL's that each post may have. So for example, you might have the blog post in different categories which would mean the post is accessible from 2 different URL's
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Ctr question with home page and product pages
do you believe that the advantage of targeting a search term on the home page is now worse off than before? as I understand it ctr is a big factor now And as far as i can see if two pages are equal on page etc the better ctr will win out, the issue with the home page is the serp stars cannot be used hence the ctr on a product page will be higher? I feel if you where able to get a home page up quicker (1 year instead of two) you still lost out in the end due to the product page winning on ctr? do you think this is correct?
Algorithm Updates | | BobAnderson0 -
Domain Migration Question
Lets say there is a brand that has one primary product type at different optional tiers. (Think something like SMB/Enterprise/Individual) Lets also say that 1 year ago this brand migrated from having everything under 1 domain (Domain A) to moving 2 of their product tiers to another domain (Domain B), a new domain. They have done some initial SEO work on this domain and had a pretty successful migration but it has also been decided that they are going to no longer offer one of these product tiers and they intend to eventually migrate everything back under the 1 domain (Domain A) They just are not sure whether they should do this now or later.
Algorithm Updates | | DRSearchEngOpt
During this next year or so there is also going to be some likely re-branding/design, etc...stemming from this decision, on the domain, meaning content changes and all that fun that goes into a migration/re-design/re-branding strategy. The timing of this has not been fully decided on. Here is the question: Should they a) Migrate back to Domain A first and then do the re-design or b) Keep 2 separate domains for now, figure out the re-design/re-branding, make content changes and then migrate Site A over in a year or so after all changes have been made? My concern with option a) is that they migrated a little less than 1 year ago and will be migrating back which I feel could have a negative impact on the content and the domain. The positive side I see here is that this impact could be just as large even if we waited so doing this now might be a better, more efficient use of our time if we can migrate and make content changes fairly close together or concurrently.
My concern with option b) is that the tier they no longer offer makes up the majority of that sites business and traffic, leaving us with not much in terms of content that ranks well and garners much traffic. Trying to optimize for the remaining product tier by itself on it's own domain could be quite hard and then having to migrate it in a year or so back to Domain A could negatively impact any small organic impact I can make on applicable pages/domain. Does anybody have any input here? I am leaning towards Option A and but wanted to get some other opinions. Thanks Everybody! Edit: So far, this has received a lot of views but no input. I am hoping to have a bit of a dialog on this so any ideas or input is welcome.0 -
Mobile Brand Markup Question
Hi Moz Community, I was searching for "Gifts for men" in Google Search on my phone and saw a few results in the 3rd (Nordstrom), 4th (Etsy) and 5th(Grommet) place that had their brand name in the area under the title tag where the green url is usually listed on desktop. One example of the green text under the title tag is Nordstrom which lookes like this: Nordstrom > Shop > Gifts Whereas the first result from UncommonGoods looks like this in the green text: www.uncommongoods.com > by recipient I'm trying to figure out what markup Nordstrom, Etsy, ect used on their site to get their brand name to show up not as a url but as a brandname Anyone know the answer to this? Thanks!
Algorithm Updates | | znotes0 -
Puzzling Penalty Question - Need Expert Help
I'm turning to the Moz Community because we're completely stumped. I actually work at a digital agency, our specialism being SEO. We've dealt with Google penalties before and have always found it fairly easy to identify the source the problem when someone comes to us with a sudden keyword/traffic drop. I'll briefly outline what we've experienced: We took on a client looking for SEO a few months ago. They had an OK site, with a small but high quality and natural link profile, but very little organic visibility. The client is an IT consultancy based in London, so there's a lot of competition for their keywords. All technical issues on the site were addressed, pages were carefully keyword targeted (obviously not in a spammy way) and on-site content, such as services pages, which were quite thin, were enriched with more user focused content. Interesting, shareable content was starting to be created and some basic outreach work had started. Things were starting to pick up. The site started showing and growing for some very relevant keywords in Google, a good range and at different levels (mostly sitting around page 3-4) depending on competition. Local keywords, particularly, were doing well, with a good number sitting on page 1-2. The keywords were starting to deliver a gentle stream of relevant traffic and user behaviour on-site looked good. Then, as of the 28th September 2015, it all went wrong. Our client's site virtually dropped from existence as far as Google was concerned. They literally lost all of their keywords. Our client even dropped hundreds of places for their own brand name. They also lost all rankings for super low competition, non-business terms they were ranking for. So, there's the problem. The keywords have not shown any sign of recovery at all yet and we're, understandably, panicking. The worst thing is that we can't identify what has caused this catastrophic drop. It looks like a Google penalty, but there's nothing we can find that would cause it. There are no messages or warnings in GWT. The link profile is small but high quality. When we started the content was a bit on the thin side, but this doesn't really look like a Panda penalty, and seems far too severe. The site is technically sound. There is no duplicate content issues or plaigarised content. The site is being indexed fine. Moz gives the site a spam score of 1 (our of 11 (i think that's right)). The site is on an ok server, which hasn't been blacklisted or anything. We've tried everything we can to identify a problem. And that's where you guys come in. Any ideas? Anyone seen anything similar around the same time? Unfortunately, we can't share our clients' site's name/URL, but feel free to ask any questions you want and we'll do our best to provide info.
Algorithm Updates | | MRSWebSolutions0 -
VRL Parameters Question - Exclude? or use a Canonical Tag?
I'm trying to figure something out, as I just finished my "new look" to an old website. It uses a custom built shopping cart, and the system worked pretty well until about a year when ranking went down. My primary traffic used to come from top level Brand pages. Each brand gets sorted by the shopping cart and a Parameter extension is added... So customers can click Page 1 , Page 2 , Page 3 etc. So for example : http://www.xyz.com/brand.html , http://www.xyz.com/brand.html?page=1 , http://www.xyz.com/brand.html?page=2 and so on... The page= is dynamic, therefore the page title, meta's, etc are the same, however the products displayed are different. I don't want to exclude the parameter page= completely, as the products are different on each page and obviously I would want the products to be indexed. However, at the same time my concern is that have these parameters might be causing some confusion an hence why I noticed a drop in google rankings. I also want to note - with my market, its not needed to break these pages up to target more specific keywords. Maybe using something like this would be the appropriate measure?
Algorithm Updates | | Southbay_Carnivorous_Plants0 -
Website evaluation questions-what to ask a client
I am looking to make a list of go-to questions when I talk to clients on the phone. What are some good questions that would help me evaluate their website/internet goals?
Algorithm Updates | | StreetwiseReports0 -
Video SEO: Youtube, Vimeo PRO, Wistia, Longtail BOTR Experience and questions
Obviously Video SEO is changing, Google is figuring out how to do it themselves. We are left wondering… Below we have tried to explain what we have learned and how the different sites work and their characteristics (links to graphics provided) Our problem is: We are not getting congruent Google site:apalytics.tv Video filter results. We are wondering how duplicate content may be affecting our results… and if so, why will Youtube not be duplicate and prevent your own site SEO efforts from working. Is Youtube special? Does that include Vimeo too? We see our own duplicate videos on multiple sites in Google results, so it seems it is not duplicate related…? We’d appreciate your experience or add to our questions and work as a community to get this figured out more definitively. Thanks! We’ve tried four video hosting solutions at quite a cost monetarily and in time. 1.) Youtube, which gets all the SEO Juice and gets our clients on to other subjects or potentially competitive content. Iframes just don’t get the results we are looking for. 2.) See Vimeo Image: Vimeo PRO, a $200 year plus solution that allows us to do many video carousels on our own domains hosted on Vimeo, but are very limited in HTML as only CSS content changes are allowed. While we were using Vimeo we allowed the Vimeo.com community to SEO our content directly and they come up often in search results. Due to duplicate content concerns we have disallowed Vimeo.com from using our content and SEOing our content to their domain. However, we have many “portfolios” (micro limited carousal sites on our domains) that continue to carry the content. The Vimeo hosted micro site shows only three videos on Google: site:apalytics.tv During our testing we are concerned that duplicate content is causing issues too, so we are getting ready to shut off the many microsite domains hosted at Vimeo. (Vimeo has an old embed code that allows a NON-iframe embed – but has discontinued it recently) That makes it difficult if not impossible to retain SEO juice for anything other than their simple micro sites that are very limited! 3.) See Wistia Image: Wistia, a $2000 year plus solution that only provides private video site hosting embedding various types of video content on one’s site/s. Wistia has a free account now for three videos and limited plays – it’s a nice interface for SEO but is still different than BOTR. We opted for BOTR because of many other advertising related options, but are again trying Wistia with the free version to see if we can figure out why our BOTR videos are not showing up as hoped. We know that Google does not promise to index and feature every video on a sitemap, but why some are there and others are not and when remains a mystery that we are hoping to get some answers about. 4.) See Longtail Image: Longtail, Bits On The Run, (JW Player author) a $1,000 year plus like Wistia provides private hosting, but it allows a one button YouTube upload for the same SEO meta data and content – isn’t that duplicate content? BOTR creates and submits video sitemaps for your content, but it has not been working for us and it has been impossible to get a definitive answer as I think they too are learning or are not wanting the expose their proprietary methods (which are not yet working for us!) 2O9w0.png 0eiPv.png O9bXV.png
Algorithm Updates | | Mark_Jay_Apsey_Jr.0