Is there a way to prevent Google Alerts from picking up old press releases?
-
I have a client that wants a lot of old press releases (pdfs) added to their news page, but they don't want these to show up in Google Alerts. Is there a way for me to prevent this?
-
Thanks for the post Keri.
Yep, the OCR option would still make the image option for hiding "moo"
-
Harder, but certainly not impossible. I had Google Alerts come up on scanned PDF copies of newsletters from the 1980s and 1990s that were images.
The files recently moved and aren't showing up for the query, but I did see something else interesting. When I went to view one of the newsletters (https://docs.google.com/file/d/0B2S0WP3ixBdTVWg3RmFadF91ek0/edit?pli=1), it said "extracting text" for a few moments, then had a search box where I could search the document. On the fly, Google was doing OCR work and seemed decently accurate in the couple of tests I had done. There's a whole bunch of these newsletters at http://www.modelwarshipcombat.com/howto.shtml#hullbusters if you want to mess around with it at all.
-
Well that is how to exclude them from an alert that they setup, but I think they are talking about anyone who would setup an alert that might find the PDFs.
One other idea I had, that I think may help. If you setup the PDFs as images vs text then it would be harder for Google to "read" the PDFs and therefore not catalog them properly for the alert, but then this would have the same net effect of not having the PDFs in the index at all.
Danielle, my other question would be - why do they give a crap about Google Alerts specifically. There has been all kinds of issues with the service and if someone is really interested in finding out info on the company, there are other ways to monitor a website than Google Alerts. I used to use services that simply monitor a page (say the news release page) and lets me know when it is updated, this was often faster than Google Alerts and I would find stuff on a page before others who did only use Google Alerts. I think they are being kind of myopic about the whole approach and that blocking for Google Alerts may not help them as much as they think. Way more people simply search on Google vs using Alerts.
-
The easiest thing to do in this situation would be to add negative keywords or advanced operators to your google alert that prevent the new pages from triggering the alert. You can do this be adding advanced operators that exclude an exact match phrase, a file type, the clients domain or just a specific directory. If all the new pdf files will be in the same directory or share a common url structure you can exclude using the "inurl:-" operator.
-
That also presumes Google Alerts is anything near accurate. I've had it come up with things that have been on the web for years and for whatever reason, Google thinks they are new.
-
That was what I was thinking would have to be done... It's a little complicated on why they don't want them showing up in Alerts. They do want them showing up on the web, just not as an Alert. I'll let them know they can't have it both ways!
-
Robots.txt and exclude those files. Note that this takes them out of the web index in general so they will not show up in searches.
You need to ask your client why they are putting things on the web if they do not want them to be found. If they do not want them found, dont put them up on the web.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best Way to Handle Multi-Language Sites
In the last year we've made a few significant changes to the structure of our site - namely adding translations for a few languages. We have historically been gaining in organic search by about 10% each month, but in the last two months we've leveled out and seen a slight dip. I am wondering if this has something to do with the addition of the second language, and namely if there's a chance we've been penalized due to duplicate content. We have almost all pages / content on the site translated by a translator, but the way the development works the site will grab the english version if a translation hasn't been added - potentially adding some duplicate content? The URL structure remains the same, other than the addiion of the language - site.com/our-tour vs site.com/de/our-tour We also haven't translated the tour name itself, so that remains the same. Just wondering if anyone has any feedback on best practices here or things I should be looking out for. Thanks in Advance.
On-Page Optimization | | mkgreyound1 -
Google Webmaster Tools Not Showing Correct Data?
Hi, I am experience a weird issue. Google webmaster tools suggested me some HTML improvements a few weeks ago. The suggestions were about duplicate Title Tags and Short Meta Descriptions. I changed the Title Tags and Meta Descriptions. But after 3 Google Updates, webmaster still shows the same suggestion. Please advise Thanks
On-Page Optimization | | Kashif-Amin0 -
Best way to upload products to Magento?
The import/export tool Magento has is really disappointing, especially when it comes to uploading images. Creating products manually is very time consuming. Could anyone recommend a method to upload products through a feed, is there any plugins etc?
On-Page Optimization | | Jseddon920 -
Creative ways to dramatically increase content on ecommerce category pages?
I need to signficantly boost the content on the category pages on my ecommerce website. Currently, they're pretty thin, with some only having approx 50 words of unique content. In the past, I've intentionally kept the content on these pages quite light, to keep the aesthetic a certain way. It's a fashion-based site, so it's very much about the visual. However, with the introduction of Panda, I need to change this mindset. But, there must be slightly more creative ways to boost the content to stop the pages looking too text heavy. I'm not talking hidden text or anything, but ways to break it up in different blocks on the page to make it look natural/relevant, while keeping it looking great. Anyone have any good ideas? Or, any links to ecommerce sites that have employed brilliant methods?
On-Page Optimization | | Coraltoes770 -
What is the best way to execute a geo redirect?
Based on what I've read, it seems like everyone agrees an IP-based, server side redirect is fine for SEO if you have content that is "geo" in nature. What I don't understand is how to actually do this. It seems like after a bit of research there are 3 options: You can do a 301 which it seems like most sites do, but that basically means if google crawls you in different US areas (which it may or may not) it essentially thinks you have multiple homepages. Does google only crawl from SF-based IPs? 302 passes no juice, so probably don't want to do that. Yelp does a 303 redirect, which it seems like nobody else does, but Yelp is obviously very SEO-savvy. Is this perhaps a better way that solves for the above issues? Thoughts on what is best approach here?
On-Page Optimization | | jcgoodrich0 -
Wrong sitelinks & landing pages in Google
I've recently launched a well-optimized website with good-content category landing pages and then I've added a blog to the website (as supporting content to the landing pages, the only links pointing to the blog are from the category landing pages) What happened is that Google is now using the Blog pages as the site - sitelinks and also as the landing pages for most keywords I only have inbound links to the reg. landing pages and none to the blog, how do I get Google to change that? I know I can demote sitelink URL's in webmaster tools, but would that help me with getting the right sitelinks, it sure wont help much with the landing pages Thanks
On-Page Optimization | | Plorex
-J0 -
Major update to site architecture (outline)-Is Google going to drop?
I'm working with a lawyer client who has a table-based, outdated site. Her nav links consist of a jumble of topics and static pages in one long sidebar list on the home page. I'm moving her site to Wordpress and I've recommended that she organize the site based on categories that roughly match the topics/keywords she wants to rank highest for in Google. The site will be much better organized and coded and the URLs for the new launch will be much stronger for SEO by being targeted and coded properly. So the site should rank better after, right? Right??? I know that when Google crawls the new architecture, it's not going to find the expected long sidebar list of internal nav links. It'll find better, more keyword targeted internal nav links. But will that keep the site from getting dropped off page 1? I'm speaking w/ the client tomorrow and if she's going to drop or get bounced around, I feel like I should prepare her and let her know roughly what might happen. I'm thinking based on my current understanding that I should tell her to expect to be bounced around for a few weeks, but in the end she should rank higher than before. What would you do/say?
On-Page Optimization | | bvrob0 -
Confirmation regarding canonical and syndication google tags
Hi, We are in the process of improving our CMS upstream to resolve our duplicate content issues. We were hit pretty hard by the Panda update. One of the steps we have taken is implementation of the canonical link tag across all domains in our site. You see, we are a news release service with muliple channels and websites to represent each. The problem is that a client will submit a release and in many cases the news item is relevant to multiple channels I.E. multiple websites under the same IP range. Site Examples:
On-Page Optimization | | jarrett.mackay
www.hotelnewsresource.com www.restaurantnewsresource.com
www.travelindustrywire.com From a user perspective, it makes sense that they should be able to access the article from the site they are browsing without being redirected to the site we feel carries the most relevance. We hope the canconical tag will resolve this issue for us. I have also read about the syndication tag and was looking for feedback or recommendations if we should implement that also, but it may be overkill as the two tags objectives seem to be similar. I guess my first question is if the syndication tag is only used by Google News. Secondly, and a little off topic is that we also offer an API and like many other sites, I have read, our content partners are now doing better in primary and long tail rankings even thought we are the original source. My assumption is that we should modify the API to force using both caconical and syndication tags as well. Lastly, I´m curious if anyone has tested the original source tag and if we should implement that as well. Thanks everyone. Jarrett0