How to handle pages I can't delete?
-
Hello Mozzers,
I am using wordpress and I have a small problem. I have two sites, I don't want but the dev of the theme told me I can't delete them.
/portfolio-items/
/faq-items/
The dev said he can't find a way to delete it because these pages just list faqs/portfolio posts. I don't have any of these posts so basically what I have are two sites with just the title "Portfolio items" and "FAQ Items". Furthermore the dev said these sites are auto-generated so he can't find a way to remove them.
I mean I don't believe that it's impossible, but if it is how should I handle them? They are indexed by search engines, should I remove them from the index and block them from robots.txt?
Thanks in advance.
-
great!
-
The code worked fine, thanks again for your help!
-
Most likely, but I would double check with your developer - since I might be unaware of anything out of the ordinary!
-
Thank you for your answer! Can I use this code in the .htaccess?:
Redirect 301 /portfolio-items/ http://mysite.com
-
Great - the info helps. To be honest, this isn't a huge issue if left alone, or if you do the redirect. But if you're worried about users landing on the wrong page, you can 301 redirect to the homepage and that should be just fine.
-
They told me this:
Salman wrote:
Hi There,
These are not pages or posts which you can remove by just clicking on a button. These are custom posts types and you can not directly remove slugs within current setup. You have following options though.
-
Add some code to unregister these custom post types.
-
Use a plugin to re-write slugs
-
Use .htaccess to redirect users when this URL is access. This might work
BEGIN REMOVE SLUG CUSTOM POST TYPE RULES
RewriteRule ^portfolio-items/(.+)/$ /$1 [R=301,L]
RewriteRule ^faq-items/(.+)/$ /$1 [R=301,L]END REMOVE SLUG CUSTOM POST TYPE RULES
I hope you understand that the solution you need is not possible with our current set of options and will require some customization of the theme. We’d love to help you with this customization but we are only able to provide support for basic theme issues and bug fixes. We do provide support for small customizations but unfortunately this request is beyond that.
There aren't any similar pages, I guess I 301 redirect them to the home page. It's three pages, will this have any negative effects?
Thanks for the help, highly appreciated!
-
-
If you can't delete them, you should probably 301 redirect them to the most similar page, or just the homepage. You can redirect them in the .htaccess file or use a redirection plugin: https://wordpress.org/plugins/redirection/
-
Thanks for all the answers! Sorry, I didn't have a chance to respond earlier.
Unfortunately, I still couldn't fix this. I talked to the original devs (on their support forums) and they couldn't help me further with this. It's about these 3 pages:
/themefusion_es_groups/group1/
/portfolio-items/
/faq-items/
I don't know what else I could try..
Best regards!
-
Hi There
Just checking, any luck with this? Can you clarify what you mean by "sites"? Those both look like sub-folders? Can you explain a little more in context how they fit into whatever domain you're on?
-
Where are they linked FROM? Are they pulled in a menu or a sidebar or are they linked internally somewhere else? If they are not being used, you simply need to delete links to them and they will disappear.
-
You an either block them via the robots.txt file or just 301 redirect them to the sites homepage or an applicable page.
I would just 301 redirect them.
Thanks,
Jake
-
Hi,
Quite interesting what you developer says
There is no such thing in Wordpress as Impossible, but let's say if there is no other way to handle it, you could download the Yoast plugin (wordpress seo) and set those pages to noindex, follow and from robots.txt disallow access to the two pages.
Still I would check what is causing the "auto-generated" pages. If it is a bought theme, go ask the original developer. I'm quite sure that there will be an easy way to fix it. (drop me a private message and we can check it out together)
Gr., Keszi
Later Edit: Hint: Maybe you (or your developer) should check into the custom post types.
-
If those pages are being created due to those being tag options you should try deleting the tag to see if it stops wordpress from generating the page.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Local Landing Pages struggling with rankings although I've done most things needed. Any idea?
Hi Mozzers, I am wondering if someone could advise if there's anything obvious here as to why my local landing pages suck ranking wise even though I have done all of the following. http://goo.gl/Lr4HXa I am trying to rank for Garden tool hire Bristol on my landing page. Main category page is garden tool hire Consitant NAP - Citations. Local branch address on Page , in title tag, H1 tag and the address is in on page content which is unique. Schema.org has been set up with address in this aswell etc. Pagination set up and view all page has concanical tag pointing to page 1 Speed not an issue as this is a fast site. Currently all the product links on the page are H3 tags but I've seen this on lots of other sites. All my NAP Citations point to the parent branch pages although I don't have any individual deep links pointing to this page. Unique Content I currently don't have internal links to relevant articles on my blog page as I have those on my main category landing page as you can see here - http://goo.gl/sO9A9U but I can add these as well to all my location specific landing pages if you think it would help. Any thoughts greatly appreciated Pete
Technical SEO | | PeteC120 -
Should I disavow links from pages that don't exist any more
Hi. Im doing a backlinks audit to two sites, one with 48k and the other with 2M backlinks. Both are very old sites and both have tons of backlinks from old pages and websites that don't exist any more, but these backlinks still exist in the Majestic Historic index. I cleaned up the obvious useless links and passed the rest through Screaming Frog to check if those old pages/sites even exist. There are tons of link sending pages that return a 0, 301, 302, 307, 404 etc errors. Should I consider all of these pages as being bad backlinks and add them to the disavow file? Just a clarification, Im not talking about l301-ing a backlink to a new target page. Im talking about the origin page generating an error at ping eg: originpage.com/page-gone sends me a link to mysite.com/product1. Screamingfrog pings originpage.com/page-gone, and returns a Status error. Do I add the originpage.com/page-gone in the disavow file or not? Hope Im making sense 🙂
Technical SEO | | IgorMateski0 -
How can I fix this home page crawl error ?
My website shows this crawl error => 612 : Home page banned by error response for robots.txt. I also did not get any page data in my account for this website ... I did get keyword rankings and traffic data, I am guessing from the analytics account. url = www.mississaugakids.com Not sure really what to do with this ! Any help is greatly appreciated.
Technical SEO | | jlane90 -
Why is this site beating mine? I can't work it out!
I feel my site is not to bad. Needs more work and could be better but what site could not, However i'm bamboozled by some of the sites that are out ranking me on key terms. e.g. Horse Rugs http://www.fasttackdirect.co.uk/products-1-122/Horse_Clothing/HORSE_RUG_SALE.html ranking 8th http://www.centralsaddlery.co.uk/horse/horse-rugs/ not ranking in first 10 pages I dont expect my site (the later) to rank well as its not up there with the bigger players in our industry yet however I can't see why a page like the one i mentioned is ranking so well when as far as i can see its not as well optimized and has very little content. This is not me ranting about it or whining about why i'm not top etc. I just can't work it out and would love sombody to explain the reasons for this? The only thing i can think of is that they have more category with the words "Horse Rugs" in them. Other than that i'm stumped! Ideas on a postcard please!
Technical SEO | | mark_baird0 -
Duplicate Titles Aren't Actually Duplicate
I am seeing duplicate title errors, but when I go to fix the problem, the titles are not actually identical. Any advice? Becky
Technical SEO | | Becky_Converge0 -
Google insists robots.txt is blocking... but it isn't.
I recently launched a new website. During development, I'd enabled the option in WordPress to prevent search engines from indexing the site. When the site went public (over 24 hours ago), I cleared that option. At that point, I added a specific robots.txt file that only disallowed a couple directories of files. You can view the robots.txt at http://photogeardeals.com/robots.txt Google (via Webmaster tools) is insisting that my robots.txt file contains a "Disallow: /" on line 2 and that it's preventing Google from indexing the site and preventing me from submitting a sitemap. These errors are showing both in the sitemap section of Webmaster tools as well as the Blocked URLs section. Bing's webmaster tools are able to read the site and sitemap just fine. Any idea why Google insists I'm disallowing everything even after telling it to re-fetch?
Technical SEO | | ahockley0 -
Can't find mistake in robots.txt
Hi all, we recently filled our robots.txt file to prevent some directories from crawling. Looks like: User-agent: * Disallow: /Views/ Disallow: /login/ Disallow: /routing/ Disallow: /Profiler/ Disallow: /LILLYPROFILER/ Disallow: /EventRweKompaktProfiler/ Disallow: /AccessIntProfiler/ Disallow: /KellyIntProfiler/ Disallow: /lilly/ now, as Google Webmaster Tools hasn't updated our robots.txt yet, I checked our robots.txt in some ckeckers. They tell me that the User agent: * contains an error. **Example:** **Line 1: Syntax error! Expected <field>:</field> <value></value> 1: User-agent: *** **`I checked other robots.txt written the same way --> they work,`** accordign to the checkers... **`Where the .... is the mistake???`** ```
Technical SEO | | accessKellyOCG0 -
What does it mean by 'blocked by Meta Robot'? How do I fix this?
When i get my crawl diagnostics, I am getting a blocked by Meta Robot, which means that my page is not being indexed in the search engines... obviously this is a major issue for organic traffic!!! What does it actually mean, and how can i fix it?
Technical SEO | | rolls1230