Organizational Research By

Surprising Reserch Topic

Question:How to properly remove URL's from Google's index?


On some of our sites, we now have several thousand pages that dilute our website's keyword density. The website is an MVC site with SEO routing.

If I submit a new sitemap with say only the 2000 or so pages that we want indexed, even though navigating to the diluting pages still works, will Google re-index the site with only those 2000 pages, dropping the superfluous ones?

For example, I want to keep roughly 2000 of the following:

www.mysite.com/some-search-term-1/some-good-keywords
www.mysite.com/some-search-term-2/some-more-good-keywords

And remove several thousand of the following that have already been indexed.

www.mysite.com/some-search-term-xx/some-poor-keywords
www.mysite.com/some-search-term-xx/some-poor-more-keywords

These pages are not actually "removed" as navigating to these URL's still renders a page. Even though there are potentially hundreds of thousands of pages, I only want say 2000 to be re-indexed and retained. The others removed (without having to do these manually).

Thanks.

asked Sep 13, 2013 in Java Interview Questions by rajesh
edited Sep 12, 2013
0 votes
35 views



Related Hot Questions

2 Answers

0 votes

You can tell Google to not crawl specific parts of your site with robots.txt, you can use the sitemap, you can use a few other technics, but in the end, if there is a way yo get to the pages following links or threads from a forums, or pages from somebody else, Google will find it.

If those pages, the good and the bad ones are related enough, you can use the canonical meta to help redirect the traffic to those pages that you want, but they have to be similar, otherwise you will be using the tag badly.

If you have a forum, it will be much better to do some clean up and remove old, inaccurate or unfinished threads, those are not useful and dilute your ranking.

If you have other kind of site, you may provide a description to it or a link so we can tell you a better approach.

 

answered Sep 13, 2013 by rajesh
edited Sep 12, 2013
0 votes

You can tell Google to not crawl specific parts of your site with robots.txt, you can use the sitemap, you can use a few other technics, but in the end, if there is a way yo get to the pages following links or threads from a forums, or pages from somebody else, Google will find it.

If those pages, the good and the bad ones are related enough, you can use the canonical meta to help redirect the traffic to those pages that you want, but they have to be similar, otherwise you will be using the tag badly.

If you have a forum, it will be much better to do some clean up and remove old, inaccurate or unfinished threads, those are not useful and dilute your ranking.

If you have other kind of site, you may provide a description to it or a link so we can tell you a better approach.

 

answered Sep 13, 2013 by rajesh
edited Sep 12, 2013

...