Are you a Google Analytics enthusiast?
More SEO Content
Banned On Google...why?
Posted 01 November 2006 - 04:20 PM
Hopefully this is the right category for this type of a question, if not I apologize in advance.
I have a client that until about two months ago was fully indexed in Google but is no longer indexed by Google. This site has had no purposeful SEO work done on it. There has never been any link recruiting, buying, exchanging and the like done for it and there has been no content or code optimization done for it.
That being said, because this is a publicly traded company the site does have quite a few backlinks and ranked well for very specific keyword terms.
From what I can tell there is no reason why this site should be banned. In an effort to get the site reindexed we have added a Google Sitemap and Google Analytics to the site. The Google Webmaster tool shows that the site map is good but says that Google has not yet indexed the site, it does not report that the site is banned. This tool also shows that the robots.txt file is not blocking access. The weirdest thing is that Google is still showing statistical results like average page ranking, ranked keywords etc... for a site they claim has never been indexed.
I have waited for about a month and a half now hoping that they would eventually fix a glitch or ommission in the index but to no avail.
I would appreciate it if anyone could give me some feedback about why this site may have been excluded/banned from the index so that I could make some changes and re-submit to the Google Reinclusion tool.
Thank you for your help.
Posted 01 November 2006 - 04:22 PM
Posted 01 November 2006 - 04:28 PM
So I guess in answer to that question I would say no I don't think this problem is related to having "no votes," for it.
Any other thoughts?
Posted 01 November 2006 - 04:31 PM
Posted 01 November 2006 - 04:56 PM
Posted 01 November 2006 - 04:58 PM
Posted 01 November 2006 - 05:02 PM
Not that this is bad, mind you. You want them to do one or the other, and you're doing it with a 301, which is the correct method.
However if all of the pages were indexed with the www. address previously, none of those pages now exist. Your server is telling the engines that.
If this is the problem will it correct itself? Yes, eventually. But it'll take 4-6 months for the engines to catch up.
If it won't mess up anything else, you could force the www.domain.com address, doing exactly the opposite of what's now being done. Then the site might pop back into the Google SERPs pretty quickly.
But whatever you do, you'll want to pick one address or the other and stick with it. Changing back and forth can cause some real heartbreak.
Posted 01 November 2006 - 06:02 PM
So this particular issue should not have caused the all out ban on Google.
Any other thoughts?
Posted 01 November 2006 - 06:10 PM
We have seen Google Sitemaps spider visit the site over the last month 24 different times.
As far as I can tell from our analytics the numbers I quoted above included all visits from GoogleBot.
Posted 01 November 2006 - 08:41 PM
Do you have all your pages or just some in Google sitemap?
Posted 02 November 2006 - 10:43 AM
I'm not sure how the engines or Google specifically would react to what's currently in your robots.txt file. It says:
I'm nowhere near and expert on the robots.txt standards (Alan!?!?) but I don't believe Allow: is a valid entry. Typically to allow all spiders you would use something like:
I seem to recall somewhere that Google does read the non-standard Allow: entry. If this is so, would if follow that your robots.txt file is actually telling them to go away because it's the opposite of an empty Disallow: ? I'm not at all sure of the logic, but it would be worth a try. So you might want to either completely remove your robots.txt file, or upload a blank robots.txt or even upload one that uses the robots.txt standard with a blank Disallow: line.
Posted 02 November 2006 - 11:56 AM
Posted 02 November 2006 - 01:47 PM
I'm just wondering aloud if the Allow: may be triggering something in Googlebot and the Google robots.txt checker isn't designed to catch such an odd misconfiguration. Or in other words, some kind of fluke.
I've never heard of it happening before, but I don't run across people using an Allow: instruction much either.
Posted 02 November 2006 - 04:15 PM
We've updated our robots.txt file to the standard disallow:
Hopefully this will help do the trick.
Yesterday evening GoogleBot hit our site for the first time in 6 months so I think fixing the 301 redirect triggered something at Google. Unfortunately, it still didn't index anything so hopefully changing this error in the robots.txt file will help get us over the hump.
If anyone else has any other good suggestions please chime in, this thread has been extremely helpful for us.
Posted 21 November 2006 - 04:28 PM
0 user(s) are reading this topic
0 members, 0 guests, 0 anonymous users