Hello, i want to disable Googlebot ONLY access to certain .html page, is there any other way except robots.txt? Issue is that i dont want to publicly reveal the fact that im disallowing Googlebot access to that html page. (robots.txt is readable for anyone)
You could probably get all the IPs that the Google bots use and then block them in .htaccess; or you could check the header server side when the request comes in and return a 403 if it identifies as googlebot. Whatever you do is going to be lame and probably unreliable. I honestly cannot think of any valid reason for that kind of nonsense instead of just using robots.txt. When I don't want a page indexed by Google I just put the meta tag in the header: <meta name="googlebot" content="noindex"> But of course anyone looking at the source will be able to see that and know your super special secret.
It's an easy fix. Just go to your Google Webmaster tools and under Google Index, use Remove URLs to request that certain pages do not get indexed.
thx, another think might be also renaming to .php and creating rule to echo the html code only if $_SERVER['HTTP_USER_AGENT'] global variable dont contain phrasse "google" (googlebot, google-mobile)
Good idea, but you don't need to rename it to .php. The existing file extension can be kept, and the server configured to execute .html files as php. For example: http://php.about.com/od/advancedphp/p/html_php.htm Martin