Robot Exclusion?
Robot Exclusion?
Is there any easy way to tell web spiders like Googlebot to not index stuff like the captcha graphics?
-
garvinhicking
- Core Developer
- Posts: 30022
- Joined: Tue Sep 16, 2003 9:45 pm
- Location: Cologne, Germany
- Contact:
Re: Robot Exclusion?
Using a robots.txt like this:
should help. I don't know if wildcards are allowed in disallow, then you could do:
To really only filter those. And, you can block any paths/referrers/googlebots via mod_rewrite rules of course.
Sadly, rel=nofollow does not work for <img> tags.
Regards,
garvin
Code: Select all
user-agent:*
disallow:/serendipity/plugin/
Code: Select all
user-agent:*
disallow:/serendipity/plugin/spamblock_captcha*
Sadly, rel=nofollow does not work for <img> tags.
Regards,
garvin
# Garvin Hicking (s9y Developer)
# Did I help you? Consider making me happy: http://wishes.garv.in/
# or use my PayPal account "paypal {at} supergarv (dot) de"
# My "other" hobby: http://flickr.garv.in/
# Did I help you? Consider making me happy: http://wishes.garv.in/
# or use my PayPal account "paypal {at} supergarv (dot) de"
# My "other" hobby: http://flickr.garv.in/
-
garvinhicking
- Core Developer
- Posts: 30022
- Joined: Tue Sep 16, 2003 9:45 pm
- Location: Cologne, Germany
- Contact:
Re: robots
Then you should contact askjeeves and tell them to honour your spiders.txt file.
Of course you could create a block on HTAccess/VirtualHost level, but this is hard to maintain.
Regards,
Garvin
Of course you could create a block on HTAccess/VirtualHost level, but this is hard to maintain.
Regards,
Garvin
# Garvin Hicking (s9y Developer)
# Did I help you? Consider making me happy: http://wishes.garv.in/
# or use my PayPal account "paypal {at} supergarv (dot) de"
# My "other" hobby: http://flickr.garv.in/
# Did I help you? Consider making me happy: http://wishes.garv.in/
# or use my PayPal account "paypal {at} supergarv (dot) de"
# My "other" hobby: http://flickr.garv.in/
-
MySchizoBuddy
- Regular
- Posts: 340
- Joined: Sun Jun 12, 2005 5:28 am
-
garvinhicking
- Core Developer
- Posts: 30022
- Joined: Tue Sep 16, 2003 9:45 pm
- Location: Cologne, Germany
- Contact:
I don't think anything should be disallowed for security reasons. It's only about files accessible via HTTP, and we already have secured those files as good as possible.
Blocking any files is only a matter if people don't like the extra traffic of robots... I have not investigated which files the robots do not need to index.
Regards,
Garvin
Blocking any files is only a matter if people don't like the extra traffic of robots... I have not investigated which files the robots do not need to index.
Regards,
Garvin
# Garvin Hicking (s9y Developer)
# Did I help you? Consider making me happy: http://wishes.garv.in/
# or use my PayPal account "paypal {at} supergarv (dot) de"
# My "other" hobby: http://flickr.garv.in/
# Did I help you? Consider making me happy: http://wishes.garv.in/
# or use my PayPal account "paypal {at} supergarv (dot) de"
# My "other" hobby: http://flickr.garv.in/
