Security Basics mailing list archives
RE: Web bots - blocking the bad
From: "Chad" <ncm () xs4all nl>
Date: Wed, 20 Aug 2003 00:26:49 +0200
From: Ben Carr Sent: 19 August 2003 19:34
The place to start is your robots.txt file. This file does/should
From what I see in my logs, compared to my robots.txt file
User-agent: * Disallow: / I still get robots that crawl the site multiple times a day. Seems that not all crawlers/robots follow this rule. Plus, I think the "bad crawlers/robots" out there will not be looking at adhering to the robots.txt standard anyway. Chad --------------------------------------------------------------------------- ----------------------------------------------------------------------------
Current thread:
- Web bots - blocking the bad Kelley Jr, Robert (Aug 19)
- RE: Web bots - blocking the bad Chad (Aug 19)
- RE: Web bots - blocking the bad Horace Pinker (Aug 19)
- <Possible follow-ups>
- Re: Web bots - blocking the bad Ben Carr (Aug 19)
- RE: Web bots - blocking the bad Chad (Aug 19)
- traceroute-like tool for UDP or TCP packets Kent James (Aug 20)
- Re: traceroute-like tool for UDP or TCP packets Sven Pfeifer (Aug 21)
- Re: traceroute-like tool for UDP or TCP packets shawnmer (Aug 21)
- RE: Web bots - blocking the bad Chad (Aug 19)
- RE: Web bots - blocking the bad Chris Santerre (Aug 19)