so what happens when I get a second internalcrawler? the problem
inherent w/ robots.txt is the administration --- it needs to be easy!
aside from the sysadmin only problem, requiring the sysadmin to keep a
detailed inventory of all the crawlers and all their behaviors is
unreasonable, and it will force new crawlers to spoof other crawlers
that are similar. (MS IE spoofing Netscape ring any alarm bells for
folks?). The behavior category is more abstract and allows for better
administration than faking it with lots of user-agents.
-- Erik Selberg "I get by with a little help selberg@cs.washington.edu from my friends." http://www.cs.washington.edu/homes/selberg _________________________________________________ This messages was sent by the robots mailing list. To unsubscribe, send mail to robots-request@webcrawler.com with the word "unsubscribe" in the body. For more info see http://info.webcrawler.com/mak/projects/robots/robots.html