I'm writing a user agent, via Netscapes OLE Automation interface, I have
a dilemma. I want to abide by some standard (I'll use the original one
until something else comes along) but at the same time I don't want my
customers to feel they are using a 'crippled' product because it can't
get at stuff they can manually.
BUT, as a spider/crawler/agent author and someone who care's about
other's I'm going to have to support robots.txt, so user education, i.e.
information supplied with the software explaining why they shouldn't
leave it running overnight with no delays between hit's, is one way I'm
going to address the contradiction of a user 'need' for information/data
now and NOT trampling all over peoples sites with rapid-fire requests.
I'd like to think I'm representative of the majority of developers in
this field.
I need help getting my User Agent: header to a server via the Navigator
interface, all attempts have failed to date, I'm afraid I may be
releasing the software and you'll only know you've been hit if you see
lots of Mozilla/??? hits in sequence.
Any help would be appreciated... of course I could put all the old
sockets stuff in, but that seems a waste when I'm using Netscape's OLE
Automation.
David.
-- [--------------------------------------------------------------------] email:dbanes@ozemail.com.au CServe - 100446.102@compuserve.com http://www.ozemail.com.au/~dbanes MSN - banes@msn.com [--------------------------------------------------------------------]_________________________________________________ This messages was sent by the robots mailing list. To unsubscribe, send mail to robots-request@webcrawler.com with the word "unsubscribe" in the body. For more info see http://info.webcrawler.com/mak/projects/robots/robots.html