My name is Henrik Fagrell and I am employed as research assistant at
the Department of Informatics at G=F6teborg University.
I am involved in a project called "The Geography of Cyberspace" (CyberGeo).
The CyberGeo project is concerned with "topographic" and "demographic"
aspects of the Internet in general, and the World Wide Web (WWW) specially.
When considering the statistical information being available for the Interne=
t,
we found a strong focus on quantitative measures such as the number of sites
accesses, the data flow through nodes and the distribution of sites among
domains, i.e. how many new companies, schools a.s.o. are entering the Intern=
et.
However, the information regarding structural par ameters of WWW-sites is
rather limited.
Does anyone know if the is robot availble that collects all the web pages
(raw HTML), from a given site, in a file.
example
bash$ collectpagesfromsite http:/info.metacrawler.com/ > outputfile &
Thanks for your help
Henrik
Henrik Fagrell Telephone: +46-31-773 27 47
Goteborgs University Fax: +46-31-773 47 54
Department of Informatics Email: fagrell@adb.gu.se
Box 3147, 400 10 Goteborg WWW: http://www.adb.gu.se/~fagrell
Sweden