On Fri, 13 Oct 2000, Wes Bateman wrote: : Hey Guys: : : I've been playing around and trying to see how large a given website is : remotely. I've been using wget and du. It can take a long time, and I : really don't want to mirror the site (I erase the data right after I do : it). Seems a shame to waste so much time and bandwidth, when all I really : want is the total space (size in bytes) that a website occupies. I've : been following infinite depth of links, but only remaining in the original : domain. : : There must be a way to get this info, without either sucking the whole : site down, or having access to the webserver? : : Anyone have any ideas, suggestions, etc.? lynx -dump -head $URL | grep Content-Length | cut -d: -f 2 --- Eric Johnson (ej@netasm.com) net.assembly http://netasm.com/ 12629 North Tatum Boulevard #191 602 996-9682 Phoenix, Arizona 85032