I asked this question, because I can only get wget to use spider if I provide
a file for it to follow. Otherwise it only reports back with the index.html
then exits... I just can't get it to work.
On Monday 28 February 2005 08:12 am, Craig White wrote:
> On Mon, 2005-02-28 at 07:51 -0700, Nathan England wrote:
> > I am looking for a way to create a list of every file on a web site.
> > Similar to the way wget will check your bookmarks, but I want something
> > that will scan a website and create a list of everything found.
> >
> > Anyone know of anything like this?
>
> ----
> wget has a '--spider' option
>
> man wget
>
> Craig
>
> ---------------------------------------------------
> PLUG-discuss mailing list - PLUG-discuss@lists.plug.phoenix.az.us
> To subscribe, unsubscribe, or to change you mail settings:
> http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss
---------------------------------------------------
PLUG-discuss mailing list -
PLUG-discuss@lists.plug.phoenix.az.us
To subscribe, unsubscribe, or to change you mail settings:
http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss