lynx-dev
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: LYNX-DEV Downloading a whole web site for local offline browsing


From: David Combs
Subject: Re: LYNX-DEV Downloading a whole web site for local offline browsing
Date: Mon, 30 Mar 1998 13:39:30 -0800 (PST)

Rather than telling everyone to go download that three-lettered
program (no, four: wget), and LEARN the darn thing, why not
simply add some of these features to LYNX?

I mean, lynx already does so much of what is necessary,
how much more work is necessary to make it do a few more
things, eg create a SUBTREE of the site, with ie the original
file names?

Asking a lynx-user to learn a whole new program for the very
few times he wants to do this is a lot.

If, as has been requested, regexps are added to the searach
command in lynx, then if that is done the code is already there
for use as a find-like pattern for file-names that should or should
not be loaded.


> From address@hidden Mon Mar 30 10:19:32 1998
> Date: Mon, 30 Mar 1998 18:23:53 +0100
> From: Peter D Read <address@hidden>
> 
> I have been searching for some while for something that will
> allow me to download a complete web site, so that I can browse
> the contents locally at will.
> There is a commercial application "Black Widow" that does this
> for Windows but I am Unix based.
> Lynx seems to be the closest I have got so far.
> Using the -crawl and - traverse options I can get all the
> files at the specified site downloaded.
> However there are two drawbacks to what I would like to
> achieve.
> Firstly the files are stored as text only, as opposed to their original
> html format, and so cannot then be viewed in their intended state
> using a suitable bowser.
> Secondly they are stored as a filename series (lnk00000000.dat
> to lnknnnnnnnn.dat), which means that their filenames give
> no clue to their contents.
> Since the original filenames and locations are known to
> Lynx (otherwise it would not be able to access them),  and
> are read in in html format, is there already a method by which I
> can achieve what I want  ie create a local mirror of the site
> such that all the files bear the original names, are in html,
> and in the same relative position so that they can be accessed
> by a browser? I realise the links might need some adjustment
> to make them local, rather than absolute.
> A third problem might be that Lynx would only access .html
> files, and not other types of file (eg  .gif), but I assume this
> could be overcome,
> I had initially thought of adapting the source code to do this,
> but though it worth discussing this in case it is already
> achievable, or alternatively if anyone knows of any other
> application that can achieve this
> 
> Thanks
> 
> Pete Read
> Systems Manager
> ESSC (Environmental Systems Science Centre)
> The University of Reading
> Harry Pitt Building
> 3 Earley Gate
> Whiteknights, PO BOX 238
> READING RG6 6AL
> U.K.
> Tel: +44 (0)118 9875123 Ext 4205
> Fax: +44 (0)118 9316413
> Home Tel: +44 (0)1264 356492
> e-mail address@hidden (preferred)
>   or   address@hidden
> http://www.nerc-essc.ac.uk
> 

reply via email to

[Prev in Thread] Current Thread [Next in Thread]