A Windows 98 & ME forum. Win98banter

If this is your first visit, be sure to check out the FAQ by clicking the link above. You may have to register before you can post: click the register link above to proceed. To start viewing messages, select the forum that you want to visit from the selection below.

Go Back   Home » Win98banter forum » Windows 98 » General
Site Map Home Authors List Search Today's Posts Mark Forums Read Web Partners

Application to download web site contents



 
 
Thread Tools Display Modes
  #1  
Old August 14th 11, 09:50 PM posted to microsoft.public.win98.gen_discussion
Franc Zabkar
External Usenet User
 
Posts: 1,702
Default Application to download web site contents

In the past I used a small utility that could download the content of
a web site starting from a particular URL. You could specify whether
or not to follow off-site links, and you could limit the number of
levels. Unfortunately I can't remember it's name.

Could anyone recommend a similar application? I'm aware of Leech, but
the free version appears to be limited in some way.

- Franc Zabkar
--
Please remove one 'i' from my address when replying by email.
  #2  
Old August 14th 11, 10:00 PM posted to microsoft.public.win98.gen_discussion
98 Guy
External Usenet User
 
Posts: 2,951
Default Application to download web site contents

Franc Zabkar wrote:

In the past I used a small utility that could download the content
of a web site starting from a particular URL.


I don't know of any "small" utility that can grab web-content, other
than wget.exe.

You could specify whether or not to follow off-site links, and you
could limit the number of levels. Unfortunately I can't remember
it's name.

Could anyone recommend a similar application? I'm aware of Leech,
but the free version appears to be limited in some way.


HT Track:

http://www.httrack.com/

==========
HTTrack is a free (GPL, libre/free software) and easy-to-use offline
browser utility.

It allows you to download a World Wide Web site from the Internet to a
local directory, building recursively all directories, getting HTML,
images, and other files from the server to your computer. HTTrack
arranges the original site's relative link-structure. Simply open a page
of the "mirrored" website in your browser, and you can browse the site
from link to link, as if you were viewing it online. HTTrack can also
update an existing mirrored site, and resume interrupted downloads.
HTTrack is fully configurable, and has an integrated help system.
===========

It might work on win-98 as is, and probably almost certainly with
kernelEx.
  #3  
Old August 14th 11, 10:07 PM posted to microsoft.public.win98.gen_discussion
J. P. Gilliver (John)
External Usenet User
 
Posts: 1,554
Default Application to download web site contents

In message , Franc Zabkar
writes:
In the past I used a small utility that could download the content of
a web site starting from a particular URL. You could specify whether
or not to follow off-site links, and you could limit the number of
levels. Unfortunately I can't remember it's name.


Webstripper? I used to have (I think) that (possibly in dial-up days!).

Could anyone recommend a similar application? I'm aware of Leech, but
the free version appears to be limited in some way.

- Franc Zabkar


(Sorry, no; there's still something called webstripper, but it costs $40
or 30 Euros.)
--
J. P. Gilliver. UMRA: 1960/1985 MB++G.5AL-IS-P--Ch++(p)Ar@T0H+Sh0!:`)DNAf

.... "Peter and out." ... "Kevin and out." (Link episode)
  #4  
Old August 14th 11, 11:03 PM posted to microsoft.public.win98.gen_discussion
Lostgallifreyan
external usenet poster
 
Posts: 1,562
Default Application to download web site contents

Franc Zabkar wrote in
:

In the past I used a small utility that could download the content of
a web site starting from a particular URL. You could specify whether
or not to follow off-site links, and you could limit the number of
levels. Unfortunately I can't remember it's name.

Could anyone recommend a similar application? I'm aware of Leech, but
the free version appears to be limited in some way.

- Franc Zabkar


Total recommendation for WGET. Commandline tool, indispensible stuff..
http://www.gnu.org/s/wget

It won't do multiesegment downloads,. but it's both polite and remorseless,
just what we need.

Whether it limits the extent of remote linkage I don't know, but I think it
does, it certainly has plenty of intelligent control on it.
  #5  
Old August 15th 11, 04:53 AM posted to microsoft.public.win98.gen_discussion
Sjouke Burry[_2_]
External Usenet User
 
Posts: 21
Default Application to download web site contents

Lostgallifreyan wrote:
Franc Zabkar wrote in
:

In the past I used a small utility that could download the content of
a web site starting from a particular URL. You could specify whether
or not to follow off-site links, and you could limit the number of
levels. Unfortunately I can't remember it's name.

Could anyone recommend a similar application? I'm aware of Leech, but
the free version appears to be limited in some way.

- Franc Zabkar


Total recommendation for WGET. Commandline tool, indispensible stuff..
http://www.gnu.org/s/wget

It won't do multiesegment downloads,. but it's both polite and remorseless,
just what we need.

Whether it limits the extent of remote linkage I don't know, but I think it
does, it certainly has plenty of intelligent control on it.


Wget does need a bit of supervision, it surprised me once by trying to
download a complete .GOV website....
I stopped it after the first 15 GB
(usage: WGET -r url)
  #6  
Old August 15th 11, 06:53 AM posted to microsoft.public.win98.gen_discussion
Hot-Text
external usenet poster
 
Posts: 81
Default Application to download web site contents

"Franc Zabkar" wrote in message
...
In the past I used a small utility that could download the content of
a web site starting from a particular URL. You could specify whether
or not to follow off-site links, and you could limit the number of
levels. Unfortunately I can't remember it's name.

Could anyone recommend a similar application? I'm aware of Leech, but
the free version appears to be limited in some way.


Mr. Franc Zabkar

Try looking here First:

http://www.majorgeeks.com/
http://www.oldapps.com/

--
http://hot-text.ath.cx

  #7  
Old August 15th 11, 07:27 AM posted to microsoft.public.win98.gen_discussion
Sanity Clause
external usenet poster
 
Posts: 22
Default Application to download web site contents


"Franc Zabkar" wrote:

In the past I used a small utility that could download the content of
a web site starting from a particular URL. You could specify whether
or not to follow off-site links, and you could limit the number of
levels. Unfortunately I can't remember it's name.

Could anyone recommend a similar application? I'm aware of Leech, but
the free version appears to be limited in some way.


In good 'ol IE, there was Favorites, Add to Favorites, Make Available Offline,
with a Customize button to set how many levels and stuff. I think it only saved
as a singe huge "archive" file, instead of making a separate page-by-page
save that could be viewed outside IE.


  #8  
Old August 15th 11, 03:09 PM posted to microsoft.public.win98.gen_discussion
Lostgallifreyan
external usenet poster
 
Posts: 1,562
Default Application to download web site contents

Sjouke Burry wrote in
:

Lostgallifreyan wrote:
Franc Zabkar wrote in
:

In the past I used a small utility that could download the content of
a web site starting from a particular URL. You could specify whether
or not to follow off-site links, and you could limit the number of
levels. Unfortunately I can't remember it's name.

Could anyone recommend a similar application? I'm aware of Leech, but
the free version appears to be limited in some way.

- Franc Zabkar


Total recommendation for WGET. Commandline tool, indispensible
stuff.. http://www.gnu.org/s/wget

It won't do multiesegment downloads,. but it's both polite and
remorseless, just what we need.

Whether it limits the extent of remote linkage I don't know, but I
think it does, it certainly has plenty of intelligent control on it.


Wget does need a bit of supervision, it surprised me once by trying to
download a complete .GOV website....
I stopped it after the first 15 GB
(usage: WGET -r url)


Ouch. Without knowing more than basic recursion switching, I'd aim to find
a small branch of the site I wanted intact, along with a few other branches,
and make a download list for them. It could be awkward if I don't know how
much I missed on any branch, but mostly I use it on lists of single files so
I never explored it enough (and the branch scheme directly extends from this
simple method). I think maybe there's a limit to how smart a tool like that
can be, it's like a proxy scanner in that there's no substitute for trying to
limit it based on perceptions we can make, but it cannot..
  #9  
Old August 15th 11, 03:12 PM posted to microsoft.public.win98.gen_discussion
Lostgallifreyan
external usenet poster
 
Posts: 1,562
Default Application to download web site contents

"Sanity Clause" wrote in
:

In good 'ol IE, there was Favorites, Add to Favorites, Make Available
Offline, with a Customize button to set how many levels and stuff. I
think it only saved as a singe huge "archive" file, instead of making a
separate page-by-page save that could be viewed outside IE.


That could be a neat method, as you can probably resave from the bulk
archive, as single web pages with images. Not that I'd ever use a browser for
bulk tasks though, I'd always want to batch script something like that.
  #10  
Old August 15th 11, 09:37 PM posted to microsoft.public.win98.gen_discussion
Franc Zabkar
External Usenet User
 
Posts: 1,702
Default Application to download web site contents

On Mon, 15 Aug 2011 09:09:59 -0500, Lostgallifreyan
put finger to keyboard and composed:

Sjouke Burry wrote in
:


Wget does need a bit of supervision, it surprised me once by trying to
download a complete .GOV website....
I stopped it after the first 15 GB
(usage: WGET -r url)


Ouch. Without knowing more than basic recursion switching, I'd aim to find
a small branch of the site I wanted intact, along with a few other branches,
and make a download list for them.


Thanks very much to all for your suggestions.

I found that the following command line worked well:

wget -r -nc -l 10 -np URL -o capture.log

... where -l = little L

I needed two attempts because the default number of levels (5) wasn't
enough.

The "-np" switch excludes parent directories, and "-nc" (no clobber)
prevents overwriting of files which have already been downloaded.

- Franc Zabkar
--
Please remove one 'i' from my address when replying by email.
 




Thread Tools
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

vB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Forum Jump

Similar Threads
Thread Thread Starter Forum Replies Last Post
Free Download Animated Screensavers from our site gv General 5 January 20th 08 07:26 AM
Unable to download only one site! Doug General 2 April 15th 07 09:08 PM
New software submission / download site [email protected] General 0 February 20th 06 01:56 AM
Unable to download updates from windows site Pinto1uk General 1 May 19th 05 01:03 AM
WIN98 Boot Disk download site needed mrbigbry General 3 February 22nd 05 07:11 PM


All times are GMT +1. The time now is 07:45 PM.


Powered by vBulletin® Version 3.6.4
Copyright ©2000 - 2024, Jelsoft Enterprises Ltd.
Copyright ©2004-2024 Win98banter.
The comments are property of their posters.