wget replacement?

Steve Buehler steve at ibapp.com
Thu Apr 1 15:43:00 UTC 2004


At 09:08 AM 4/1/2004, you wrote:
>At 08:42 4/1/2004, you wrote:
>>Does anybody know of a replacement for wget that will work as well as 
>>wget, but will not have the file size limit problem?  wget can't get a 
>>file that is bigger than 2gigs in size.  On the wget mail list, it is 
>>reported as a bug by some and as just a feature request by others.  I am 
>>trying to mirror an ftp directory for a client so they can have a backup, 
>>but one file stops the wget download process.  I can't find a way to 
>>exclude that one file from the wget download so now I have to see if 
>>there is another program out there that can work as well.  Here is the 
>>command that I use.  Yes, I have replaced the server IP with a fictious 
>>one.  The actual IP is for an internet IP.
>>
>>wget --passive-ftp --mirror --no-host-directories --cut-dirs=1 
>>--directory-prefix=/home/SHARE1/ 'ftp://login:password@192.168.1.1/SHARE1/'
>
>rsync (communicating over ssh) should be a perfect solution for you and 
>provide better security and functionality than wget in this case. 
>Something like:
>
># rsync -ave ssh user at remotehost:/path/to/files/* /local/path/
>
>is the basic command. You can use exclude and include directives to finely 
>tune what is or is not mirrored, and rsync will transfer only changed 
>files. Simply an amazing program. Read the man page for more details, 
>since it has *lots* of power and flexibility.
>
>You could also look at curl as an alternate, but I am not very familiar 
>with it.

Well, I can't seem to use rsync.  They have a snap server that doesn't have 
it installed.  Plus, that Snap server version also doesn't allow for ssh 
either.  I guess I will have to try "curl".

thanks
Steve






More information about the redhat-list mailing list