Hi folks, I've turned a bunch of Seagate 160gb IDE disks into a large software RAID5 volume and am trying to mirror the raw fasta data from ftp://ftp.ncbi.nlm.nih.gov/blast/db/FASTA/ for use on a personal development project. The 'wget' utility (Redhat 9 on a dual PIII system w/ ext3 filesystem) is bombing out on a few of the remote files which even when compressed are greater than 2gb in size. My kernel and ext3 filesystem support large filesizes but 'wget' or my shell seem to have issues. I've recompiled the wget .src.rpm with the usual compiler flags to add large file support and wget _seems_ to be working but I don't really trust it as it is reporting negative filesizes like this now: > RETR nt.Z...done > Length: -1,668,277,957 [-1,705,001,669 to go] What are others doing? Would 'curl' be better? Any recommendations would be appreciated. -Chris -- Chris Dagdigian, <dag@sonsorol.org> BioTeam Inc. - Independent Bio-IT & Informatics consulting Office: 617-666-6454, Mobile: 617-877-5498, Fax: 425-699-0193 PGP KeyID: 83D4310E iChat/AIM: 'bioteamdag' Web: http://bioteam.net