[Bioclusters] ack. I'm getting bitten by the 2gb filesize problem on a linux cluster...

Jeremy Mann bioclusters@bioinformatics.org
Thu, 30 Jan 2003 15:17:02 -0600 (CST)


On Thu, 30 Jan 2003, Jeffrey B. Layton wrote:

> > Now however I've got a problem that the compressed archive file that
> > someone is trying to download is greater than 2gb in size :)
> >
> > The database in question is:
> >
> > ftp://ftp.ncbi.nlm.nih.gov/blast/db/FormattedDatabases/htgs.tar.gz
> >
> > The file is mirrored via 'wget' and a cron script and has recently
> > started core dumping. A ftp session for this file also seemed to bomb
> > out but I have not verified this fully.
> >
> > I did the usual things that one does; verified that the wget binary
> > core dumps regardless of what shell one is using (Joe Landman found
> > this issue a while ago...). I also verified that the error occurs when
> > downloading to a NFS mounted NetApp filesystem as well as a local ext3
> > formatted filesystem.  The node is running Redhat 7.2 with a
> > 2.4.18-18.7 kernel.
> >
> > Next step was to recompile 'wget' from the source tarball with the
> > usual  "-D_ENABLE_64_BIT_OFFSET" and "-D_LARGE_FILES"  compiler
> > directives.
> >
> > Still no love. The wget binary still fails once the downloaded file
> > gets a little larger than 2gb in size.
>
>
> What shell are you running? What filesystem? (was it built under a 2.2
> kernel?)
>
> Jeff
>
> >
> >
> > Anyone seen this before? What FTP or HTTP download clients are people
> > using to download large files?

There is a patch floating around (from Debian) that cures this problem.
The other way to do it is with NCFTP (which does support <2Gb files).



-- 
Jeremy Mann
jeremy@biochem.uthscsa.edu