[BioBrew Users] Unable to get mpiblast running
Glen Otero
glen at callident.com
Tue Apr 11 00:50:21 EDT 2006
Bastian-
Any progress running mpiBLAST?
Glen
On Apr 6, 2006, at 2:27 PM, Bastian Friedrich wrote:
> Hi Glen,
>
> thank you for your quick response.
>
> On Thursday 06 April 2006 15:53, Glen Otero wrote:
>>
>> I usually don't see these types of errors. Here are a few questions:
>>
>> How did you format the database for mpiblast?
>
> /usr/local/bin/mpiformatdb --nfrags=28 -i Hs.seq.uniq
> was the latest call, but I had used
> mpiformatdb -N 28 -i Hs.seq.uniq
> earlier.
>
>> Is the mpiblast database on a shared filesystem, like NFS (I don't
>> think symlinks will work)?
>
> Currently, I have created a /export/data/blastdb/ on the frontend;
> this
> was rsynced to /state/partition1/blastdb on the compute nodes; on the
> frontend, I had a directory /state/partition1 (on the root
> partition...) containing a symlink to /export/data/blastdb.
>
> I have just used a bind mount on the frontend (no more symlinking),
> but
> this was not successful, either.
>
> The first tests were done via NFS, which did not work either.
>
>> How did you launch the job, SGE?
>
> In the future, we surely want to use mpiblast in an SGE environment;
> currently, it was started from the command line.
>
>> Can you try a smaller job using just the 6 compute nodes (and
>> formatting the db into 6 pieces)?
>
> Wow, I get a new one now:
> ===============================
> bastian at frontend:/state/partition1/blastdb> mpiformatdb -N 6 -i
> Hs.seq.uniq
> [...]
> [... semi-manual distributing data to /state/partiton1/blastdb of all
> nodes ...]
> bastian at frontend:/state/partition1/blastdb> cd ~/tmp03
> bastian at frontend:~/tmp03> /opt/mpich/gnu/sbin/cleanipcs
> bastian at frontend:~/tmp03> cluster-fork /opt/mpich/gnu/sbin/cleanipcs
> [...]
> bastian at frontend:~/tmp03> mpirun -np 6 /usr/local/bin/mpiblast -p
> blastn
> -d Hs.seq.uniq -i IL2RA -o blast_results
> 54p3_2934: p4_error: : 0
> 3 0.078125 Bailing out with signal 11
> [3] MPI Abort by user Aborting program !
> [3] Aborting program!
> 2p1_28697: p4_error: interrupt SIGx: 13
> p5_17962: p4_error: : 0
> 0.0742188 Bailing out with signal 11
> [5] MPI Abort by user Aborting program !
> [5] Aborting program!
> p4_21219: p4_error: : 0
> rm_l_4_21279: (0.367188) net_send: could not write to fd=5, errno =
> 104
> 0.078125 Bailing out with signal 11
> [4] MPI Abort by user Aborting program !
> [4] Aborting program!
> p2_13443: p4_error: : 0
> 0.078125 Bailing out with signal 11
> [2] MPI Abort by user Aborting program !
> [2] Aborting program!
> rm_l_3_2994: (0.644531) net_send: could not write to fd=5, errno = 104
>
> p1_28697: (7.242188) net_send: could not write to fd=5, errno = 32
> rm_l_2_13503: (6.929688) net_send: could not write to fd=5, errno =
> 104
> p2_13443: (6.929688) net_send: could not write to fd=5, errno = 32
> p5_17962: (6.093750) net_send: could not write to fd=5, errno = 32
> ===============================
>
> Signal 11 seems to be a segfault? Something's going awfully wrong
> here...
>
>> Can you try a smaller blast job using p53, p53db from ftp://
>> ftp.bioinformatics.org/pub/biobrew/ and blastp?
>
> This works! :)) The first time I see mpiblast actually working :)
>
> Unfortunately, we are looking forward to blasting against the 17 GB
> genebank... Any more ideas?
>
> Thx again,
> Bastian
>
> --
> Bastian Friedrich bastian at bastian-friedrich.de
> Adress & Fon available on my HP http://www.bastian-friedrich.de/
> \~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\
> \ Computers make very fast, very accurate mistakes.
> _______________________________________________
> BioBrew-Users mailing list
> BioBrew-Users at bioinformatics.org
> https://bioinformatics.org/mailman/listinfo/BioBrew-Users
More information about the BioBrew-Users
mailing list