[BioBrew Users] Unable to get mpiblast running

Bastian Friedrich bastian at bastian-friedrich.de
Tue Apr 11 04:20:51 EDT 2006


Hi,

On Tuesday 11 April 2006 06:50, Glen Otero wrote:
> Bastian-
>
> Any progress running mpiBLAST?

yesterday, I compiled mpiblast 1.3, linked against LAM/MPI (lam-gnu 
7.1.1 as included in the Rocks "hpc" roll). This seems to be running 
fine so far.

As I said, your mpiblast 1.4 with MPICH worked for the "mini data set" 
as described by you (p53/p53db). I was not successful getting it 
running with any other data set I tested - the Hs.seq.uniq/IL2RA 
combination as described in your tutorial, as well as genbank. This 
seems to be an MPICH problem I'm encountering (although it is unclear 
whether openmpi would run if this problem did not exist... :)

When compiling mpiblast 1.4 with LAM/MPI, I get an interesting result: 
The Hs.seq.uniq/IL2RA combination _does_ run (I just found out about 
the --debug flag yesterday), but it takes more than 8 hours (!) to 
finish. This is a little too long, I think ;) Here is a little snippet 
of it's output:
=============================
[...]
[9]     26.7763 rank 9 exiting successfully
[9]     26.7763 MPI startup  time is 0
[0]     26.9766 Processing result_id: 0
[0]     26.9767 Query 0 now has 2 / 28 fragments searched
[0]     26.9767 Got result count of 1
[0]     26.9767 receiveResults to read 1 records
[0]     27.4961 Read SeqAnnot
[0]     27.4961 Reading bioseqs for frag/query 4/0
[0]     108.928 Receive was successful -- about to merge
[0]     108.94  Query results have been merged
[0]     108.952 Processing result_id: 0
[0]     108.952 Query 0 now has 3 / 28 fragments searched
[0]     108.952 Got result count of 1
[0]     108.953 receiveResults to read 1 records
[0]     109.473 Read SeqAnnot
[0]     109.473 Reading bioseqs for frag/query 5/0
[0]     245.66  Receive was successful -- about to merge
[...]
=============================

The concrete blast jobs seem to finish in a reasonable time (the 26 
seconds are more or less the same as when running mpiblast 1.3), but 
the receiving and merging of results takes hours. The load on the 
compute nodes drops to zero, but stays high (1) on the head node. I 
have not fully investigated this topic, as even a rather small job 
takes hours to finish :)

This topic - or a similar one - is currently discussed on the mpiblast 
mailinglist, to which I subscribed yesterday.

When compiling mpiblast 1.4 with OpenMPI, load on all nodes (head and 
compute nodes) stays high for hours; I have not fully investigated this 
topic, but there seems to be some incompatibility between mpiblast and 
OpenMPI... mpiblast 1.3 does not work, either; OpenMPI is not actively 
supported by mpiblast, so I'll stick with the alternatives for now.

So, once more: mpiblast 1.3.0 with LAM/MPI works well; all alternatives 
failed so far. We will see if mpiblast 1.3.0 is sufficient for our 
needs later today.

If there is any more information I can provide, feel free to ask.

Thx again, Regards,
   Bastian

-- 
 Bastian Friedrich                  bastian at bastian-friedrich.de
 Adress & Fon available on my HP   http://www.bastian-friedrich.de/
\~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\
\ ... All the world's a stage, and I missed rehearsal.
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 189 bytes
Desc: not available
Url : http://bioinformatics.org/pipermail/biobrew-users/attachments/20060411/f7c3e0e9/attachment.bin


More information about the BioBrew-Users mailing list