Re: [OMPI users] Strange "All-to-All" behavior

2013-04-30 Thread Number Cruncher
Sorry, I seem to have misread your post. You're not actually invoking MPI_Alltoall or MPI_Alltoallv. Please disregard my last post. Simon. On 26/04/2013 23:14, Stephan Wolf wrote: Hi, I have encountered really bad performance when all the nodes send data to all the other nodes. I use Isen

Re: [OMPI users] Strange "All-to-All" behavior

2013-04-30 Thread Number Cruncher
This sounds a bit like the All_to_allv algorithm change I complained about when 1.6.1 was released. Original post: http://www.open-mpi.org/community/lists/users/2012/11/20722.php Everything waits for "rank 0" observation: http://www.open-mpi.org/community/lists/users/2013/01/21219.php Does s

Re: [OMPI users] job termination on grid

2013-04-30 Thread Ralph Castain
On Apr 30, 2013, at 1:54 PM, Vladimir Yamshchikov wrote: > This is the question I am trying to answer - how many threads I can use with > blastx on a grid? If I could request resources by_node, use -pernode option > to have one process per node, and then specify the correct number of threads

Re: [OMPI users] job termination on grid

2013-04-30 Thread Vladimir Yamshchikov
This is the question I am trying to answer - how many threads I can use with blastx on a grid? If I could request resources by_node, use -pernode option to have one process per node, and then specify the correct number of threads for each node. But I cannot, resurces (slots) are requested per-core

Re: [OMPI users] job termination on grid

2013-04-30 Thread Ralph Castain
On Apr 30, 2013, at 1:34 PM, Vladimir Yamshchikov wrote: > I asked grid IT and they said they had to kill it as the job was overloading > nodes. They saw loads up to 180 instead of close to 12 on 12-core nodes. They > think that blastx is not an openmpi application, so openMPI is spawning > b

Re: [OMPI users] job termination on grid

2013-04-30 Thread Vladimir Yamshchikov
I asked grid IT and they said they had to kill it as the job was overloading nodes. They saw loads up to 180 instead of close to 12 on 12-core nodes. They think that blastx is not an openmpi application, so openMPI is spawning between 64-96 blastx processes, each of which is then starting up 96 wor

Re: [OMPI users] job termination on grid

2013-04-30 Thread Reuti
Hi, Am 30.04.2013 um 21:26 schrieb Vladimir Yamshchikov: > My recent job started normally but after a few hours of running died with the > following message: > > -- > A daemon (pid 19390) died unexpectedly with status 137

[OMPI users] job termination on grid

2013-04-30 Thread Vladimir Yamshchikov
Hello, My recent job started normally but after a few hours of running died with the following message: -- A daemon (pid 19390) died unexpectedly with status 137 while attempting to launch so we are aborting. There m

Re: [OMPI users] multithreaded jobs

2013-04-30 Thread Ralph Castain
On Apr 30, 2013, at 7:52 AM, Dave Love wrote: > Ralph Castain writes: > >> On Apr 25, 2013, at 5:33 PM, Vladimir Yamshchikov wrote: >> >>> $NSLOTS is what requested by -pe openmpi in the script, my >>> understanding that by default it is threads. > > Is there something in the documentatio

Re: [OMPI users] multithreaded jobs

2013-04-30 Thread Dave Love
Ralph Castain writes: > On Apr 25, 2013, at 5:33 PM, Vladimir Yamshchikov wrote: > >> $NSLOTS is what requested by -pe openmpi in the script, my >> understanding that by default it is threads. Is there something in the documentation that suggest

Re: [OMPI users] Problem with Openmpi-1.4.0 and qlogic-ofed-1.5.4.1

2013-04-30 Thread Dave Love
Padma Pavani writes: > Hi Team, > > I am facing some problem while running HPL benchmark. > > > > I am using Intel mpi -4.0.1 with Qlogic-OFED-1.5.4.1 to run benchmark and > also tried with openmpi-1.4.0 but getting same error. > > > Error File : > > [compute-0-1.local:06936] [[14544,1],25] ORTE

Re: [OMPI users] QLogic HCA random crash after prolonged use

2013-04-30 Thread Dave Love
Ralph Castain writes: >> Dropped CR is definitely reason not to use OMPI past 1.6. [By the way, >> the release notes are confusing, saying that DMTCP is supported, but CR >> is dropped.] I'd have hoped a vendor who needs to support CR would >> contribute, but I suppose changes just become propr

Re: [OMPI users] Broadcast problem

2013-04-30 Thread George Bosilca
You can't use gather either. Same for gatherv, as you need to know the amount you will receive in advance. If I understand correctly your scenario (a random process is doing a broadcast at random time steps), using MPI collectives is not the best approach as they need global knowledge (there is

Re: [OMPI users] Broadcast and root process

2013-04-30 Thread giggzounet
ok. Thx for your answer. The documentation was not clear on this subject. Cheers Guillaume Le 29/04/2013 17:49, George Bosilca a écrit : > No, the root processor can be different for every broadcast, but for a same > broadcast every process involved must know who the root is. That's the only >

Re: [OMPI users] Broadcast problem

2013-04-30 Thread Randolph Pullen
Oops,I think I meant gather not scatter...

[OMPI users] Broadcast problem

2013-04-30 Thread Randolph Pullen
I have a number of processes split into sender and receivers. Senders read large quantities of randomly organised data into buffers for transmission to receivers. When a buffer is full it needs to be transmitted to all receivers this repeats until all the data is transmitted. Problem is that MPI