Re: [OMPI users] Error when using more than 88 processors for a specific executable -Abyss

2011-10-14 Thread Ashwani Kumar Mishra
Hi Ralph, No idea how much this program consumes the numbers of file descriptors :( Best Regards, Ashwani On Sat, Oct 15, 2011 at 12:08 AM, Ralph Castain wrote: > Should be plenty for us - does your program consume a lot? > > > On Oct 14, 2011, at 12:25 PM, Ashwani Kumar Mishra wrote: > > Hi Ra

Re: [OMPI users] MPI_Comm_accept - Busy wait

2011-10-14 Thread Thatyene Louise Alves de Souza Ramos
Thank you for the explanation! I use "-mca mpi_yield_when_idle 1" already! Thank you again! --- Thatyene Ramos On Fri, Oct 14, 2011 at 3:43 PM, Ralph Castain wrote: > Sorry - been occupied. This is normal behavior. As has been discussed on > this list before, OMPI made a design decision to mini

Re: [OMPI users] MPI_Comm_accept - Busy wait

2011-10-14 Thread Ralph Castain
Sorry - been occupied. This is normal behavior. As has been discussed on this list before, OMPI made a design decision to minimize latency. This means we aggressively poll for connections. Only thing you can do is tell it to yield the processor when idle so, if something else is trying to run, w

Re: [OMPI users] Error when using more than 88 processors for a specific executable -Abyss

2011-10-14 Thread Ralph Castain
Should be plenty for us - does your program consume a lot? On Oct 14, 2011, at 12:25 PM, Ashwani Kumar Mishra wrote: > Hi Ralph, > fs.file-max = 10 > is this ok or less? > > Best Regards, > Ashwani > > > On Fri, Oct 14, 2011 at 11:45 PM, Ralph Castain wrote: > Can't offer much about the

Re: [OMPI users] Error when using more than 88 processors for a specific executable -Abyss

2011-10-14 Thread Ashwani Kumar Mishra
Hi Ralph, fs.file-max = 10 is this ok or less? Best Regards, Ashwani On Fri, Oct 14, 2011 at 11:45 PM, Ralph Castain wrote: > Can't offer much about the qsub job. On the first one, what is your limit > on the number of file descriptors? Could be your sys admin has it too low. > > > On Oc

Re: [OMPI users] MPI_Comm_accept - Busy wait

2011-10-14 Thread Thatyene Louise Alves de Souza Ramos
Does anyone have any idea? --- Thatyene Ramos On Fri, Oct 7, 2011 at 12:01 PM, Thatyene Louise Alves de Souza Ramos < thaty...@gmail.com> wrote: > Hi there! > > In my code I use MPI_Comm_accept in a server-client communication. I > noticed that the server remains on busy wait whereas waiting for

Re: [OMPI users] Error when using more than 88 processors for a specific executable -Abyss

2011-10-14 Thread Ralph Castain
Can't offer much about the qsub job. On the first one, what is your limit on the number of file descriptors? Could be your sys admin has it too low. On Oct 14, 2011, at 12:07 PM, Ashwani Kumar Mishra wrote: > Hello, > When i try to run the following command i receive the following error when i

[OMPI users] Error when using more than 88 processors for a specific executable -Abyss

2011-10-14 Thread Ashwani Kumar Mishra
Hello, When i try to run the following command i receive the following error when i try to submit this job on the cluster having 40 nodes with each node having 8 processor & 8 GB RAM: Both the command work well, as long as i use only upto 88 processors in the cluster, but the moment i allocate mor

Re: [OMPI users] MPI_Waitany segfaults or (maybe) hangs

2011-10-14 Thread Francesco Salvadore
Dear MPI users, using Valgrind I found that the possibile error (which leads to segfault or hanging) comes from: ==10334== Conditional jump or move depends on uninitialised value(s) ==10334==    at 0xB150740: btl_openib_handle_incoming (btl_openib_component.c:2888) ==10334==    by 0xB1525A2: h

Re: [OMPI users] gpudirect p2p?

2011-10-14 Thread Rolf vandeVaart
>-Original Message- >From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] >On Behalf Of Chris Cooper >Sent: Friday, October 14, 2011 1:28 AM >To: us...@open-mpi.org >Subject: [OMPI users] gpudirect p2p? > >Hi, > >Are the recent peer to peer capabilities of cuda leveraged by

[OMPI users] gpudirect p2p?

2011-10-14 Thread Chris Cooper
Hi, Are the recent peer to peer capabilities of cuda leveraged by Open MPI when eg you're running a rank per gpu on the one workstation? It seems in my testing that I only get in the order of about 1GB/s as per http://www.open-mpi.org/community/lists/users/2011/03/15823.php, whereas nvidia's simp