Re: [OMPI users] How to force eager behavior during Isend?

2008-12-08 Thread Barry Rountree
On Monday 08 December 2008 02:44:42 pm George Bosilca wrote: > Barry, > > If you set the eager size large enough, the isend will not return > until the data is pushed into the network layer. That's exactly what I want it to do -- good. I've set the eagerness to 2MB, but for messages 64k and up,

Re: [OMPI users] How to force eager behavior during Isend?

2008-12-08 Thread Barry Rountree
Original message >Date: Mon, 8 Dec 2008 11:47:19 -0500 >From: George Bosilca >Subject: Re: [OMPI users] How to force eager behavior during Isend? >To: Open MPI Users > >Barry, > >These values are used deep inside the Open MPI library, in order to >define how we handle the messag

[OMPI users] How to force eager behavior during Isend?

2008-12-08 Thread Barry Rountree
would accomplish this, but that's didn't work on this system. What's the correct way of doing this? Thanks much, Barry Rountree University of Georgia OpenMPI 2.8 Linux+gcc, self+tcp. Relevant bit of the hostfile looks like: opt00 slots=1 max_slots=1 opt01 slo

[OMPI users] How to force eager behavior during Isend?

2008-12-08 Thread Barry Rountree
would accomplish this, but that's didn't work on this system. What's the correct way of doing this? Thanks much, Barry Rountree University of Georgia OpenMPI 2.8 Linux+gcc, self+tcp. Relevant bit of the hostfile looks like: opt00 slots=1 max_slots=1 opt01 slo

Re: [OMPI users] A simple mpi code hangs at MPI_Init?

2008-05-07 Thread Barry Rountree
On Wed, May 07, 2008 at 03:47:24PM -0400, Sang Chul Choi wrote: > Hi, > > I tried to run a hello world example of mpi code. It just hangs at > MPI_Init. My machine is installed with Ubuntu linux and I install some > package of open mpi. Compiling was okay, but running the code hangs at > MPI_Init.

Re: [OMPI users] Application logging

2008-05-07 Thread Barry Rountree
On Wed, May 07, 2008 at 01:51:03PM -0400, Alberto Giannetti wrote: > > On May 7, 2008, at 1:32 PM, Barry Rountree wrote: > > > On Wed, May 07, 2008 at 12:33:59PM -0400, Alberto Giannetti wrote: > >> I need to log application-level messages on disk to trace my program &

Re: [OMPI users] Application logging

2008-05-07 Thread Barry Rountree
On Wed, May 07, 2008 at 12:33:59PM -0400, Alberto Giannetti wrote: > I need to log application-level messages on disk to trace my program > activity. For better performances, one solution is to dedicate one > processor to the actual I/O logging, while the other working > processors would trac

Re: [OMPI users] Busy waiting [was Re: (no subject)]

2008-04-24 Thread Barry Rountree
On Thu, Apr 24, 2008 at 11:17:30AM -0400, George Bosilca wrote: > Well, blocking or not blocking this is the question !!! Unfortunately, it's > more complex than this thread seems to indicate. It's not that we didn't > want to implement it in Open MPI, it's that at one point we had to make a > c

Re: [OMPI users] Busy waiting [was Re: (no subject)]

2008-04-24 Thread Barry Rountree
milliseconds, and then use some other method that > sleeps not for a fixed time, but until new messages arrive. Well, it sounds like you can get to this before I can. Post your patch here and I'll test it on the NAS suite, UMT2K, Paradis, and a few synthetic benchmarks I've written.

[OMPI users] Busy waiting [was Re: (no subject)]

2008-04-23 Thread Barry Rountree
or energy savings. Are you volunteering to test a patch? (I've got four other papers I need to get finished up, so it'll be a few weeks before I start coding.) Barry Rountree Ph.D. Candidate, Computer Science University of Georgia > > > > > Jeff Squyres schrieb: >

Re: [OMPI users] Occasional mpirun hang on completion

2008-01-25 Thread Barry Rountree
On Thu, Jan 24, 2008 at 10:09:51PM -0500, Barry Rountree wrote: > On Thu, Jan 24, 2008 at 04:03:49PM -0500, Tim Mattox wrote: > > Hello Barry, > > I am guessing you are trying to use a threaded build of Open MPI... > > > > Unfortunately, the threading support in Ope

Re: [OMPI users] Occasional mpirun hang on completion

2008-01-24 Thread Barry Rountree
ave threading support actually working. Well, I'm happy the problem is obvious and the workaround is easy. I'll compile that version tonight and try it out when I get some time on the cluster tomorrow. Thanks for the help! Barry > > On Jan 24, 2008 3:25 AM, Barry Rountree wrote

Re: [OMPI users] Occasional mpirun hang on completion

2008-01-24 Thread Barry Rountree
On Thu, Jan 24, 2008 at 03:01:40AM -0500, Barry Rountree wrote: > On Fri, Jan 18, 2008 at 08:33:10PM -0500, Jeff Squyres wrote: > > Barry -- > > > > Could you check what apps are still running when it hangs? I.e., I > > assume that all the uptime's are dead;

Re: [OMPI users] Occasional mpirun hang on completion

2008-01-24 Thread Barry Rountree
d at the end of time (look for > function names like iof_flush or similar). We thought we had fixed > all of those on the 1.2 branch, but perhaps there's some other weird > race condition happening that doesn't happen on our test machines... > > > > On Jan 1

Re: [OMPI users] Occasional mpirun hang on completion

2008-01-18 Thread Barry Rountree
xed > all of those on the 1.2 branch, but perhaps there's some other weird > race condition happening that doesn't happen on our test machines... I'm happy to help. I've got a paper submission deadline on Tuesday, so it might not be until midweek. Thanks for the rep

Re: [OMPI users] odd network behavior

2008-01-16 Thread Barry Rountree
d ompi directory. > > Any assistance on this matter would be appreciated, > > Mark E. Kosmowski I'd posted a message earlier about intermittent hangs -- perhaps it's the same issue. If you run a hundred instances or so of "

Re: [OMPI users] Occasional mpirun hang on completion

2008-01-13 Thread Barry Rountree
On Sun, Jan 13, 2008 at 09:54:47AM -0500, Barry Rountree wrote: > Hello, > > The following command > > mpirun -np 2 -hostfile ~/hostfile uptime > > will occasionally hang after completing. The expected output appears on > the screen, but mpirun needs a SIGKILL

[OMPI users] Occasional mpirun hang on completion

2008-01-13 Thread Barry Rountree
9.1.045) and gcc 4.1.0 20060304 (aka Red Hat 4.1.0-3). I have also tried earlier versions of OpenMPI and found the same bug (1.1.2 and 1.2.2). Using -verbose didn't provide any additional output. I'm happy to help tracking down whatever is causing this. Many thanks, Barry Rou