How much memory is available on that quad core machine? The minimum requirements for MPIM2007 are: 16GB of memory for the whole system or 1GB of memory per rank, whichever is larger. For MPIL2007 you need to use at least 64 processes and a minimum of 128GB (2GB/process) is required.
Cheers, Martin -- Martin Siegert Head, Research Computing WestGrid Site Lead IT Services phone: 778 782-4691 Simon Fraser University fax: 778 782-4242 Burnaby, British Columbia email: sieg...@sfu.ca Canada V5A 1S6 On Wed, Apr 28, 2010 at 05:32:12AM -0500, Jeff Squyres (jsquyres) wrote: > > I don't know much about specmpi, but it seems like it is choosing to > abort. Maybe the "no room for lattice" has some meaning...? > -jms > Sent from my PDA. No type good. > _______________________________________________________________________ > > From: users-boun...@open-mpi.org <users-boun...@open-mpi.org> > To: us...@open-mpi.org <us...@open-mpi.org> > Sent: Wed Apr 28 01:47:01 2010 > Subject: [OMPI users] MPI_ABORT was invoked on rank 0 in > communicatorMPI_COMM_WORLD with errorcode 0. > > Hi, > I am trying to run SPEC MPI 2007 workload on a quad-core machine. > However getting this error message. I also tried to use hostfile option > by specifying localhost slots=4, but still getting the following error. > Please help me. > $mpirun --mca btl tcp,sm,self -np 4 su3imp_base.solaris > SU3 with improved KS action > Microcanonical simulation with refreshing > MIMD version 6 > Machine = > R algorithm > type 0 for no prompts or 1 for prompts > nflavors 2 > nx 30 > ny 30 > nz 56 > nt 84 > iseed 1234 > LAYOUT = Hypercubes, options = EVENFIRST, > NODE 0: no room for lattice > termination: Tue Apr 27 23:41:44 2010 > Termination: node 0, status = 1 > ----------------------------------------------------------------------- > --- > MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD > with errorcode 0. > NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. > You may or may not see output from other processes, depending on > exactly when Open MPI kills them. > ----------------------------------------------------------------------- > --- > ----------------------------------------------------------------------- > --- > mpirun has exited due to process rank 0 with PID 17239 on > node cache-aware exiting without calling "finalize". This may > have caused other processes in the application to be > terminated by signals sent by mpirun (as reported here). > Best, > Kishore Kumar Pusukuri > [1]http://www.cs.ucr.edu/~kishore > > References > > 1. http://www.cs.ucr.edu/~kishore > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users