[OMPI users] Query:File locking failed in ADIOI_Set_lock

2008-05-06 Thread rajlaxmi
Hello,

I am getting this following error when I run my
application on more than 1 nodes.

File locking failed in ADIOI_Set_lock. If the file
system is NFS, you need to use NFS version 3, ensure
that the lockd daemon is running on all the machines,
and mount the directory with the 'noac' option (no
attribute caching).

I have NFS version 4, and we have mounted the dir with
noac option.

Will it work only with NFS 3 version only?

Any help will be appreciated.

Thanks
Raj


Send instant messages to your online friends http://uk.messenger.yahoo.com 


Re: [OMPI users] Problem Installing OpenMPI with Lahey 64-bit Fortran compiler

2008-05-06 Thread Jeff Squyres
This post was accidentally sent to the list owner rather than to the  
entire list.  Re-sending to the list...


On May 1, 2008, at 11:28 AM, cfdm...@aim.com wrote:

I am encountering a difficulty when installing OpenMPI with the  
Lahey 64-bit (LF64 V8.0) Fortran Compiler for use on a dual quad- 
core machine. I am currently trying to install this particular Lahey  
compiler with openmpi-1.2.6 on a 64-bit intel machine running Red  
Hat Linux Enterprise 5. In the past, I have not had any difficulty  
installing 32-bit versions of the Lahey compiler with the current or  
previous versions of OpenMPI, so this issue appears to be specific  
to the 64-bit version of the Lahey compiler. I have attached the  
install script that I used to install OpenMPI. I have also attached  
the output stream that results when I execute this install script.



The error message that it is showing is:

/usr/bin/ld: .libs/mpi_wtick_f90.o: relocation R_X86_64_32S against `a  
local symbol' can not be used when making a shared object; recompile  
with -fPIC


We don't test with the Laheay compiler, so I'm not familiar with the  
exact problem here -- perhaps try adding -fPIC to the FCFLAGS when  
running configure?


./configure FCFLAGS=-fPIC ...

--
Jeff Squyres
Cisco Systems


config.log.bz2
Description: Binary data


configure.log.bz2
Description: Binary data


make.log.bz2
Description: Binary data


Re: [OMPI users] Install BLACS and ScaLAPACK on Leopard

2008-05-06 Thread Jeff Squyres
FWIW, I'm not a fortran expert, but if you built your Fortran  
libraries with g77 and then tried to link against them with gfortran,  
you might run into problems.


My advice would be to use a single fortran compiler for building  
everything: Open MPI, your libraries, your apps.  I prefer gfortran  
because it's more modern, but I have not done any performance  
evaluations of gfortran vs. g77 -- I have heard [unverified] anecdotes  
that gfortran is "slower" than g77 -- google around and see what the  
recent buzz is.


FWIW: I tend to use the gnu suite from http://hpc.sourceforge.net/ --  
it contains pre-built gcc/g++/gfortran binaries and libraries for  
Leopard.



On May 5, 2008, at 2:59 PM, Linwei Wang wrote:


Dear Reeder,

   It does not work. I do think they are from the fortran programs
I'm using (they are files included from the BLACS installation
package, not written by my own.

   The thing is last time when I was using g77, it caused no  
problem...


   thanks for  your help.

   Linwei.

On May 5, 2008, at 2:33 PM, Doug Reeder wrote:


_s_wsle, _e_wsle


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



--
Jeff Squyres
Cisco Systems



Re: [OMPI users] Install BLACS and ScaLAPACK on Leopard

2008-05-06 Thread Linwei Wang
Thanks, Jeff, I finally choose to use the ScaLAPACK installer provided  
on the website~
with gcc/gfortran and openmpi...it is installed successfully but I  
haven't tested the applications..



Thanks for your help.

Linwei


On May 6, 2008, at 10:09 AM, Jeff Squyres wrote:


FWIW, I'm not a fortran expert, but if you built your Fortran
libraries with g77 and then tried to link against them with gfortran,
you might run into problems.

My advice would be to use a single fortran compiler for building
everything: Open MPI, your libraries, your apps.  I prefer gfortran
because it's more modern, but I have not done any performance
evaluations of gfortran vs. g77 -- I have heard [unverified] anecdotes
that gfortran is "slower" than g77 -- google around and see what the
recent buzz is.

FWIW: I tend to use the gnu suite from http://hpc.sourceforge.net/ --
it contains pre-built gcc/g++/gfortran binaries and libraries for
Leopard.


On May 5, 2008, at 2:59 PM, Linwei Wang wrote:


Dear Reeder,

  It does not work. I do think they are from the fortran programs
I'm using (they are files included from the BLACS installation
package, not written by my own.

  The thing is last time when I was using g77, it caused no
problem...

  thanks for  your help.

  Linwei.

On May 5, 2008, at 2:33 PM, Doug Reeder wrote:


_s_wsle, _e_wsle


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



--
Jeff Squyres
Cisco Systems

___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users




Re: [OMPI users] Install BLACS and ScaLAPACK on Leopard

2008-05-06 Thread Gregory John Orris
Points to clarify if I may, having gone through this relatively  
recently:

g77 and gfortran are NOT one and the same.
gfortran from sourceforge works well, but it is based on gnu gcc 4.3  
and not on the gnu gcc 4.0.1 that comes with Leopard.
Your best bet is to download the ENTIRE gcc package from sourceforge  
and install it into /usr/local. This includes gcc, g++, and gfortran.


Then you will need to do a number of things to actually get a reliable  
set of packages all compiled from the same version of gcc 4.3.
Why? Because 4.3 seems to be notoriously faster. AND, I had a lot of  
problems integrating the 4.0.1 libs with the 4.3 libs without errors

1. download CLAPACK-3.1.1 from netlib And compile
2. Download ATLAS-1.8 from dourceforge (netlib is a little behind  
here) and configure it with the --with-netlib-lapack=your just  
compiled lapack from CLAPACK
3. Download OpenMPI 1.2.6 and install it also so that openMPI will  
have the fortran not installed with Leopard.

4. NOW you can compile BLACS and ScaLAPACK

In all of this you will need to do a couple of additional things like  
set the env's

setenv LDFLAGS "-L/usr/local/lib/x86_64"
setenv DYLD_LIBRARY_PATH "your openmpi path"
setenv LD_LIBRARY_PATH "your openmpi path"

Do all this right and make sure you compile with the -m64 -mtune=core2  
flags and you will be golden.


So what will you have---
A new cblas, atlas, lapack, openmpi, fortran, c, c++, blacs, and  
scalapack.

All on the same version of gnu c.

Alternatively you can buy and use the intel compiler. It is  
significantly faster than gfortran, but it has a host of other  
problems associated with it.
But if you follow the outline above,  you will be left with the best  
that's available. I have lots more info on this, but time is short.


FINALLY, and this is important, DO NOT FORGET ABOUT THE small STACK  
size on Mac's when using gfortran. It's so small that it's useless for  
large parallel jobs.



On May 6, 2008, at 10:09 AM, Jeff Squyres wrote:


FWIW, I'm not a fortran expert, but if you built your Fortran
libraries with g77 and then tried to link against them with gfortran,
you might run into problems.

My advice would be to use a single fortran compiler for building
everything: Open MPI, your libraries, your apps.  I prefer gfortran
because it's more modern, but I have not done any performance
evaluations of gfortran vs. g77 -- I have heard [unverified] anecdotes
that gfortran is "slower" than g77 -- google around and see what the
recent buzz is.

FWIW: I tend to use the gnu suite from http://hpc.sourceforge.net/ --
it contains pre-built gcc/g++/gfortran binaries and libraries for
Leopard.


On May 5, 2008, at 2:59 PM, Linwei Wang wrote:


Dear Reeder,

  It does not work. I do think they are from the fortran programs
I'm using (they are files included from the BLACS installation
package, not written by my own.

  The thing is last time when I was using g77, it caused no
problem...

  thanks for  your help.

  Linwei.

On May 5, 2008, at 2:33 PM, Doug Reeder wrote:


_s_wsle, _e_wsle


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



--
Jeff Squyres
Cisco Systems

___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users





[OMPI users] Problem with AlltoAll routine

2008-05-06 Thread Gabriele FATIGATI
Hi,
i tried to run SkaMPI 5.0.4 benchmark on IBM-BladeCenterLS21 system with 256 
processors over Infiniband 5 Gb/s, but test has stopped on "AlltoAll-length" 
routine, with count=2048 for some reason.

I have launched test with:
--mca btl_openib_eager_limit 1024

Same tests with 128 processor or less, have finished successful. 

Different values of eager limit don't solve the problem. Version of OpenMPI 
involved is 1.2.5. There's someone with this kind of problem over Infiniband?
Thanks in advance.
-- 
Gabriele Fatigati

CINECA Systems & Tecnologies Department

Supercomputing  Group

Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy

www.cineca.itTel:39 051 6171722

g.fatig...@cineca.it   

Re: [OMPI users] MPI_Lookup_name error

2008-05-06 Thread Alberto Giannetti
Correct, I am using 1.2.6 and not running a persistent deamon (thank you for
the link Pak by the way).
However, in the Client/Server test thread I'm showing a complete example
where I tried to run both applications through the same mpirun command and
still having an internal error:
http://www.open-mpi.org/community/lists/users/2008/04/5537.php

On Mon, May 5, 2008 at 10:18 AM, Ralph Castain  wrote:

> I assume you are using some variant of OMPI 1.2?
>
> When last I checked, which admittedly was quite a while ago, this worked
> on
> the 1.2.x series. However, I note something here that may be a problem. In
> the 1.2.x series, we do not have a global publish/lookup service - the
> application doing the publish must be launched by the same mpirun as the
> application doing the lookup.
>
> The code below only does the lookup, and appears to be asking that you
> provide some server name. I assume you are somehow looking up the name of
> the mpirun that launched the application that did the publish, and hoping
> the two will cross-connect? Unfortunately, I don't believe the 1.2.x code
> is
> smart enough to figure out how to do that.
>
> This is cleaned up in the upcoming 1.3 release and should work much
> smoother.
>
> Ralph
>
>
>
> On 4/27/08 6:58 PM, "Alberto Giannetti" 
> wrote:
>
> > I am having error using MPI_Lookup_name. The same program works fine
> > when using MPICH:
> >
> >
> > /usr/local/bin/mpiexec -np 2 ./client myfriend
> > Processor 0 (662, Sender) initialized
> > Processor 0 looking for service myfriend-0
> > Processor 1 (664, Sender) initialized
> > Processor 1 looking for service myfriend-1
> > [local:00662] *** An error occurred in MPI_Lookup_name
> > [local:00662] *** on communicator MPI_COMM_WORLD
> > [local:00662] *** MPI_ERR_NAME: invalid name argument
> > [local:00662] *** MPI_ERRORS_ARE_FATAL (goodbye)
> > [local:00664] *** An error occurred in MPI_Lookup_name
> > [local:00664] *** on communicator MPI_COMM_WORLD
> > [local:00664] *** MPI_ERR_NAME: invalid name argument
> > [local:00664] *** MPI_ERRORS_ARE_FATAL (goodbye)
> >
> >
> > int main(int argc, char* argv[])
> > {
> >int rank, i;
> >float data[100];
> >char cdata[64];
> >char myport[MPI_MAX_PORT_NAME];
> >char myservice[64];
> >MPI_Comm intercomm;
> >MPI_Status status;
> >int intercomm_size;
> >
> >MPI_Init(&argc, &argv);
> >MPI_Comm_rank(MPI_COMM_WORLD, &rank);
> >printf("Processor %d (%d, Sender) initialized\n", rank, getpid());
> >
> >if( argc < 2 ) {
> >  fprintf(stderr, "Require server name\n");
> >  MPI_Finalize();
> >  exit(-1);
> >}
> >
> >for( i = 0; i < 100; i++ )
> >  data[i] = i;
> >
> >sprintf(myservice, "%s-%d", argv[1], rank);
> >printf("Processor %d looking for service %s\n", rank, myservice);
> >MPI_Lookup_name(myservice, MPI_INFO_NULL, myport);
> >printf("Processor %d found port %s looking for service %s\n",
> > rank, myport, myservice);
> >
> >while( 1 ) {
> >  printf("Processor %d connecting to '%s'\n", rank, myport);
> >  if( MPI_Comm_connect(myport, MPI_INFO_NULL, 0, MPI_COMM_SELF,
> > &intercomm) == MPI_SUCCESS )
> >break;
> >  sleep(1);
> >}
> >printf("Processor %d connected\n", rank);
> >
> >MPI_Comm_remote_size(intercomm, &intercomm_size);
> >printf("Processor %d remote comm size is %d\n", rank,
> > intercomm_size);
> >
> >printf("Processor %d sending data through intercomm to rank 0...
> > \n", rank);
> >MPI_Send(data, 100, MPI_FLOAT, 0, rank, intercomm);
> >printf("Processor %d data sent!\n", rank);
> >MPI_Recv(cdata, 12, MPI_CHAR, MPI_ANY_SOURCE, MPI_ANY_TAG,
> > intercomm, &status);
> >printf("Processor %d received string data '%s' from rank %d, tag %d
> > \n", rank, cdata, status.MPI_SOURCE, status.MPI_TAG);
> >
> >sleep(5);
> >
> >printf("Processor %d disconnecting communicator\n", rank);
> >MPI_Comm_disconnect(&intercomm);
> >printf("Processor %d finalizing\n", rank);
> >
> >MPI_Finalize();
> >printf("Processor %d Goodbye!\n", rank);
> > }
> >
> > ___
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>


Re: [OMPI users] Problem with AlltoAll routine

2008-05-06 Thread Brad Benton
Hello Gabriele,

To help track down this problem, could I ask you to take a look at the Open
MPI "Getting Help" page?
  http://www.open-mpi.org/community/help/

In particular, if you could collect and send the information requested on
that page to the list, it will help us to better understand your
configuration and how others might reproduce the problem.

Thanks & Regards,
--Brad

Brad Benton
IBM


On Tue, May 6, 2008 at 10:35 AM, Gabriele FATIGATI 
wrote:

> Hi,
> i tried to run SkaMPI 5.0.4 benchmark on IBM-BladeCenterLS21 system with
> 256 processors over Infiniband 5 Gb/s, but test has stopped on
> "AlltoAll-length" routine, with count=2048 for some reason.
>
> I have launched test with:
> --mca btl_openib_eager_limit 1024
>
> Same tests with 128 processor or less, have finished successful.
>
> Different values of eager limit don't solve the problem. Version of
> OpenMPI involved is 1.2.5. There's someone with this kind of problem over
> Infiniband?
> Thanks in advance.
> --
> Gabriele Fatigati
>
> CINECA Systems & Tecnologies Department
>
> Supercomputing  Group
>
> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>
> www.cineca.itTel:39 051 6171722
>
> g.fatig...@cineca.it
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>