[OMPI users] Absoft fortran detected as g77?

2006-03-23 Thread Michael Kluskens
I have Absoft version 8.2a installed on my OS X 10.4.5 system and in  
order to do some testing I was trying to build OpenMPI 1.1a1r9364  
with it and got the following funny result:


*** Fortran 77 compiler
checking whether we are using the GNU Fortran 77 compiler... yes
checking whether f95 accepts -g... yes
checking if Fortran 77 compiler works... yes
checking f95 external symbol convention... upper case
checking if Fortran 77 compiler supports LOGICAL... yes
checking size of Fortran 77 LOGICAL...
checking for C type corresponding to LOGICAL... not found
configure: WARNING: *** Did not find corresponding C type
configure: error: Cannot continue

Note that f95 is/should be my Absoft compiler but the configure  
script flags it as the GNU Fortran 77 compiler.


I had set FC and F77 to f95.

The config.log is attached and the terminal display is pasted below:

checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... yes
checking for gawk... no
checking for mawk... no
checking for nawk... no
checking for awk... awk
checking whether make sets $(MAKE)... yes

 


== Configuring Open MPI
 



*** Checking versions
checking Open MPI version... 1.1a1r9364
checking Open MPI Subversion repository version... r9364
checking Open Run-Time Environment version... 1.1a1r9364
checking Open Run-Time Environment Subversion repository version...  
r9364

checking Open Portable Access Layer version... 1.1a1r9364
checking Open Portable Access Layer Subversion repository version...  
r9364


*** Initialization, setup
configure: builddir: /Users/mkluskens/Public/MPI/OpenMPI/ 
openmpi-1.1a1r9364
configure: srcdir: /Users/mkluskens/Public/MPI/OpenMPI/ 
openmpi-1.1a1r9364

checking build system type... powerpc-apple-darwin8.5.0
checking host system type... powerpc-apple-darwin8.5.0
checking for prefix by checking for ompi_clean... no
installing to directory "/usr/local"

*** Configuration options
checking Whether to run code coverage... no
checking whether to debug memory usage... no
checking whether to profile memory usage... no
checking if want developer-level compiler pickyness... no
checking if want developer-level debugging code... no
checking if want Fortran 77 bindings... yes
checking if want Fortran 90 bindings... yes
checking whether to enable PMPI... yes
checking if want C++ bindings... yes
checking if want to enable weak symbol support... yes
checking if want run-time MPI parameter checking... runtime
checking if want to install OMPI header files... no
checking if want pretty-print stacktrace... yes
checking if want deprecated executable names... no
checking max supported array dimension in F90 MPI bindings... 4
checking if pty support should be enabled... yes
checking if user wants dlopen support... yes
checking if heterogeneous support should be enabled... yes
checking if want trace file debugging... no

 


== Compiler and preprocessor tests
 



*** C compiler and preprocessor
checking for style of include used by make... GNU
checking for gcc... gcc
checking for C compiler default output file name... a.out
checking whether the C compiler works... yes
checking whether we are cross compiling... no
checking for suffix of executables...
checking for suffix of object files... o
checking whether we are using the GNU C compiler... yes
checking whether gcc accepts -g... yes
checking for gcc option to accept ANSI C... none needed
checking dependency style of gcc... gcc3
checking whether gcc and cc understand -c and -o together... yes
checking for the C compiler vendor... gnu
checking if gcc supports -finline-functions... yes
checking if gcc supports -fno-strict-aliasing... yes
configure: WARNING:  -fno-strict-aliasing has been added to CFLAGS
checking for C optimization flags... -O3 -DNDEBUG -fno-strict-aliasing
checking how to run the C preprocessor... gcc -E
checking for egrep... grep -E
checking for ANSI C header files... yes
checking for sys/types.h... yes
checking for sys/stat.h... yes
checking for stdlib.h... yes
checking for string.h... yes
checking for memory.h... yes
checking for strings.h... yes
checking for inttypes.h... yes
checking for stdint.h... yes
checking for unistd.h... yes
checking for long long... yes
checking for long double... yes
checking for int8_t... yes
checking for uint8_t... yes
checking for int16_t... yes
checking for uint16_t... yes
checking for int32_t... yes
checking for uint32_t... yes
checking for int64_t... yes
checking for uint64_t... yes
checking for intptr_t... yes
checking for uintptr_t... yes
checking for mode_t... yes
checking for char... yes
checking size of char... 1
checking for short... yes
checking size of short... 2
checking for 

Re: [OMPI users] problems with OpenMPI-1.0.1 on SunOS 5.9; problems on heterogeneous cluster

2006-03-23 Thread Ralph Castain




This is now fixed in the trunk with r9393. Should show up in the
nightly build, I think. We had inadvertently set a maximum limit on the
number of app_contexts we could support that came in through the
app_file interface - this limitation has been removed.

Thanks for catching it!

Ralph


Ravi Manumachu wrote:

  Hi Brian,

I have installed OpenMPI-1.1a1r9260 on my SunOS machines. It has solved
the problems. However there is one more issue that I found in my testing
and that I failed to report. This concerns Linux machines too.

My host file is

hosts.txt
-
csultra06
csultra02
csultra05
csultra08

My app file is 

mpiinit_appfile
---
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit
-np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit

My application program is

mpiinit.c
-

#include 

int main(int argc, char** argv)
{
int rc, me;
char pname[MPI_MAX_PROCESSOR_NAME];
int plen;

MPI_Init(
   &argc,
   &argv
);

rc = MPI_Comm_rank(
MPI_COMM_WORLD,
&me
);

if (rc != MPI_SUCCESS)
{
   return rc;
}

MPI_Get_processor_name(
   pname,
   &plen
);

printf("%s:Hello world from %d\n", pname, me);

MPI_Finalize();

return 0;
}

Compilation is successful

csultra06$ mpicc -o mpiinit mpiinit.c

However mpirun prints just 6 statements instead of 8.

csultra06$ mpirun --hostfile hosts.txt --app mpiinit_appfile
csultra02:Hello world from 5
csultra06:Hello world from 0
csultra06:Hello world from 4
csultra02:Hello world from 1
csultra08:Hello world from 3
csultra05:Hello world from 2

The following two more statements are not printed.

csultra05:Hello world from 6
csultra08:Hello world from 7

This behavior I observed on my Linux cluster too.

I have attached the log for "-d" option for your debugging purposes.

Regards,
Ravi.

- Original Message -
From: Brian Barrett 
List-Post: users@lists.open-mpi.org
Date: Monday, March 13, 2006 7:56 pm
Subject: Re: [OMPI users] problems with OpenMPI-1.0.1 on SunOS 5.9;
problems on heterogeneous cluster
To: Open MPI Users 

  
  
Hi Ravi -

With the help of another Open MPI user, I spent the weekend finding 
a  
couple of issues with Open MPI on Solaris.  I believe you are 
running  
into the same problems.  We're in the process of certifying the  
changes for release as part of 1.0.2, but it's Monday morning and 
the  
release manager hasn't gotten them into the release branch just 
yet.   
Could you give the nightly tarball from our development trunk a try 

and let us know if it solves your problems on Solaris?  You 
probably  
want last night's 1.1a1r9260 release.

http://www.open-mpi.org/nightly/trunk/

Thanks,

Brian


On Mar 12, 2006, at 11:23 PM, Ravi Manumachu wrote:



   Hi Brian,

 Thank you for your help. I have attached all the files you have 
  

asked>  for in a tar file.


   Please find attached the 'config.log' and 'libmpi.la' for my 
  

Solaris>  installation.


   The output from 'mpicc -showme' is

 sunos$ mpicc -showme
 gcc -I/home/cs/manredd/OpenMPI/openmpi-1.0.1/OpenMPI-SunOS-5.9/ 
include
 -I/home/cs/manredd/OpenMPI/openmpi-1.0.1/OpenMPI-SunOS-
 5.9/include/openmpi/ompi-L/home/cs/manredd/OpenMPI/openmpi-
 1.0.1/OpenMPI-SunOS-5.9/lib -lmpi
 -lorte -lopal -lnsl -lsocket -lthread -laio -lm -lnsl -lsocket -
 lthread -ldl

 There are serious issues when running on just solaris machines.

 I am using the host file and app file shown below. Both the
 machines are
 SunOS and are similar.

 hosts.txt
 -
 csultra01 slots=1
 csultra02 slots=1

 mpiinit_appfile
 ---
 -np 1 /home/cs/manredd/OpenMPI/openmpi-1.0.1/MPITESTS/mpiinit_sunos
 -np 1 /home/cs/manredd/OpenMPI/openmpi-1.0.1/MPITESTS/mpiinit_sunos

 Running mpirun without -d option hangs.

 csultra01$ mpirun --hostfile hosts.txt --app mpiinit_appfile
 hangs

 Running mpirun with -d option dumps core with output in the file
 "mpirun_output_d_option.txt", which is attached. The core is also
 attached.
 Running just on only one host is also not working. The output from
 mpirun using "-d" option for this scenario is attached in file
 "mpirun_output_d_option_one_host.txt".

 I have also attached the list of packages installed on my solaris
 machine in "pkginfo.txt"

 I hope these will help you to resolve the issue.

 Regards,
 Ravi.

  
  
- Original Message -
From: Brian Barrett 
List-Post: users@lists.open-mpi.org
Date: Friday, March 10, 2006 7:

Re: [OMPI users] Absoft fortran detected as g77?

2006-03-23 Thread Brian Barrett

On Mar 23, 2006, at 5:32 PM, Michael Kluskens wrote:

I have Absoft version 8.2a installed on my OS X 10.4.5 system and  
in order to do some testing I was trying to build OpenMPI  
1.1a1r9364 with it and got the following funny result:


*** Fortran 77 compiler
checking whether we are using the GNU Fortran 77 compiler... yes
checking whether f95 accepts -g... yes
checking if Fortran 77 compiler works... yes
checking f95 external symbol convention... upper case
checking if Fortran 77 compiler supports LOGICAL... yes
checking size of Fortran 77 LOGICAL...
checking for C type corresponding to LOGICAL... not found
configure: WARNING: *** Did not find corresponding C type
configure: error: Cannot continue

Note that f95 is/should be my Absoft compiler but the configure  
script flags it as the GNU Fortran 77 compiler.


The GNU Fortran thing just means that Absoft sets some flags that g77/ 
gfortran set.  No worries there.


I unfortunately can't replicate your issue with our Absoft compilers,  
but we're at 9.2, so perhaps something has changed.  The test code in  
configure appears to have run correctly (according to your  
config.log), but it doesn't look like it can read anything from the  
file that is supposed to contain the sizeof result for the test.


Could you try to build and run the attached test code (should just  
have to untar, run make, and run ./fsize) and let me know what the  
result is?  If all works properly, you should see something like:


[21:23] brbarret@frood:ttyp3 ~/f77test% ./fsize
size is 4

Thanks,

Brian

--
  Brian Barrett
  Open MPI developer
  http://www.open-mpi.org/



f77size.tar.gz
Description: GNU Zip compressed data




Re: [OMPI users] Error message about libopal.so

2006-03-23 Thread Brian Barrett

On Mar 22, 2006, at 1:47 PM, Michael Kluskens wrote:

Trying to find the cause of one or more errors, might involve  
libopal.so


Built openmpi-1.1a1r9351 on Debian Linux on Operton with PGI 6.1-3
using "./configure --with-gnu-ld F77=pgf77 FFLAGS=-fastsse FC=pgf90
FCFLAGS=-fastsse"

My program generates the following error which I do not understand:

Signal:11 info.si_errno:0(Success) si_code:1(SEGV_MAPERR)
Failing at addr:0x4
[0] func:/usr/local/lib/libopal.so.0 [0x2a959927dd]
*** End of error message ***

Is it possible I'm over running the OpenMPI buffers, my test program
works fine other than the "GPR data corruption" errors (uses
MPI_SPAWN and posted previously); the basic MPI difference between my
test program and the real program is massive amount of data being
distributed via BCAST and SEND/RECV.


It worries me that the call stack only goes that deep - there should  
be more functions listed there (if nothing else, the main()  
function).  Can you run your application in a debugger and try to get  
a full stack trace?  Typically, segmentation faults point to  
overwriting user buffers, but without more detail, it's hard to pin- 
point the issue.



Thanks,

Brian

--
  Brian Barrett
  Open MPI developer
  http://www.open-mpi.org/