[OMPI users] Absoft fortran detected as g77?
I have Absoft version 8.2a installed on my OS X 10.4.5 system and in order to do some testing I was trying to build OpenMPI 1.1a1r9364 with it and got the following funny result: *** Fortran 77 compiler checking whether we are using the GNU Fortran 77 compiler... yes checking whether f95 accepts -g... yes checking if Fortran 77 compiler works... yes checking f95 external symbol convention... upper case checking if Fortran 77 compiler supports LOGICAL... yes checking size of Fortran 77 LOGICAL... checking for C type corresponding to LOGICAL... not found configure: WARNING: *** Did not find corresponding C type configure: error: Cannot continue Note that f95 is/should be my Absoft compiler but the configure script flags it as the GNU Fortran 77 compiler. I had set FC and F77 to f95. The config.log is attached and the terminal display is pasted below: checking for a BSD-compatible install... /usr/bin/install -c checking whether build environment is sane... yes checking for gawk... no checking for mawk... no checking for nawk... no checking for awk... awk checking whether make sets $(MAKE)... yes == Configuring Open MPI *** Checking versions checking Open MPI version... 1.1a1r9364 checking Open MPI Subversion repository version... r9364 checking Open Run-Time Environment version... 1.1a1r9364 checking Open Run-Time Environment Subversion repository version... r9364 checking Open Portable Access Layer version... 1.1a1r9364 checking Open Portable Access Layer Subversion repository version... r9364 *** Initialization, setup configure: builddir: /Users/mkluskens/Public/MPI/OpenMPI/ openmpi-1.1a1r9364 configure: srcdir: /Users/mkluskens/Public/MPI/OpenMPI/ openmpi-1.1a1r9364 checking build system type... powerpc-apple-darwin8.5.0 checking host system type... powerpc-apple-darwin8.5.0 checking for prefix by checking for ompi_clean... no installing to directory "/usr/local" *** Configuration options checking Whether to run code coverage... no checking whether to debug memory usage... no checking whether to profile memory usage... no checking if want developer-level compiler pickyness... no checking if want developer-level debugging code... no checking if want Fortran 77 bindings... yes checking if want Fortran 90 bindings... yes checking whether to enable PMPI... yes checking if want C++ bindings... yes checking if want to enable weak symbol support... yes checking if want run-time MPI parameter checking... runtime checking if want to install OMPI header files... no checking if want pretty-print stacktrace... yes checking if want deprecated executable names... no checking max supported array dimension in F90 MPI bindings... 4 checking if pty support should be enabled... yes checking if user wants dlopen support... yes checking if heterogeneous support should be enabled... yes checking if want trace file debugging... no == Compiler and preprocessor tests *** C compiler and preprocessor checking for style of include used by make... GNU checking for gcc... gcc checking for C compiler default output file name... a.out checking whether the C compiler works... yes checking whether we are cross compiling... no checking for suffix of executables... checking for suffix of object files... o checking whether we are using the GNU C compiler... yes checking whether gcc accepts -g... yes checking for gcc option to accept ANSI C... none needed checking dependency style of gcc... gcc3 checking whether gcc and cc understand -c and -o together... yes checking for the C compiler vendor... gnu checking if gcc supports -finline-functions... yes checking if gcc supports -fno-strict-aliasing... yes configure: WARNING: -fno-strict-aliasing has been added to CFLAGS checking for C optimization flags... -O3 -DNDEBUG -fno-strict-aliasing checking how to run the C preprocessor... gcc -E checking for egrep... grep -E checking for ANSI C header files... yes checking for sys/types.h... yes checking for sys/stat.h... yes checking for stdlib.h... yes checking for string.h... yes checking for memory.h... yes checking for strings.h... yes checking for inttypes.h... yes checking for stdint.h... yes checking for unistd.h... yes checking for long long... yes checking for long double... yes checking for int8_t... yes checking for uint8_t... yes checking for int16_t... yes checking for uint16_t... yes checking for int32_t... yes checking for uint32_t... yes checking for int64_t... yes checking for uint64_t... yes checking for intptr_t... yes checking for uintptr_t... yes checking for mode_t... yes checking for char... yes checking size of char... 1 checking for short... yes checking size of short... 2 checking for
Re: [OMPI users] problems with OpenMPI-1.0.1 on SunOS 5.9; problems on heterogeneous cluster
This is now fixed in the trunk with r9393. Should show up in the nightly build, I think. We had inadvertently set a maximum limit on the number of app_contexts we could support that came in through the app_file interface - this limitation has been removed. Thanks for catching it! Ralph Ravi Manumachu wrote: Hi Brian, I have installed OpenMPI-1.1a1r9260 on my SunOS machines. It has solved the problems. However there is one more issue that I found in my testing and that I failed to report. This concerns Linux machines too. My host file is hosts.txt - csultra06 csultra02 csultra05 csultra08 My app file is mpiinit_appfile --- -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit -np 1 /home/cs/manredd/OpenMPI/openmpi-1.1a1r9260/MPITESTS/mpiinit My application program is mpiinit.c - #include int main(int argc, char** argv) { int rc, me; char pname[MPI_MAX_PROCESSOR_NAME]; int plen; MPI_Init( &argc, &argv ); rc = MPI_Comm_rank( MPI_COMM_WORLD, &me ); if (rc != MPI_SUCCESS) { return rc; } MPI_Get_processor_name( pname, &plen ); printf("%s:Hello world from %d\n", pname, me); MPI_Finalize(); return 0; } Compilation is successful csultra06$ mpicc -o mpiinit mpiinit.c However mpirun prints just 6 statements instead of 8. csultra06$ mpirun --hostfile hosts.txt --app mpiinit_appfile csultra02:Hello world from 5 csultra06:Hello world from 0 csultra06:Hello world from 4 csultra02:Hello world from 1 csultra08:Hello world from 3 csultra05:Hello world from 2 The following two more statements are not printed. csultra05:Hello world from 6 csultra08:Hello world from 7 This behavior I observed on my Linux cluster too. I have attached the log for "-d" option for your debugging purposes. Regards, Ravi. - Original Message - From: Brian Barrett List-Post: users@lists.open-mpi.org Date: Monday, March 13, 2006 7:56 pm Subject: Re: [OMPI users] problems with OpenMPI-1.0.1 on SunOS 5.9; problems on heterogeneous cluster To: Open MPI Users Hi Ravi - With the help of another Open MPI user, I spent the weekend finding a couple of issues with Open MPI on Solaris. I believe you are running into the same problems. We're in the process of certifying the changes for release as part of 1.0.2, but it's Monday morning and the release manager hasn't gotten them into the release branch just yet. Could you give the nightly tarball from our development trunk a try and let us know if it solves your problems on Solaris? You probably want last night's 1.1a1r9260 release. http://www.open-mpi.org/nightly/trunk/ Thanks, Brian On Mar 12, 2006, at 11:23 PM, Ravi Manumachu wrote: Hi Brian, Thank you for your help. I have attached all the files you have asked> for in a tar file. Please find attached the 'config.log' and 'libmpi.la' for my Solaris> installation. The output from 'mpicc -showme' is sunos$ mpicc -showme gcc -I/home/cs/manredd/OpenMPI/openmpi-1.0.1/OpenMPI-SunOS-5.9/ include -I/home/cs/manredd/OpenMPI/openmpi-1.0.1/OpenMPI-SunOS- 5.9/include/openmpi/ompi-L/home/cs/manredd/OpenMPI/openmpi- 1.0.1/OpenMPI-SunOS-5.9/lib -lmpi -lorte -lopal -lnsl -lsocket -lthread -laio -lm -lnsl -lsocket - lthread -ldl There are serious issues when running on just solaris machines. I am using the host file and app file shown below. Both the machines are SunOS and are similar. hosts.txt - csultra01 slots=1 csultra02 slots=1 mpiinit_appfile --- -np 1 /home/cs/manredd/OpenMPI/openmpi-1.0.1/MPITESTS/mpiinit_sunos -np 1 /home/cs/manredd/OpenMPI/openmpi-1.0.1/MPITESTS/mpiinit_sunos Running mpirun without -d option hangs. csultra01$ mpirun --hostfile hosts.txt --app mpiinit_appfile hangs Running mpirun with -d option dumps core with output in the file "mpirun_output_d_option.txt", which is attached. The core is also attached. Running just on only one host is also not working. The output from mpirun using "-d" option for this scenario is attached in file "mpirun_output_d_option_one_host.txt". I have also attached the list of packages installed on my solaris machine in "pkginfo.txt" I hope these will help you to resolve the issue. Regards, Ravi. - Original Message - From: Brian Barrett List-Post: users@lists.open-mpi.org Date: Friday, March 10, 2006 7:
Re: [OMPI users] Absoft fortran detected as g77?
On Mar 23, 2006, at 5:32 PM, Michael Kluskens wrote: I have Absoft version 8.2a installed on my OS X 10.4.5 system and in order to do some testing I was trying to build OpenMPI 1.1a1r9364 with it and got the following funny result: *** Fortran 77 compiler checking whether we are using the GNU Fortran 77 compiler... yes checking whether f95 accepts -g... yes checking if Fortran 77 compiler works... yes checking f95 external symbol convention... upper case checking if Fortran 77 compiler supports LOGICAL... yes checking size of Fortran 77 LOGICAL... checking for C type corresponding to LOGICAL... not found configure: WARNING: *** Did not find corresponding C type configure: error: Cannot continue Note that f95 is/should be my Absoft compiler but the configure script flags it as the GNU Fortran 77 compiler. The GNU Fortran thing just means that Absoft sets some flags that g77/ gfortran set. No worries there. I unfortunately can't replicate your issue with our Absoft compilers, but we're at 9.2, so perhaps something has changed. The test code in configure appears to have run correctly (according to your config.log), but it doesn't look like it can read anything from the file that is supposed to contain the sizeof result for the test. Could you try to build and run the attached test code (should just have to untar, run make, and run ./fsize) and let me know what the result is? If all works properly, you should see something like: [21:23] brbarret@frood:ttyp3 ~/f77test% ./fsize size is 4 Thanks, Brian -- Brian Barrett Open MPI developer http://www.open-mpi.org/ f77size.tar.gz Description: GNU Zip compressed data
Re: [OMPI users] Error message about libopal.so
On Mar 22, 2006, at 1:47 PM, Michael Kluskens wrote: Trying to find the cause of one or more errors, might involve libopal.so Built openmpi-1.1a1r9351 on Debian Linux on Operton with PGI 6.1-3 using "./configure --with-gnu-ld F77=pgf77 FFLAGS=-fastsse FC=pgf90 FCFLAGS=-fastsse" My program generates the following error which I do not understand: Signal:11 info.si_errno:0(Success) si_code:1(SEGV_MAPERR) Failing at addr:0x4 [0] func:/usr/local/lib/libopal.so.0 [0x2a959927dd] *** End of error message *** Is it possible I'm over running the OpenMPI buffers, my test program works fine other than the "GPR data corruption" errors (uses MPI_SPAWN and posted previously); the basic MPI difference between my test program and the real program is massive amount of data being distributed via BCAST and SEND/RECV. It worries me that the call stack only goes that deep - there should be more functions listed there (if nothing else, the main() function). Can you run your application in a debugger and try to get a full stack trace? Typically, segmentation faults point to overwriting user buffers, but without more detail, it's hard to pin- point the issue. Thanks, Brian -- Brian Barrett Open MPI developer http://www.open-mpi.org/