Hi Gustav, I did read Price's email:
When I do "which mpif90", i get: /opt/openmpi/intel/bin/mpif90 which is the desired directory/binary As I mentioned, the config log file indicated it was using ifort, and had no mention of gfortran. Below is the output from ompi_info. It shows reference to the correct ifort compiler. But, yet the mpif90 compiler, still yeilds a gfortran compiler. --> barells@ip-10-17-153-123:~> ompi_info Package: Open MPI barells@ip-10-17-148-204 Distribution Open MPI: 1.4.4 Open MPI SVN revision: r25188 Open MPI release date: Sep 27, 2011 Open RTE: 1.4.4 Open RTE SVN revision: r25188 Open RTE release date: Sep 27, 2011 OPAL: 1.4.4 OPAL SVN revision: r25188 OPAL release date: Sep 27, 2011 Ident string: 1.4.4 Prefix: /usr/lib64/mpi/gcc/openmpi Configured architecture: x86_64-unknown-linux-gnu Configure host: ip-10-17-148-204 Configured by: barells Configured on: Wed Dec 14 14:22:43 UTC 2011 Configure host: ip-10-17-148-204 Built by: barells Built on: Wed Dec 14 14:27:56 UTC 2011 Built host: ip-10-17-148-204 C bindings: yes C++ bindings: yes Fortran77 bindings: yes (all) Fortran90 bindings: yes Fortran90 bindings size: small C compiler: gcc C compiler absolute: /usr/bin/gcc C++ compiler: g++ C++ compiler absolute: /usr/bin/g++ Fortran77 compiler: ifort Fortran77 compiler abs: /opt/intel/fce/9.1.040/bin/ifort Fortran90 compiler: ifort Fortran90 compiler abs: /opt/intel/fce/9.1.040/bin/ifort C profiling: yes C++ profiling: yes Fortran77 profiling: yes Fortran90 profiling: yes C++ exceptions: no Thread support: posix (mpi: no, progress: no) Sparse Groups: no Internal debug support: no MPI parameter check: runtime Memory profiling support: no Memory debugging support: no libltdl support: yes Heterogeneous support: no mpirun default --prefix: no MPI I/O support: yes MPI_WTIME support: gettimeofday Symbol visibility support: yes FT Checkpoint support: no (checkpoint thread: no) MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.4.2) MCA memory: ptmalloc2 (MCA v2.0, API v2.0, Component v1.4.2) MCA paffinity: linux (MCA v2.0, API v2.0, Component v1.4.2) MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.4.2) MCA carto: file (MCA v2.0, API v2.0, Component v1.4.2) MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.4.2) MCA timer: linux (MCA v2.0, API v2.0, Component v1.4.2) MCA installdirs: env (MCA v2.0, API v2.0, Component v1.4.2) MCA installdirs: config (MCA v2.0, API v2.0, Component v1.4.2) MCA dpm: orte (MCA v2.0, API v2.0, Component v1.4.2) MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.4.2) MCA allocator: basic (MCA v2.0, API v2.0, Component v1.4.2) MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.4.2) MCA coll: basic (MCA v2.0, API v2.0, Component v1.4.2) MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.4.2) MCA coll: inter (MCA v2.0, API v2.0, Component v1.4.2) MCA coll: self (MCA v2.0, API v2.0, Component v1.4.2) MCA coll: sm (MCA v2.0, API v2.0, Component v1.4.2) MCA coll: sync (MCA v2.0, API v2.0, Component v1.4.2) MCA coll: tuned (MCA v2.0, API v2.0, Component v1.4.2) MCA io: romio (MCA v2.0, API v2.0, Component v1.4.2) MCA mpool: fake (MCA v2.0, API v2.0, Component v1.4.2) MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.4.2) MCA mpool: sm (MCA v2.0, API v2.0, Component v1.4.2) MCA pml: cm (MCA v2.0, API v2.0, Component v1.4.2) MCA pml: csum (MCA v2.0, API v2.0, Component v1.4.2) MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.4.2) MCA pml: v (MCA v2.0, API v2.0, Component v1.4.2) MCA bml: r2 (MCA v2.0, API v2.0, Component v1.4.2) MCA rcache: vma (MCA v2.0, API v2.0, Component v1.4.2) MCA btl: ofud (MCA v2.0, API v2.0, Component v1.4.2) MCA btl: openib (MCA v2.0, API v2.0, Component v1.4.2) MCA btl: self (MCA v2.0, API v2.0, Component v1.4.2) MCA btl: sm (MCA v2.0, API v2.0, Component v1.4.2) MCA btl: tcp (MCA v2.0, API v2.0, Component v1.4.2) MCA btl: udapl (MCA v2.0, API v2.0, Component v1.4.2) MCA topo: unity (MCA v2.0, API v2.0, Component v1.4.2) MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.4.2) MCA osc: rdma (MCA v2.0, API v2.0, Component v1.4.2) MCA iof: hnp (MCA v2.0, API v2.0, Component v1.4.2) MCA iof: orted (MCA v2.0, API v2.0, Component v1.4.2) MCA iof: tool (MCA v2.0, API v2.0, Component v1.4.2) MCA oob: tcp (MCA v2.0, API v2.0, Component v1.4.2) MCA odls: default (MCA v2.0, API v2.0, Component v1.4.2) MCA ras: slurm (MCA v2.0, API v2.0, Component v1.4.2) MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.4.2) MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.4.2) MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.4.2) MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.4.2) MCA rml: oob (MCA v2.0, API v2.0, Component v1.4.2) MCA routed: binomial (MCA v2.0, API v2.0, Component v1.4.2) MCA routed: direct (MCA v2.0, API v2.0, Component v1.4.2) MCA routed: linear (MCA v2.0, API v2.0, Component v1.4.2) MCA plm: rsh (MCA v2.0, API v2.0, Component v1.4.2) MCA plm: slurm (MCA v2.0, API v2.0, Component v1.4.2) MCA filem: rsh (MCA v2.0, API v2.0, Component v1.4.2) MCA errmgr: default (MCA v2.0, API v2.0, Component v1.4.2) MCA ess: env (MCA v2.0, API v2.0, Component v1.4.2) MCA ess: hnp (MCA v2.0, API v2.0, Component v1.4.2) MCA ess: singleton (MCA v2.0, API v2.0, Component v1.4.2) MCA ess: slurm (MCA v2.0, API v2.0, Component v1.4.2) MCA ess: tool (MCA v2.0, API v2.0, Component v1.4.2) MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.4.2) MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.4.2) On Wed, Dec 14, 2011 at 12:11 PM, Gustavo Correa <g...@ldeo.columbia.edu>wrote: > Hi Micah > > Did you read Tim Prince's email to you? Check it out. > > Best thing is to set your environment variables [PATH, LD_LIBRARY_PATH, > intel setup] > in your initialization file, .profile/.bashrc or .[t]cshrc. > > What is the output of 'ompi_info'? [From your ifort-built OpenMPI.] > Does it show ifort or gfortran? > > I hope this helps, > Gus Correa > > On Dec 14, 2011, at 11:21 AM, Micah Sklut wrote: > > > Thanks for your thoughts, > > > > It would certainly appear that it is a PATH issue, but I still haven't > figured it out. > > > > When I type the ifort command, ifort does run. > > The intel path is in my PATH and is the first directory listed. > > > > Looking at the configure.log, there is nothing indicating use or > mentioning of "gfortran". > > > > gfortran is in the /usr/bin directory, which is in the PATH as well. > > > > Any other suggestions of things to look for? > > > > Thank you, > > > > On Wed, Dec 14, 2011 at 11:05 AM, Gustavo Correa <g...@ldeo.columbia.edu> > wrote: > > Hi Micah > > > > Is ifort in your PATH? > > If not, the OpenMPI configure script will use any fortran compiler it > finds first, which may be gfortran. > > You need to run the Intel compiler startup script before you run the > OpenMPI configure. > > The easy thing to do is to source the Intel script inside your > .profile/.bashrc or .[t]cshrc file. > > I hope this helps, > > > > Gus Correa > > > > On Dec 14, 2011, at 9:49 AM, Micah Sklut wrote: > > > > > Hi All, > > > > > > I have installed openmpi for gfortran, but am now attempting to > install openmpi as ifort. > > > > > > I have run the following configuration: > > > ./configure --prefix=/opt/openmpi/intel CC=gcc CXX=g++ F77=ifort > FC=ifort > > > > > > The install works successfully, but when I run > /opt/openmpi/intel/bin/mpif90, it runs as gfortran. > > > Oddly, when I am user: root, the same mpif90 runs as ifort. > > > > > > Can someone please alleviate my confusion as to why I mpif90 is not > running as ifort? > > > > > > Thank you for your suggestions, > > > > > > -- > > > Micah > > > > > > > > > _______________________________________________ > > > users mailing list > > > us...@open-mpi.org > > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > > > > > -- > > Micah Sklut > > > > > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > -- Micah Sklut