Hi Micah,

I do not know if it is exactly what you need but I know that there are
environment variables to use with intel mpi. They are: I_MPI_CC,
I_MPI_CXX, I_MPI_F77, I_MPI_F90. So, you can set this using 'export'
for bash, for instance or directly when you run.

I use in my bashrc:

export I_MPI_CC=icc
export I_MPI_CXX=icpc
export I_MPI_F77=ifort
export I_MPI_F90=ifort


Let me know if it helps.
Fernanda Oliveira


2011/12/14 Micah Sklut <micahs2...@gmail.com>:
> Hi Gustavo,
>
> Here is the output of :
> barells@ip-10-17-153-123:~> /opt/openmpi/intel/bin/mpif90 -showme
> gfortran -I/usr/lib64/mpi/gcc/openmpi/include -pthread
> -I/usr/lib64/mpi/gcc/openmpi/lib64 -L/usr/lib64/mpi/gcc/openmpi/lib64
> -lmpi_f90 -lmpi_f77 -lmpi -lopen-rte -lopen-pal -ldl -Wl,--export-dynamic
> -lnsl -lutil -lm -ldl
>
> This points to gfortran.
>
> I do see what you are saying about the 1.4.2 and 1.4.4 components.
> I'm not sure why that is, but there seems to be some conflict with the
> existing openmpi, before recently installed 1.4.4 and trying to install with
> ifort.
>
>
> On Wed, Dec 14, 2011 at 12:43 PM, Gustavo Correa <g...@ldeo.columbia.edu>
> wrote:
>>
>> How about the output of this?
>>
>> /opt/openmpi/intel/bin/mpif90 -showme
>>
>> Anyway, something seems to be wrong with your OpenMPI installation.
>> Just read the output of your ompi_info in your email below.
>> You will see that the OpenMPI version is 1.4.4.
>> However, most components are version 1.4.2.
>> Do you agree?
>>
>> I would download the OpenMPI 1.4.4 tarball again and start fresh.
>> Untar the tarball in a brand new directory, don't overwrite old stuff.
>> Also, every time your OpenMPI build fails, or if you want to change
>> compilers
>> [say from gfortran to ifort],
>> do a 'make distclean' to cleanup any leftovers of previous builds,
>> and change the destination directory in --prefix= , to install in a
>> different location.
>>
>> I hope this helps,
>> Gus Correa
>>
>> On Dec 14, 2011, at 12:21 PM, Micah Sklut wrote:
>>
>> > Hi Gustav,
>> >
>> > I did read Price's email:
>> >
>> > When I do "which mpif90", i get:
>> > /opt/openmpi/intel/bin/mpif90
>> > which is the desired directory/binary
>> >
>> > As I mentioned, the config log file indicated it was using ifort, and
>> > had no mention of gfortran.
>> > Below is the output from ompi_info. It shows reference to the correct
>> > ifort compiler. But, yet the mpif90 compiler, still yeilds a gfortran
>> > compiler.
>> > -->
>> > barells@ip-10-17-153-123:~> ompi_info
>> >                  Package: Open MPI barells@ip-10-17-148-204 Distribution
>> >                 Open MPI: 1.4.4
>> >    Open MPI SVN revision: r25188
>> >    Open MPI release date: Sep 27, 2011
>> >                 Open RTE: 1.4.4
>> >    Open RTE SVN revision: r25188
>> >    Open RTE release date: Sep 27, 2011
>> >                     OPAL: 1.4.4
>> >        OPAL SVN revision: r25188
>> >        OPAL release date: Sep 27, 2011
>> >             Ident string: 1.4.4
>> >                   Prefix: /usr/lib64/mpi/gcc/openmpi
>> >  Configured architecture: x86_64-unknown-linux-gnu
>> >           Configure host: ip-10-17-148-204
>> >            Configured by: barells
>> >            Configured on: Wed Dec 14 14:22:43 UTC 2011
>> >           Configure host: ip-10-17-148-204
>> >                 Built by: barells
>> >                 Built on: Wed Dec 14 14:27:56 UTC 2011
>> >               Built host: ip-10-17-148-204
>> >               C bindings: yes
>> >             C++ bindings: yes
>> >       Fortran77 bindings: yes (all)
>> >       Fortran90 bindings: yes
>> >  Fortran90 bindings size: small
>> >               C compiler: gcc
>> >      C compiler absolute: /usr/bin/gcc
>> >             C++ compiler: g++
>> >    C++ compiler absolute: /usr/bin/g++
>> >       Fortran77 compiler: ifort
>> >   Fortran77 compiler abs: /opt/intel/fce/9.1.040/bin/ifort
>> >       Fortran90 compiler: ifort
>> >   Fortran90 compiler abs: /opt/intel/fce/9.1.040/bin/ifort
>> >              C profiling: yes
>> >            C++ profiling: yes
>> >      Fortran77 profiling: yes
>> >      Fortran90 profiling: yes
>> >           C++ exceptions: no
>> >           Thread support: posix (mpi: no, progress: no)
>> >            Sparse Groups: no
>> >   Internal debug support: no
>> >      MPI parameter check: runtime
>> > Memory profiling support: no
>> > Memory debugging support: no
>> >          libltdl support: yes
>> >    Heterogeneous support: no
>> >  mpirun default --prefix: no
>> >          MPI I/O support: yes
>> >        MPI_WTIME support: gettimeofday
>> > Symbol visibility support: yes
>> >    FT Checkpoint support: no  (checkpoint thread: no)
>> >            MCA backtrace: execinfo (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >               MCA memory: ptmalloc2 (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >            MCA paffinity: linux (MCA v2.0, API v2.0, Component v1.4.2)
>> >                MCA carto: auto_detect (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >                MCA carto: file (MCA v2.0, API v2.0, Component v1.4.2)
>> >            MCA maffinity: first_use (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >                MCA timer: linux (MCA v2.0, API v2.0, Component v1.4.2)
>> >          MCA installdirs: env (MCA v2.0, API v2.0, Component v1.4.2)
>> >          MCA installdirs: config (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA dpm: orte (MCA v2.0, API v2.0, Component v1.4.2)
>> >               MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.4.2)
>> >            MCA allocator: basic (MCA v2.0, API v2.0, Component v1.4.2)
>> >            MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA coll: basic (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA coll: hierarch (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >                 MCA coll: inter (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA coll: self (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA coll: sm (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA coll: sync (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA coll: tuned (MCA v2.0, API v2.0, Component v1.4.2)
>> >                   MCA io: romio (MCA v2.0, API v2.0, Component v1.4.2)
>> >                MCA mpool: fake (MCA v2.0, API v2.0, Component v1.4.2)
>> >                MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.4.2)
>> >                MCA mpool: sm (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA pml: cm (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA pml: csum (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA pml: v (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA bml: r2 (MCA v2.0, API v2.0, Component v1.4.2)
>> >               MCA rcache: vma (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA btl: ofud (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA btl: openib (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA btl: self (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA btl: sm (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA btl: tcp (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA btl: udapl (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA topo: unity (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA osc: rdma (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA iof: hnp (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA iof: orted (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA iof: tool (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA oob: tcp (MCA v2.0, API v2.0, Component v1.4.2)
>> >                 MCA odls: default (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA ras: slurm (MCA v2.0, API v2.0, Component v1.4.2)
>> >                MCA rmaps: load_balance (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >                MCA rmaps: rank_file (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >                MCA rmaps: round_robin (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >                MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA rml: oob (MCA v2.0, API v2.0, Component v1.4.2)
>> >               MCA routed: binomial (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >               MCA routed: direct (MCA v2.0, API v2.0, Component v1.4.2)
>> >               MCA routed: linear (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA plm: rsh (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA plm: slurm (MCA v2.0, API v2.0, Component v1.4.2)
>> >                MCA filem: rsh (MCA v2.0, API v2.0, Component v1.4.2)
>> >               MCA errmgr: default (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA ess: env (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA ess: hnp (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA ess: singleton (MCA v2.0, API v2.0, Component
>> > v1.4.2)
>> >                  MCA ess: slurm (MCA v2.0, API v2.0, Component v1.4.2)
>> >                  MCA ess: tool (MCA v2.0, API v2.0, Component v1.4.2)
>> >              MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.4.2)
>> >              MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.4.2)
>> >
>> >
>> >
>> >
>> > On Wed, Dec 14, 2011 at 12:11 PM, Gustavo Correa <g...@ldeo.columbia.edu>
>> > wrote:
>> > Hi Micah
>> >
>> > Did you read Tim Prince's email to you?  Check it out.
>> >
>> > Best thing is to set your environment variables [PATH, LD_LIBRARY_PATH,
>> > intel setup]
>> > in your initialization file, .profile/.bashrc or .[t]cshrc.
>> >
>> > What is the output of 'ompi_info'? [From your ifort-built OpenMPI.]
>> > Does it show ifort or gfortran?
>> >
>> > I hope this helps,
>> > Gus Correa
>> >
>> > On Dec 14, 2011, at 11:21 AM, Micah Sklut wrote:
>> >
>> > > Thanks for your thoughts,
>> > >
>> > > It would certainly appear that it is a PATH issue, but I still haven't
>> > > figured it out.
>> > >
>> > > When I type the ifort command, ifort does run.
>> > > The intel path is in my PATH and is the first directory listed.
>> > >
>> > > Looking at the configure.log, there is nothing indicating use or
>> > > mentioning of "gfortran".
>> > >
>> > > gfortran is in the /usr/bin directory, which is in the PATH as well.
>> > >
>> > > Any other suggestions of things to look for?
>> > >
>> > > Thank you,
>> > >
>> > > On Wed, Dec 14, 2011 at 11:05 AM, Gustavo Correa
>> > > <g...@ldeo.columbia.edu> wrote:
>> > > Hi Micah
>> > >
>> > > Is  ifort in your PATH?
>> > > If not, the OpenMPI configure script will use any fortran compiler it
>> > > finds first, which may be gfortran.
>> > > You need to run the Intel compiler startup script before you run the
>> > > OpenMPI configure.
>> > > The easy thing to do is to source the Intel script inside your
>> > > .profile/.bashrc or .[t]cshrc file.
>> > > I hope this helps,
>> > >
>> > > Gus Correa
>> > >
>> > > On Dec 14, 2011, at 9:49 AM, Micah Sklut wrote:
>> > >
>> > > > Hi All,
>> > > >
>> > > > I have installed openmpi for gfortran, but am now attempting to
>> > > > install openmpi as ifort.
>> > > >
>> > > > I have run the following configuration:
>> > > > ./configure --prefix=/opt/openmpi/intel CC=gcc CXX=g++ F77=ifort
>> > > > FC=ifort
>> > > >
>> > > > The install works successfully, but when I run
>> > > > /opt/openmpi/intel/bin/mpif90, it runs as gfortran.
>> > > > Oddly, when I am user: root, the same mpif90 runs as ifort.
>> > > >
>> > > > Can someone please alleviate my confusion as to why I mpif90 is not
>> > > > running as ifort?
>> > > >
>> > > > Thank you for your suggestions,
>> > > >
>> > > > --
>> > > > Micah
>> > > >
>> > > >
>> > > > _______________________________________________
>> > > > users mailing list
>> > > > us...@open-mpi.org
>> > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> > >
>> > >
>> > > _______________________________________________
>> > > users mailing list
>> > > us...@open-mpi.org
>> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> > >
>> > >
>> > >
>> > > --
>> > > Micah Sklut
>> > >
>> > >
>> > > _______________________________________________
>> > > users mailing list
>> > > us...@open-mpi.org
>> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >
>> >
>> > _______________________________________________
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >
>> >
>> >
>> > --
>> > Micah Sklut
>> >
>> >
>> > _______________________________________________
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
>
> --
> Micah Sklut
>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to