Thank you very much guys. Now a more serious issue: I am using mpi with laamps (a molecular dynamics package) on a single Rack Dell server Poweredge R810 (4 eight-core processors, 128 Gb RAM memory). I am now potentially interested into buying the Intel MPI 4.1 libraries, and I am trying them by exploiting the 30 days trial. However, I am not experiencing any significant improved performance by using the Intel MPI libraries with respect to the Open Mpi (compiled with the Itel compilers).
Here there is the (makefile) working configuration for the Intel MPI 4.1 compilers: CC = /opt/intel/impi/4.1.0.024/intel64/bin/mpiicpc CCFLAGS = -O -DMPICH_IGNORE_CXX_SEEK -DMPICH_SKIP_MPICXX And here there is the Open Mpi one: CC = /usr/local/bin/mpicc CCFLAGS = -O -mpicc I also tried the flag -O3 but I detected no significant differences in performance. Now, I would be considering the Intel Mpi libraries, provided this would bring to a significant increase in performance with respect to Open Mpi. I have evidence that there is room to improve because laamps under the same conditions and on an HP Z650 with two 6-core processors (the clock frequency is the same though, and tests for comparison were done on parallel runs using 8 cores), improves of nearly the 70% by using the proprietary HP MPI libraries. Kind regards Giuseppe 2012/10/27 Ralph Castain <r...@open-mpi.org> > The reason is that you aren't actually running Open MPI - those error > messages are coming from MPICH. Check your path and ensure you put the OMPI > install location first, or use the absolute path to the OMPI mpirun > > On Oct 27, 2012, at 8:46 AM, Giuseppe P. <istruzi...@gmail.com> wrote: > > Hello! > > I have built open mpi 1.6 with Intel compilers (2013 versions). > Compilation was smooth, however even when I try to execute > the simple program hello.c: > > mpirun -np 4 ./hello_c.x > [mpie...@claudio.ukzn] HYDU_create_process (./utils/launch/launch.c:102): > execvp error on file > /opt/intel/composer_xe_2013.0.079/mpirt/bin/intel64/pmi_proxy (No such file > or directory) > [mpie...@claudio.ukzn] HYD_pmcd_pmiserv_proxy_init_cb > (./pm/pmiserv/pmiserv_cb.c:1177): assert (!closed) failed > [mpie...@claudio.ukzn] HYDT_dmxu_poll_wait_for_event > (./tools/demux/demux_poll.c:77): callback returned error status > [mpie...@claudio.ukzn] HYD_pmci_wait_for_completion > (./pm/pmiserv/pmiserv_pmci.c:358): error waiting for event > [mpie...@claudio.ukzn] main (./ui/mpich/mpiexec.c:689): process manager > error waiting for completion > > Before that, there was an additional error, since also the file mpivars.sh > was not present in /opt/intel/composer_xe_2013.0.079/mpirt/bin/intel64/. > Even though I managed to create one and it worked: > > #!/bin/bash > > if [ -z "`echo $PATH | grep /usr/local/bin`" ]; then > export PATH=/usr/local/bin:$PATH > fi > > if [ -z "`echo $LD_LIBRARY_PATH | grep /usr/local/lib`" ]; then > if [ -n "$LD_LIBRARY_PATH" ]; then > export LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH > else > export LD_LIBRARY_PATH=/usr/local/lib > fi > fi > > I do not have any clue about how to generate the file pmi_proxy. > > Thank you in advance for your help! > _______________________________________________ > > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >