Yes, Empire does the fluid structure coupling. It couples OpenFoam (fluid 
analysis) and Abaqus (structural analysis).
Does all the software need to have the same MPI architecture in order to 
communicate ?
Regards,Islem 

    Le Mardi 24 mai 2016 1h02, Gilles Gouaillardet <gil...@rist.or.jp> a écrit :
 

  what do you mean by coupling ? does Empire and OpenFoam communicate via MPI ? 
wouldn't it be much easier if you rebuild OpenFoam with mpich or intelmpi ? 
  Cheers, 
  Gilles
  
 On 5/24/2016 8:44 AM, Megdich Islem wrote:
  
  "Open MPI does not work when MPICH or intel MPI are installed" 
  Thank you for your suggestion. But I need to run OpenFoam and Empire at the 
same time. In fact, Empire couples OpenFoam with another software. 
  Is there any solution for this case ? 
  
  Regards, Islem 
 
      Le Lundi 23 mai 2016 17h00, "users-requ...@open-mpi.org" 
<users-requ...@open-mpi.org> a écrit :
  
 
 Send users mailing list submissions to
     us...@open-mpi.org
 
 To subscribe or unsubscribe via the World Wide Web, visit
     https://www.open-mpi.org/mailman/listinfo.cgi/users
 or, via email, send a message with subject or body 'help' to
     users-requ...@open-mpi.org
 
 You can reach the person managing the list at
     users-ow...@open-mpi.org
 
 When replying, please edit your Subject line so it is more specific
 than "Re: Contents of users digest..."
 
 
 Today's Topics:
 
   1. Re: Open MPI does not work when MPICH or intel MPI are
       installed (Andy Riebs)
   2. segmentation fault for slot-list and openmpi-1.10.3rc2
       (Siegmar Gross)
   3. Re: problem about mpirun on two nodes (Jeff Squyres (jsquyres))
   4. Re: Open MPI does not work when MPICH or intel MPI are
       installed (Gilles Gouaillardet)
   5. Re: segmentation fault for slot-list and    openmpi-1.10.3rc2
       (Ralph Castain)
   6. mpirun java (Claudio Stamile)
 
 
----------------------------------------------------------------------
 
 [Message discarded by content filter]
 ------------------------------
 
 Message: 2
 Date: Mon, 23 May 2016 15:26:52 +0200
 From: Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de>
 To: Open MPI Users <us...@open-mpi.org>
 Subject: [OMPI users] segmentation fault for slot-list and
     openmpi-1.10.3rc2
 Message-ID:
     <241613b1-ada6-292f-eeb9-722fc8fa2...@informatik.hs-fulda.de>
 Content-Type: text/plain; charset=utf-8; format=flowed
 
 Hi,
 
 I installed openmpi-1.10.3rc2 on my "SUSE Linux Enterprise Server
 12 (x86_64)" with Sun C 5.13  and gcc-6.1.0. Unfortunately I get
 a segmentation fault for "--slot-list" for one of my small programs.
 
 
 loki spawn 119 ompi_info | grep -e "OPAL repo revision:" -e "C compiler 
absolute:"
       OPAL repo revision: v1.10.2-201-gd23dda8
       C compiler absolute: /usr/local/gcc-6.1.0/bin/gcc
 
 
 loki spawn 120 mpiexec -np 1 --host loki,loki,loki,loki,loki spawn_master
 
 Parent process 0 running on loki
   I create 4 slave processes
 
 Parent process 0: tasks in MPI_COMM_WORLD:                    1
                   tasks in COMM_CHILD_PROCESSES local group:  1
                   tasks in COMM_CHILD_PROCESSES remote group: 4
 
 Slave process 0 of 4 running on loki
 Slave process 1 of 4 running on loki
 Slave process 2 of 4 running on loki
 spawn_slave 2: argv[0]: spawn_slave
 Slave process 3 of 4 running on loki
 spawn_slave 0: argv[0]: spawn_slave
 spawn_slave 1: argv[0]: spawn_slave
 spawn_slave 3: argv[0]: spawn_slave
 
 
 
 
 loki spawn 121 mpiexec -np 1 --host loki --slot-list 0:0-5,1:0-5 spawn_master
 
 Parent process 0 running on loki
   I create 4 slave processes
 
 [loki:17326] *** Process received signal ***
 [loki:17326] Signal: Segmentation fault (11)
 [loki:17326] Signal code: Address not mapped (1)
 [loki:17326] Failing at address: 0x8
 [loki:17326] [ 0] /lib64/libpthread.so.0(+0xf870)[0x7f4e469b3870]
 [loki:17326] [ 1] *** An error occurred in MPI_Init
 *** on a NULL communicator
 *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
 ***    and potentially your MPI job)
 [loki:17324] Local abort before MPI_INIT completed successfully; not able to 
 aggregate error messages, and not able to guarantee that all other processes 
 were killed!
/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(ompi_proc_self+0x35)[0x7f4e46c165b0]
 [loki:17326] [ 2] 
/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(ompi_comm_init+0x68b)[0x7f4e46bf5b08]
 [loki:17326] [ 3] *** An error occurred in MPI_Init
 *** on a NULL communicator
 *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
 ***    and potentially your MPI job)
 [loki:17325] Local abort before MPI_INIT completed successfully; not able to 
 aggregate error messages, and not able to guarantee that all other processes 
 were killed!
/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(ompi_mpi_init+0xa90)[0x7f4e46c1be8a]
 [loki:17326] [ 4] 
/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(MPI_Init+0x180)[0x7f4e46c5828e]
 [loki:17326] [ 5] spawn_slave[0x40097e]
 [loki:17326] [ 6] /lib64/libc.so.6(__libc_start_main+0xf5)[0x7f4e4661db05]
 [loki:17326] [ 7] spawn_slave[0x400a54]
 [loki:17326] *** End of error message ***
 -------------------------------------------------------
 Child job 2 terminated normally, but 1 process returned
 a non-zero exit code.. Per user-direction, the job has been aborted.
 -------------------------------------------------------
--------------------------------------------------------------------------
 mpiexec detected that one or more processes exited with non-zero status, thus 
 causing
 the job to be terminated. The first process to do so was:
 
   Process name: [[56340,2],0]
   Exit code:    1
--------------------------------------------------------------------------
 loki spawn 122
 
 
 
 
 I would be grateful, if somebody can fix the problem. Thank you
 very much for any help in advance.
 
 
 Kind regards
 
 Siegmar
 
 
 ------------------------------
 
 Message: 3
 Date: Mon, 23 May 2016 14:13:11 +0000
 From: "Jeff Squyres (jsquyres)" <jsquy...@cisco.com>
 To: "Open MPI User's List" <us...@open-mpi.org>
 Subject: Re: [OMPI users] problem about mpirun on two nodes
 Message-ID: <b2033c1d-8aa4-4823-b984-92756dc1e...@cisco.com>
 Content-Type: text/plain; charset="us-ascii"
 
 On May 21, 2016, at 11:31 PM, dour...@aol.com wrote:
 > 
 > I encountered a problem about mpirun and SSH when using OMPI 1.10.0 compiled 
 > with gcc, running on centos7.2.
 > When I execute mpirun on my 2 node cluster, I get the following errors 
 > pasted below.
 > 
 > [douraku@master home]$ mpirun -np 12 a.out
 > Permission denied (publickey,gssapi-keyex,gssapi-with-mic).
 
 This is the key right here: you got a permission denied error when you 
(assumedly) tried to execute on the remote server.
 
 Triple check your ssh settings to ensure that you can run on the remote 
server(s) without a password or interactive passphrase entry.
 
 -- 
 Jeff Squyres
 jsquy...@cisco.com
 For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/
 
 
 
 ------------------------------
 
 Message: 4
 Date: Mon, 23 May 2016 23:31:30 +0900
 From: Gilles Gouaillardet <gilles.gouaillar...@gmail.com>
 To: Open MPI Users <us...@open-mpi.org>
 Subject: Re: [OMPI users] Open MPI does not work when MPICH or intel
     MPI are    installed
 Message-ID:
     <CAAkFZ5u86Q0ev=ospehnKvd0kumYBoeMD8WF=j+taduh3xy...@mail.gmail.com>
 Content-Type: text/plain; charset="utf-8"
 
 modules are way more friendly that manually setting and exporting your
 environment.
 the issue here is you are setting your environment in your .bashrc, and
 that cannot work if your account is used with various MPI implementations.
 (unless your .bashrc checks a third party variable to select the
 appropriate mpi, in this case, simply extend the logic to select openmpi)
 
 if you configure'd with --enable-mpirun-prefix-by-default, you should not
 need anything in your environment.
 
 Cheers,
 
 Gilles
 
 On Monday, May 23, 2016, Andy Riebs <andy.ri...@hpe.com> wrote:
 
 > Hi,
 >
 > The short answer: Environment module files are probably the best solution
 > for your problem.
 >
 > The long answer: See
 > <http://www.admin-magazine.com/HPC/Articles/Environment-Modules>
 > <http://www.admin-magazine.com/HPC/Articles/Environment-Modules>, which
 > pretty much addresses your question.
 >
 > Andy
 >
 > On 05/23/2016 07:40 AM, Megdich Islem wrote:
 >
 > Hi,
 >
 > I am using 2 software, one is called Open Foam and the other called EMPIRE
 > that need to run together at the same time.
 > Open Foam uses  Open MPI implementation and EMPIRE uses either MPICH or
 > intel mpi.
 > The version of Open MPI that comes with Open Foam is 1.6.5.
 > I am using Intel (R) MPI Library for linux * OS, version 5.1.3 and MPICH
 > 3.0.4.
 >
 > My problem is when I have the environment variables of  either mpich or
 > Intel MPI  sourced to bashrc, I fail to run a case of Open Foam with
 > parallel processing ( You find attached a picture of the error I got )
 > This is an example of a command line I use to run Open Foam
 > mpirun -np 4 interFoam -parallel
 >
 > Once I keep the environment variable of OpenFoam only, the parallel
 > processing works without any problem, so I won't be able to run EMPIRE.
 >
 > I am sourcing the environment variables in this way:
 >
 > For Open Foam:
 > source /opt/openfoam30/etc/bashrc
 >
 > For MPICH 3.0.4
 >
 > export PATH=/home/islem/Desktop/mpich/bin:$PATH
 > export LD_LIBRARY_PATH="/home/islem/Desktop/mpich/lib/:$LD_LIBRARY_PATH"
 > export MPICH_F90=gfortran
 > export MPICH_CC=/opt/intel/bin/icc
 > export MPICH_CXX=/opt/intel/bin/icpc
 > export MPICH-LINK_CXX="-L/home/islem/Desktop/mpich/lib/ -Wl,-rpath
 > -Wl,/home/islem/Desktop/mpich/lib -lmpichcxx -lmpich -lopa -lmpl -lrt
 > -lpthread"
 >
 > For intel
 >
 > export PATH=$PATH:/opt/intel/bin/
 > LD_LIBRARY_PATH="/opt/intel/lib/intel64:$LD_LIBRARY_PATH"
 > export LD_LIBRARY_PATH
 > source
 >/opt/intel/compilers_and_libraries_2016.3.210/linux/mpi/intel64/bin/mpivars.sh
 > intel64
 >
 > If Only Open Foam is sourced, mpirun --version gives OPEN MPI (1.6.5)
 > If Open Foam and MPICH are sourced, mpirun --version gives mpich 3.0.1
 > If Open Foam and intel MPI are sourced, mpirun --version gives intel (R)
 > MPI libarary for linux, version 5.1.3
 >
 > My question is why I can't have two MPI implementation installed and
 > sourced together. How can I solve the problem ?
 >
 > Regards,
 > Islem Megdiche
 >
 >
 >
 >
 >
 >
 > _______________________________________________
 > users mailing listus...@open-mpi.org 
 > <javascript:_e(%7B%7D,'cvml','us...@open-mpi.org');>
 > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
 > Link to this post: 
 > http://www.open-mpi.org/community/lists/users/2016/05/29279.php
 >
 >
 >
 -------------- next part --------------
 HTML attachment scrubbed and removed
 
 ------------------------------
 
 Message: 5
 Date: Mon, 23 May 2016 08:45:53 -0700
 From: Ralph Castain <r...@open-mpi.org>
 To: Open MPI Users <us...@open-mpi.org>
 Subject: Re: [OMPI users] segmentation fault for slot-list and
     openmpi-1.10.3rc2
 Message-ID: <73195d72-cea7-4afc-9527-8f725c8b1...@open-mpi.org>
 Content-Type: text/plain; charset="utf-8"
 
 I cannot replicate the problem - both scenarios work fine for me. I?m not 
convinced your test code is correct, however, as you call Comm_free the 
inter-communicator but didn?t call Comm_disconnect. Checkout the attached for a 
correct code and see if it works for you.
 
 FWIW: I don?t know how many cores you have on your sockets, but if you have 6 
cores/socket, then your slot-list is equivalent to ??bind-to none? as the 
slot-list applies to every process being launched
 
 -------------- next part --------------
 A non-text attachment was scrubbed...
 Name: simple_spawn.c
 Type: application/octet-stream
 Size: 1926 bytes
 Desc: not available
 URL: 
<http://www.open-mpi.org/MailArchives/users/attachments/20160523/7554b3ec/attachment.obj>
 -------------- next part --------------
 
 
 > On May 23, 2016, at 6:26 AM, Siegmar Gross 
 > <siegmar.gr...@informatik.hs-fulda.de> wrote:
 > 
 > Hi,
 > 
 > I installed openmpi-1.10.3rc2 on my "SUSE Linux Enterprise Server
 > 12 (x86_64)" with Sun C 5.13  and gcc-6.1.0. Unfortunately I get
 > a segmentation fault for "--slot-list" for one of my small programs.
 > 
 > 
 > loki spawn 119 ompi_info | grep -e "OPAL repo revision:" -e "C compiler 
 > absolute:"
 >      OPAL repo revision: v1.10.2-201-gd23dda8
 >    C compiler absolute: /usr/local/gcc-6.1.0/bin/gcc
 > 
 > 
 > loki spawn 120 mpiexec -np 1 --host loki,loki,loki,loki,loki spawn_master
 > 
 > Parent process 0 running on loki
 >  I create 4 slave processes
 > 
 > Parent process 0: tasks in MPI_COMM_WORLD:                    1
 >                  tasks in COMM_CHILD_PROCESSES local group:  1
 >                  tasks in COMM_CHILD_PROCESSES remote group: 4
 > 
 > Slave process 0 of 4 running on loki
 > Slave process 1 of 4 running on loki
 > Slave process 2 of 4 running on loki
 > spawn_slave 2: argv[0]: spawn_slave
 > Slave process 3 of 4 running on loki
 > spawn_slave 0: argv[0]: spawn_slave
 > spawn_slave 1: argv[0]: spawn_slave
 > spawn_slave 3: argv[0]: spawn_slave
 > 
 > 
 > 
 > 
 > loki spawn 121 mpiexec -np 1 --host loki --slot-list 0:0-5,1:0-5 spawn_master
 > 
 > Parent process 0 running on loki
 >  I create 4 slave processes
 > 
 > [loki:17326] *** Process received signal ***
 > [loki:17326] Signal: Segmentation fault (11)
 > [loki:17326] Signal code: Address not mapped (1)
 > [loki:17326] Failing at address: 0x8
 > [loki:17326] [ 0] /lib64/libpthread.so.0(+0xf870)[0x7f4e469b3870]
 > [loki:17326] [ 1] *** An error occurred in MPI_Init
 > *** on a NULL communicator
 > *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
 > ***    and potentially your MPI job)
 > [loki:17324] Local abort before MPI_INIT completed successfully; not able to 
 > aggregate error messages, and not able to guarantee that all other processes 
 > were killed!
 >/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(ompi_proc_self+0x35)[0x7f4e46c165b0]
 > [loki:17326] [ 
 > 2]/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(ompi_comm_init+0x68b)[0x7f4e46bf5b08]
 > [loki:17326] [ 3] *** An error occurred in MPI_Init
 > *** on a NULL communicator
 > *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
 > ***    and potentially your MPI job)
 > [loki:17325] Local abort before MPI_INIT completed successfully; not able to 
 > aggregate error messages, and not able to guarantee that all other processes 
 > were killed!
 >/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(ompi_mpi_init+0xa90)[0x7f4e46c1be8a]
 > [loki:17326] [ 
 > 4]/usr/local/openmpi-1.10.3_64_gcc/lib64/libmpi.so.12(MPI_Init+0x180)[0x7f4e46c5828e]
 > [loki:17326] [ 5] spawn_slave[0x40097e]
 > [loki:17326] [ 6] /lib64/libc.so.6(__libc_start_main+0xf5)[0x7f4e4661db05]
 > [loki:17326] [ 7] spawn_slave[0x400a54]
 > [loki:17326] *** End of error message ***
 > -------------------------------------------------------
 > Child job 2 terminated normally, but 1 process returned
 > a non-zero exit code.. Per user-direction, the job has been aborted.
 > -------------------------------------------------------
 >--------------------------------------------------------------------------
 > mpiexec detected that one or more processes exited with non-zero status, 
 > thus causing
 > the job to be terminated. The first process to do so was:
 > 
 >  Process name: [[56340,2],0]
 >  Exit code:    1
 >--------------------------------------------------------------------------
 > loki spawn 122
 > 
 > 
 > 
 > 
 > I would be grateful, if somebody can fix the problem. Thank you
 > very much for any help in advance.
 > 
 > 
 > Kind regards
 > 
 > Siegmar
 > _______________________________________________
 > users mailing list
 > us...@open-mpi.org
 > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
 > Link to this post: 
 > http://www.open-mpi.org/community/lists/users/2016/05/29281.php
 
 
 ------------------------------
 
 Message: 6
 Date: Mon, 23 May 2016 17:47:53 +0200
 From: Claudio Stamile <claudiostam...@gmail.com>
 To: us...@open-mpi.org
 Subject: [OMPI users] mpirun java
 Message-ID:
     <caadd79zz1waonmr5hod3jp51qeduhmp5ww8tp7eujldfshe...@mail.gmail.com>
 Content-Type: text/plain; charset="utf-8"
 
 Dear all,
 
 I'm using openmpi for Java.
 I've a problem when I try to use more option parameters in my java command.
 More in detail I run mpirun as follow:
 
 mpirun -n 5 java -cp path1:path2 -Djava.library.path=pathLibs
 classification.MyClass
 
 It seems that the option "-Djava.library.path" is ignored when i execute
 the command.
 
 Is it normal ?
 
 Do you know how to solve this problem ?
 
 Thank you.
 
 Best,
 Claudio
 
 -- 
 C.
 -------------- next part --------------
 HTML attachment scrubbed and removed
 
 ------------------------------
 
 Subject: Digest Footer
 
 _______________________________________________
 users mailing list
 us...@open-mpi.org
 https://www.open-mpi.org/mailman/listinfo.cgi/users
 
 ------------------------------
 
 End of users Digest, Vol 3510, Issue 2
 **************************************
 
 
      
  
 _______________________________________________
users mailing list
us...@open-mpi.org
Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2016/05/29295.php 
 
 

Reply via email to