Sorry about that - I hadn’t brought it over to the 1.8 branch yet. I’ve done so 
now, which means the ERROR_LOG shouldn’t show up any more. It won’t fix the 
memheap problem, though.

You might try adding “--mca memheap_base_verbose 100” to your cmd line so we 
can see why none of the memheap components are being selected.


> On Apr 12, 2015, at 11:30 AM, Andy Riebs <andy.ri...@hp.com> wrote:
> 
> Hi Ralph,
> 
> Here's the output with openmpi-v1.8.4-202-gc2da6a5.tar.bz2 
> <https://www.open-mpi.org/nightly/v1.8/openmpi-v1.8.4-202-gc2da6a5.tar.bz2>:
> 
> $ shmemrun -H localhost -N 2 --mca sshmem mmap  --mca plm_base_verbose 5 
> $PWD/mic.out
> [atl1-01-mic0:190189] mca:base:select:(  plm) Querying component [rsh]
> [atl1-01-mic0:190189] [[INVALID],INVALID] plm:rsh_lookup on agent ssh : rsh 
> path NULL
> [atl1-01-mic0:190189] mca:base:select:(  plm) Query of component [rsh] set 
> priority to 10
> [atl1-01-mic0:190189] mca:base:select:(  plm) Querying component [isolated]
> [atl1-01-mic0:190189] mca:base:select:(  plm) Query of component [isolated] 
> set priority to 0
> [atl1-01-mic0:190189] mca:base:select:(  plm) Querying component [slurm]
> [atl1-01-mic0:190189] mca:base:select:(  plm) Skipping component [slurm]. 
> Query failed to return a module
> [atl1-01-mic0:190189] mca:base:select:(  plm) Selected component [rsh]
> [atl1-01-mic0:190189] plm:base:set_hnp_name: initial bias 190189 nodename 
> hash 4121194178
> [atl1-01-mic0:190189] plm:base:set_hnp_name: final jobfam 32137
> [atl1-01-mic0:190189] [[32137,0],0] plm:rsh_setup on agent ssh : rsh path NULL
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:receive start comm
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:setup_job
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:setup_vm
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:setup_vm creating map
> [atl1-01-mic0:190189] [[32137,0],0] setup:vm: working unmanaged allocation
> [atl1-01-mic0:190189] [[32137,0],0] using dash_host
> [atl1-01-mic0:190189] [[32137,0],0] checking node atl1-01-mic0
> [atl1-01-mic0:190189] [[32137,0],0] ignoring myself
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:setup_vm only HNP in allocation
> [atl1-01-mic0:190189] [[32137,0],0] complete_setup on job [32137,1]
> [atl1-01-mic0:190189] [[32137,0],0] ORTE_ERROR_LOG: Not found in file 
> base/plm_base_launch_support.c at line 440
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:launch_apps for job [32137,1]
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:launch wiring up iof for job 
> [32137,1]
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:launch [32137,1] registered
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:launch job [32137,1] is not a 
> dynamic spawn
> --------------------------------------------------------------------------
> It looks like SHMEM_INIT failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during SHMEM_INIT; some of which are due to configuration or environment
> problems.  This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open SHMEM
> developer):
> 
>   mca_memheap_base_select() failed
>   --> Returned "Error" (-1) instead of "Success" (0)
> --------------------------------------------------------------------------
> [atl1-01-mic0:190191] Error: pshmem_init.c:61 - shmem_init() SHMEM failed to 
> initialize - aborting
> [atl1-01-mic0:190192] Error: pshmem_init.c:61 - shmem_init() SHMEM failed to 
> initialize - aborting
> --------------------------------------------------------------------------
> SHMEM_ABORT was invoked on rank 1 (pid 190192, host=atl1-01-mic0) with 
> errorcode -1.
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> A SHMEM process is aborting at a time when it cannot guarantee that all
> of its peer processes in the job will be killed properly.  You should
> double check that everything has shut down cleanly.
> 
> Local host: atl1-01-mic0
> PID:        190192
> --------------------------------------------------------------------------
> -------------------------------------------------------
> Primary job  terminated normally, but 1 process returned
> a non-zero exit code.. Per user-direction, the job has been aborted.
> -------------------------------------------------------
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:orted_cmd sending orted_exit 
> commands
> --------------------------------------------------------------------------
> shmemrun detected that one or more processes exited with non-zero status, 
> thus causing
> the job to be terminated. The first process to do so was:
> 
>   Process name: [[32137,1],0]
>   Exit code:    255
> --------------------------------------------------------------------------
> [atl1-01-mic0:190189] 1 more process has sent help message 
> help-shmem-runtime.txt / shmem_init:startup:internal-failure
> [atl1-01-mic0:190189] Set MCA parameter "orte_base_help_aggregate" to 0 to 
> see all help / error messages
> [atl1-01-mic0:190189] 1 more process has sent help message help-shmem-api.txt 
> / shmem-abort
> [atl1-01-mic0:190189] 1 more process has sent help message 
> help-shmem-runtime.txt / oshmem shmem abort:cannot guarantee all killed
> [atl1-01-mic0:190189] [[32137,0],0] plm:base:receive stop comm
> 
> 
> On 04/11/2015 07:41 PM, Ralph Castain wrote:
>> Got it - thanks. I fixed that ERROR_LOG issue (I think- please verify). I 
>> suspect the memheap issue relates to something else, but I probably need to 
>> let the OSHMEM folks comment on it
>> 
>> 
>>> On Apr 11, 2015, at 9:52 AM, Andy Riebs <andy.ri...@hp.com 
>>> <mailto:andy.ri...@hp.com>> wrote:
>>> 
>>> Everything is built on the Xeon side, with the icc "-mmic" switch. I then 
>>> ssh into one of the PHIs, and run shmemrun from there.
>>> 
>>> 
>>> On 04/11/2015 12:00 PM, Ralph Castain wrote:
>>>> Let me try to understand the setup a little better. Are you running 
>>>> shmemrun on the PHI itself? Or is it running on the host processor, and 
>>>> you are trying to spawn a process onto the Phi?
>>>> 
>>>> 
>>>>> On Apr 11, 2015, at 7:55 AM, Andy Riebs <andy.ri...@hp.com 
>>>>> <mailto:andy.ri...@hp.com>> wrote:
>>>>> 
>>>>> Hi Ralph,
>>>>> 
>>>>> Yes, this is attempting to get OSHMEM to run on the Phi.
>>>>> 
>>>>> I grabbed openmpi-dev-1484-g033418f.tar.bz2 and configured it with
>>>>> 
>>>>> $ ./configure --prefix=/home/ariebs/mic/mpi-nightly    CC=icc -mmic 
>>>>> CXX=icpc -mmic    \
>>>>>     --build=x86_64-unknown-linux-gnu --host=x86_64-k1om-linux    \
>>>>>      AR=x86_64-k1om-linux-ar RANLIB=x86_64-k1om-linux-ranlib  
>>>>> LD=x86_64-k1om-linux-ld   \
>>>>>      --enable-mpirun-prefix-by-default --disable-io-romio     
>>>>> --disable-mpi-fortran    \
>>>>>      --enable-debug     
>>>>> --enable-mca-no-build=btl-usnic,btl-openib,common-verbs,oob-ud
>>>>> 
>>>>> (Note that I had to add "oob-ud" to the "--enable-mca-no-build" option, 
>>>>> as the build complained that mca oob/ud needed mca common-verbs.)
>>>>> 
>>>>> With that configuration, here is what I am seeing now...
>>>>> 
>>>>> $ export SHMEM_SYMMETRIC_HEAP_SIZE=1G
>>>>> $ shmemrun -H localhost -N 2 --mca sshmem mmap  --mca plm_base_verbose 5 
>>>>> $PWD/mic.out
>>>>> [atl1-01-mic0:189895] mca:base:select:(  plm) Querying component [rsh]
>>>>> [atl1-01-mic0:189895] [[INVALID],INVALID] plm:rsh_lookup on agent ssh : 
>>>>> rsh path NULL
>>>>> [atl1-01-mic0:189895] mca:base:select:(  plm) Query of component [rsh] 
>>>>> set priority to 10
>>>>> [atl1-01-mic0:189895] mca:base:select:(  plm) Querying component 
>>>>> [isolated]
>>>>> [atl1-01-mic0:189895] mca:base:select:(  plm) Query of component 
>>>>> [isolated] set priority to 0
>>>>> [atl1-01-mic0:189895] mca:base:select:(  plm) Querying component [slurm]
>>>>> [atl1-01-mic0:189895] mca:base:select:(  plm) Skipping component [slurm]. 
>>>>> Query failed to return a module
>>>>> [atl1-01-mic0:189895] mca:base:select:(  plm) Selected component [rsh]
>>>>> [atl1-01-mic0:189895] plm:base:set_hnp_name: initial bias 189895 nodename 
>>>>> hash 4121194178
>>>>> [atl1-01-mic0:189895] plm:base:set_hnp_name: final jobfam 32419
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:rsh_setup on agent ssh : rsh path 
>>>>> NULL
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:receive start comm
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:setup_job
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:setup_vm
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:setup_vm creating map
>>>>> [atl1-01-mic0:189895] [[32419,0],0] setup:vm: working unmanaged allocation
>>>>> [atl1-01-mic0:189895] [[32419,0],0] using dash_host
>>>>> [atl1-01-mic0:189895] [[32419,0],0] checking node atl1-01-mic0
>>>>> [atl1-01-mic0:189895] [[32419,0],0] ignoring myself
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:setup_vm only HNP in 
>>>>> allocation
>>>>> [atl1-01-mic0:189895] [[32419,0],0] complete_setup on job [32419,1]
>>>>> [atl1-01-mic0:189895] [[32419,0],0] ORTE_ERROR_LOG: Not found in file 
>>>>> base/plm_base_launch_support.c at line 440
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:launch_apps for job [32419,1]
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:launch wiring up iof for job 
>>>>> [32419,1]
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:launch [32419,1] registered
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:launch job [32419,1] is not 
>>>>> a dynamic spawn
>>>>> [atl1-01-mic0:189899] Error: pshmem_init.c:61 - shmem_init() SHMEM failed 
>>>>> to initialize - aborting
>>>>> [atl1-01-mic0:189898] Error: pshmem_init.c:61 - shmem_init() SHMEM failed 
>>>>> to initialize - aborting
>>>>> --------------------------------------------------------------------------
>>>>> It looks like SHMEM_INIT failed for some reason; your parallel process is
>>>>> likely to abort.  There are many reasons that a parallel process can
>>>>> fail during SHMEM_INIT; some of which are due to configuration or 
>>>>> environment
>>>>> problems.  This failure appears to be an internal failure; here's some
>>>>> additional information (which may only be relevant to an Open SHMEM
>>>>> developer):
>>>>> 
>>>>>   mca_memheap_base_select() failed
>>>>>   --> Returned "Error" (-1) instead of "Success" (0)
>>>>> --------------------------------------------------------------------------
>>>>> --------------------------------------------------------------------------
>>>>> SHMEM_ABORT was invoked on rank 1 (pid 189899, host=atl1-01-mic0) with 
>>>>> errorcode -1.
>>>>> --------------------------------------------------------------------------
>>>>> --------------------------------------------------------------------------
>>>>> A SHMEM process is aborting at a time when it cannot guarantee that all
>>>>> of its peer processes in the job will be killed properly.  You should
>>>>> double check that everything has shut down cleanly.
>>>>> 
>>>>> Local host: atl1-01-mic0
>>>>> PID:        189899
>>>>> --------------------------------------------------------------------------
>>>>> -------------------------------------------------------
>>>>> Primary job  terminated normally, but 1 process returned
>>>>> a non-zero exit code.. Per user-direction, the job has been aborted.
>>>>> -------------------------------------------------------
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:orted_cmd sending orted_exit 
>>>>> commands
>>>>> --------------------------------------------------------------------------
>>>>> shmemrun detected that one or more processes exited with non-zero status, 
>>>>> thus causing
>>>>> the job to be terminated. The first process to do so was:
>>>>> 
>>>>>   Process name: [[32419,1],1]
>>>>>   Exit code:    255
>>>>> --------------------------------------------------------------------------
>>>>> [atl1-01-mic0:189895] 1 more process has sent help message 
>>>>> help-shmem-runtime.txt / shmem_init:startup:internal-failure
>>>>> [atl1-01-mic0:189895] Set MCA parameter "orte_base_help_aggregate" to 0 
>>>>> to see all help / error messages
>>>>> [atl1-01-mic0:189895] 1 more process has sent help message 
>>>>> help-shmem-api.txt / shmem-abort
>>>>> [atl1-01-mic0:189895] 1 more process has sent help message 
>>>>> help-shmem-runtime.txt / oshmem shmem abort:cannot guarantee all killed
>>>>> [atl1-01-mic0:189895] [[32419,0],0] plm:base:receive stop comm
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> On 04/10/2015 06:37 PM, Ralph Castain wrote:
>>>>>> Andy - could you please try the current 1.8.5 nightly tarball and see if 
>>>>>> it helps? The error log indicates that it is failing to get the topology 
>>>>>> from some daemon, I�m assuming the one on the Phi?
>>>>>> 
>>>>>> You might also add �enable-debug to that configure line and then put 
>>>>>> -mca plm_base_verbose on the shmemrun cmd to get more help
>>>>>> 
>>>>>> 
>>>>>>> On Apr 10, 2015, at 11:55 AM, Andy Riebs <andy.ri...@hp.com 
>>>>>>> <mailto:andy.ri...@hp.com>> wrote:
>>>>>>> 
>>>>>>> Summary: MPI jobs work fine, SHMEM jobs work just often enough to be 
>>>>>>> tantalizing, on an Intel Xeon Phi/MIC system.
>>>>>>> 
>>>>>>> Longer version
>>>>>>> 
>>>>>>> Thanks to the excellent write-up last June 
>>>>>>> (<https://www.open-mpi.org/community/lists/users/2014/06/24711.php> 
>>>>>>> <https://www.open-mpi.org/community/lists/users/2014/06/24711.php>), I 
>>>>>>> have been able to build a version of Open MPI for the Xeon Phi 
>>>>>>> coprocessor that runs MPI jobs on the Phi coprocessor with no problem, 
>>>>>>> but not SHMEM jobs.  Just at the point where I was about to document 
>>>>>>> the problems I was having with SHMEM, my trivial SHMEM job worked. And 
>>>>>>> then failed when I tried to run it again, immediately afterwards. I 
>>>>>>> have a feeling I may be in uncharted  territory here.
>>>>>>> 
>>>>>>> Environment
>>>>>>> RHEL 6.5
>>>>>>> Intel Composer XE 2015
>>>>>>> Xeon Phi/MIC
>>>>>>> ----------------
>>>>>>> 
>>>>>>> 
>>>>>>> Configuration
>>>>>>> 
>>>>>>> $ export PATH=/usr/linux-k1om-4.7/bin/:$PATH
>>>>>>> $ source /opt/intel/15.0/composer_xe_2015/bin/compilervars.sh intel64
>>>>>>> $ ./configure --prefix=/home/ariebs/mic/mpi \
>>>>>>>    CC="icc -mmic" CXX="icpc -mmic" \
>>>>>>>    --build=x86_64-unknown-linux-gnu --host=x86_64-k1om-linux \
>>>>>>>     AR=x86_64-k1om-linux-ar RANLIB=x86_64-k1om-linux-ranlib \
>>>>>>>     LD=x86_64-k1om-linux-ld \
>>>>>>>     --enable-mpirun-prefix-by-default --disable-io-romio \
>>>>>>>     --disable-vt --disable-mpi-fortran \
>>>>>>>     --enable-mca-no-build=btl-usnic,btl-openib,common-verbs
>>>>>>> $ make
>>>>>>> $ make install
>>>>>>> 
>>>>>>> ----------------
>>>>>>> 
>>>>>>> Test program
>>>>>>> 
>>>>>>> #include <stdio.h>
>>>>>>> #include <stdlib.h>
>>>>>>> #include <shmem.h>
>>>>>>> int main(int argc, char **argv)
>>>>>>> {
>>>>>>>         int me, num_pe;
>>>>>>>         shmem_init();
>>>>>>>         num_pe = num_pes();
>>>>>>>         me = my_pe();
>>>>>>>         printf("Hello World from process %ld of %ld\n", me, num_pe);
>>>>>>>         exit(0);
>>>>>>> }
>>>>>>> 
>>>>>>> ----------------
>>>>>>> 
>>>>>>> Building the program
>>>>>>> 
>>>>>>> export PATH=/home/ariebs/mic/mpi/bin:$PATH
>>>>>>> export PATH=/usr/linux-k1om-4.7/bin/:$PATH
>>>>>>> source /opt/intel/15.0/composer_xe_2015/bin/compilervars.sh intel64
>>>>>>> export 
>>>>>>> LD_LIBRARY_PATH=/opt/intel/15.0/composer_xe_2015.2.164/compiler/lib/mic:$LD_LIBRARY_PATH
>>>>>>> 
>>>>>>> icc -mmic -std=gnu99 -I/home/ariebs/mic/mpi/include -pthread \
>>>>>>>         -Wl,-rpath -Wl,/home/ariebs/mic/mpi/lib -Wl,--enable-new-dtags \
>>>>>>>         -L/home/ariebs/mic/mpi/lib -loshmem -lmpi -lopen-rte -lopen-pal 
>>>>>>> \
>>>>>>>         -lm -ldl -lutil \
>>>>>>>         -Wl,-rpath 
>>>>>>> -Wl,/opt/intel/15.0/composer_xe_2015.2.164/compiler/lib/mic \
>>>>>>>         -L/opt/intel/15.0/composer_xe_2015.2.164/compiler/lib/mic \
>>>>>>>         -o mic.out  shmem_hello.c
>>>>>>> 
>>>>>>> ----------------
>>>>>>> 
>>>>>>> Running the program
>>>>>>> 
>>>>>>> (Note that the program had been consistently failing. Then, when I 
>>>>>>> logged back into the system to capture the results, it worked once,  
>>>>>>> and then immediately failed when I tried again, as shown below. Logging 
>>>>>>> in and out isn't sufficient to correct the problem. Overall, I think I 
>>>>>>> had 3 successful runs in 30-40 attempts.)
>>>>>>> 
>>>>>>> $ shmemrun -H localhost -N 2 --mca sshmem mmap ./mic.out
>>>>>>> [atl1-01-mic0:189372] [[30936,0],0] ORTE_ERROR_LOG: Not found in file 
>>>>>>> base/plm_base_launch_support.c at line 426
>>>>>>> Hello World from process 0 of 2
>>>>>>> Hello World from process 1 of 2
>>>>>>> $ shmemrun -H localhost -N 2 --mca sshmem mmap ./mic.out
>>>>>>> [atl1-01-mic0:189381] [[30881,0],0] ORTE_ERROR_LOG: Not found in file 
>>>>>>> base/plm_base_launch_support.c at line 426
>>>>>>> [atl1-01-mic0:189383] Error: pshmem_init.c:61 - shmem_init() SHMEM 
>>>>>>> failed to initialize - aborting
>>>>>>> --------------------------------------------------------------------------
>>>>>>> It looks like SHMEM_INIT failed for some reason; your parallel process 
>>>>>>> is
>>>>>>> likely to abort.  There are many reasons that a parallel process can
>>>>>>> fail during SHMEM_INIT; some of which are due to configuration or 
>>>>>>> environment
>>>>>>> problems.  This failure appears to be an internal failure; here's some
>>>>>>> additional information (which may only be relevant to an Open SHMEM
>>>>>>> developer):
>>>>>>> 
>>>>>>>   mca_memheap_base_select() failed
>>>>>>>   --> Returned "Error" (-1) instead of "Success" (0)
>>>>>>> --------------------------------------------------------------------------
>>>>>>> --------------------------------------------------------------------------
>>>>>>> SHMEM_ABORT was invoked on rank 0 (pid 189383, host=atl1-01-mic0) with 
>>>>>>> errorcode -1.
>>>>>>> --------------------------------------------------------------------------
>>>>>>> --------------------------------------------------------------------------
>>>>>>> A SHMEM process is aborting at a time when it cannot guarantee that all
>>>>>>> of its peer processes in the job will be killed properly.  You should
>>>>>>> double check that everything has shut down cleanly.
>>>>>>> 
>>>>>>> Local host: atl1-01-mic0
>>>>>>> PID:        189383
>>>>>>> --------------------------------------------------------------------------
>>>>>>> -------------------------------------------------------
>>>>>>> Primary job  terminated normally, but 1 process returned
>>>>>>> a non-zero exit code.. Per user-direction, the job has been aborted.
>>>>>>> -------------------------------------------------------
>>>>>>> --------------------------------------------------------------------------
>>>>>>> shmemrun detected that one or more processes exited with non-zero 
>>>>>>> status, thus causing
>>>>>>> the job to be terminated. The first process to do so was:
>>>>>>> 
>>>>>>>   Process name: [[30881,1],0]
>>>>>>>   Exit code:    255
>>>>>>> --------------------------------------------------------------------------
>>>>>>> 
>>>>>>> Any thoughts about where to go from here?
>>>>>>> 
>>>>>>> Andy
>>>>>>> 
>>>>>>> -- 
>>>>>>> Andy Riebs
>>>>>>> Hewlett-Packard Company
>>>>>>> High Performance Computing
>>>>>>> +1 404 648 9024
>>>>>>> My opinions are not necessarily those of HP
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>>>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>>>>>> Link to this post: 
>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26670.php 
>>>>>>> <http://www.open-mpi.org/community/lists/users/2015/04/26670.php>
>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>>>>> Link to this post: 
>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26676.php 
>>>>>> <http://www.open-mpi.org/community/lists/users/2015/04/26676.php>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>>>> Link to this post: 
>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26678.php 
>>>>> <http://www.open-mpi.org/community/lists/users/2015/04/26678.php>
>>>> 
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>>> Link to this post: 
>>>> http://www.open-mpi.org/community/lists/users/2015/04/26679.php 
>>>> <http://www.open-mpi.org/community/lists/users/2015/04/26679.php>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2015/04/26680.php 
>>> <http://www.open-mpi.org/community/lists/users/2015/04/26680.php>
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2015/04/26682.php 
>> <http://www.open-mpi.org/community/lists/users/2015/04/26682.php>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/04/26683.php

Reply via email to