Hmmm...well, we certainly will ke a point to give you a better error
message! Probably won't get it into 1.2.3, but should make later releases.

Thanks for letting me know
Ralph


On 6/7/07 1:22 PM, "Guillaume THOMAS-COLLIGNON"
<guillaume.thomas-collig...@cggveritas.com> wrote:

> You're right, the --wdir option works fine !
> Thanks !
> 
> I just tried an older version we had compiled (1.2b3), and the error
> was more explicit than the seg fault we get with 1.2.2 :
> 
> Could not chdir to home directory /rdu/thomasco: No such file or
> directory
> ------------------------------------------------------------------------
> --
> Failed to change to the working directory:
> <...>
> 
> -Guillaume
> 
> On Jun 7, 2007, at 12:57 PM, Ralph Castain wrote:
> 
>> Have you tried the --wdir option yet? It should let you set your
>> working
>> directory to anywhere. I don't believe it will require you to have
>> a home
>> directory on the backend nodes, though I can't sweathat ssh will be
>> happy
>> if you don't.
>> 
>> Just do "mpirun -h" for a full list of options - it will describe
>> the exact
>> format of the wdir one plusthers you might find useful.
>> 
>> Ralph
>> 
>> 
>> 
>> On 6/7/07 11:12 AM, "Guillaume THOMAS-COLLIGNON<guillaume.thomas-
>> collig...@cggveritas.com> wrote:
>> 
>>> I am trying to switch to OpenI, and I ran into a problem : my home
>>> directory must exist on all the nodes, or ted will crash.
>>> 
>>> I have a "master" machine where I initiate the mpirun command>
>>> Then I have a bunch of slave machines, which will ao execute the
>>> MPI job.
>>> My user exists on all the machines, but the home directory is not
>>> mounted on the slaves,o it's only visible on the master node. I can
>>> log on a slave node, but don't have a home there. Of course the
>>> binary I'm running exists on all the machines (not in my home !). And
>>> the problem can be reproduced by running a shell command too, to make
>>> things simpler.
>>> 
>>> We have thousands of slave nod and we don't want to mount the
>>> user's homedirs on all the slaves, so a fix would be really really
>>> nice.
>>> 
>>> Example :
>>> 
>>> I have 3 hosts, master, slave1, slave2. My home directory exists only
>>> on master.
>>> 
>>> If I log on master and run "mpirun -host master,slave1uname -a" I get
>>> a segfault.
>>> If I log on slave1 and run "mpirun -host slave1,slave2 uname -a", it
>>> runs fine. My home directory does not exist on either slave1 or
>>> slave2.
>>> If I log on master and run "mpirun -host master uname -a" it runs
>>> fine. I can run across several master nodes, it's fine too.
>>> 
>>> So it runs fine if my home directory exists everywhere, or if it does
>>> not exist at all. If it exists only on some nodes and not others,
>>> orted crashes.
>>> I thought it could be related to my environment but I created a new
>>> user with an empty home and it does the same thing. As soon as I
>>> create the homedir on slave1 and slave2 it works fine.
>>> 
>>> 
>>> 
>>> 
>>> I'm using OpenMPI 1.2.2, here is the error message and the result of
>>> ompi_info.
>>> 
>>> Short version (rnd04 is the master, r137n001 is a slave node).
>>> 
>>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001
>>> uname -a
>>> Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64
>>> x86_64 x86_64 GNU/Linux
>>> [r137n001:31533] *** Process received signal ***
>>> [r137n001:31533] Signal: Segmentation fault (11)
>>> [r137n001:31533] Signal code: Address not mapped (1)
>>> [r137n001:533] Failing at address: 0x1
>>> [r137n001:31533] [ 0] [0xffffe600]
>>> [r137n001:31533] [ 1] /lib/tls/libc.so.6 [0xbf3bfc]
>>> [r137n001:31533] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb)
>>> [0xbf3e3b]
>>> [r137n001:31533] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0
>>> (opal_show_help+0x263) [0xf7f78de3]
>>> [r137n001:31533] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0
>>> (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef]
>>> [r137n001:31533] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f)
>>> [0xf7ea041f]
>>> [r137n001:31533] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea]
>>> [r137n001:31533] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136)
>>> [0xf7ef65c6]
>>> [r137n001:31533] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68]
>>> [r137n001:31533] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0
>>> [0xf7fd9a18]
>>> [r137n001:31533] [10] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc]
>>> [r137n001:31533] [11] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_oob_tcp.so [0xf7f07f61]
>>> [r137n001:31533] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0
>>> (opal_event_base_loop+0x388) [0xf7f67dd8]
>>> [r137n001:31533] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0
>>> (opal_event_loop+0x29) [0xf7f67fb9]
>>> [r137n001:31533] [14] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_oob_tcp.so(mca_oob_tcp_msg_wait+0x37) [0xf7f053c7]
>>> [r137n001:31533] [15] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_oob_tcp.so(mca_oob_tcp_recv+0x374) [0xf7f09a04]
>>> [r137n001:31533] [16] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0
>>> (mca_oob_recv_packed+0x4d) [0xf7fd980d]
>>> [r137n001:31533] [17] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_gpr_proxy.so(orte_gpr_proxy_exec_compound_cmd+0x137) [0xf7ef55e7]
>>> [r137n001:31533] [18] /usr/local/openmpi-1.2.2/bin/orted(main+0x99d)
>>> [0x8049d0d]
>>> [r137n001:31533] [19] /lib/tls/libc.so.6(__libc_start_main+0xd3)
>>> [0xbcee23]
>>> [r137n001:31533] [20] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1]
>>> [r137n001:31533] *** End of error message ***
>>> mpirun noticed that job rank 1 with PID 31533 on node r137n001 exited
>>> on signal 11 (Segmentation fault).
>>> 
>>> 
>>> 
>>> If I create /home/toto on r137n001, it works fine :
>>> (as root on r137n001: "mkdir /home/toto && chown toto:users /home/
>>> toto")
>>> 
>>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001
>>> uname -a
>>> Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64
>>> x86_64 x86_64 GNU/Linux
>>> Linux r137n001 2.6.9-34.ELsmp #1 SMP Fri Feb 24 16:56:28 EST 2006
>>> x86_64 x86_64 x86_64 GNU/Linux
>>> 
>>> 
>>> I tried to use ssh instead of rsh, it crashes too.
>>> 
>>> If anyone knows a way to run OpenMPI jobs in this configuration where
>>> the home directory does not exist on all the nodes, it woud really
>>> help !
>>> 
>>> Or is there a way to fix orted so that it won't crash ?
>>> 
>>> 
>>> Here is the crash with -d option :
>>> 
>>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -d -host
>>> rnd04,r137n001  uname -a
>>> [rnd04:10736] connect_uni: connection not allowed
>>> [rnd04:10736] [0,0,0] setting up session dir with
>>> [rnd04:10736]   universe default-universe-10736
>>> [rnd04:10736]   user toto
>>> [rnd04:10736]   host rnd04
>>> [rnd04:10736]   jobid 0
>>> [rnd04:10736]   procid 0
>>> [rnd04:10736] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default-
>>> universe-10736/0/0
>>> [rnd04:10736] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default-
>>> universe-10736/0
>>> [rnd04:10736] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default-
>>> universe-10736
>>> [rnd04:10736] top: openmpi-sessions-toto@rnd04_0
>>> [rnd04:10736] tmp: /tmp
>>> [rnd04:10736] [0,0,0] contact_file /tmp/openmpi-sessions-
>>> toto@rnd04_0/
>>> default-universe-10736/universe-setup.txt
>>> [rnd04:10736] [0,0,0] wrote setup file
>>> [rnd04:10736] pls:rsh: local csh: 0, local sh: 1
>>> [rnd04:10736] pls:rsh: assuming same remote shell as local shell
>>> [rnd04:10736] pls:rsh: remote csh: 0, remote sh: 1
>>> [rnd04:10736] pls:rsh: final template argv:
>>> [rnd04:10736] pls:rsh:     /usr/bin/rsh <template> orted --debug --
>>> bootproxy 1 --name <template> --num_procs 3 --vpid_start 0 --nodename
>>> <template> --universe toto@rnd04:default-universe-10736 --nsreplica
>>> "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" --
>>> gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029"
>>> [rnd04:10736] pls:rsh: launching on node rnd04
>>> [rnd04:10736] pls:rsh: rnd04 is a LOCAL node
>>> [rnd04:10736] pls:rsh: reset PATH: /usr/local/openmpi-1.2.2/bin:/usr/
>>> local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin
>>> [rnd04:10736] pls:rsh: reset LD_LIBRARY_PATH: /usr/local/
>>> openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib:/usr/local/
>>> openmpi-1.2.2/lib64
>>> [rnd04:10736] pls:rsh: changing to directory /home/toto
>>> [rnd04:10736] pls:rsh: executing: (/usr/local/openmpi-1.2.2/bin/
>>> orted) orted --debug --bootproxy 1 --name 0.0.1 --num_procs 3 --
>>> vpid_start 0 --nodename rnd04 --universe toto@rnd04:default-
>>> universe-10736 --nsreplica "0.0.0;tcp://172.28.20.143:33029;tcp://
>>> 10.3.254.105:33029" --gprreplica "0.0.0;tcp://
>>> 172.28.20.143:33029;tcp://10.3.254.105:33029" --set-sid
>>> [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color HISTSIZE=1000
>>> USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:/usr/local/
>>> openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib64
>>> LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01
>>> :cd
>>> =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01
>>> ;32
>>> :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01;
>>> 31:
>>> *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31
>>> :*.
>>> Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cp
>>> io=
>>> 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.p
>>> ng=
>>> 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/
>>> local/openmpi-1.2.2/bin:/usr/local/bin:/bin:/usr/bin:/usr/X11R6/bin:/
>>> usr/kerberos/bin INPUTRC=/etc/inputrc PWD=/home/toto LANG=en_US.UTF-8
>>> SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass SHLVL=1 HOME=/
>>> home/
>>> toto LOGNAME=toto LESSOPEN=|/usr/bin/lesspipe.sh %s
>>> G_BROKEN_FILENAMES=1 _=/usr/local/openmpi-1.2.2/bin/mpirun
>>> OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0]
>>> [rnd04:10736] pls:rsh: launching on node r137n001
>>> [rnd04:10736] pls:rsh: r137n001 is a REMOTE node
>>> [rnd04:10736] pls:rsh: executing: (//usr/bin/rsh) /usr/bin/rsh
>>> r137n001  PATH=/usr/local/openmpi-1.2.2/bin:$PATH ; export PATH ;
>>> LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:$LD_LIBRARY_PATH ;
>>> export LD_LIBRARY_PATH ; /usr/local/openmpi-1.2.2/bin/orted --
>>> debug --
>>> bootproxy 1 --name 0.0.2 --num_procs 3 --vpid_start 0 --nodename
>>> r137n001 --universe toto@rnd04:default-universe-10736 --nsreplica
>>> "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" --
>>> gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp://
>>> 10.3.254.105:33029" [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color
>>> HISTSIZE=1000 USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/
>>> lib:/
>>> usr/local/openmpi-1.2.2/lib64
>>> LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01
>>> :cd
>>> =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01
>>> ;32
>>> :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01;
>>> 31:
>>> *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31
>>> :*.
>>> Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cp
>>> io=
>>> 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.p
>>> ng=
>>> 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/
>>> local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin INPUTRC=/
>>> etc/
>>> inputrc PWD=/home/toto LANG=en_US.UTF-8 SSH_ASKPASS=/usr/libexec/
>>> openssh/gnome-ssh-askpass SHLVL=1 HOME=/home/toto LOGNAME=toto
>>> LESSOPEN=|/usr/bin/lesspipe.sh %s G_BROKEN_FILENAMES=1 _=/usr/local/
>>> openmpi-1.2.2/bin/mpirun OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0]
>>> [rnd04:10737] [0,0,1] setting up session dir with
>>> [rnd04:10737]   universe default-universe-10736
>>> [rnd04:10737]   user toto
>>> [rnd04:10737]   host rnd04
>>> [rnd04:10737]   jobid 0
>>> [rnd04:10737]   procid 1
>>> [rnd04:10737] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default-
>>> universe-10736/0/1
>>> [rnd04:10737] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default-
>>> universe-10736/0
>>> [rnd04:10737] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default-
>>> universe-10736
>>> [rnd04:10737] top: openmpi-sessions-toto@rnd04_0
>>> [rnd04:10737] tmp: /tmp
>>> [r137n001:31527] [0,0,2] setting up session dir with
>>> [r137n001:31527]        universe default-universe-10736
>>> [r137n001:31527]        user toto
>>> [r137n001:31527]        host r137n001
>>> [r137n001:31527]        jobid 0
>>> [r137n001:31527]        procid 2
>>> [r137n001:31527] procdir: /tmp/openmpi-sessions-toto@r137n001_0/
>>> default-universe-10736/0/2
>>> [r137n001:31527] jobdir: /tmp/openmpi-sessions-toto@r137n001_0/
>>> default-universe-10736/0
>>> [r137n001:31527] unidir: /tmp/openmpi-sessions-toto@r137n001_0/
>>> default-universe-10736
>>> [r137n001:31527] top: openmpi-sessions-toto@r137n001_0
>>> [r137n001:31527] tmp: /tmp
>>> Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64
>>> x86_64 x86_64 GNU/Linux
>>> [rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving
>>> [r137n001:31528] *** Process received signal ***
>>> [r137n001:31528] Signal: Segmentation fault (11)
>>> [r137n001:31528] Signal code: Address not mapped (1)
>>> [r137n001:31528] Failing at address: 0x1
>>> [r137n001:31528] [ 0] [0xffffe600]
>>> [r137n001:31528] [ 1] /lib/tls/libc.so.6 [0xbf3bfc]
>>> [r137n001:31528] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb)
>>> [0xbf3e3b]
>>> [r137n001:31528] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0
>>> (opal_show_help+0x263) [0xf7f78de3]
>>> [r137n001:31528] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0
>>> (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef]
>>> [r137n001:31528] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f)
>>> [0xf7ea041f]
>>> [r137n001:31528] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea]
>>> [r137n001:31528] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136)
>>> [0xf7ef65c6]
>>> [r137n001:31528] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68]
>>> [r137n001:31528] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0
>>> [0xf7fd9a18]
>>> [r137n001:31528] [10] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc]
>>> [r137n001:31528] [11] /usr/local/openmpi-1.2.2/lib/openmpi/
>>> mca_oob_tcp.so [0xf7f07f61]
>>> [r137n001:31528] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0
>>> (opal_event_base_loop+0x388) [0xf7f67dd8]
>>> [r137n001:31528] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0
>>> (opal_event_loop+0x29) [0xf7f67fb9]
>>> [r137n001:31528] [14] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0
>>> (opal_progress+0xbe) [0xf7f6123e]
>>> [r137n001:31528] [15] /usr/local/openmpi-1.2.2/bin/orted(main+0xd74)
>>> [0x804a0e4]
>>> [r137n001:31528] [16] /lib/tls/libc.so.6(__libc_start_main+0xd3)
>>> [0xbcee23]
>>> [r137n001:31528] [17] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1]
>>> [r137n001:31528] *** End of error message ***
>>> [r137n001:31527] sess_dir_finalize: proc session dir not empty -
>>> leaving
>>> [rnd04:10736] spawn: in job_state_callback(jobid = 1, state = 0x80)
>>> mpirun noticed that job rank 1 with PID 31528 on node r137n001 exited
>>> on signal 11 (Segmentation fault).
>>> [rnd04:10737] sess_dir_finalize: job session dir not empty - leaving
>>> [r137n001:31527] sess_dir_finalize: job session dir not empty -
>>> leaving
>>> [rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving
>>> [rnd04:10736] sess_dir_finalize: proc session dir not empty - leaving
>>> -bash-3.00$ [r137n001:31527] sess_dir_finalize: proc session dir not
>>> empty - leaving
>>> 
>>> 
>>> 
>>> 
>>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/ompi_info --all
>>>                  Open MPI: 1.2.2
>>>     Open MPI SVN revision: r14613
>>>                  Open RTE: 1.2.2
>>>     Open RTE SVN revision: r14613
>>>                      OPAL: 1.2.2
>>>         OPAL SVN revision: r14613
>>>             MCA backtrace: execinfo (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>             MCA paffinity: linux (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>             MCA maffinity: first_use (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                 MCA timer: linux (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>           MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.2)
>>>           MCA installdirs: config (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>             MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
>>>             MCA allocator: bucket (MCA v1.0, API v1.0, Component
>>> v1.0)
>>>                  MCA coll: basic (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                  MCA coll: self (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                  MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.2)
>>>                  MCA coll: tuned (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                    MCA io: romio (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                 MCA mpool: rdma (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                 MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.2)
>>>                   MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.2)
>>>                   MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.2)
>>>                   MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.2)
>>>                MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.2)
>>>                   MCA btl: self (MCA v1.0, API v1.0.1, Component
>>> v1.2.2)
>>>                   MCA btl: sm (MCA v1.0, API v1.0.1, Component
>>> v1.2.2)
>>>                   MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0)
>>>                  MCA topo: unity (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA osc: pt2pt (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.2)
>>>                MCA errmgr: orted (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                MCA errmgr: proxy (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA gpr: null (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA gpr: proxy (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA gpr: replica (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA iof: proxy (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA iof: svc (MCA v1.0, API v1.0, Component v1.2.2)
>>>                    MCA ns: proxy (MCA v1.0, API v2.0, Component
>>> v1.2.2)
>>>                    MCA ns: replica (MCA v1.0, API v2.0, Component
>>> v1.2.2)
>>>                   MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
>>>                   MCA ras: dash_host (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA ras: gridengine (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA ras: localhost (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA ras: slurm (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA rds: hostfile (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA rds: proxy (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA rds: resfile (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                 MCA rmaps: round_robin (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                  MCA rmgr: proxy (MCA v1.0, API v2.0, Component
>>> v1.2.2)
>>>                  MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2.2)
>>>                   MCA rml: oob (MCA v1.0, API v1.0, Component v1.2.2)
>>>                   MCA pls: gridengine (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA pls: proxy (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2.2)
>>>                   MCA pls: slurm (MCA v1.0, API v1.3, Component
>>> v1.2.2)
>>>                   MCA sds: env (MCA v1.0, API v1.0, Component v1.2.2)
>>>                   MCA sds: pipe (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA sds: seed (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA sds: singleton (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                   MCA sds: slurm (MCA v1.0, API v1.0, Component
>>> v1.2.2)
>>>                    Prefix: /usr/local/openmpi-1.2.2
>>>                    Bindir: /usr/local/openmpi-1.2.2/bin
>>>                    Libdir: /usr/local/openmpi-1.2.2/lib
>>>                    Incdir: /usr/local/openmpi-1.2.2/include
>>>                 Pkglibdir: /usr/local/openmpi-1.2.2/lib/openmpi
>>>                Sysconfdir: /usr/local/openmpi-1.2.2/etc
>>> Configured architecture: x86_64-unknown-linux-gnu
>>>             Configured by: root
>>>             Configured on: Tue Jun  5 14:32:20 CDT 2007
>>>            Configure host: qpac171
>>>                  Built by: root
>>>                  Built on: Tue Jun  5 14:39:38 CDT 2007
>>>                Built host: qpac171
>>>                C bindings: yes
>>>              C++ bindings: yes
>>>        Fortran77 bindings: yes (all)
>>>        Fortran90 bindings: no
>>> Fortran90 bindings size: na
>>>                C compiler: gcc
>>>       C compiler absolute: /usr/bin/gcc
>>>               C char size: 1
>>>               C bool size: 1
>>>              C short size: 2
>>>                C int size: 4
>>>               C long size: 4
>>>              C float size: 4
>>>             C double size: 8
>>>            C pointer size: 4
>>>              C char align: 1
>>>              C bool align: 1
>>>               C int align: 4
>>>             C float align: 4
>>>            C double align: 4
>>>              C++ compiler: g++
>>>     C++ compiler absolute: /usr/bin/g++
>>>        Fortran77 compiler: g77
>>>    Fortran77 compiler abs: /usr/bin/g77
>>>        Fortran90 compiler: none
>>>    Fortran90 compiler abs: none
>>>         Fort integer size: 4
>>>         Fort logical size: 4
>>> Fort logical value true: 1
>>>        Fort have integer1: yes
>>>        Fort have integer2: yes
>>>        Fort have integer4: yes
>>>        Fort have integer8: yes
>>>       Fort have integer16: no
>>>           Fort have real4: yes
>>>           Fort have real8: yes
>>>          Fort have real16: no
>>>        Fort have complex8: yes
>>>       Fort have complex16: yes
>>>       Fort have complex32: no
>>>        Fort integer1 size: 1
>>>        Fort integer2 size: 2
>>>        Fort integer4 size: 4
>>>        Fort integer8 size: 8
>>>       Fort integer16 size: -1
>>>            Fort real size: 4
>>>           Fort real4 size: 4
>>>           Fort real8 size: 8
>>>          Fort real16 size: -1
>>>        Fort dbl prec size: 4
>>>            Fort cplx size: 4
>>>        Fort dbl cplx size: 4
>>>           Fort cplx8 size: 8
>>>          Fort cplx16 size: 16
>>>          Fort cplx32 size: -1
>>>        Fort integer align: 4
>>>       Fort integer1 align: 1
>>>       Fort integer2 align: 2
>>>       Fort integer4 align: 4
>>>       Fort integer8 align: 8
>>>      Fort integer16 align: -1
>>>           Fort real align: 4
>>>          Fort real4 align: 4
>>>          Fort real8 align: 8
>>>         Fort real16 align: -1
>>>       Fort dbl prec align: 4
>>>           Fort cplx align: 4
>>>       Fort dbl cplx align: 4
>>>          Fort cplx8 align: 4
>>>         Fort cplx16 align: 8
>>>         Fort cplx32 align: -1
>>>               C profiling: yes
>>>             C++ profiling: yes
>>>       Fortran77 profiling: yes
>>>       Fortran90 profiling: no
>>>            C++ exceptions: no
>>>            Thread support: posix (mpi: no, progress: no)
>>>              Build CFLAGS: -O3 -DNDEBUG -m32 -finline-functions -fno-
>>> strict-aliasing -pthread
>>>            Build CXXFLAGS: -O3 -DNDEBUG -m32 -finline-functions -
>>> pthread
>>>              Build FFLAGS: -m32
>>>             Build FCFLAGS: -m32
>>>             Build LDFLAGS: -export-dynamic
>>>                Build LIBS: -lnsl -lutil  -lm
>>>      Wrapper extra CFLAGS: -pthread -m32
>>>    Wrapper extra CXXFLAGS: -pthread -m32
>>>      Wrapper extra FFLAGS: -pthread -m32
>>>     Wrapper extra FCFLAGS: -pthread -m32
>>>     Wrapper extra LDFLAGS:
>>>        Wrapper extra LIBS:   -ldl   -Wl,--export-dynamic -lnsl -lutil
>>> -lm -ldl
>>>    Internal debug support: no
>>>       MPI parameter check: runtime
>>> Memory profiling support: no
>>> Memory debugging support: no
>>>           libltdl support: yes
>>>     Heterogeneous support: yes
>>> mpirun default --prefix: no
>>>                   MCA mca: parameter "mca_param_files" (current
>>> value: "/home/toto/.openmpi/mca-params.conf:/usr/local/openmpi-1.2.2/
>>> etc/openmpi-mca-params.conf")
>>>                            Path for MCA configuration files
>>> containing default parameter values
>>>                   MCA mca: parameter "mca_component_path" (current
>>> value: "/usr/local/openmpi-1.2.2/lib/openmpi:/home/toto/.openmpi/
>>> components")
>>>                            Path where to look for Open MPI and ORTE
>>> components
>>>                   MCA mca: parameter "mca_verbose" (current value:
>>> <none>)
>>>                            Top-level verbosity parameter
>>>                   MCA mca: parameter
>>> "mca_component_show_load_errors" (current value: "1")
>>>                            Whether to show errors for components that
>>> failed to load or not
>>>                   MCA mca: parameter
>>> "mca_component_disable_dlopen" (current value: "0")
>>>                            Whether to attempt to disable opening
>>> dynamic components or not
>>>                   MCA mpi: parameter "mpi_param_check" (current
>>> value: "1")
>>>                            Whether you want MPI API parameters
>>> checked at run-time or not.  Possible values are 0 (no checking) and
>>> 1 (perform checking at run-time)
>>>                   MCA mpi: parameter "mpi_yield_when_idle" (current
>>> value: "0")
>>>                            Yield the processor when waiting for MPI
>>> communication (for MPI processes, will default to 1 when
>>> oversubscribing nodes)
>>>                   MCA mpi: parameter "mpi_event_tick_rate" (current
>>> value: "-1")
>>>                            How often to progress TCP communications
>>> (0 = never, otherwise specified in microseconds)
>>>                   MCA mpi: parameter "mpi_show_handle_leaks" (current
>>> value: "0")
>>>                            Whether MPI_FINALIZE shows all MPI handles
>>> that were not freed or not
>>>                   MCA mpi: parameter "mpi_no_free_handles" (current
>>> value: "0")
>>>                            Whether to actually free MPI objects when
>>> their handles are freed
>>>                   MCA mpi: parameter "mpi_show_mca_params" (current
>>> value: "0")
>>>                            Whether to show all MCA parameter value
>>> during MPI_INIT or not (good for reproducability of MPI jobs)
>>>                   MCA mpi: parameter
>>> "mpi_show_mca_params_file" (current value: <none>)
>>>                            If mpi_show_mca_params is true, setting
>>> this string to a valid filename tells Open MPI to dump all the MCA
>>> parameter values into a file suitable for reading via the
>>> mca_param_files parameter (good for reproducability of MPI jobs)
>>>                   MCA mpi: parameter "mpi_paffinity_alone" (current
>>> value: "0")
>>>                            If nonzero, assume that this job is the
>>> only (set of) process(es) running on each node and bind processes to
>>> processors, starting with processor ID 0
>>>                   MCA mpi: parameter
>>> "mpi_keep_peer_hostnames" (current value: "1")
>>>                            If nonzero, save the string hostnames of
>>> all MPI peer processes (mostly for error / debugging output
>>> messages).  This can add quite a bit of memory usage to each MPI
>>> process.
>>>                   MCA mpi: parameter "mpi_abort_delay" (current
>>> value: "0")
>>>                            If nonzero, print out an identifying
>>> message when MPI_ABORT is invoked (hostname, PID of the process that
>>> called MPI_ABORT) and delay for that many seconds before exiting (a
>>> negative delay value means to never abort).  This allows
>>>                            attaching of a debugger before quitting
>>> the job.
>>>                   MCA mpi: parameter "mpi_abort_print_stack" (current
>>> value: "0")
>>>                            If nonzero, print out a stack trace when
>>> MPI_ABORT is invoked
>>>                   MCA mpi: parameter "mpi_preconnect_all" (current
>>> value: "0")
>>>                            Whether to force MPI processes to create
>>> connections / warmup with *all* peers during MPI_INIT (vs. making
>>> connections lazily -- upon the first MPI traffic between each process
>>> peer pair)
>>>                   MCA mpi: parameter "mpi_preconnect_oob" (current
>>> value: "0")
>>>                            Whether to force MPI processes to fully
>>> wire-up the OOB system between MPI processes.
>>>                   MCA mpi: parameter "mpi_leave_pinned" (current
>>> value: "0")
>>>                            Whether to use the "leave pinned" protocol
>>> or not.  Enabling this setting can help bandwidth performance when
>>> repeatedly sending and receiving large messages with the same buffers
>>> over RDMA-based networks.
>>>                   MCA mpi: parameter
>>> "mpi_leave_pinned_pipeline" (current value: "0")
>>>                            Whether to use the "leave pinned pipeline"
>>> protocol or not.
>>>                  MCA orte: parameter "orte_debug" (current value:
>>> "0")
>>>                            Top-level ORTE debug switch
>>>                  MCA orte: parameter "orte_no_daemonize" (current
>>> value: "0")
>>>                            Whether to properly daemonize the ORTE
>>> daemons or not
>>>                  MCA orte: parameter
>>> "orte_base_user_debugger" (current value: "totalview @mpirun@ -a
>>> @mpirun_args@ : fxp @mpirun@ -a @mpirun_args@")
>>>                            Sequence of user-level debuggers to search
>>> for in orterun
>>>                  MCA orte: parameter "orte_abort_timeout" (current
>>> value: "10")
>>>                            Time to wait [in seconds] before giving up
>>> on aborting an ORTE operation
>>>                  MCA orte: parameter "orte_timing" (current value:
>>> "0")
>>>                            Request that critical timing loops be
>>> measured
>>>                  MCA opal: parameter "opal_signal" (current value:
>>> "6,7,8,11")
>>>                            If a signal is received, display the stack
>>> trace frame
>>>             MCA backtrace: parameter "backtrace" (current value:
>>> <none>)
>>>                            Default selection set of components for
>>> the backtrace framework (<none> means "use all components that can be
>>> found")
>>>             MCA backtrace: parameter
>>> "backtrace_base_verbose" (current value: "0")
>>>                            Verbosity level for the backtrace
>>> framework (0 = no verbosity)
>>>             MCA backtrace: parameter
>>> "backtrace_execinfo_priority" (current value: "0")
>>>                MCA memory: parameter "memory" (current value: <none>)
>>>                            Default selection set of components for
>>> the memory framework (<none> means "use all components that can be
>>> found")
>>>                MCA memory: parameter "memory_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the memory framework
>>> (0 = no verbosity)
>>>                MCA memory: parameter
>>> "memory_ptmalloc2_priority" (current value: "0")
>>>             MCA paffinity: parameter "paffinity" (current value:
>>> <none>)
>>>                            Default selection set of components for
>>> the paffinity framework (<none> means "use all components that can be
>>> found")
>>>             MCA paffinity: parameter
>>> "paffinity_linux_priority" (current value: "10")
>>>                            Priority of the linux paffinity component
>>>             MCA paffinity: information
>>> "paffinity_linux_have_cpu_set_t" (value: "1")
>>>                            Whether this component was compiled on a
>>> system with the type cpu_set_t or not (1 = yes, 0 = no)
>>>             MCA paffinity: information
>>> "paffinity_linux_CPU_ZERO_ok" (value: "1")
>>>                            Whether this component was compiled on a
>>> system where CPU_ZERO() is functional or broken (1 = functional, 0 =
>>> broken/not available)
>>>             MCA paffinity: information
>>> "paffinity_linux_sched_setaffinity_num_params" (value: "3")
>>>                            The number of parameters that
>>> sched_set_affinity() takes on the machine where this component was
>>> compiled
>>>             MCA maffinity: parameter "maffinity" (current value:  
>>> <none>)
>>>                            Default selection set of components for
>>> the maffinity framework (<none> means "use all components that can be
>>> found")
>>>             MCA maffinity: parameter
>>> "maffinity_first_use_priority" (current value: "10")
>>>                            Priority of the first_use maffinity  
>>> component
>>>                 MCA timer: parameter "timer" (current value: <none>)
>>>                            Default selection set of components for
>>> the timer framework (<none> means "use all components that can be
>>> found")
>>>                 MCA timer: parameter "timer_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the timer framework (0
>>> = no verbosity)
>>>                 MCA timer: parameter "timer_linux_priority" (current
>>> value: "0")
>>>             MCA allocator: parameter "allocator" (current value:  
>>> <none>)
>>>                            Default selection set of components for
>>> the allocator framework (<none> means "use all components that can be
>>> found")
>>>             MCA allocator: parameter
>>> "allocator_base_verbose" (current value: "0")
>>>                            Verbosity level for the allocator
>>> framework (0 = no verbosity)
>>>             MCA allocator: parameter
>>> "allocator_basic_priority" (current value: "0")
>>>             MCA allocator: parameter
>>> "allocator_bucket_num_buckets" (current value: "30")
>>>             MCA allocator: parameter
>>> "allocator_bucket_priority" (current value: "0")
>>>                  MCA coll: parameter "coll" (current value: <none>)
>>>                            Default selection set of components for
>>> the coll framework (<none> means "use all components that can be  
>>> found")
>>>                  MCA coll: parameter "coll_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the coll framework (0
>>> = no verbosity)
>>>                  MCA coll: parameter "coll_basic_priority" (current
>>> value: "10")
>>>                            Priority of the basic coll component
>>>                  MCA coll: parameter "coll_basic_crossover" (current
>>> value: "4")
>>>                            Minimum number of processes in a
>>> communicator before using the logarithmic algorithms
>>>                  MCA coll: parameter "coll_self_priority" (current
>>> value: "75")
>>>                  MCA coll: parameter "coll_sm_priority" (current
>>> value: "0")
>>>                            Priority of the sm coll component
>>>                  MCA coll: parameter "coll_sm_control_size" (current
>>> value: "4096")
>>>                            Length of the control data -- should
>>> usually be either the length of a cache line on most SMPs, or the
>>> size of a page on machines that support direct memory affinity page
>>> placement (in bytes)
>>>                  MCA coll: parameter
>>> "coll_sm_bootstrap_filename" (current value:  
>>> "shared_mem_sm_bootstrap")
>>>                            Filename (in the Open MPI session
>>> directory) of the coll sm component bootstrap rendezvous mmap file
>>>                  MCA coll: parameter
>>> "coll_sm_bootstrap_num_segments" (current value: "8")
>>>                            Number of segments in the bootstrap file
>>>                  MCA coll: parameter "coll_sm_fragment_size" (current
>>> value: "8192")
>>>                            Fragment size (in bytes) used for passing
>>> data through shared memory (will be rounded up to the nearest
>>> control_size size)
>>>                  MCA coll: parameter "coll_sm_mpool" (current value:
>>> "sm")
>>>                            Name of the mpool component to use
>>>                  MCA coll: parameter
>>> "coll_sm_comm_in_use_flags" (current value: "2")
>>>                            Number of "in use" flags, used to mark a
>>> message passing area segment as currently being used or not (must be
>>>> = 2 and <= comm_num_segments)
>>>                  MCA coll: parameter
>>> "coll_sm_comm_num_segments" (current value: "8")
>>>                            Number of segments in each communicator's
>>> shared memory message passing area (must be >= 2, and must be a
>>> multiple of comm_in_use_flags)
>>>                  MCA coll: parameter "coll_sm_tree_degree" (current
>>> value: "4")
>>>                            Degree of the tree for tree-based
>>> operations (must be => 1 and <= min(control_size, 255))
>>>                  MCA coll: information
>>> "coll_sm_shared_mem_used_bootstrap" (value: "160")
>>>                            Amount of shared memory used in the shared
>>> memory bootstrap area (in bytes)
>>>                  MCA coll: parameter
>>> "coll_sm_info_num_procs" (current value: "4")
>>>                            Number of processes to use for the
>>> calculation of the shared_mem_size MCA information parameter (must be
>>> => 2)
>>>                  MCA coll: information
>>> "coll_sm_shared_mem_used_data" (value: "548864")
>>>                            Amount of shared memory used in the shared
>>> memory data area for info_num_procs processes (in bytes)
>>>                  MCA coll: parameter "coll_tuned_priority" (current
>>> value: "30")
>>>                            Priority of the tuned coll component
>>>                  MCA coll: parameter
>>> "coll_tuned_pre_allocate_memory_comm_size_limit" (current value:
>>> "32768")
>>>                            Size of communicator were we stop pre-
>>> allocating memory for the fixed internal buffer used for message
>>> requests etc that is hung off the communicator data segment. I.e. if
>>> you have a 100'000 nodes you might not want to pre-allocate
>>>                            200'000 request handle slots per
>>> communicator instance!
>>>                  MCA coll: parameter
>>> "coll_tuned_init_tree_fanout" (current value: "4")
>>>                            Inital fanout used in the tree topologies
>>> for each communicator. This is only an initial guess, if a tuned
>>> collective needs a different fanout for an operation, it build it
>>> dynamically. This parameter is only for the first guess and might
>>>                            save a little time
>>>                  MCA coll: parameter
>>> "coll_tuned_init_chain_fanout" (current value: "4")
>>>                            Inital fanout used in the chain (fanout
>>> followed by pipeline) topologies for each communicator. This is only
>>> an initial guess, if a tuned collective needs a different fanout for
>>> an operation, it build it dynamically. This parameter is
>>>                            only for the first guess and might save a
>>> little time
>>>                  MCA coll: parameter
>>> "coll_tuned_use_dynamic_rules" (current value: "0")
>>>                            Switch used to decide if we use static
>>> (compiled/if statements) or dynamic (built at runtime) decision
>>> function rules
>>>                    MCA io: parameter
>>> "io_base_freelist_initial_size" (current value: "16")
>>>                            Initial MPI-2 IO request freelist size
>>>                    MCA io: parameter
>>> "io_base_freelist_max_size" (current value: "64")
>>>                            Max size of the MPI-2 IO request freelist
>>>                    MCA io: parameter
>>> "io_base_freelist_increment" (current value: "16")
>>>                            Increment size of the MPI-2 IO request
>>> freelist
>>>                    MCA io: parameter "io" (current value: <none>)
>>>                            Default selection set of components for
>>> the io framework (<none> means "use all components that can be  
>>> found")
>>>                    MCA io: parameter "io_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the io framework (0 =
>>> no verbosity)
>>>                    MCA io: parameter "io_romio_priority" (current
>>> value: "10")
>>>                            Priority of the io romio component
>>>                    MCA io: parameter
>>> "io_romio_delete_priority" (current value: "10")
>>>                            Delete priority of the io romio component
>>>                    MCA io: parameter
>>> "io_romio_enable_parallel_optimizations" (current value: "0")
>>>                            Enable set of Open MPI-added options to
>>> improve collective file i/o performance
>>>                 MCA mpool: parameter "mpool" (current value: <none>)
>>>                            Default selection set of components for
>>> the mpool framework (<none> means "use all components that can be
>>> found")
>>>                 MCA mpool: parameter "mpool_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the mpool framework (0
>>> = no verbosity)
>>>                 MCA mpool: parameter
>>> "mpool_rdma_rcache_name" (current value: "vma")
>>>                            The name of the registration cache the
>>> mpool should use
>>>                 MCA mpool: parameter
>>> "mpool_rdma_rcache_size_limit" (current value: "0")
>>>                            the maximum size of registration cache in
>>> bytes. 0 is unlimited (default 0)
>>>                 MCA mpool: parameter
>>> "mpool_rdma_print_stats" (current value: "0")
>>>                            print pool usage statistics at the end of
>>> the run
>>>                 MCA mpool: parameter "mpool_rdma_priority" (current
>>> value: "0")
>>>                 MCA mpool: parameter "mpool_sm_allocator" (current
>>> value: "bucket")
>>>                            Name of allocator component to use with sm
>>> mpool
>>>                 MCA mpool: parameter "mpool_sm_max_size" (current
>>> value: "536870912")
>>>                            Maximum size of the sm mpool shared memory
>>> file
>>>                 MCA mpool: parameter "mpool_sm_min_size" (current
>>> value: "134217728")
>>>                            Minimum size of the sm mpool shared memory
>>> file
>>>                 MCA mpool: parameter
>>> "mpool_sm_per_peer_size" (current value: "33554432")
>>>                            Size (in bytes) to allocate per local peer
>>> in the sm mpool shared memory file, bounded by min_size and max_size
>>>                 MCA mpool: parameter "mpool_sm_priority" (current
>>> value: "0")
>>>                 MCA mpool: parameter
>>> "mpool_base_use_mem_hooks" (current value: "0")
>>>                            use memory hooks for deregistering freed
>>> memory
>>>                 MCA mpool: parameter "mpool_use_mem_hooks" (current
>>> value: "0")
>>>                            (deprecated, use mpool_base_use_mem_hooks)
>>>                 MCA mpool: parameter
>>> "mpool_base_disable_sbrk" (current value: "0")
>>>                            use mallopt to override calling sbrk
>>> (doesn't return memory to OS!)
>>>                 MCA mpool: parameter "mpool_disable_sbrk" (current
>>> value: "0")
>>>                            (deprecated, use  
>>> mca_mpool_base_disable_sbrk)
>>>                   MCA pml: parameter "pml" (current value: <none>)
>>>                            Default selection set of components for
>>> the pml framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA pml: parameter "pml_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the pml framework (0 =
>>> no verbosity)
>>>                   MCA pml: parameter "pml_cm_free_list_num" (current
>>> value: "4")
>>>                            Initial size of request free lists
>>>                   MCA pml: parameter "pml_cm_free_list_max" (current
>>> value: "-1")
>>>                            Maximum size of request free lists
>>>                   MCA pml: parameter "pml_cm_free_list_inc" (current
>>> value: "64")
>>>                            Number of elements to add when growing
>>> request free lists
>>>                   MCA pml: parameter "pml_cm_priority" (current
>>> value: "30")
>>>                            CM PML selection priority
>>>                   MCA pml: parameter "pml_ob1_free_list_num" (current
>>> value: "4")
>>>                   MCA pml: parameter "pml_ob1_free_list_max" (current
>>> value: "-1")
>>>                   MCA pml: parameter "pml_ob1_free_list_inc" (current
>>> value: "64")
>>>                   MCA pml: parameter "pml_ob1_priority" (current
>>> value: "20")
>>>                   MCA pml: parameter "pml_ob1_eager_limit" (current
>>> value: "131072")
>>>                   MCA pml: parameter
>>> "pml_ob1_send_pipeline_depth" (current value: "3")
>>>                   MCA pml: parameter
>>> "pml_ob1_recv_pipeline_depth" (current value: "4")
>>>                   MCA bml: parameter "bml" (current value: <none>)
>>>                            Default selection set of components for
>>> the bml framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA bml: parameter "bml_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the bml framework (0 =
>>> no verbosity)
>>>                   MCA bml: parameter
>>> "bml_r2_show_unreach_errors" (current value: "1")
>>>                            Show error message when procs are  
>>> unreachable
>>>                   MCA bml: parameter "bml_r2_priority" (current
>>> value: "0")
>>>                MCA rcache: parameter "rcache" (current value: <none>)
>>>                            Default selection set of components for
>>> the rcache framework (<none> means "use all components that can be
>>> found")
>>>                MCA rcache: parameter "rcache_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the rcache framework
>>> (0 = no verbosity)
>>>                MCA rcache: parameter "rcache_vma_priority" (current
>>> value: "0")
>>>                   MCA btl: parameter "btl_base_debug" (current value:
>>> "0")
>>>                            If btl_base_debug is 1 standard debug is
>>> output, if > 1 verbose debug is output
>>>                   MCA btl: parameter "btl" (current value: <none>)
>>>                            Default selection set of components for
>>> the btl framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA btl: parameter "btl_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the btl framework (0 =
>>> no verbosity)
>>>                   MCA btl: parameter
>>> "btl_self_free_list_num" (current value: "0")
>>>                            Number of fragments by default
>>>                   MCA btl: parameter
>>> "btl_self_free_list_max" (current value: "-1")
>>>                            Maximum number of fragments
>>>                   MCA btl: parameter
>>> "btl_self_free_list_inc" (current value: "32")
>>>                            Increment by this number of fragments
>>>                   MCA btl: parameter "btl_self_eager_limit" (current
>>> value: "131072")
>>>                            Eager size fragmeng (before the rendez-
>>> vous ptotocol)
>>>                   MCA btl: parameter
>>> "btl_self_min_send_size" (current value: "262144")
>>>                            Minimum fragment size after the rendez- 
>>> vous
>>>                   MCA btl: parameter
>>> "btl_self_max_send_size" (current value: "262144")
>>>                            Maximum fragment size after the rendez- 
>>> vous
>>>                   MCA btl: parameter
>>> "btl_self_min_rdma_size" (current value: "2147483647")
>>>                            Maximum fragment size for the RDMA  
>>> transfer
>>>                   MCA btl: parameter
>>> "btl_self_max_rdma_size" (current value: "2147483647")
>>>                            Maximum fragment size for the RDMA  
>>> transfer
>>>                   MCA btl: parameter "btl_self_exclusivity" (current
>>> value: "65536")
>>>                            Device exclusivity
>>>                   MCA btl: parameter "btl_self_flags" (current value:
>>> "10")
>>>                            Active behavior flags
>>>                   MCA btl: parameter "btl_self_priority" (current
>>> value: "0")
>>>                   MCA btl: parameter "btl_sm_free_list_num" (current
>>> value: "8")
>>>                   MCA btl: parameter "btl_sm_free_list_max" (current
>>> value: "-1")
>>>                   MCA btl: parameter "btl_sm_free_list_inc" (current
>>> value: "64")
>>>                   MCA btl: parameter "btl_sm_exclusivity" (current
>>> value: "65535")
>>>                   MCA btl: parameter "btl_sm_latency" (current value:
>>> "100")
>>>                   MCA btl: parameter "btl_sm_max_procs" (current
>>> value: "-1")
>>>                   MCA btl: parameter "btl_sm_sm_extra_procs" (current
>>> value: "2")
>>>                   MCA btl: parameter "btl_sm_mpool" (current value:
>>> "sm")
>>>                   MCA btl: parameter "btl_sm_eager_limit" (current
>>> value: "4096")
>>>                   MCA btl: parameter "btl_sm_max_frag_size" (current
>>> value: "32768")
>>>                   MCA btl: parameter
>>> "btl_sm_size_of_cb_queue" (current value: "128")
>>>                   MCA btl: parameter
>>> "btl_sm_cb_lazy_free_freq" (current value: "120")
>>>                   MCA btl: parameter "btl_sm_priority" (current
>>> value: "0")
>>>                   MCA btl: parameter "btl_tcp_if_include" (current
>>> value: <none>)
>>>                   MCA btl: parameter "btl_tcp_if_exclude" (current
>>> value: "lo")
>>>                   MCA btl: parameter "btl_tcp_free_list_num" (current
>>> value: "8")
>>>                   MCA btl: parameter "btl_tcp_free_list_max" (current
>>> value: "-1")
>>>                   MCA btl: parameter "btl_tcp_free_list_inc" (current
>>> value: "32")
>>>                   MCA btl: parameter "btl_tcp_sndbuf" (current value:
>>> "131072")
>>>                   MCA btl: parameter "btl_tcp_rcvbuf" (current value:
>>> "131072")
>>>                   MCA btl: parameter
>>> "btl_tcp_endpoint_cache" (current value: "30720")
>>>                   MCA btl: parameter "btl_tcp_exclusivity" (current
>>> value: "0")
>>>                   MCA btl: parameter "btl_tcp_eager_limit" (current
>>> value: "65536")
>>>                   MCA btl: parameter "btl_tcp_min_send_size" (current
>>> value: "65536")
>>>                   MCA btl: parameter "btl_tcp_max_send_size" (current
>>> value: "131072")
>>>                   MCA btl: parameter "btl_tcp_min_rdma_size" (current
>>> value: "131072")
>>>                   MCA btl: parameter "btl_tcp_max_rdma_size" (current
>>> value: "2147483647")
>>>                   MCA btl: parameter "btl_tcp_flags" (current value:
>>> "122")
>>>                   MCA btl: parameter "btl_tcp_priority" (current
>>> value: "0")
>>>                   MCA btl: parameter "btl_base_include" (current
>>> value: <none>)
>>>                   MCA btl: parameter "btl_base_exclude" (current
>>> value: <none>)
>>>                   MCA btl: parameter
>>> "btl_base_warn_component_unused" (current value: "1")
>>>                            This parameter is used to turn on warning
>>> messages when certain NICs are not used
>>>                   MCA mtl: parameter "mtl" (current value: <none>)
>>>                            Default selection set of components for
>>> the mtl framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA mtl: parameter "mtl_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the mtl framework (0 =
>>> no verbosity)
>>>                  MCA topo: parameter "topo" (current value: <none>)
>>>                            Default selection set of components for
>>> the topo framework (<none> means "use all components that can be  
>>> found")
>>>                  MCA topo: parameter "topo_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the topo framework (0
>>> = no verbosity)
>>>                   MCA osc: parameter "osc" (current value: <none>)
>>>                            Default selection set of components for
>>> the osc framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA osc: parameter "osc_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the osc framework (0 =
>>> no verbosity)
>>>                   MCA osc: parameter "osc_pt2pt_no_locks" (current
>>> value: "0")
>>>                            Enable optimizations available only if
>>> MPI_LOCK is not used.
>>>                   MCA osc: parameter "osc_pt2pt_eager_limit" (current
>>> value: "16384")
>>>                            Max size of eagerly sent data
>>>                   MCA osc: parameter "osc_pt2pt_priority" (current
>>> value: "0")
>>>                MCA errmgr: parameter "errmgr" (current value: <none>)
>>>                            Default selection set of components for
>>> the errmgr framework (<none> means "use all components that can be
>>> found")
>>>                MCA errmgr: parameter "errmgr_hnp_debug" (current
>>> value: "0")
>>>                MCA errmgr: parameter "errmgr_hnp_priority" (current
>>> value: "0")
>>>                MCA errmgr: parameter "errmgr_orted_debug" (current
>>> value: "0")
>>>                MCA errmgr: parameter "errmgr_orted_priority" (current
>>> value: "0")
>>>                MCA errmgr: parameter "errmgr_proxy_debug" (current
>>> value: "0")
>>>                MCA errmgr: parameter "errmgr_proxy_priority" (current
>>> value: "0")
>>>                   MCA gpr: parameter "gpr_base_maxsize" (current
>>> value: "2147483647")
>>>                   MCA gpr: parameter "gpr_base_blocksize" (current
>>> value: "512")
>>>                   MCA gpr: parameter "gpr" (current value: <none>)
>>>                            Default selection set of components for
>>> the gpr framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA gpr: parameter "gpr_null_priority" (current
>>> value: "0")
>>>                   MCA gpr: parameter "gpr_proxy_debug" (current
>>> value: "0")
>>>                   MCA gpr: parameter "gpr_proxy_priority" (current
>>> value: "0")
>>>                   MCA gpr: parameter "gpr_replica_debug" (current
>>> value: "0")
>>>                   MCA gpr: parameter "gpr_replica_isolate" (current
>>> value: "0")
>>>                   MCA gpr: parameter "gpr_replica_priority" (current
>>> value: "0")
>>>                   MCA iof: parameter "iof_base_window_size" (current
>>> value: "4096")
>>>                   MCA iof: parameter "iof_base_service" (current
>>> value: "0.0.0")
>>>                   MCA iof: parameter "iof" (current value: <none>)
>>>                            Default selection set of components for
>>> the iof framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA iof: parameter "iof_proxy_debug" (current
>>> value: "1")
>>>                   MCA iof: parameter "iof_proxy_priority" (current
>>> value: "0")
>>>                   MCA iof: parameter "iof_svc_debug" (current value:
>>> "1")
>>>                   MCA iof: parameter "iof_svc_priority" (current
>>> value: "0")
>>>                    MCA ns: parameter "ns" (current value: <none>)
>>>                            Default selection set of components for
>>> the ns framework (<none> means "use all components that can be  
>>> found")
>>>                    MCA ns: parameter "ns_proxy_debug" (current value:
>>> "0")
>>>                    MCA ns: parameter "ns_proxy_maxsize" (current
>>> value: "2147483647")
>>>                    MCA ns: parameter "ns_proxy_blocksize" (current
>>> value: "512")
>>>                    MCA ns: parameter "ns_proxy_priority" (current
>>> value: "0")
>>>                    MCA ns: parameter "ns_replica_debug" (current
>>> value: "0")
>>>                    MCA ns: parameter "ns_replica_isolate" (current
>>> value: "0")
>>>                    MCA ns: parameter "ns_replica_maxsize" (current
>>> value: "2147483647")
>>>                    MCA ns: parameter "ns_replica_blocksize" (current
>>> value: "512")
>>>                    MCA ns: parameter "ns_replica_priority" (current
>>> value: "0")
>>>                   MCA oob: parameter "oob" (current value: <none>)
>>>                            Default selection set of components for
>>> the oob framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA oob: parameter "oob_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the oob framework (0 =
>>> no verbosity)
>>>                   MCA oob: parameter "oob_tcp_peer_limit" (current
>>> value: "-1")
>>>                   MCA oob: parameter "oob_tcp_peer_retries" (current
>>> value: "60")
>>>                   MCA oob: parameter "oob_tcp_debug" (current value:
>>> "0")
>>>                   MCA oob: parameter "oob_tcp_include" (current
>>> value: <none>)
>>>                   MCA oob: parameter "oob_tcp_exclude" (current
>>> value: <none>)
>>>                   MCA oob: parameter "oob_tcp_sndbuf" (current value:
>>> "131072")
>>>                   MCA oob: parameter "oob_tcp_rcvbuf" (current value:
>>> "131072")
>>>                   MCA oob: parameter "oob_tcp_connect_sleep" (current
>>> value: "1")
>>>                            Enable (1) /Disable (0)  random sleep for
>>> connection wireup
>>>                   MCA oob: parameter "oob_tcp_listen_mode" (current
>>> value: "event")
>>>                            Mode for HNP to accept incoming
>>> connections: event, listen_thread
>>>                   MCA oob: parameter
>>> "oob_tcp_listen_thread_max_queue" (current value: "10")
>>>                            High water mark for queued accepted socket
>>> list size
>>>                   MCA oob: parameter
>>> "oob_tcp_listen_thread_max_time" (current value: "10")
>>>                            Maximum amount of time (in milliseconds)
>>> to wait between processing accepted socket list
>>>                   MCA oob: parameter
>>> "oob_tcp_accept_spin_count" (current value: "10")
>>>                            Number of times to let accept return
>>> EWOULDBLOCK before updating accepted socket list
>>>                   MCA oob: parameter "oob_tcp_priority" (current
>>> value: "0")
>>>                   MCA ras: parameter "ras" (current value: <none>)
>>>                   MCA ras: parameter
>>> "ras_dash_host_priority" (current value: "5")
>>>                            Selection priority for the dash_host RAS
>>> component
>>>                   MCA ras: parameter "ras_gridengine_debug" (current
>>> value: "0")
>>>                            Enable debugging output for the gridengine
>>> ras component
>>>                   MCA ras: parameter
>>> "ras_gridengine_priority" (current value: "100")
>>>                            Priority of the gridengine ras component
>>>                   MCA ras: parameter
>>> "ras_gridengine_verbose" (current value: "0")
>>>                            Enable verbose output for the gridengine
>>> ras component
>>>                   MCA ras: parameter
>>> "ras_gridengine_show_jobid" (current value: "0")
>>>                            Show the JOB_ID of the Grid Engine job
>>>                   MCA ras: parameter
>>> "ras_localhost_priority" (current value: "0")
>>>                            Selection priority for the localhost RAS
>>> component
>>>                   MCA ras: parameter "ras_slurm_priority" (current
>>> value: "75")
>>>                            Priority of the slurm ras component
>>>                   MCA rds: parameter "rds" (current value: <none>)
>>>                   MCA rds: parameter "rds_hostfile_debug" (current
>>> value: "0")
>>>                            Toggle debug output for hostfile RDS
>>> component
>>>                   MCA rds: parameter "rds_hostfile_path" (current
>>> value: "/usr/local/openmpi-1.2.2/etc/openmpi-default-hostfile")
>>>                            ORTE Host filename
>>>                   MCA rds: parameter "rds_hostfile_priority" (current
>>> value: "0")
>>>                   MCA rds: parameter "rds_proxy_priority" (current
>>> value: "0")
>>>                   MCA rds: parameter "rds_resfile_debug" (current
>>> value: "0")
>>>                            Toggle debug output for resfile RDS  
>>> component
>>>                   MCA rds: parameter "rds_resfile_name" (current
>>> value: <none>)
>>>                            ORTE Resource filename
>>>                   MCA rds: parameter "rds_resfile_priority" (current
>>> value: "0")
>>>                 MCA rmaps: parameter "rmaps_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the rmaps framework
>>>                 MCA rmaps: parameter
>>> "rmaps_base_schedule_policy" (current value: "unspec")
>>>                            Scheduling Policy for RMAPS. [slot | node]
>>>                 MCA rmaps: parameter "rmaps_base_pernode" (current
>>> value: "0")
>>>                            Launch one ppn as directed
>>>                 MCA rmaps: parameter "rmaps_base_n_pernode" (current
>>> value: "-1")
>>>                            Launch n procs/node
>>>                 MCA rmaps: parameter
>>> "rmaps_base_no_schedule_local" (current value: "0")
>>>                            If false, allow scheduling MPI
>>> applications on the same node as mpirun (default).  If true, do not
>>> schedule any MPI applications on the same node as mpirun
>>>                 MCA rmaps: parameter
>>> "rmaps_base_no_oversubscribe" (current value: "0")
>>>                            If true, then do not allow
>>> oversubscription of nodes - mpirun will return an error if there
>>> aren't enough nodes to launch all processes without oversubscribing
>>>                 MCA rmaps: parameter "rmaps" (current value: <none>)
>>>                            Default selection set of components for
>>> the rmaps framework (<none> means "use all components that can be
>>> found")
>>>                 MCA rmaps: parameter
>>> "rmaps_round_robin_debug" (current value: "1")
>>>                            Toggle debug output for Round Robin RMAPS
>>> component
>>>                 MCA rmaps: parameter
>>> "rmaps_round_robin_priority" (current value: "1")
>>>                            Selection priority for Round Robin RMAPS
>>> component
>>>                  MCA rmgr: parameter "rmgr" (current value: <none>)
>>>                            Default selection set of components for
>>> the rmgr framework (<none> means "use all components that can be  
>>> found")
>>>                  MCA rmgr: parameter "rmgr_proxy_priority" (current
>>> value: "0")
>>>                  MCA rmgr: parameter "rmgr_urm_priority" (current
>>> value: "0")
>>>                   MCA rml: parameter "rml" (current value: <none>)
>>>                            Default selection set of components for
>>> the rml framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA rml: parameter "rml_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the rml framework (0 =
>>> no verbosity)
>>>                   MCA rml: parameter "rml_oob_priority" (current
>>> value: "0")
>>>                   MCA pls: parameter
>>> "pls_base_reuse_daemons" (current value: "0")
>>>                            If nonzero, reuse daemons to launch
>>> dynamically spawned processes.  If zero, do not reuse daemons  
>>> (default)
>>>                   MCA pls: parameter "pls" (current value: <none>)
>>>                            Default selection set of components for
>>> the pls framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA pls: parameter "pls_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the pls framework (0 =
>>> no verbosity)
>>>                   MCA pls: parameter "pls_gridengine_debug" (current
>>> value: "0")
>>>                            Enable debugging of gridengine pls  
>>> component
>>>                   MCA pls: parameter
>>> "pls_gridengine_verbose" (current value: "0")
>>>                            Enable verbose output of the gridengine
>>> qrsh -inherit command
>>>                   MCA pls: parameter
>>> "pls_gridengine_priority" (current value: "100")
>>>                            Priority of the gridengine pls component
>>>                   MCA pls: parameter "pls_gridengine_orted" (current
>>> value: "orted")
>>>                            The command name that the gridengine pls
>>> component will invoke for the ORTE daemon
>>>                   MCA pls: parameter "pls_proxy_priority" (current
>>> value: "0")
>>>                   MCA pls: parameter "pls_rsh_debug" (current value:
>>> "0")
>>>                            Whether or not to enable debugging output
>>> for the rsh pls component (0 or 1)
>>>                   MCA pls: parameter
>>> "pls_rsh_num_concurrent" (current value: "128")
>>>                            How many pls_rsh_agent instances to invoke
>>> concurrently (must be > 0)
>>>                   MCA pls: parameter "pls_rsh_force_rsh" (current
>>> value: "0")
>>>                            Force the launcher to always use rsh, even
>>> for local daemons
>>>                   MCA pls: parameter "pls_rsh_orted" (current value:
>>> "orted")
>>>                            The command name that the rsh pls
>>> component will invoke for the ORTE daemon
>>>                   MCA pls: parameter "pls_rsh_priority" (current
>>> value: "10")
>>>                            Priority of the rsh pls component
>>>                   MCA pls: parameter "pls_rsh_delay" (current value:
>>> "1")
>>>                            Delay (in seconds) between invocations of
>>> the remote agent, but only used when the "debug" MCA parameter is
>>> true, or the top-level MCA debugging is enabled (otherwise this value
>>> is ignored)
>>>                   MCA pls: parameter "pls_rsh_reap" (current  
>>> value: "1")
>>>                            If set to 1, wait for all the processes to
>>> complete before exiting.  Otherwise, quit immediately -- without
>>> waiting for confirmation that all other processes in the job have
>>> completed.
>>>                   MCA pls: parameter
>>> "pls_rsh_assume_same_shell" (current value: "1")
>>>                            If set to 1, assume that the shell on the
>>> remote node is the same as the shell on the local node.  Otherwise,
>>> probe for what the remote shell.
>>>                   MCA pls: parameter "pls_rsh_agent" (current value:
>>> "rsh")
>>>                            The command used to launch executables on
>>> remote nodes (typically either "ssh" or "rsh")
>>>                   MCA pls: parameter "pls_slurm_debug" (current
>>> value: "0")
>>>                            Enable debugging of slurm pls
>>>                   MCA pls: parameter "pls_slurm_priority" (current
>>> value: "75")
>>>                            Default selection priority
>>>                   MCA pls: parameter "pls_slurm_orted" (current
>>> value: "orted")
>>>                            Command to use to start proxy orted
>>>                   MCA pls: parameter "pls_slurm_args" (current value:
>>> <none>)
>>>                            Custom arguments to srun
>>>                   MCA sds: parameter "sds" (current value: <none>)
>>>                            Default selection set of components for
>>> the sds framework (<none> means "use all components that can be  
>>> found")
>>>                   MCA sds: parameter "sds_base_verbose" (current
>>> value: "0")
>>>                            Verbosity level for the sds framework (0 =
>>> no verbosity)
>>>                   MCA sds: parameter "sds_env_priority" (current
>>> value: "0")
>>>                   MCA sds: parameter "sds_pipe_priority" (current
>>> value: "0")
>>>                   MCA sds: parameter "sds_seed_priority" (current
>>> value: "0")
>>>                   MCA sds: parameter
>>> "sds_singleton_priority" (current value: "0")
>>>                   MCA sds: parameter "sds_slurm_priority" (current
>>> value: "0")
>>> -bash-3.00$
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


Reply via email to