Have you tried the --wdir option yet? It should let you set your working directory to anywhere. I don't believe it will require you to have a home directory on the backend nodes, though I can't sweathat ssh will be happy if you don't.
Just do "mpirun -h" for a full list of options - it will describe the exact format of the wdir one plusthers you might find useful. Ralph On 6/7/07 11:12 AM, "Guillaume THOMAS-COLLIGNON<guillaume.thomas-collig...@cggveritas.com> wrote: > I am trying to switch to OpenI, and I ran into a problem : my home > directory must exist on all the nodes, or ted will crash. > > I have a "master" machine where I initiate the mpirun command> Then I have a > bunch of slave machines, which will ao execute the > MPI job. > My user exists on all the machines, but the home directory is not > mounted on the slaves,o it's only visible on the master node. I can > log on a slave node, but don't have a home there. Of course the > binary I'm running exists on all the machines (not in my home !). And > the problem can be reproduced by running a shell command too, to make > things simpler. > > We have thousands of slave nodes and we don't want to mount the > user's homedirs on all the slaves, so a fix would be really really nice. > > Example : > > I have 3 hosts, master, slave1, slave2. My home directory exists only > on master. > > If I log on master and run "mpirun -host master,slave1uname -a" I get > a segfault. > If I log on slave1 and run "mpirun -host slave1,slave2 uname -a", it > runs fine. My home directory does not exist on either slave1 or slave2. > If I log on master and run "mpirun -host master uname -a" it runs > fine. I can run across several master nodes, it's fine too. > > So it runs fine if my home directory exists everywhere, or if it does > not exist at all. If it exists only on some nodes and not others, > orted crashes. > I thought it could be related to my environment but I created a new > user with an empty home and it does the same thing. As soon as I > create the homedir on slave1 and slave2 it works fine. > > > > > I'm using OpenMPI 1.2.2, here is the error message and the result of > ompi_info. > > Short version (rnd04 is the master, r137n001 is a slave node). > > -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001 > uname -a > Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 > x86_64 x86_64 GNU/Linux > [r137n001:31533] *** Process received signal *** > [r137n001:31533] Signal: Segmentation fault (11) > [r137n001:31533] Signal code: Address not mapped (1) > [r137n001:31533] Failing at address: 0x1 > [r137n001:31533] [ 0] [0xffffe600] > [r137n001:31533] [ 1] /lib/tls/libc.so.6 [0xbf3bfc] > [r137n001:31533] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb) [0xbf3e3b] > [r137n001:31533] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 > (opal_show_help+0x263) [0xf7f78de3] > [r137n001:31533] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 > (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef] > [r137n001:31533] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f) > [0xf7ea041f] > [r137n001:31533] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea] > [r137n001:31533] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136) [0xf7ef65c6] > [r137n001:31533] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68] > [r137n001:31533] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 > [0xf7fd9a18] > [r137n001:31533] [10] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc] > [r137n001:31533] [11] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_oob_tcp.so [0xf7f07f61] > [r137n001:31533] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 > (opal_event_base_loop+0x388) [0xf7f67dd8] > [r137n001:31533] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 > (opal_event_loop+0x29) [0xf7f67fb9] > [r137n001:31533] [14] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_oob_tcp.so(mca_oob_tcp_msg_wait+0x37) [0xf7f053c7] > [r137n001:31533] [15] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_oob_tcp.so(mca_oob_tcp_recv+0x374) [0xf7f09a04] > [r137n001:31533] [16] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 > (mca_oob_recv_packed+0x4d) [0xf7fd980d] > [r137n001:31533] [17] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_gpr_proxy.so(orte_gpr_proxy_exec_compound_cmd+0x137) [0xf7ef55e7] > [r137n001:31533] [18] /usr/local/openmpi-1.2.2/bin/orted(main+0x99d) > [0x8049d0d] > [r137n001:31533] [19] /lib/tls/libc.so.6(__libc_start_main+0xd3) > [0xbcee23] > [r137n001:31533] [20] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1] > [r137n001:31533] *** End of error message *** > mpirun noticed that job rank 1 with PID 31533 on node r137n001 exited > on signal 11 (Segmentation fault). > > > > If I create /home/toto on r137n001, it works fine : > (as root on r137n001: "mkdir /home/toto && chown toto:users /home/toto") > > -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001 > uname -a > Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 > x86_64 x86_64 GNU/Linux > Linux r137n001 2.6.9-34.ELsmp #1 SMP Fri Feb 24 16:56:28 EST 2006 > x86_64 x86_64 x86_64 GNU/Linux > > > I tried to use ssh instead of rsh, it crashes too. > > If anyone knows a way to run OpenMPI jobs in this configuration where > the home directory does not exist on all the nodes, it woud really > help ! > > Or is there a way to fix orted so that it won't crash ? > > > Here is the crash with -d option : > > -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -d -host > rnd04,r137n001 uname -a > [rnd04:10736] connect_uni: connection not allowed > [rnd04:10736] [0,0,0] setting up session dir with > [rnd04:10736] universe default-universe-10736 > [rnd04:10736] user toto > [rnd04:10736] host rnd04 > [rnd04:10736] jobid 0 > [rnd04:10736] procid 0 > [rnd04:10736] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default- > universe-10736/0/0 > [rnd04:10736] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default- > universe-10736/0 > [rnd04:10736] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default- > universe-10736 > [rnd04:10736] top: openmpi-sessions-toto@rnd04_0 > [rnd04:10736] tmp: /tmp > [rnd04:10736] [0,0,0] contact_file /tmp/openmpi-sessions-toto@rnd04_0/ > default-universe-10736/universe-setup.txt > [rnd04:10736] [0,0,0] wrote setup file > [rnd04:10736] pls:rsh: local csh: 0, local sh: 1 > [rnd04:10736] pls:rsh: assuming same remote shell as local shell > [rnd04:10736] pls:rsh: remote csh: 0, remote sh: 1 > [rnd04:10736] pls:rsh: final template argv: > [rnd04:10736] pls:rsh: /usr/bin/rsh <template> orted --debug -- > bootproxy 1 --name <template> --num_procs 3 --vpid_start 0 --nodename > <template> --universe toto@rnd04:default-universe-10736 --nsreplica > "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" -- > gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" > [rnd04:10736] pls:rsh: launching on node rnd04 > [rnd04:10736] pls:rsh: rnd04 is a LOCAL node > [rnd04:10736] pls:rsh: reset PATH: /usr/local/openmpi-1.2.2/bin:/usr/ > local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin > [rnd04:10736] pls:rsh: reset LD_LIBRARY_PATH: /usr/local/ > openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib:/usr/local/ > openmpi-1.2.2/lib64 > [rnd04:10736] pls:rsh: changing to directory /home/toto > [rnd04:10736] pls:rsh: executing: (/usr/local/openmpi-1.2.2/bin/ > orted) orted --debug --bootproxy 1 --name 0.0.1 --num_procs 3 -- > vpid_start 0 --nodename rnd04 --universe toto@rnd04:default- > universe-10736 --nsreplica "0.0.0;tcp://172.28.20.143:33029;tcp:// > 10.3.254.105:33029" --gprreplica "0.0.0;tcp:// > 172.28.20.143:33029;tcp://10.3.254.105:33029" --set-sid > [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color HISTSIZE=1000 > USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:/usr/local/ > openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib64 > LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01:cd > =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01;32 > :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01;31: > *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31:*. > Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cpio= > 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.png= > 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/ > local/openmpi-1.2.2/bin:/usr/local/bin:/bin:/usr/bin:/usr/X11R6/bin:/ > usr/kerberos/bin INPUTRC=/etc/inputrc PWD=/home/toto LANG=en_US.UTF-8 > SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass SHLVL=1 HOME=/home/ > toto LOGNAME=toto LESSOPEN=|/usr/bin/lesspipe.sh %s > G_BROKEN_FILENAMES=1 _=/usr/local/openmpi-1.2.2/bin/mpirun > OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0] > [rnd04:10736] pls:rsh: launching on node r137n001 > [rnd04:10736] pls:rsh: r137n001 is a REMOTE node > [rnd04:10736] pls:rsh: executing: (//usr/bin/rsh) /usr/bin/rsh > r137n001 PATH=/usr/local/openmpi-1.2.2/bin:$PATH ; export PATH ; > LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:$LD_LIBRARY_PATH ; > export LD_LIBRARY_PATH ; /usr/local/openmpi-1.2.2/bin/orted --debug -- > bootproxy 1 --name 0.0.2 --num_procs 3 --vpid_start 0 --nodename > r137n001 --universe toto@rnd04:default-universe-10736 --nsreplica > "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" -- > gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp:// > 10.3.254.105:33029" [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color > HISTSIZE=1000 USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:/ > usr/local/openmpi-1.2.2/lib64 > LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01:cd > =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01;32 > :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01;31: > *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31:*. > Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cpio= > 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.png= > 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/ > local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin INPUTRC=/etc/ > inputrc PWD=/home/toto LANG=en_US.UTF-8 SSH_ASKPASS=/usr/libexec/ > openssh/gnome-ssh-askpass SHLVL=1 HOME=/home/toto LOGNAME=toto > LESSOPEN=|/usr/bin/lesspipe.sh %s G_BROKEN_FILENAMES=1 _=/usr/local/ > openmpi-1.2.2/bin/mpirun OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0] > [rnd04:10737] [0,0,1] setting up session dir with > [rnd04:10737] universe default-universe-10736 > [rnd04:10737] user toto > [rnd04:10737] host rnd04 > [rnd04:10737] jobid 0 > [rnd04:10737] procid 1 > [rnd04:10737] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default- > universe-10736/0/1 > [rnd04:10737] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default- > universe-10736/0 > [rnd04:10737] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default- > universe-10736 > [rnd04:10737] top: openmpi-sessions-toto@rnd04_0 > [rnd04:10737] tmp: /tmp > [r137n001:31527] [0,0,2] setting up session dir with > [r137n001:31527] universe default-universe-10736 > [r137n001:31527] user toto > [r137n001:31527] host r137n001 > [r137n001:31527] jobid 0 > [r137n001:31527] procid 2 > [r137n001:31527] procdir: /tmp/openmpi-sessions-toto@r137n001_0/ > default-universe-10736/0/2 > [r137n001:31527] jobdir: /tmp/openmpi-sessions-toto@r137n001_0/ > default-universe-10736/0 > [r137n001:31527] unidir: /tmp/openmpi-sessions-toto@r137n001_0/ > default-universe-10736 > [r137n001:31527] top: openmpi-sessions-toto@r137n001_0 > [r137n001:31527] tmp: /tmp > Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 > x86_64 x86_64 GNU/Linux > [rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving > [r137n001:31528] *** Process received signal *** > [r137n001:31528] Signal: Segmentation fault (11) > [r137n001:31528] Signal code: Address not mapped (1) > [r137n001:31528] Failing at address: 0x1 > [r137n001:31528] [ 0] [0xffffe600] > [r137n001:31528] [ 1] /lib/tls/libc.so.6 [0xbf3bfc] > [r137n001:31528] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb) [0xbf3e3b] > [r137n001:31528] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 > (opal_show_help+0x263) [0xf7f78de3] > [r137n001:31528] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 > (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef] > [r137n001:31528] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f) > [0xf7ea041f] > [r137n001:31528] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea] > [r137n001:31528] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136) [0xf7ef65c6] > [r137n001:31528] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68] > [r137n001:31528] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 > [0xf7fd9a18] > [r137n001:31528] [10] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc] > [r137n001:31528] [11] /usr/local/openmpi-1.2.2/lib/openmpi/ > mca_oob_tcp.so [0xf7f07f61] > [r137n001:31528] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 > (opal_event_base_loop+0x388) [0xf7f67dd8] > [r137n001:31528] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 > (opal_event_loop+0x29) [0xf7f67fb9] > [r137n001:31528] [14] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 > (opal_progress+0xbe) [0xf7f6123e] > [r137n001:31528] [15] /usr/local/openmpi-1.2.2/bin/orted(main+0xd74) > [0x804a0e4] > [r137n001:31528] [16] /lib/tls/libc.so.6(__libc_start_main+0xd3) > [0xbcee23] > [r137n001:31528] [17] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1] > [r137n001:31528] *** End of error message *** > [r137n001:31527] sess_dir_finalize: proc session dir not empty - leaving > [rnd04:10736] spawn: in job_state_callback(jobid = 1, state = 0x80) > mpirun noticed that job rank 1 with PID 31528 on node r137n001 exited > on signal 11 (Segmentation fault). > [rnd04:10737] sess_dir_finalize: job session dir not empty - leaving > [r137n001:31527] sess_dir_finalize: job session dir not empty - leaving > [rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving > [rnd04:10736] sess_dir_finalize: proc session dir not empty - leaving > -bash-3.00$ [r137n001:31527] sess_dir_finalize: proc session dir not > empty - leaving > > > > > -bash-3.00$ /usr/local/openmpi-1.2.2/bin/ompi_info --all > Open MPI: 1.2.2 > Open MPI SVN revision: r14613 > Open RTE: 1.2.2 > Open RTE SVN revision: r14613 > OPAL: 1.2.2 > OPAL SVN revision: r14613 > MCA backtrace: execinfo (MCA v1.0, API v1.0, Component > v1.2.2) > MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component > v1.2.2) > MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2.2) > MCA maffinity: first_use (MCA v1.0, API v1.0, Component > v1.2.2) > MCA timer: linux (MCA v1.0, API v1.0, Component v1.2.2) > MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.2) > MCA installdirs: config (MCA v1.0, API v1.0, Component v1.2.2) > MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0) > MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0) > MCA coll: basic (MCA v1.0, API v1.0, Component v1.2.2) > MCA coll: self (MCA v1.0, API v1.0, Component v1.2.2) > MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.2) > MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2.2) > MCA io: romio (MCA v1.0, API v1.0, Component v1.2.2) > MCA mpool: rdma (MCA v1.0, API v1.0, Component v1.2.2) > MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.2) > MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.2) > MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.2) > MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.2) > MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.2) > MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2.2) > MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2.2) > MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0) > MCA topo: unity (MCA v1.0, API v1.0, Component v1.2.2) > MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2.2) > MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.2) > MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2.2) > MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2.2) > MCA gpr: null (MCA v1.0, API v1.0, Component v1.2.2) > MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.2.2) > MCA gpr: replica (MCA v1.0, API v1.0, Component > v1.2.2) > MCA iof: proxy (MCA v1.0, API v1.0, Component v1.2.2) > MCA iof: svc (MCA v1.0, API v1.0, Component v1.2.2) > MCA ns: proxy (MCA v1.0, API v2.0, Component v1.2.2) > MCA ns: replica (MCA v1.0, API v2.0, Component > v1.2.2) > MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0) > MCA ras: dash_host (MCA v1.0, API v1.3, Component > v1.2.2) > MCA ras: gridengine (MCA v1.0, API v1.3, Component > v1.2.2) > MCA ras: localhost (MCA v1.0, API v1.3, Component > v1.2.2) > MCA ras: slurm (MCA v1.0, API v1.3, Component v1.2.2) > MCA rds: hostfile (MCA v1.0, API v1.3, Component > v1.2.2) > MCA rds: proxy (MCA v1.0, API v1.3, Component v1.2.2) > MCA rds: resfile (MCA v1.0, API v1.3, Component > v1.2.2) > MCA rmaps: round_robin (MCA v1.0, API v1.3, Component > v1.2.2) > MCA rmgr: proxy (MCA v1.0, API v2.0, Component v1.2.2) > MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2.2) > MCA rml: oob (MCA v1.0, API v1.0, Component v1.2.2) > MCA pls: gridengine (MCA v1.0, API v1.3, Component > v1.2.2) > MCA pls: proxy (MCA v1.0, API v1.3, Component v1.2.2) > MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2.2) > MCA pls: slurm (MCA v1.0, API v1.3, Component v1.2.2) > MCA sds: env (MCA v1.0, API v1.0, Component v1.2.2) > MCA sds: pipe (MCA v1.0, API v1.0, Component v1.2.2) > MCA sds: seed (MCA v1.0, API v1.0, Component v1.2.2) > MCA sds: singleton (MCA v1.0, API v1.0, Component > v1.2.2) > MCA sds: slurm (MCA v1.0, API v1.0, Component v1.2.2) > Prefix: /usr/local/openmpi-1.2.2 > Bindir: /usr/local/openmpi-1.2.2/bin > Libdir: /usr/local/openmpi-1.2.2/lib > Incdir: /usr/local/openmpi-1.2.2/include > Pkglibdir: /usr/local/openmpi-1.2.2/lib/openmpi > Sysconfdir: /usr/local/openmpi-1.2.2/etc > Configured architecture: x86_64-unknown-linux-gnu > Configured by: root > Configured on: Tue Jun 5 14:32:20 CDT 2007 > Configure host: qpac171 > Built by: root > Built on: Tue Jun 5 14:39:38 CDT 2007 > Built host: qpac171 > C bindings: yes > C++ bindings: yes > Fortran77 bindings: yes (all) > Fortran90 bindings: no > Fortran90 bindings size: na > C compiler: gcc > C compiler absolute: /usr/bin/gcc > C char size: 1 > C bool size: 1 > C short size: 2 > C int size: 4 > C long size: 4 > C float size: 4 > C double size: 8 > C pointer size: 4 > C char align: 1 > C bool align: 1 > C int align: 4 > C float align: 4 > C double align: 4 > C++ compiler: g++ > C++ compiler absolute: /usr/bin/g++ > Fortran77 compiler: g77 > Fortran77 compiler abs: /usr/bin/g77 > Fortran90 compiler: none > Fortran90 compiler abs: none > Fort integer size: 4 > Fort logical size: 4 > Fort logical value true: 1 > Fort have integer1: yes > Fort have integer2: yes > Fort have integer4: yes > Fort have integer8: yes > Fort have integer16: no > Fort have real4: yes > Fort have real8: yes > Fort have real16: no > Fort have complex8: yes > Fort have complex16: yes > Fort have complex32: no > Fort integer1 size: 1 > Fort integer2 size: 2 > Fort integer4 size: 4 > Fort integer8 size: 8 > Fort integer16 size: -1 > Fort real size: 4 > Fort real4 size: 4 > Fort real8 size: 8 > Fort real16 size: -1 > Fort dbl prec size: 4 > Fort cplx size: 4 > Fort dbl cplx size: 4 > Fort cplx8 size: 8 > Fort cplx16 size: 16 > Fort cplx32 size: -1 > Fort integer align: 4 > Fort integer1 align: 1 > Fort integer2 align: 2 > Fort integer4 align: 4 > Fort integer8 align: 8 > Fort integer16 align: -1 > Fort real align: 4 > Fort real4 align: 4 > Fort real8 align: 8 > Fort real16 align: -1 > Fort dbl prec align: 4 > Fort cplx align: 4 > Fort dbl cplx align: 4 > Fort cplx8 align: 4 > Fort cplx16 align: 8 > Fort cplx32 align: -1 > C profiling: yes > C++ profiling: yes > Fortran77 profiling: yes > Fortran90 profiling: no > C++ exceptions: no > Thread support: posix (mpi: no, progress: no) > Build CFLAGS: -O3 -DNDEBUG -m32 -finline-functions -fno- > strict-aliasing -pthread > Build CXXFLAGS: -O3 -DNDEBUG -m32 -finline-functions -pthread > Build FFLAGS: -m32 > Build FCFLAGS: -m32 > Build LDFLAGS: -export-dynamic > Build LIBS: -lnsl -lutil -lm > Wrapper extra CFLAGS: -pthread -m32 > Wrapper extra CXXFLAGS: -pthread -m32 > Wrapper extra FFLAGS: -pthread -m32 > Wrapper extra FCFLAGS: -pthread -m32 > Wrapper extra LDFLAGS: > Wrapper extra LIBS: -ldl -Wl,--export-dynamic -lnsl -lutil > -lm -ldl > Internal debug support: no > MPI parameter check: runtime > Memory profiling support: no > Memory debugging support: no > libltdl support: yes > Heterogeneous support: yes > mpirun default --prefix: no > MCA mca: parameter "mca_param_files" (current > value: "/home/toto/.openmpi/mca-params.conf:/usr/local/openmpi-1.2.2/ > etc/openmpi-mca-params.conf") > Path for MCA configuration files > containing default parameter values > MCA mca: parameter "mca_component_path" (current > value: "/usr/local/openmpi-1.2.2/lib/openmpi:/home/toto/.openmpi/ > components") > Path where to look for Open MPI and ORTE > components > MCA mca: parameter "mca_verbose" (current value: > <none>) > Top-level verbosity parameter > MCA mca: parameter > "mca_component_show_load_errors" (current value: "1") > Whether to show errors for components that > failed to load or not > MCA mca: parameter > "mca_component_disable_dlopen" (current value: "0") > Whether to attempt to disable opening > dynamic components or not > MCA mpi: parameter "mpi_param_check" (current > value: "1") > Whether you want MPI API parameters > checked at run-time or not. Possible values are 0 (no checking) and > 1 (perform checking at run-time) > MCA mpi: parameter "mpi_yield_when_idle" (current > value: "0") > Yield the processor when waiting for MPI > communication (for MPI processes, will default to 1 when > oversubscribing nodes) > MCA mpi: parameter "mpi_event_tick_rate" (current > value: "-1") > How often to progress TCP communications > (0 = never, otherwise specified in microseconds) > MCA mpi: parameter "mpi_show_handle_leaks" (current > value: "0") > Whether MPI_FINALIZE shows all MPI handles > that were not freed or not > MCA mpi: parameter "mpi_no_free_handles" (current > value: "0") > Whether to actually free MPI objects when > their handles are freed > MCA mpi: parameter "mpi_show_mca_params" (current > value: "0") > Whether to show all MCA parameter value > during MPI_INIT or not (good for reproducability of MPI jobs) > MCA mpi: parameter > "mpi_show_mca_params_file" (current value: <none>) > If mpi_show_mca_params is true, setting > this string to a valid filename tells Open MPI to dump all the MCA > parameter values into a file suitable for reading via the > mca_param_files parameter (good for reproducability of MPI jobs) > MCA mpi: parameter "mpi_paffinity_alone" (current > value: "0") > If nonzero, assume that this job is the > only (set of) process(es) running on each node and bind processes to > processors, starting with processor ID 0 > MCA mpi: parameter > "mpi_keep_peer_hostnames" (current value: "1") > If nonzero, save the string hostnames of > all MPI peer processes (mostly for error / debugging output > messages). This can add quite a bit of memory usage to each MPI > process. > MCA mpi: parameter "mpi_abort_delay" (current > value: "0") > If nonzero, print out an identifying > message when MPI_ABORT is invoked (hostname, PID of the process that > called MPI_ABORT) and delay for that many seconds before exiting (a > negative delay value means to never abort). This allows > attaching of a debugger before quitting > the job. > MCA mpi: parameter "mpi_abort_print_stack" (current > value: "0") > If nonzero, print out a stack trace when > MPI_ABORT is invoked > MCA mpi: parameter "mpi_preconnect_all" (current > value: "0") > Whether to force MPI processes to create > connections / warmup with *all* peers during MPI_INIT (vs. making > connections lazily -- upon the first MPI traffic between each process > peer pair) > MCA mpi: parameter "mpi_preconnect_oob" (current > value: "0") > Whether to force MPI processes to fully > wire-up the OOB system between MPI processes. > MCA mpi: parameter "mpi_leave_pinned" (current > value: "0") > Whether to use the "leave pinned" protocol > or not. Enabling this setting can help bandwidth performance when > repeatedly sending and receiving large messages with the same buffers > over RDMA-based networks. > MCA mpi: parameter > "mpi_leave_pinned_pipeline" (current value: "0") > Whether to use the "leave pinned pipeline" > protocol or not. > MCA orte: parameter "orte_debug" (current value: "0") > Top-level ORTE debug switch > MCA orte: parameter "orte_no_daemonize" (current > value: "0") > Whether to properly daemonize the ORTE > daemons or not > MCA orte: parameter > "orte_base_user_debugger" (current value: "totalview @mpirun@ -a > @mpirun_args@ : fxp @mpirun@ -a @mpirun_args@") > Sequence of user-level debuggers to search > for in orterun > MCA orte: parameter "orte_abort_timeout" (current > value: "10") > Time to wait [in seconds] before giving up > on aborting an ORTE operation > MCA orte: parameter "orte_timing" (current value: "0") > Request that critical timing loops be > measured > MCA opal: parameter "opal_signal" (current value: > "6,7,8,11") > If a signal is received, display the stack > trace frame > MCA backtrace: parameter "backtrace" (current value: <none>) > Default selection set of components for > the backtrace framework (<none> means "use all components that can be > found") > MCA backtrace: parameter > "backtrace_base_verbose" (current value: "0") > Verbosity level for the backtrace > framework (0 = no verbosity) > MCA backtrace: parameter > "backtrace_execinfo_priority" (current value: "0") > MCA memory: parameter "memory" (current value: <none>) > Default selection set of components for > the memory framework (<none> means "use all components that can be > found") > MCA memory: parameter "memory_base_verbose" (current > value: "0") > Verbosity level for the memory framework > (0 = no verbosity) > MCA memory: parameter > "memory_ptmalloc2_priority" (current value: "0") > MCA paffinity: parameter "paffinity" (current value: <none>) > Default selection set of components for > the paffinity framework (<none> means "use all components that can be > found") > MCA paffinity: parameter > "paffinity_linux_priority" (current value: "10") > Priority of the linux paffinity component > MCA paffinity: information > "paffinity_linux_have_cpu_set_t" (value: "1") > Whether this component was compiled on a > system with the type cpu_set_t or not (1 = yes, 0 = no) > MCA paffinity: information > "paffinity_linux_CPU_ZERO_ok" (value: "1") > Whether this component was compiled on a > system where CPU_ZERO() is functional or broken (1 = functional, 0 = > broken/not available) > MCA paffinity: information > "paffinity_linux_sched_setaffinity_num_params" (value: "3") > The number of parameters that > sched_set_affinity() takes on the machine where this component was > compiled > MCA maffinity: parameter "maffinity" (current value: <none>) > Default selection set of components for > the maffinity framework (<none> means "use all components that can be > found") > MCA maffinity: parameter > "maffinity_first_use_priority" (current value: "10") > Priority of the first_use maffinity component > MCA timer: parameter "timer" (current value: <none>) > Default selection set of components for > the timer framework (<none> means "use all components that can be > found") > MCA timer: parameter "timer_base_verbose" (current > value: "0") > Verbosity level for the timer framework (0 > = no verbosity) > MCA timer: parameter "timer_linux_priority" (current > value: "0") > MCA allocator: parameter "allocator" (current value: <none>) > Default selection set of components for > the allocator framework (<none> means "use all components that can be > found") > MCA allocator: parameter > "allocator_base_verbose" (current value: "0") > Verbosity level for the allocator > framework (0 = no verbosity) > MCA allocator: parameter > "allocator_basic_priority" (current value: "0") > MCA allocator: parameter > "allocator_bucket_num_buckets" (current value: "30") > MCA allocator: parameter > "allocator_bucket_priority" (current value: "0") > MCA coll: parameter "coll" (current value: <none>) > Default selection set of components for > the coll framework (<none> means "use all components that can be found") > MCA coll: parameter "coll_base_verbose" (current > value: "0") > Verbosity level for the coll framework (0 > = no verbosity) > MCA coll: parameter "coll_basic_priority" (current > value: "10") > Priority of the basic coll component > MCA coll: parameter "coll_basic_crossover" (current > value: "4") > Minimum number of processes in a > communicator before using the logarithmic algorithms > MCA coll: parameter "coll_self_priority" (current > value: "75") > MCA coll: parameter "coll_sm_priority" (current > value: "0") > Priority of the sm coll component > MCA coll: parameter "coll_sm_control_size" (current > value: "4096") > Length of the control data -- should > usually be either the length of a cache line on most SMPs, or the > size of a page on machines that support direct memory affinity page > placement (in bytes) > MCA coll: parameter > "coll_sm_bootstrap_filename" (current value: "shared_mem_sm_bootstrap") > Filename (in the Open MPI session > directory) of the coll sm component bootstrap rendezvous mmap file > MCA coll: parameter > "coll_sm_bootstrap_num_segments" (current value: "8") > Number of segments in the bootstrap file > MCA coll: parameter "coll_sm_fragment_size" (current > value: "8192") > Fragment size (in bytes) used for passing > data through shared memory (will be rounded up to the nearest > control_size size) > MCA coll: parameter "coll_sm_mpool" (current value: > "sm") > Name of the mpool component to use > MCA coll: parameter > "coll_sm_comm_in_use_flags" (current value: "2") > Number of "in use" flags, used to mark a > message passing area segment as currently being used or not (must be >> = 2 and <= comm_num_segments) > MCA coll: parameter > "coll_sm_comm_num_segments" (current value: "8") > Number of segments in each communicator's > shared memory message passing area (must be >= 2, and must be a > multiple of comm_in_use_flags) > MCA coll: parameter "coll_sm_tree_degree" (current > value: "4") > Degree of the tree for tree-based > operations (must be => 1 and <= min(control_size, 255)) > MCA coll: information > "coll_sm_shared_mem_used_bootstrap" (value: "160") > Amount of shared memory used in the shared > memory bootstrap area (in bytes) > MCA coll: parameter > "coll_sm_info_num_procs" (current value: "4") > Number of processes to use for the > calculation of the shared_mem_size MCA information parameter (must be > => 2) > MCA coll: information > "coll_sm_shared_mem_used_data" (value: "548864") > Amount of shared memory used in the shared > memory data area for info_num_procs processes (in bytes) > MCA coll: parameter "coll_tuned_priority" (current > value: "30") > Priority of the tuned coll component > MCA coll: parameter > "coll_tuned_pre_allocate_memory_comm_size_limit" (current value: > "32768") > Size of communicator were we stop pre- > allocating memory for the fixed internal buffer used for message > requests etc that is hung off the communicator data segment. I.e. if > you have a 100'000 nodes you might not want to pre-allocate > 200'000 request handle slots per > communicator instance! > MCA coll: parameter > "coll_tuned_init_tree_fanout" (current value: "4") > Inital fanout used in the tree topologies > for each communicator. This is only an initial guess, if a tuned > collective needs a different fanout for an operation, it build it > dynamically. This parameter is only for the first guess and might > save a little time > MCA coll: parameter > "coll_tuned_init_chain_fanout" (current value: "4") > Inital fanout used in the chain (fanout > followed by pipeline) topologies for each communicator. This is only > an initial guess, if a tuned collective needs a different fanout for > an operation, it build it dynamically. This parameter is > only for the first guess and might save a > little time > MCA coll: parameter > "coll_tuned_use_dynamic_rules" (current value: "0") > Switch used to decide if we use static > (compiled/if statements) or dynamic (built at runtime) decision > function rules > MCA io: parameter > "io_base_freelist_initial_size" (current value: "16") > Initial MPI-2 IO request freelist size > MCA io: parameter > "io_base_freelist_max_size" (current value: "64") > Max size of the MPI-2 IO request freelist > MCA io: parameter > "io_base_freelist_increment" (current value: "16") > Increment size of the MPI-2 IO request > freelist > MCA io: parameter "io" (current value: <none>) > Default selection set of components for > the io framework (<none> means "use all components that can be found") > MCA io: parameter "io_base_verbose" (current > value: "0") > Verbosity level for the io framework (0 = > no verbosity) > MCA io: parameter "io_romio_priority" (current > value: "10") > Priority of the io romio component > MCA io: parameter > "io_romio_delete_priority" (current value: "10") > Delete priority of the io romio component > MCA io: parameter > "io_romio_enable_parallel_optimizations" (current value: "0") > Enable set of Open MPI-added options to > improve collective file i/o performance > MCA mpool: parameter "mpool" (current value: <none>) > Default selection set of components for > the mpool framework (<none> means "use all components that can be > found") > MCA mpool: parameter "mpool_base_verbose" (current > value: "0") > Verbosity level for the mpool framework (0 > = no verbosity) > MCA mpool: parameter > "mpool_rdma_rcache_name" (current value: "vma") > The name of the registration cache the > mpool should use > MCA mpool: parameter > "mpool_rdma_rcache_size_limit" (current value: "0") > the maximum size of registration cache in > bytes. 0 is unlimited (default 0) > MCA mpool: parameter > "mpool_rdma_print_stats" (current value: "0") > print pool usage statistics at the end of > the run > MCA mpool: parameter "mpool_rdma_priority" (current > value: "0") > MCA mpool: parameter "mpool_sm_allocator" (current > value: "bucket") > Name of allocator component to use with sm > mpool > MCA mpool: parameter "mpool_sm_max_size" (current > value: "536870912") > Maximum size of the sm mpool shared memory > file > MCA mpool: parameter "mpool_sm_min_size" (current > value: "134217728") > Minimum size of the sm mpool shared memory > file > MCA mpool: parameter > "mpool_sm_per_peer_size" (current value: "33554432") > Size (in bytes) to allocate per local peer > in the sm mpool shared memory file, bounded by min_size and max_size > MCA mpool: parameter "mpool_sm_priority" (current > value: "0") > MCA mpool: parameter > "mpool_base_use_mem_hooks" (current value: "0") > use memory hooks for deregistering freed > memory > MCA mpool: parameter "mpool_use_mem_hooks" (current > value: "0") > (deprecated, use mpool_base_use_mem_hooks) > MCA mpool: parameter > "mpool_base_disable_sbrk" (current value: "0") > use mallopt to override calling sbrk > (doesn't return memory to OS!) > MCA mpool: parameter "mpool_disable_sbrk" (current > value: "0") > (deprecated, use mca_mpool_base_disable_sbrk) > MCA pml: parameter "pml" (current value: <none>) > Default selection set of components for > the pml framework (<none> means "use all components that can be found") > MCA pml: parameter "pml_base_verbose" (current > value: "0") > Verbosity level for the pml framework (0 = > no verbosity) > MCA pml: parameter "pml_cm_free_list_num" (current > value: "4") > Initial size of request free lists > MCA pml: parameter "pml_cm_free_list_max" (current > value: "-1") > Maximum size of request free lists > MCA pml: parameter "pml_cm_free_list_inc" (current > value: "64") > Number of elements to add when growing > request free lists > MCA pml: parameter "pml_cm_priority" (current > value: "30") > CM PML selection priority > MCA pml: parameter "pml_ob1_free_list_num" (current > value: "4") > MCA pml: parameter "pml_ob1_free_list_max" (current > value: "-1") > MCA pml: parameter "pml_ob1_free_list_inc" (current > value: "64") > MCA pml: parameter "pml_ob1_priority" (current > value: "20") > MCA pml: parameter "pml_ob1_eager_limit" (current > value: "131072") > MCA pml: parameter > "pml_ob1_send_pipeline_depth" (current value: "3") > MCA pml: parameter > "pml_ob1_recv_pipeline_depth" (current value: "4") > MCA bml: parameter "bml" (current value: <none>) > Default selection set of components for > the bml framework (<none> means "use all components that can be found") > MCA bml: parameter "bml_base_verbose" (current > value: "0") > Verbosity level for the bml framework (0 = > no verbosity) > MCA bml: parameter > "bml_r2_show_unreach_errors" (current value: "1") > Show error message when procs are unreachable > MCA bml: parameter "bml_r2_priority" (current > value: "0") > MCA rcache: parameter "rcache" (current value: <none>) > Default selection set of components for > the rcache framework (<none> means "use all components that can be > found") > MCA rcache: parameter "rcache_base_verbose" (current > value: "0") > Verbosity level for the rcache framework > (0 = no verbosity) > MCA rcache: parameter "rcache_vma_priority" (current > value: "0") > MCA btl: parameter "btl_base_debug" (current value: > "0") > If btl_base_debug is 1 standard debug is > output, if > 1 verbose debug is output > MCA btl: parameter "btl" (current value: <none>) > Default selection set of components for > the btl framework (<none> means "use all components that can be found") > MCA btl: parameter "btl_base_verbose" (current > value: "0") > Verbosity level for the btl framework (0 = > no verbosity) > MCA btl: parameter > "btl_self_free_list_num" (current value: "0") > Number of fragments by default > MCA btl: parameter > "btl_self_free_list_max" (current value: "-1") > Maximum number of fragments > MCA btl: parameter > "btl_self_free_list_inc" (current value: "32") > Increment by this number of fragments > MCA btl: parameter "btl_self_eager_limit" (current > value: "131072") > Eager size fragmeng (before the rendez- > vous ptotocol) > MCA btl: parameter > "btl_self_min_send_size" (current value: "262144") > Minimum fragment size after the rendez-vous > MCA btl: parameter > "btl_self_max_send_size" (current value: "262144") > Maximum fragment size after the rendez-vous > MCA btl: parameter > "btl_self_min_rdma_size" (current value: "2147483647") > Maximum fragment size for the RDMA transfer > MCA btl: parameter > "btl_self_max_rdma_size" (current value: "2147483647") > Maximum fragment size for the RDMA transfer > MCA btl: parameter "btl_self_exclusivity" (current > value: "65536") > Device exclusivity > MCA btl: parameter "btl_self_flags" (current value: > "10") > Active behavior flags > MCA btl: parameter "btl_self_priority" (current > value: "0") > MCA btl: parameter "btl_sm_free_list_num" (current > value: "8") > MCA btl: parameter "btl_sm_free_list_max" (current > value: "-1") > MCA btl: parameter "btl_sm_free_list_inc" (current > value: "64") > MCA btl: parameter "btl_sm_exclusivity" (current > value: "65535") > MCA btl: parameter "btl_sm_latency" (current value: > "100") > MCA btl: parameter "btl_sm_max_procs" (current > value: "-1") > MCA btl: parameter "btl_sm_sm_extra_procs" (current > value: "2") > MCA btl: parameter "btl_sm_mpool" (current value: > "sm") > MCA btl: parameter "btl_sm_eager_limit" (current > value: "4096") > MCA btl: parameter "btl_sm_max_frag_size" (current > value: "32768") > MCA btl: parameter > "btl_sm_size_of_cb_queue" (current value: "128") > MCA btl: parameter > "btl_sm_cb_lazy_free_freq" (current value: "120") > MCA btl: parameter "btl_sm_priority" (current > value: "0") > MCA btl: parameter "btl_tcp_if_include" (current > value: <none>) > MCA btl: parameter "btl_tcp_if_exclude" (current > value: "lo") > MCA btl: parameter "btl_tcp_free_list_num" (current > value: "8") > MCA btl: parameter "btl_tcp_free_list_max" (current > value: "-1") > MCA btl: parameter "btl_tcp_free_list_inc" (current > value: "32") > MCA btl: parameter "btl_tcp_sndbuf" (current value: > "131072") > MCA btl: parameter "btl_tcp_rcvbuf" (current value: > "131072") > MCA btl: parameter > "btl_tcp_endpoint_cache" (current value: "30720") > MCA btl: parameter "btl_tcp_exclusivity" (current > value: "0") > MCA btl: parameter "btl_tcp_eager_limit" (current > value: "65536") > MCA btl: parameter "btl_tcp_min_send_size" (current > value: "65536") > MCA btl: parameter "btl_tcp_max_send_size" (current > value: "131072") > MCA btl: parameter "btl_tcp_min_rdma_size" (current > value: "131072") > MCA btl: parameter "btl_tcp_max_rdma_size" (current > value: "2147483647") > MCA btl: parameter "btl_tcp_flags" (current value: > "122") > MCA btl: parameter "btl_tcp_priority" (current > value: "0") > MCA btl: parameter "btl_base_include" (current > value: <none>) > MCA btl: parameter "btl_base_exclude" (current > value: <none>) > MCA btl: parameter > "btl_base_warn_component_unused" (current value: "1") > This parameter is used to turn on warning > messages when certain NICs are not used > MCA mtl: parameter "mtl" (current value: <none>) > Default selection set of components for > the mtl framework (<none> means "use all components that can be found") > MCA mtl: parameter "mtl_base_verbose" (current > value: "0") > Verbosity level for the mtl framework (0 = > no verbosity) > MCA topo: parameter "topo" (current value: <none>) > Default selection set of components for > the topo framework (<none> means "use all components that can be found") > MCA topo: parameter "topo_base_verbose" (current > value: "0") > Verbosity level for the topo framework (0 > = no verbosity) > MCA osc: parameter "osc" (current value: <none>) > Default selection set of components for > the osc framework (<none> means "use all components that can be found") > MCA osc: parameter "osc_base_verbose" (current > value: "0") > Verbosity level for the osc framework (0 = > no verbosity) > MCA osc: parameter "osc_pt2pt_no_locks" (current > value: "0") > Enable optimizations available only if > MPI_LOCK is not used. > MCA osc: parameter "osc_pt2pt_eager_limit" (current > value: "16384") > Max size of eagerly sent data > MCA osc: parameter "osc_pt2pt_priority" (current > value: "0") > MCA errmgr: parameter "errmgr" (current value: <none>) > Default selection set of components for > the errmgr framework (<none> means "use all components that can be > found") > MCA errmgr: parameter "errmgr_hnp_debug" (current > value: "0") > MCA errmgr: parameter "errmgr_hnp_priority" (current > value: "0") > MCA errmgr: parameter "errmgr_orted_debug" (current > value: "0") > MCA errmgr: parameter "errmgr_orted_priority" (current > value: "0") > MCA errmgr: parameter "errmgr_proxy_debug" (current > value: "0") > MCA errmgr: parameter "errmgr_proxy_priority" (current > value: "0") > MCA gpr: parameter "gpr_base_maxsize" (current > value: "2147483647") > MCA gpr: parameter "gpr_base_blocksize" (current > value: "512") > MCA gpr: parameter "gpr" (current value: <none>) > Default selection set of components for > the gpr framework (<none> means "use all components that can be found") > MCA gpr: parameter "gpr_null_priority" (current > value: "0") > MCA gpr: parameter "gpr_proxy_debug" (current > value: "0") > MCA gpr: parameter "gpr_proxy_priority" (current > value: "0") > MCA gpr: parameter "gpr_replica_debug" (current > value: "0") > MCA gpr: parameter "gpr_replica_isolate" (current > value: "0") > MCA gpr: parameter "gpr_replica_priority" (current > value: "0") > MCA iof: parameter "iof_base_window_size" (current > value: "4096") > MCA iof: parameter "iof_base_service" (current > value: "0.0.0") > MCA iof: parameter "iof" (current value: <none>) > Default selection set of components for > the iof framework (<none> means "use all components that can be found") > MCA iof: parameter "iof_proxy_debug" (current > value: "1") > MCA iof: parameter "iof_proxy_priority" (current > value: "0") > MCA iof: parameter "iof_svc_debug" (current value: > "1") > MCA iof: parameter "iof_svc_priority" (current > value: "0") > MCA ns: parameter "ns" (current value: <none>) > Default selection set of components for > the ns framework (<none> means "use all components that can be found") > MCA ns: parameter "ns_proxy_debug" (current value: > "0") > MCA ns: parameter "ns_proxy_maxsize" (current > value: "2147483647") > MCA ns: parameter "ns_proxy_blocksize" (current > value: "512") > MCA ns: parameter "ns_proxy_priority" (current > value: "0") > MCA ns: parameter "ns_replica_debug" (current > value: "0") > MCA ns: parameter "ns_replica_isolate" (current > value: "0") > MCA ns: parameter "ns_replica_maxsize" (current > value: "2147483647") > MCA ns: parameter "ns_replica_blocksize" (current > value: "512") > MCA ns: parameter "ns_replica_priority" (current > value: "0") > MCA oob: parameter "oob" (current value: <none>) > Default selection set of components for > the oob framework (<none> means "use all components that can be found") > MCA oob: parameter "oob_base_verbose" (current > value: "0") > Verbosity level for the oob framework (0 = > no verbosity) > MCA oob: parameter "oob_tcp_peer_limit" (current > value: "-1") > MCA oob: parameter "oob_tcp_peer_retries" (current > value: "60") > MCA oob: parameter "oob_tcp_debug" (current value: > "0") > MCA oob: parameter "oob_tcp_include" (current > value: <none>) > MCA oob: parameter "oob_tcp_exclude" (current > value: <none>) > MCA oob: parameter "oob_tcp_sndbuf" (current value: > "131072") > MCA oob: parameter "oob_tcp_rcvbuf" (current value: > "131072") > MCA oob: parameter "oob_tcp_connect_sleep" (current > value: "1") > Enable (1) /Disable (0) random sleep for > connection wireup > MCA oob: parameter "oob_tcp_listen_mode" (current > value: "event") > Mode for HNP to accept incoming > connections: event, listen_thread > MCA oob: parameter > "oob_tcp_listen_thread_max_queue" (current value: "10") > High water mark for queued accepted socket > list size > MCA oob: parameter > "oob_tcp_listen_thread_max_time" (current value: "10") > Maximum amount of time (in milliseconds) > to wait between processing accepted socket list > MCA oob: parameter > "oob_tcp_accept_spin_count" (current value: "10") > Number of times to let accept return > EWOULDBLOCK before updating accepted socket list > MCA oob: parameter "oob_tcp_priority" (current > value: "0") > MCA ras: parameter "ras" (current value: <none>) > MCA ras: parameter > "ras_dash_host_priority" (current value: "5") > Selection priority for the dash_host RAS > component > MCA ras: parameter "ras_gridengine_debug" (current > value: "0") > Enable debugging output for the gridengine > ras component > MCA ras: parameter > "ras_gridengine_priority" (current value: "100") > Priority of the gridengine ras component > MCA ras: parameter > "ras_gridengine_verbose" (current value: "0") > Enable verbose output for the gridengine > ras component > MCA ras: parameter > "ras_gridengine_show_jobid" (current value: "0") > Show the JOB_ID of the Grid Engine job > MCA ras: parameter > "ras_localhost_priority" (current value: "0") > Selection priority for the localhost RAS > component > MCA ras: parameter "ras_slurm_priority" (current > value: "75") > Priority of the slurm ras component > MCA rds: parameter "rds" (current value: <none>) > MCA rds: parameter "rds_hostfile_debug" (current > value: "0") > Toggle debug output for hostfile RDS > component > MCA rds: parameter "rds_hostfile_path" (current > value: "/usr/local/openmpi-1.2.2/etc/openmpi-default-hostfile") > ORTE Host filename > MCA rds: parameter "rds_hostfile_priority" (current > value: "0") > MCA rds: parameter "rds_proxy_priority" (current > value: "0") > MCA rds: parameter "rds_resfile_debug" (current > value: "0") > Toggle debug output for resfile RDS component > MCA rds: parameter "rds_resfile_name" (current > value: <none>) > ORTE Resource filename > MCA rds: parameter "rds_resfile_priority" (current > value: "0") > MCA rmaps: parameter "rmaps_base_verbose" (current > value: "0") > Verbosity level for the rmaps framework > MCA rmaps: parameter > "rmaps_base_schedule_policy" (current value: "unspec") > Scheduling Policy for RMAPS. [slot | node] > MCA rmaps: parameter "rmaps_base_pernode" (current > value: "0") > Launch one ppn as directed > MCA rmaps: parameter "rmaps_base_n_pernode" (current > value: "-1") > Launch n procs/node > MCA rmaps: parameter > "rmaps_base_no_schedule_local" (current value: "0") > If false, allow scheduling MPI > applications on the same node as mpirun (default). If true, do not > schedule any MPI applications on the same node as mpirun > MCA rmaps: parameter > "rmaps_base_no_oversubscribe" (current value: "0") > If true, then do not allow > oversubscription of nodes - mpirun will return an error if there > aren't enough nodes to launch all processes without oversubscribing > MCA rmaps: parameter "rmaps" (current value: <none>) > Default selection set of components for > the rmaps framework (<none> means "use all components that can be > found") > MCA rmaps: parameter > "rmaps_round_robin_debug" (current value: "1") > Toggle debug output for Round Robin RMAPS > component > MCA rmaps: parameter > "rmaps_round_robin_priority" (current value: "1") > Selection priority for Round Robin RMAPS > component > MCA rmgr: parameter "rmgr" (current value: <none>) > Default selection set of components for > the rmgr framework (<none> means "use all components that can be found") > MCA rmgr: parameter "rmgr_proxy_priority" (current > value: "0") > MCA rmgr: parameter "rmgr_urm_priority" (current > value: "0") > MCA rml: parameter "rml" (current value: <none>) > Default selection set of components for > the rml framework (<none> means "use all components that can be found") > MCA rml: parameter "rml_base_verbose" (current > value: "0") > Verbosity level for the rml framework (0 = > no verbosity) > MCA rml: parameter "rml_oob_priority" (current > value: "0") > MCA pls: parameter > "pls_base_reuse_daemons" (current value: "0") > If nonzero, reuse daemons to launch > dynamically spawned processes. If zero, do not reuse daemons (default) > MCA pls: parameter "pls" (current value: <none>) > Default selection set of components for > the pls framework (<none> means "use all components that can be found") > MCA pls: parameter "pls_base_verbose" (current > value: "0") > Verbosity level for the pls framework (0 = > no verbosity) > MCA pls: parameter "pls_gridengine_debug" (current > value: "0") > Enable debugging of gridengine pls component > MCA pls: parameter > "pls_gridengine_verbose" (current value: "0") > Enable verbose output of the gridengine > qrsh -inherit command > MCA pls: parameter > "pls_gridengine_priority" (current value: "100") > Priority of the gridengine pls component > MCA pls: parameter "pls_gridengine_orted" (current > value: "orted") > The command name that the gridengine pls > component will invoke for the ORTE daemon > MCA pls: parameter "pls_proxy_priority" (current > value: "0") > MCA pls: parameter "pls_rsh_debug" (current value: > "0") > Whether or not to enable debugging output > for the rsh pls component (0 or 1) > MCA pls: parameter > "pls_rsh_num_concurrent" (current value: "128") > How many pls_rsh_agent instances to invoke > concurrently (must be > 0) > MCA pls: parameter "pls_rsh_force_rsh" (current > value: "0") > Force the launcher to always use rsh, even > for local daemons > MCA pls: parameter "pls_rsh_orted" (current value: > "orted") > The command name that the rsh pls > component will invoke for the ORTE daemon > MCA pls: parameter "pls_rsh_priority" (current > value: "10") > Priority of the rsh pls component > MCA pls: parameter "pls_rsh_delay" (current value: > "1") > Delay (in seconds) between invocations of > the remote agent, but only used when the "debug" MCA parameter is > true, or the top-level MCA debugging is enabled (otherwise this value > is ignored) > MCA pls: parameter "pls_rsh_reap" (current value: "1") > If set to 1, wait for all the processes to > complete before exiting. Otherwise, quit immediately -- without > waiting for confirmation that all other processes in the job have > completed. > MCA pls: parameter > "pls_rsh_assume_same_shell" (current value: "1") > If set to 1, assume that the shell on the > remote node is the same as the shell on the local node. Otherwise, > probe for what the remote shell. > MCA pls: parameter "pls_rsh_agent" (current value: > "rsh") > The command used to launch executables on > remote nodes (typically either "ssh" or "rsh") > MCA pls: parameter "pls_slurm_debug" (current > value: "0") > Enable debugging of slurm pls > MCA pls: parameter "pls_slurm_priority" (current > value: "75") > Default selection priority > MCA pls: parameter "pls_slurm_orted" (current > value: "orted") > Command to use to start proxy orted > MCA pls: parameter "pls_slurm_args" (current value: > <none>) > Custom arguments to srun > MCA sds: parameter "sds" (current value: <none>) > Default selection set of components for > the sds framework (<none> means "use all components that can be found") > MCA sds: parameter "sds_base_verbose" (current > value: "0") > Verbosity level for the sds framework (0 = > no verbosity) > MCA sds: parameter "sds_env_priority" (current > value: "0") > MCA sds: parameter "sds_pipe_priority" (current > value: "0") > MCA sds: parameter "sds_seed_priority" (current > value: "0") > MCA sds: parameter > "sds_singleton_priority" (current value: "0") > MCA sds: parameter "sds_slurm_priority" (current > value: "0") > -bash-3.00$ > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users