Hmmm...well, we certainly will ke a point to give you a better error message! Probably won't get it into 1.2.3, but should make later releases.
Thanks for letting me know Ralph On 6/7/07 1:22 PM, "Guillaume THOMAS-COLLIGNON" <guillaume.thomas-collig...@cggveritas.com> wrote: > You're right, the --wdir option works fine ! > Thanks ! > > I just tried an older version we had compiled (1.2b3), and the error > was more explicit than the seg fault we get with 1.2.2 : > > Could not chdir to home directory /rdu/thomasco: No such file or > directory > ------------------------------------------------------------------------ > -- > Failed to change to the working directory: > <...> > > -Guillaume > > On Jun 7, 2007, at 12:57 PM, Ralph Castain wrote: > >> Have you tried the --wdir option yet? It should let you set your >> working >> directory to anywhere. I don't believe it will require you to have >> a home >> directory on the backend nodes, though I can't sweathat ssh will be >> happy >> if you don't. >> >> Just do "mpirun -h" for a full list of options - it will describe >> the exact >> format of the wdir one plusthers you might find useful. >> >> Ralph >> >> >> >> On 6/7/07 11:12 AM, "Guillaume THOMAS-COLLIGNON<guillaume.thomas- >> collig...@cggveritas.com> wrote: >> >>> I am trying to switch to OpenI, and I ran into a problem : my home >>> directory must exist on all the nodes, or ted will crash. >>> >>> I have a "master" machine where I initiate the mpirun command> >>> Then I have a bunch of slave machines, which will ao execute the >>> MPI job. >>> My user exists on all the machines, but the home directory is not >>> mounted on the slaves,o it's only visible on the master node. I can >>> log on a slave node, but don't have a home there. Of course the >>> binary I'm running exists on all the machines (not in my home !). And >>> the problem can be reproduced by running a shell command too, to make >>> things simpler. >>> >>> We have thousands of slave nod and we don't want to mount the >>> user's homedirs on all the slaves, so a fix would be really really >>> nice. >>> >>> Example : >>> >>> I have 3 hosts, master, slave1, slave2. My home directory exists only >>> on master. >>> >>> If I log on master and run "mpirun -host master,slave1uname -a" I get >>> a segfault. >>> If I log on slave1 and run "mpirun -host slave1,slave2 uname -a", it >>> runs fine. My home directory does not exist on either slave1 or >>> slave2. >>> If I log on master and run "mpirun -host master uname -a" it runs >>> fine. I can run across several master nodes, it's fine too. >>> >>> So it runs fine if my home directory exists everywhere, or if it does >>> not exist at all. If it exists only on some nodes and not others, >>> orted crashes. >>> I thought it could be related to my environment but I created a new >>> user with an empty home and it does the same thing. As soon as I >>> create the homedir on slave1 and slave2 it works fine. >>> >>> >>> >>> >>> I'm using OpenMPI 1.2.2, here is the error message and the result of >>> ompi_info. >>> >>> Short version (rnd04 is the master, r137n001 is a slave node). >>> >>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001 >>> uname -a >>> Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 >>> x86_64 x86_64 GNU/Linux >>> [r137n001:31533] *** Process received signal *** >>> [r137n001:31533] Signal: Segmentation fault (11) >>> [r137n001:31533] Signal code: Address not mapped (1) >>> [r137n001:533] Failing at address: 0x1 >>> [r137n001:31533] [ 0] [0xffffe600] >>> [r137n001:31533] [ 1] /lib/tls/libc.so.6 [0xbf3bfc] >>> [r137n001:31533] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb) >>> [0xbf3e3b] >>> [r137n001:31533] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 >>> (opal_show_help+0x263) [0xf7f78de3] >>> [r137n001:31533] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 >>> (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef] >>> [r137n001:31533] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f) >>> [0xf7ea041f] >>> [r137n001:31533] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea] >>> [r137n001:31533] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136) >>> [0xf7ef65c6] >>> [r137n001:31533] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68] >>> [r137n001:31533] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 >>> [0xf7fd9a18] >>> [r137n001:31533] [10] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc] >>> [r137n001:31533] [11] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_oob_tcp.so [0xf7f07f61] >>> [r137n001:31533] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 >>> (opal_event_base_loop+0x388) [0xf7f67dd8] >>> [r137n001:31533] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 >>> (opal_event_loop+0x29) [0xf7f67fb9] >>> [r137n001:31533] [14] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_oob_tcp.so(mca_oob_tcp_msg_wait+0x37) [0xf7f053c7] >>> [r137n001:31533] [15] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_oob_tcp.so(mca_oob_tcp_recv+0x374) [0xf7f09a04] >>> [r137n001:31533] [16] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 >>> (mca_oob_recv_packed+0x4d) [0xf7fd980d] >>> [r137n001:31533] [17] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_gpr_proxy.so(orte_gpr_proxy_exec_compound_cmd+0x137) [0xf7ef55e7] >>> [r137n001:31533] [18] /usr/local/openmpi-1.2.2/bin/orted(main+0x99d) >>> [0x8049d0d] >>> [r137n001:31533] [19] /lib/tls/libc.so.6(__libc_start_main+0xd3) >>> [0xbcee23] >>> [r137n001:31533] [20] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1] >>> [r137n001:31533] *** End of error message *** >>> mpirun noticed that job rank 1 with PID 31533 on node r137n001 exited >>> on signal 11 (Segmentation fault). >>> >>> >>> >>> If I create /home/toto on r137n001, it works fine : >>> (as root on r137n001: "mkdir /home/toto && chown toto:users /home/ >>> toto") >>> >>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -host rnd04,r137n001 >>> uname -a >>> Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 >>> x86_64 x86_64 GNU/Linux >>> Linux r137n001 2.6.9-34.ELsmp #1 SMP Fri Feb 24 16:56:28 EST 2006 >>> x86_64 x86_64 x86_64 GNU/Linux >>> >>> >>> I tried to use ssh instead of rsh, it crashes too. >>> >>> If anyone knows a way to run OpenMPI jobs in this configuration where >>> the home directory does not exist on all the nodes, it woud really >>> help ! >>> >>> Or is there a way to fix orted so that it won't crash ? >>> >>> >>> Here is the crash with -d option : >>> >>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/mpirun -d -host >>> rnd04,r137n001 uname -a >>> [rnd04:10736] connect_uni: connection not allowed >>> [rnd04:10736] [0,0,0] setting up session dir with >>> [rnd04:10736] universe default-universe-10736 >>> [rnd04:10736] user toto >>> [rnd04:10736] host rnd04 >>> [rnd04:10736] jobid 0 >>> [rnd04:10736] procid 0 >>> [rnd04:10736] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default- >>> universe-10736/0/0 >>> [rnd04:10736] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default- >>> universe-10736/0 >>> [rnd04:10736] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default- >>> universe-10736 >>> [rnd04:10736] top: openmpi-sessions-toto@rnd04_0 >>> [rnd04:10736] tmp: /tmp >>> [rnd04:10736] [0,0,0] contact_file /tmp/openmpi-sessions- >>> toto@rnd04_0/ >>> default-universe-10736/universe-setup.txt >>> [rnd04:10736] [0,0,0] wrote setup file >>> [rnd04:10736] pls:rsh: local csh: 0, local sh: 1 >>> [rnd04:10736] pls:rsh: assuming same remote shell as local shell >>> [rnd04:10736] pls:rsh: remote csh: 0, remote sh: 1 >>> [rnd04:10736] pls:rsh: final template argv: >>> [rnd04:10736] pls:rsh: /usr/bin/rsh <template> orted --debug -- >>> bootproxy 1 --name <template> --num_procs 3 --vpid_start 0 --nodename >>> <template> --universe toto@rnd04:default-universe-10736 --nsreplica >>> "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" -- >>> gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" >>> [rnd04:10736] pls:rsh: launching on node rnd04 >>> [rnd04:10736] pls:rsh: rnd04 is a LOCAL node >>> [rnd04:10736] pls:rsh: reset PATH: /usr/local/openmpi-1.2.2/bin:/usr/ >>> local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin >>> [rnd04:10736] pls:rsh: reset LD_LIBRARY_PATH: /usr/local/ >>> openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib:/usr/local/ >>> openmpi-1.2.2/lib64 >>> [rnd04:10736] pls:rsh: changing to directory /home/toto >>> [rnd04:10736] pls:rsh: executing: (/usr/local/openmpi-1.2.2/bin/ >>> orted) orted --debug --bootproxy 1 --name 0.0.1 --num_procs 3 -- >>> vpid_start 0 --nodename rnd04 --universe toto@rnd04:default- >>> universe-10736 --nsreplica "0.0.0;tcp://172.28.20.143:33029;tcp:// >>> 10.3.254.105:33029" --gprreplica "0.0.0;tcp:// >>> 172.28.20.143:33029;tcp://10.3.254.105:33029" --set-sid >>> [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color HISTSIZE=1000 >>> USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:/usr/local/ >>> openmpi-1.2.2/lib:/usr/local/openmpi-1.2.2/lib64 >>> LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01 >>> :cd >>> =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01 >>> ;32 >>> :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01; >>> 31: >>> *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31 >>> :*. >>> Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cp >>> io= >>> 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.p >>> ng= >>> 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/ >>> local/openmpi-1.2.2/bin:/usr/local/bin:/bin:/usr/bin:/usr/X11R6/bin:/ >>> usr/kerberos/bin INPUTRC=/etc/inputrc PWD=/home/toto LANG=en_US.UTF-8 >>> SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass SHLVL=1 HOME=/ >>> home/ >>> toto LOGNAME=toto LESSOPEN=|/usr/bin/lesspipe.sh %s >>> G_BROKEN_FILENAMES=1 _=/usr/local/openmpi-1.2.2/bin/mpirun >>> OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0] >>> [rnd04:10736] pls:rsh: launching on node r137n001 >>> [rnd04:10736] pls:rsh: r137n001 is a REMOTE node >>> [rnd04:10736] pls:rsh: executing: (//usr/bin/rsh) /usr/bin/rsh >>> r137n001 PATH=/usr/local/openmpi-1.2.2/bin:$PATH ; export PATH ; >>> LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/lib:$LD_LIBRARY_PATH ; >>> export LD_LIBRARY_PATH ; /usr/local/openmpi-1.2.2/bin/orted -- >>> debug -- >>> bootproxy 1 --name 0.0.2 --num_procs 3 --vpid_start 0 --nodename >>> r137n001 --universe toto@rnd04:default-universe-10736 --nsreplica >>> "0.0.0;tcp://172.28.20.143:33029;tcp://10.3.254.105:33029" -- >>> gprreplica "0.0.0;tcp://172.28.20.143:33029;tcp:// >>> 10.3.254.105:33029" [HOSTNAME=rnd04 SHELL=/bin/bash TERM=xterm-color >>> HISTSIZE=1000 USER=toto LD_LIBRARY_PATH=/usr/local/openmpi-1.2.2/ >>> lib:/ >>> usr/local/openmpi-1.2.2/lib64 >>> LS_COLORS=no=00:fi=00:di=01;34:ln=01;36:pi=40;33:so=01;35:bd=40;33;01 >>> :cd >>> =40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=01;32:*.cmd=01;32:*.exe=01 >>> ;32 >>> :*.com=01;32:*.btm=01;32:*.bat=01;32:*.sh=01;32:*.csh=01;32:*.tar=01; >>> 31: >>> *.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.zip=01;31:*.z=01;31 >>> :*. >>> Z=01;31:*.gz=01;31:*.bz2=01;31:*.bz=01;31:*.tz=01;31:*.rpm=01;31:*.cp >>> io= >>> 01;31:*.jpg=01;35:*.gif=01;35:*.bmp=01;35:*.xbm=01;35:*.xpm=01;35:*.p >>> ng= >>> 01;35:*.tif=01;35: KDEDIR=/usr MAIL=/var/spool/mail/toto PATH=/usr/ >>> local/bin:/bin:/usr/bin:/usr/X11R6/bin:/usr/kerberos/bin INPUTRC=/ >>> etc/ >>> inputrc PWD=/home/toto LANG=en_US.UTF-8 SSH_ASKPASS=/usr/libexec/ >>> openssh/gnome-ssh-askpass SHLVL=1 HOME=/home/toto LOGNAME=toto >>> LESSOPEN=|/usr/bin/lesspipe.sh %s G_BROKEN_FILENAMES=1 _=/usr/local/ >>> openmpi-1.2.2/bin/mpirun OMPI_MCA_orte_debug=1 OMPI_MCA_seed=0] >>> [rnd04:10737] [0,0,1] setting up session dir with >>> [rnd04:10737] universe default-universe-10736 >>> [rnd04:10737] user toto >>> [rnd04:10737] host rnd04 >>> [rnd04:10737] jobid 0 >>> [rnd04:10737] procid 1 >>> [rnd04:10737] procdir: /tmp/openmpi-sessions-toto@rnd04_0/default- >>> universe-10736/0/1 >>> [rnd04:10737] jobdir: /tmp/openmpi-sessions-toto@rnd04_0/default- >>> universe-10736/0 >>> [rnd04:10737] unidir: /tmp/openmpi-sessions-toto@rnd04_0/default- >>> universe-10736 >>> [rnd04:10737] top: openmpi-sessions-toto@rnd04_0 >>> [rnd04:10737] tmp: /tmp >>> [r137n001:31527] [0,0,2] setting up session dir with >>> [r137n001:31527] universe default-universe-10736 >>> [r137n001:31527] user toto >>> [r137n001:31527] host r137n001 >>> [r137n001:31527] jobid 0 >>> [r137n001:31527] procid 2 >>> [r137n001:31527] procdir: /tmp/openmpi-sessions-toto@r137n001_0/ >>> default-universe-10736/0/2 >>> [r137n001:31527] jobdir: /tmp/openmpi-sessions-toto@r137n001_0/ >>> default-universe-10736/0 >>> [r137n001:31527] unidir: /tmp/openmpi-sessions-toto@r137n001_0/ >>> default-universe-10736 >>> [r137n001:31527] top: openmpi-sessions-toto@r137n001_0 >>> [r137n001:31527] tmp: /tmp >>> Linux rnd04 2.6.9-55.ELsmp #1 SMP Fri Apr 20 16:36:54 EDT 2007 x86_64 >>> x86_64 x86_64 GNU/Linux >>> [rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving >>> [r137n001:31528] *** Process received signal *** >>> [r137n001:31528] Signal: Segmentation fault (11) >>> [r137n001:31528] Signal code: Address not mapped (1) >>> [r137n001:31528] Failing at address: 0x1 >>> [r137n001:31528] [ 0] [0xffffe600] >>> [r137n001:31528] [ 1] /lib/tls/libc.so.6 [0xbf3bfc] >>> [r137n001:31528] [ 2] /lib/tls/libc.so.6(_IO_vfprintf+0xcb) >>> [0xbf3e3b] >>> [r137n001:31528] [ 3] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 >>> (opal_show_help+0x263) [0xf7f78de3] >>> [r137n001:31528] [ 4] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 >>> (orte_rmgr_base_check_context_cwd+0xff) [0xf7fea7ef] >>> [r137n001:31528] [ 5] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_odls_default.so(orte_odls_default_launch_local_procs+0xe7f) >>> [0xf7ea041f] >>> [r137n001:31528] [ 6] /usr/local/openmpi-1.2.2/bin/orted [0x804a1ea] >>> [r137n001:31528] [ 7] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_gpr_proxy.so(orte_gpr_proxy_deliver_notify_msg+0x136) >>> [0xf7ef65c6] >>> [r137n001:31528] [ 8] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_gpr_proxy.so(orte_gpr_proxy_notify_recv+0x108) [0xf7ef4f68] >>> [r137n001:31528] [ 9] /usr/local/openmpi-1.2.2/lib/libopen-rte.so.0 >>> [0xf7fd9a18] >>> [r137n001:31528] [10] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_oob_tcp.so(mca_oob_tcp_msg_recv_complete+0x24c) [0xf7f05fdc] >>> [r137n001:31528] [11] /usr/local/openmpi-1.2.2/lib/openmpi/ >>> mca_oob_tcp.so [0xf7f07f61] >>> [r137n001:31528] [12] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 >>> (opal_event_base_loop+0x388) [0xf7f67dd8] >>> [r137n001:31528] [13] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 >>> (opal_event_loop+0x29) [0xf7f67fb9] >>> [r137n001:31528] [14] /usr/local/openmpi-1.2.2/lib/libopen-pal.so.0 >>> (opal_progress+0xbe) [0xf7f6123e] >>> [r137n001:31528] [15] /usr/local/openmpi-1.2.2/bin/orted(main+0xd74) >>> [0x804a0e4] >>> [r137n001:31528] [16] /lib/tls/libc.so.6(__libc_start_main+0xd3) >>> [0xbcee23] >>> [r137n001:31528] [17] /usr/local/openmpi-1.2.2/bin/orted [0x80492e1] >>> [r137n001:31528] *** End of error message *** >>> [r137n001:31527] sess_dir_finalize: proc session dir not empty - >>> leaving >>> [rnd04:10736] spawn: in job_state_callback(jobid = 1, state = 0x80) >>> mpirun noticed that job rank 1 with PID 31528 on node r137n001 exited >>> on signal 11 (Segmentation fault). >>> [rnd04:10737] sess_dir_finalize: job session dir not empty - leaving >>> [r137n001:31527] sess_dir_finalize: job session dir not empty - >>> leaving >>> [rnd04:10737] sess_dir_finalize: proc session dir not empty - leaving >>> [rnd04:10736] sess_dir_finalize: proc session dir not empty - leaving >>> -bash-3.00$ [r137n001:31527] sess_dir_finalize: proc session dir not >>> empty - leaving >>> >>> >>> >>> >>> -bash-3.00$ /usr/local/openmpi-1.2.2/bin/ompi_info --all >>> Open MPI: 1.2.2 >>> Open MPI SVN revision: r14613 >>> Open RTE: 1.2.2 >>> Open RTE SVN revision: r14613 >>> OPAL: 1.2.2 >>> OPAL SVN revision: r14613 >>> MCA backtrace: execinfo (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA paffinity: linux (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA maffinity: first_use (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA timer: linux (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA installdirs: config (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0) >>> MCA allocator: bucket (MCA v1.0, API v1.0, Component >>> v1.0) >>> MCA coll: basic (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA coll: self (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA coll: tuned (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA io: romio (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA mpool: rdma (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA btl: self (MCA v1.0, API v1.0.1, Component >>> v1.2.2) >>> MCA btl: sm (MCA v1.0, API v1.0.1, Component >>> v1.2.2) >>> MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0) >>> MCA topo: unity (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA osc: pt2pt (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.2) >>> MCA errmgr: orted (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA errmgr: proxy (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA gpr: null (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA gpr: proxy (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA gpr: replica (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA iof: proxy (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA iof: svc (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA ns: proxy (MCA v1.0, API v2.0, Component >>> v1.2.2) >>> MCA ns: replica (MCA v1.0, API v2.0, Component >>> v1.2.2) >>> MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0) >>> MCA ras: dash_host (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA ras: gridengine (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA ras: localhost (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA ras: slurm (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA rds: hostfile (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA rds: proxy (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA rds: resfile (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA rmaps: round_robin (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA rmgr: proxy (MCA v1.0, API v2.0, Component >>> v1.2.2) >>> MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2.2) >>> MCA rml: oob (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA pls: gridengine (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA pls: proxy (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2.2) >>> MCA pls: slurm (MCA v1.0, API v1.3, Component >>> v1.2.2) >>> MCA sds: env (MCA v1.0, API v1.0, Component v1.2.2) >>> MCA sds: pipe (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA sds: seed (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA sds: singleton (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> MCA sds: slurm (MCA v1.0, API v1.0, Component >>> v1.2.2) >>> Prefix: /usr/local/openmpi-1.2.2 >>> Bindir: /usr/local/openmpi-1.2.2/bin >>> Libdir: /usr/local/openmpi-1.2.2/lib >>> Incdir: /usr/local/openmpi-1.2.2/include >>> Pkglibdir: /usr/local/openmpi-1.2.2/lib/openmpi >>> Sysconfdir: /usr/local/openmpi-1.2.2/etc >>> Configured architecture: x86_64-unknown-linux-gnu >>> Configured by: root >>> Configured on: Tue Jun 5 14:32:20 CDT 2007 >>> Configure host: qpac171 >>> Built by: root >>> Built on: Tue Jun 5 14:39:38 CDT 2007 >>> Built host: qpac171 >>> C bindings: yes >>> C++ bindings: yes >>> Fortran77 bindings: yes (all) >>> Fortran90 bindings: no >>> Fortran90 bindings size: na >>> C compiler: gcc >>> C compiler absolute: /usr/bin/gcc >>> C char size: 1 >>> C bool size: 1 >>> C short size: 2 >>> C int size: 4 >>> C long size: 4 >>> C float size: 4 >>> C double size: 8 >>> C pointer size: 4 >>> C char align: 1 >>> C bool align: 1 >>> C int align: 4 >>> C float align: 4 >>> C double align: 4 >>> C++ compiler: g++ >>> C++ compiler absolute: /usr/bin/g++ >>> Fortran77 compiler: g77 >>> Fortran77 compiler abs: /usr/bin/g77 >>> Fortran90 compiler: none >>> Fortran90 compiler abs: none >>> Fort integer size: 4 >>> Fort logical size: 4 >>> Fort logical value true: 1 >>> Fort have integer1: yes >>> Fort have integer2: yes >>> Fort have integer4: yes >>> Fort have integer8: yes >>> Fort have integer16: no >>> Fort have real4: yes >>> Fort have real8: yes >>> Fort have real16: no >>> Fort have complex8: yes >>> Fort have complex16: yes >>> Fort have complex32: no >>> Fort integer1 size: 1 >>> Fort integer2 size: 2 >>> Fort integer4 size: 4 >>> Fort integer8 size: 8 >>> Fort integer16 size: -1 >>> Fort real size: 4 >>> Fort real4 size: 4 >>> Fort real8 size: 8 >>> Fort real16 size: -1 >>> Fort dbl prec size: 4 >>> Fort cplx size: 4 >>> Fort dbl cplx size: 4 >>> Fort cplx8 size: 8 >>> Fort cplx16 size: 16 >>> Fort cplx32 size: -1 >>> Fort integer align: 4 >>> Fort integer1 align: 1 >>> Fort integer2 align: 2 >>> Fort integer4 align: 4 >>> Fort integer8 align: 8 >>> Fort integer16 align: -1 >>> Fort real align: 4 >>> Fort real4 align: 4 >>> Fort real8 align: 8 >>> Fort real16 align: -1 >>> Fort dbl prec align: 4 >>> Fort cplx align: 4 >>> Fort dbl cplx align: 4 >>> Fort cplx8 align: 4 >>> Fort cplx16 align: 8 >>> Fort cplx32 align: -1 >>> C profiling: yes >>> C++ profiling: yes >>> Fortran77 profiling: yes >>> Fortran90 profiling: no >>> C++ exceptions: no >>> Thread support: posix (mpi: no, progress: no) >>> Build CFLAGS: -O3 -DNDEBUG -m32 -finline-functions -fno- >>> strict-aliasing -pthread >>> Build CXXFLAGS: -O3 -DNDEBUG -m32 -finline-functions - >>> pthread >>> Build FFLAGS: -m32 >>> Build FCFLAGS: -m32 >>> Build LDFLAGS: -export-dynamic >>> Build LIBS: -lnsl -lutil -lm >>> Wrapper extra CFLAGS: -pthread -m32 >>> Wrapper extra CXXFLAGS: -pthread -m32 >>> Wrapper extra FFLAGS: -pthread -m32 >>> Wrapper extra FCFLAGS: -pthread -m32 >>> Wrapper extra LDFLAGS: >>> Wrapper extra LIBS: -ldl -Wl,--export-dynamic -lnsl -lutil >>> -lm -ldl >>> Internal debug support: no >>> MPI parameter check: runtime >>> Memory profiling support: no >>> Memory debugging support: no >>> libltdl support: yes >>> Heterogeneous support: yes >>> mpirun default --prefix: no >>> MCA mca: parameter "mca_param_files" (current >>> value: "/home/toto/.openmpi/mca-params.conf:/usr/local/openmpi-1.2.2/ >>> etc/openmpi-mca-params.conf") >>> Path for MCA configuration files >>> containing default parameter values >>> MCA mca: parameter "mca_component_path" (current >>> value: "/usr/local/openmpi-1.2.2/lib/openmpi:/home/toto/.openmpi/ >>> components") >>> Path where to look for Open MPI and ORTE >>> components >>> MCA mca: parameter "mca_verbose" (current value: >>> <none>) >>> Top-level verbosity parameter >>> MCA mca: parameter >>> "mca_component_show_load_errors" (current value: "1") >>> Whether to show errors for components that >>> failed to load or not >>> MCA mca: parameter >>> "mca_component_disable_dlopen" (current value: "0") >>> Whether to attempt to disable opening >>> dynamic components or not >>> MCA mpi: parameter "mpi_param_check" (current >>> value: "1") >>> Whether you want MPI API parameters >>> checked at run-time or not. Possible values are 0 (no checking) and >>> 1 (perform checking at run-time) >>> MCA mpi: parameter "mpi_yield_when_idle" (current >>> value: "0") >>> Yield the processor when waiting for MPI >>> communication (for MPI processes, will default to 1 when >>> oversubscribing nodes) >>> MCA mpi: parameter "mpi_event_tick_rate" (current >>> value: "-1") >>> How often to progress TCP communications >>> (0 = never, otherwise specified in microseconds) >>> MCA mpi: parameter "mpi_show_handle_leaks" (current >>> value: "0") >>> Whether MPI_FINALIZE shows all MPI handles >>> that were not freed or not >>> MCA mpi: parameter "mpi_no_free_handles" (current >>> value: "0") >>> Whether to actually free MPI objects when >>> their handles are freed >>> MCA mpi: parameter "mpi_show_mca_params" (current >>> value: "0") >>> Whether to show all MCA parameter value >>> during MPI_INIT or not (good for reproducability of MPI jobs) >>> MCA mpi: parameter >>> "mpi_show_mca_params_file" (current value: <none>) >>> If mpi_show_mca_params is true, setting >>> this string to a valid filename tells Open MPI to dump all the MCA >>> parameter values into a file suitable for reading via the >>> mca_param_files parameter (good for reproducability of MPI jobs) >>> MCA mpi: parameter "mpi_paffinity_alone" (current >>> value: "0") >>> If nonzero, assume that this job is the >>> only (set of) process(es) running on each node and bind processes to >>> processors, starting with processor ID 0 >>> MCA mpi: parameter >>> "mpi_keep_peer_hostnames" (current value: "1") >>> If nonzero, save the string hostnames of >>> all MPI peer processes (mostly for error / debugging output >>> messages). This can add quite a bit of memory usage to each MPI >>> process. >>> MCA mpi: parameter "mpi_abort_delay" (current >>> value: "0") >>> If nonzero, print out an identifying >>> message when MPI_ABORT is invoked (hostname, PID of the process that >>> called MPI_ABORT) and delay for that many seconds before exiting (a >>> negative delay value means to never abort). This allows >>> attaching of a debugger before quitting >>> the job. >>> MCA mpi: parameter "mpi_abort_print_stack" (current >>> value: "0") >>> If nonzero, print out a stack trace when >>> MPI_ABORT is invoked >>> MCA mpi: parameter "mpi_preconnect_all" (current >>> value: "0") >>> Whether to force MPI processes to create >>> connections / warmup with *all* peers during MPI_INIT (vs. making >>> connections lazily -- upon the first MPI traffic between each process >>> peer pair) >>> MCA mpi: parameter "mpi_preconnect_oob" (current >>> value: "0") >>> Whether to force MPI processes to fully >>> wire-up the OOB system between MPI processes. >>> MCA mpi: parameter "mpi_leave_pinned" (current >>> value: "0") >>> Whether to use the "leave pinned" protocol >>> or not. Enabling this setting can help bandwidth performance when >>> repeatedly sending and receiving large messages with the same buffers >>> over RDMA-based networks. >>> MCA mpi: parameter >>> "mpi_leave_pinned_pipeline" (current value: "0") >>> Whether to use the "leave pinned pipeline" >>> protocol or not. >>> MCA orte: parameter "orte_debug" (current value: >>> "0") >>> Top-level ORTE debug switch >>> MCA orte: parameter "orte_no_daemonize" (current >>> value: "0") >>> Whether to properly daemonize the ORTE >>> daemons or not >>> MCA orte: parameter >>> "orte_base_user_debugger" (current value: "totalview @mpirun@ -a >>> @mpirun_args@ : fxp @mpirun@ -a @mpirun_args@") >>> Sequence of user-level debuggers to search >>> for in orterun >>> MCA orte: parameter "orte_abort_timeout" (current >>> value: "10") >>> Time to wait [in seconds] before giving up >>> on aborting an ORTE operation >>> MCA orte: parameter "orte_timing" (current value: >>> "0") >>> Request that critical timing loops be >>> measured >>> MCA opal: parameter "opal_signal" (current value: >>> "6,7,8,11") >>> If a signal is received, display the stack >>> trace frame >>> MCA backtrace: parameter "backtrace" (current value: >>> <none>) >>> Default selection set of components for >>> the backtrace framework (<none> means "use all components that can be >>> found") >>> MCA backtrace: parameter >>> "backtrace_base_verbose" (current value: "0") >>> Verbosity level for the backtrace >>> framework (0 = no verbosity) >>> MCA backtrace: parameter >>> "backtrace_execinfo_priority" (current value: "0") >>> MCA memory: parameter "memory" (current value: <none>) >>> Default selection set of components for >>> the memory framework (<none> means "use all components that can be >>> found") >>> MCA memory: parameter "memory_base_verbose" (current >>> value: "0") >>> Verbosity level for the memory framework >>> (0 = no verbosity) >>> MCA memory: parameter >>> "memory_ptmalloc2_priority" (current value: "0") >>> MCA paffinity: parameter "paffinity" (current value: >>> <none>) >>> Default selection set of components for >>> the paffinity framework (<none> means "use all components that can be >>> found") >>> MCA paffinity: parameter >>> "paffinity_linux_priority" (current value: "10") >>> Priority of the linux paffinity component >>> MCA paffinity: information >>> "paffinity_linux_have_cpu_set_t" (value: "1") >>> Whether this component was compiled on a >>> system with the type cpu_set_t or not (1 = yes, 0 = no) >>> MCA paffinity: information >>> "paffinity_linux_CPU_ZERO_ok" (value: "1") >>> Whether this component was compiled on a >>> system where CPU_ZERO() is functional or broken (1 = functional, 0 = >>> broken/not available) >>> MCA paffinity: information >>> "paffinity_linux_sched_setaffinity_num_params" (value: "3") >>> The number of parameters that >>> sched_set_affinity() takes on the machine where this component was >>> compiled >>> MCA maffinity: parameter "maffinity" (current value: >>> <none>) >>> Default selection set of components for >>> the maffinity framework (<none> means "use all components that can be >>> found") >>> MCA maffinity: parameter >>> "maffinity_first_use_priority" (current value: "10") >>> Priority of the first_use maffinity >>> component >>> MCA timer: parameter "timer" (current value: <none>) >>> Default selection set of components for >>> the timer framework (<none> means "use all components that can be >>> found") >>> MCA timer: parameter "timer_base_verbose" (current >>> value: "0") >>> Verbosity level for the timer framework (0 >>> = no verbosity) >>> MCA timer: parameter "timer_linux_priority" (current >>> value: "0") >>> MCA allocator: parameter "allocator" (current value: >>> <none>) >>> Default selection set of components for >>> the allocator framework (<none> means "use all components that can be >>> found") >>> MCA allocator: parameter >>> "allocator_base_verbose" (current value: "0") >>> Verbosity level for the allocator >>> framework (0 = no verbosity) >>> MCA allocator: parameter >>> "allocator_basic_priority" (current value: "0") >>> MCA allocator: parameter >>> "allocator_bucket_num_buckets" (current value: "30") >>> MCA allocator: parameter >>> "allocator_bucket_priority" (current value: "0") >>> MCA coll: parameter "coll" (current value: <none>) >>> Default selection set of components for >>> the coll framework (<none> means "use all components that can be >>> found") >>> MCA coll: parameter "coll_base_verbose" (current >>> value: "0") >>> Verbosity level for the coll framework (0 >>> = no verbosity) >>> MCA coll: parameter "coll_basic_priority" (current >>> value: "10") >>> Priority of the basic coll component >>> MCA coll: parameter "coll_basic_crossover" (current >>> value: "4") >>> Minimum number of processes in a >>> communicator before using the logarithmic algorithms >>> MCA coll: parameter "coll_self_priority" (current >>> value: "75") >>> MCA coll: parameter "coll_sm_priority" (current >>> value: "0") >>> Priority of the sm coll component >>> MCA coll: parameter "coll_sm_control_size" (current >>> value: "4096") >>> Length of the control data -- should >>> usually be either the length of a cache line on most SMPs, or the >>> size of a page on machines that support direct memory affinity page >>> placement (in bytes) >>> MCA coll: parameter >>> "coll_sm_bootstrap_filename" (current value: >>> "shared_mem_sm_bootstrap") >>> Filename (in the Open MPI session >>> directory) of the coll sm component bootstrap rendezvous mmap file >>> MCA coll: parameter >>> "coll_sm_bootstrap_num_segments" (current value: "8") >>> Number of segments in the bootstrap file >>> MCA coll: parameter "coll_sm_fragment_size" (current >>> value: "8192") >>> Fragment size (in bytes) used for passing >>> data through shared memory (will be rounded up to the nearest >>> control_size size) >>> MCA coll: parameter "coll_sm_mpool" (current value: >>> "sm") >>> Name of the mpool component to use >>> MCA coll: parameter >>> "coll_sm_comm_in_use_flags" (current value: "2") >>> Number of "in use" flags, used to mark a >>> message passing area segment as currently being used or not (must be >>>> = 2 and <= comm_num_segments) >>> MCA coll: parameter >>> "coll_sm_comm_num_segments" (current value: "8") >>> Number of segments in each communicator's >>> shared memory message passing area (must be >= 2, and must be a >>> multiple of comm_in_use_flags) >>> MCA coll: parameter "coll_sm_tree_degree" (current >>> value: "4") >>> Degree of the tree for tree-based >>> operations (must be => 1 and <= min(control_size, 255)) >>> MCA coll: information >>> "coll_sm_shared_mem_used_bootstrap" (value: "160") >>> Amount of shared memory used in the shared >>> memory bootstrap area (in bytes) >>> MCA coll: parameter >>> "coll_sm_info_num_procs" (current value: "4") >>> Number of processes to use for the >>> calculation of the shared_mem_size MCA information parameter (must be >>> => 2) >>> MCA coll: information >>> "coll_sm_shared_mem_used_data" (value: "548864") >>> Amount of shared memory used in the shared >>> memory data area for info_num_procs processes (in bytes) >>> MCA coll: parameter "coll_tuned_priority" (current >>> value: "30") >>> Priority of the tuned coll component >>> MCA coll: parameter >>> "coll_tuned_pre_allocate_memory_comm_size_limit" (current value: >>> "32768") >>> Size of communicator were we stop pre- >>> allocating memory for the fixed internal buffer used for message >>> requests etc that is hung off the communicator data segment. I.e. if >>> you have a 100'000 nodes you might not want to pre-allocate >>> 200'000 request handle slots per >>> communicator instance! >>> MCA coll: parameter >>> "coll_tuned_init_tree_fanout" (current value: "4") >>> Inital fanout used in the tree topologies >>> for each communicator. This is only an initial guess, if a tuned >>> collective needs a different fanout for an operation, it build it >>> dynamically. This parameter is only for the first guess and might >>> save a little time >>> MCA coll: parameter >>> "coll_tuned_init_chain_fanout" (current value: "4") >>> Inital fanout used in the chain (fanout >>> followed by pipeline) topologies for each communicator. This is only >>> an initial guess, if a tuned collective needs a different fanout for >>> an operation, it build it dynamically. This parameter is >>> only for the first guess and might save a >>> little time >>> MCA coll: parameter >>> "coll_tuned_use_dynamic_rules" (current value: "0") >>> Switch used to decide if we use static >>> (compiled/if statements) or dynamic (built at runtime) decision >>> function rules >>> MCA io: parameter >>> "io_base_freelist_initial_size" (current value: "16") >>> Initial MPI-2 IO request freelist size >>> MCA io: parameter >>> "io_base_freelist_max_size" (current value: "64") >>> Max size of the MPI-2 IO request freelist >>> MCA io: parameter >>> "io_base_freelist_increment" (current value: "16") >>> Increment size of the MPI-2 IO request >>> freelist >>> MCA io: parameter "io" (current value: <none>) >>> Default selection set of components for >>> the io framework (<none> means "use all components that can be >>> found") >>> MCA io: parameter "io_base_verbose" (current >>> value: "0") >>> Verbosity level for the io framework (0 = >>> no verbosity) >>> MCA io: parameter "io_romio_priority" (current >>> value: "10") >>> Priority of the io romio component >>> MCA io: parameter >>> "io_romio_delete_priority" (current value: "10") >>> Delete priority of the io romio component >>> MCA io: parameter >>> "io_romio_enable_parallel_optimizations" (current value: "0") >>> Enable set of Open MPI-added options to >>> improve collective file i/o performance >>> MCA mpool: parameter "mpool" (current value: <none>) >>> Default selection set of components for >>> the mpool framework (<none> means "use all components that can be >>> found") >>> MCA mpool: parameter "mpool_base_verbose" (current >>> value: "0") >>> Verbosity level for the mpool framework (0 >>> = no verbosity) >>> MCA mpool: parameter >>> "mpool_rdma_rcache_name" (current value: "vma") >>> The name of the registration cache the >>> mpool should use >>> MCA mpool: parameter >>> "mpool_rdma_rcache_size_limit" (current value: "0") >>> the maximum size of registration cache in >>> bytes. 0 is unlimited (default 0) >>> MCA mpool: parameter >>> "mpool_rdma_print_stats" (current value: "0") >>> print pool usage statistics at the end of >>> the run >>> MCA mpool: parameter "mpool_rdma_priority" (current >>> value: "0") >>> MCA mpool: parameter "mpool_sm_allocator" (current >>> value: "bucket") >>> Name of allocator component to use with sm >>> mpool >>> MCA mpool: parameter "mpool_sm_max_size" (current >>> value: "536870912") >>> Maximum size of the sm mpool shared memory >>> file >>> MCA mpool: parameter "mpool_sm_min_size" (current >>> value: "134217728") >>> Minimum size of the sm mpool shared memory >>> file >>> MCA mpool: parameter >>> "mpool_sm_per_peer_size" (current value: "33554432") >>> Size (in bytes) to allocate per local peer >>> in the sm mpool shared memory file, bounded by min_size and max_size >>> MCA mpool: parameter "mpool_sm_priority" (current >>> value: "0") >>> MCA mpool: parameter >>> "mpool_base_use_mem_hooks" (current value: "0") >>> use memory hooks for deregistering freed >>> memory >>> MCA mpool: parameter "mpool_use_mem_hooks" (current >>> value: "0") >>> (deprecated, use mpool_base_use_mem_hooks) >>> MCA mpool: parameter >>> "mpool_base_disable_sbrk" (current value: "0") >>> use mallopt to override calling sbrk >>> (doesn't return memory to OS!) >>> MCA mpool: parameter "mpool_disable_sbrk" (current >>> value: "0") >>> (deprecated, use >>> mca_mpool_base_disable_sbrk) >>> MCA pml: parameter "pml" (current value: <none>) >>> Default selection set of components for >>> the pml framework (<none> means "use all components that can be >>> found") >>> MCA pml: parameter "pml_base_verbose" (current >>> value: "0") >>> Verbosity level for the pml framework (0 = >>> no verbosity) >>> MCA pml: parameter "pml_cm_free_list_num" (current >>> value: "4") >>> Initial size of request free lists >>> MCA pml: parameter "pml_cm_free_list_max" (current >>> value: "-1") >>> Maximum size of request free lists >>> MCA pml: parameter "pml_cm_free_list_inc" (current >>> value: "64") >>> Number of elements to add when growing >>> request free lists >>> MCA pml: parameter "pml_cm_priority" (current >>> value: "30") >>> CM PML selection priority >>> MCA pml: parameter "pml_ob1_free_list_num" (current >>> value: "4") >>> MCA pml: parameter "pml_ob1_free_list_max" (current >>> value: "-1") >>> MCA pml: parameter "pml_ob1_free_list_inc" (current >>> value: "64") >>> MCA pml: parameter "pml_ob1_priority" (current >>> value: "20") >>> MCA pml: parameter "pml_ob1_eager_limit" (current >>> value: "131072") >>> MCA pml: parameter >>> "pml_ob1_send_pipeline_depth" (current value: "3") >>> MCA pml: parameter >>> "pml_ob1_recv_pipeline_depth" (current value: "4") >>> MCA bml: parameter "bml" (current value: <none>) >>> Default selection set of components for >>> the bml framework (<none> means "use all components that can be >>> found") >>> MCA bml: parameter "bml_base_verbose" (current >>> value: "0") >>> Verbosity level for the bml framework (0 = >>> no verbosity) >>> MCA bml: parameter >>> "bml_r2_show_unreach_errors" (current value: "1") >>> Show error message when procs are >>> unreachable >>> MCA bml: parameter "bml_r2_priority" (current >>> value: "0") >>> MCA rcache: parameter "rcache" (current value: <none>) >>> Default selection set of components for >>> the rcache framework (<none> means "use all components that can be >>> found") >>> MCA rcache: parameter "rcache_base_verbose" (current >>> value: "0") >>> Verbosity level for the rcache framework >>> (0 = no verbosity) >>> MCA rcache: parameter "rcache_vma_priority" (current >>> value: "0") >>> MCA btl: parameter "btl_base_debug" (current value: >>> "0") >>> If btl_base_debug is 1 standard debug is >>> output, if > 1 verbose debug is output >>> MCA btl: parameter "btl" (current value: <none>) >>> Default selection set of components for >>> the btl framework (<none> means "use all components that can be >>> found") >>> MCA btl: parameter "btl_base_verbose" (current >>> value: "0") >>> Verbosity level for the btl framework (0 = >>> no verbosity) >>> MCA btl: parameter >>> "btl_self_free_list_num" (current value: "0") >>> Number of fragments by default >>> MCA btl: parameter >>> "btl_self_free_list_max" (current value: "-1") >>> Maximum number of fragments >>> MCA btl: parameter >>> "btl_self_free_list_inc" (current value: "32") >>> Increment by this number of fragments >>> MCA btl: parameter "btl_self_eager_limit" (current >>> value: "131072") >>> Eager size fragmeng (before the rendez- >>> vous ptotocol) >>> MCA btl: parameter >>> "btl_self_min_send_size" (current value: "262144") >>> Minimum fragment size after the rendez- >>> vous >>> MCA btl: parameter >>> "btl_self_max_send_size" (current value: "262144") >>> Maximum fragment size after the rendez- >>> vous >>> MCA btl: parameter >>> "btl_self_min_rdma_size" (current value: "2147483647") >>> Maximum fragment size for the RDMA >>> transfer >>> MCA btl: parameter >>> "btl_self_max_rdma_size" (current value: "2147483647") >>> Maximum fragment size for the RDMA >>> transfer >>> MCA btl: parameter "btl_self_exclusivity" (current >>> value: "65536") >>> Device exclusivity >>> MCA btl: parameter "btl_self_flags" (current value: >>> "10") >>> Active behavior flags >>> MCA btl: parameter "btl_self_priority" (current >>> value: "0") >>> MCA btl: parameter "btl_sm_free_list_num" (current >>> value: "8") >>> MCA btl: parameter "btl_sm_free_list_max" (current >>> value: "-1") >>> MCA btl: parameter "btl_sm_free_list_inc" (current >>> value: "64") >>> MCA btl: parameter "btl_sm_exclusivity" (current >>> value: "65535") >>> MCA btl: parameter "btl_sm_latency" (current value: >>> "100") >>> MCA btl: parameter "btl_sm_max_procs" (current >>> value: "-1") >>> MCA btl: parameter "btl_sm_sm_extra_procs" (current >>> value: "2") >>> MCA btl: parameter "btl_sm_mpool" (current value: >>> "sm") >>> MCA btl: parameter "btl_sm_eager_limit" (current >>> value: "4096") >>> MCA btl: parameter "btl_sm_max_frag_size" (current >>> value: "32768") >>> MCA btl: parameter >>> "btl_sm_size_of_cb_queue" (current value: "128") >>> MCA btl: parameter >>> "btl_sm_cb_lazy_free_freq" (current value: "120") >>> MCA btl: parameter "btl_sm_priority" (current >>> value: "0") >>> MCA btl: parameter "btl_tcp_if_include" (current >>> value: <none>) >>> MCA btl: parameter "btl_tcp_if_exclude" (current >>> value: "lo") >>> MCA btl: parameter "btl_tcp_free_list_num" (current >>> value: "8") >>> MCA btl: parameter "btl_tcp_free_list_max" (current >>> value: "-1") >>> MCA btl: parameter "btl_tcp_free_list_inc" (current >>> value: "32") >>> MCA btl: parameter "btl_tcp_sndbuf" (current value: >>> "131072") >>> MCA btl: parameter "btl_tcp_rcvbuf" (current value: >>> "131072") >>> MCA btl: parameter >>> "btl_tcp_endpoint_cache" (current value: "30720") >>> MCA btl: parameter "btl_tcp_exclusivity" (current >>> value: "0") >>> MCA btl: parameter "btl_tcp_eager_limit" (current >>> value: "65536") >>> MCA btl: parameter "btl_tcp_min_send_size" (current >>> value: "65536") >>> MCA btl: parameter "btl_tcp_max_send_size" (current >>> value: "131072") >>> MCA btl: parameter "btl_tcp_min_rdma_size" (current >>> value: "131072") >>> MCA btl: parameter "btl_tcp_max_rdma_size" (current >>> value: "2147483647") >>> MCA btl: parameter "btl_tcp_flags" (current value: >>> "122") >>> MCA btl: parameter "btl_tcp_priority" (current >>> value: "0") >>> MCA btl: parameter "btl_base_include" (current >>> value: <none>) >>> MCA btl: parameter "btl_base_exclude" (current >>> value: <none>) >>> MCA btl: parameter >>> "btl_base_warn_component_unused" (current value: "1") >>> This parameter is used to turn on warning >>> messages when certain NICs are not used >>> MCA mtl: parameter "mtl" (current value: <none>) >>> Default selection set of components for >>> the mtl framework (<none> means "use all components that can be >>> found") >>> MCA mtl: parameter "mtl_base_verbose" (current >>> value: "0") >>> Verbosity level for the mtl framework (0 = >>> no verbosity) >>> MCA topo: parameter "topo" (current value: <none>) >>> Default selection set of components for >>> the topo framework (<none> means "use all components that can be >>> found") >>> MCA topo: parameter "topo_base_verbose" (current >>> value: "0") >>> Verbosity level for the topo framework (0 >>> = no verbosity) >>> MCA osc: parameter "osc" (current value: <none>) >>> Default selection set of components for >>> the osc framework (<none> means "use all components that can be >>> found") >>> MCA osc: parameter "osc_base_verbose" (current >>> value: "0") >>> Verbosity level for the osc framework (0 = >>> no verbosity) >>> MCA osc: parameter "osc_pt2pt_no_locks" (current >>> value: "0") >>> Enable optimizations available only if >>> MPI_LOCK is not used. >>> MCA osc: parameter "osc_pt2pt_eager_limit" (current >>> value: "16384") >>> Max size of eagerly sent data >>> MCA osc: parameter "osc_pt2pt_priority" (current >>> value: "0") >>> MCA errmgr: parameter "errmgr" (current value: <none>) >>> Default selection set of components for >>> the errmgr framework (<none> means "use all components that can be >>> found") >>> MCA errmgr: parameter "errmgr_hnp_debug" (current >>> value: "0") >>> MCA errmgr: parameter "errmgr_hnp_priority" (current >>> value: "0") >>> MCA errmgr: parameter "errmgr_orted_debug" (current >>> value: "0") >>> MCA errmgr: parameter "errmgr_orted_priority" (current >>> value: "0") >>> MCA errmgr: parameter "errmgr_proxy_debug" (current >>> value: "0") >>> MCA errmgr: parameter "errmgr_proxy_priority" (current >>> value: "0") >>> MCA gpr: parameter "gpr_base_maxsize" (current >>> value: "2147483647") >>> MCA gpr: parameter "gpr_base_blocksize" (current >>> value: "512") >>> MCA gpr: parameter "gpr" (current value: <none>) >>> Default selection set of components for >>> the gpr framework (<none> means "use all components that can be >>> found") >>> MCA gpr: parameter "gpr_null_priority" (current >>> value: "0") >>> MCA gpr: parameter "gpr_proxy_debug" (current >>> value: "0") >>> MCA gpr: parameter "gpr_proxy_priority" (current >>> value: "0") >>> MCA gpr: parameter "gpr_replica_debug" (current >>> value: "0") >>> MCA gpr: parameter "gpr_replica_isolate" (current >>> value: "0") >>> MCA gpr: parameter "gpr_replica_priority" (current >>> value: "0") >>> MCA iof: parameter "iof_base_window_size" (current >>> value: "4096") >>> MCA iof: parameter "iof_base_service" (current >>> value: "0.0.0") >>> MCA iof: parameter "iof" (current value: <none>) >>> Default selection set of components for >>> the iof framework (<none> means "use all components that can be >>> found") >>> MCA iof: parameter "iof_proxy_debug" (current >>> value: "1") >>> MCA iof: parameter "iof_proxy_priority" (current >>> value: "0") >>> MCA iof: parameter "iof_svc_debug" (current value: >>> "1") >>> MCA iof: parameter "iof_svc_priority" (current >>> value: "0") >>> MCA ns: parameter "ns" (current value: <none>) >>> Default selection set of components for >>> the ns framework (<none> means "use all components that can be >>> found") >>> MCA ns: parameter "ns_proxy_debug" (current value: >>> "0") >>> MCA ns: parameter "ns_proxy_maxsize" (current >>> value: "2147483647") >>> MCA ns: parameter "ns_proxy_blocksize" (current >>> value: "512") >>> MCA ns: parameter "ns_proxy_priority" (current >>> value: "0") >>> MCA ns: parameter "ns_replica_debug" (current >>> value: "0") >>> MCA ns: parameter "ns_replica_isolate" (current >>> value: "0") >>> MCA ns: parameter "ns_replica_maxsize" (current >>> value: "2147483647") >>> MCA ns: parameter "ns_replica_blocksize" (current >>> value: "512") >>> MCA ns: parameter "ns_replica_priority" (current >>> value: "0") >>> MCA oob: parameter "oob" (current value: <none>) >>> Default selection set of components for >>> the oob framework (<none> means "use all components that can be >>> found") >>> MCA oob: parameter "oob_base_verbose" (current >>> value: "0") >>> Verbosity level for the oob framework (0 = >>> no verbosity) >>> MCA oob: parameter "oob_tcp_peer_limit" (current >>> value: "-1") >>> MCA oob: parameter "oob_tcp_peer_retries" (current >>> value: "60") >>> MCA oob: parameter "oob_tcp_debug" (current value: >>> "0") >>> MCA oob: parameter "oob_tcp_include" (current >>> value: <none>) >>> MCA oob: parameter "oob_tcp_exclude" (current >>> value: <none>) >>> MCA oob: parameter "oob_tcp_sndbuf" (current value: >>> "131072") >>> MCA oob: parameter "oob_tcp_rcvbuf" (current value: >>> "131072") >>> MCA oob: parameter "oob_tcp_connect_sleep" (current >>> value: "1") >>> Enable (1) /Disable (0) random sleep for >>> connection wireup >>> MCA oob: parameter "oob_tcp_listen_mode" (current >>> value: "event") >>> Mode for HNP to accept incoming >>> connections: event, listen_thread >>> MCA oob: parameter >>> "oob_tcp_listen_thread_max_queue" (current value: "10") >>> High water mark for queued accepted socket >>> list size >>> MCA oob: parameter >>> "oob_tcp_listen_thread_max_time" (current value: "10") >>> Maximum amount of time (in milliseconds) >>> to wait between processing accepted socket list >>> MCA oob: parameter >>> "oob_tcp_accept_spin_count" (current value: "10") >>> Number of times to let accept return >>> EWOULDBLOCK before updating accepted socket list >>> MCA oob: parameter "oob_tcp_priority" (current >>> value: "0") >>> MCA ras: parameter "ras" (current value: <none>) >>> MCA ras: parameter >>> "ras_dash_host_priority" (current value: "5") >>> Selection priority for the dash_host RAS >>> component >>> MCA ras: parameter "ras_gridengine_debug" (current >>> value: "0") >>> Enable debugging output for the gridengine >>> ras component >>> MCA ras: parameter >>> "ras_gridengine_priority" (current value: "100") >>> Priority of the gridengine ras component >>> MCA ras: parameter >>> "ras_gridengine_verbose" (current value: "0") >>> Enable verbose output for the gridengine >>> ras component >>> MCA ras: parameter >>> "ras_gridengine_show_jobid" (current value: "0") >>> Show the JOB_ID of the Grid Engine job >>> MCA ras: parameter >>> "ras_localhost_priority" (current value: "0") >>> Selection priority for the localhost RAS >>> component >>> MCA ras: parameter "ras_slurm_priority" (current >>> value: "75") >>> Priority of the slurm ras component >>> MCA rds: parameter "rds" (current value: <none>) >>> MCA rds: parameter "rds_hostfile_debug" (current >>> value: "0") >>> Toggle debug output for hostfile RDS >>> component >>> MCA rds: parameter "rds_hostfile_path" (current >>> value: "/usr/local/openmpi-1.2.2/etc/openmpi-default-hostfile") >>> ORTE Host filename >>> MCA rds: parameter "rds_hostfile_priority" (current >>> value: "0") >>> MCA rds: parameter "rds_proxy_priority" (current >>> value: "0") >>> MCA rds: parameter "rds_resfile_debug" (current >>> value: "0") >>> Toggle debug output for resfile RDS >>> component >>> MCA rds: parameter "rds_resfile_name" (current >>> value: <none>) >>> ORTE Resource filename >>> MCA rds: parameter "rds_resfile_priority" (current >>> value: "0") >>> MCA rmaps: parameter "rmaps_base_verbose" (current >>> value: "0") >>> Verbosity level for the rmaps framework >>> MCA rmaps: parameter >>> "rmaps_base_schedule_policy" (current value: "unspec") >>> Scheduling Policy for RMAPS. [slot | node] >>> MCA rmaps: parameter "rmaps_base_pernode" (current >>> value: "0") >>> Launch one ppn as directed >>> MCA rmaps: parameter "rmaps_base_n_pernode" (current >>> value: "-1") >>> Launch n procs/node >>> MCA rmaps: parameter >>> "rmaps_base_no_schedule_local" (current value: "0") >>> If false, allow scheduling MPI >>> applications on the same node as mpirun (default). If true, do not >>> schedule any MPI applications on the same node as mpirun >>> MCA rmaps: parameter >>> "rmaps_base_no_oversubscribe" (current value: "0") >>> If true, then do not allow >>> oversubscription of nodes - mpirun will return an error if there >>> aren't enough nodes to launch all processes without oversubscribing >>> MCA rmaps: parameter "rmaps" (current value: <none>) >>> Default selection set of components for >>> the rmaps framework (<none> means "use all components that can be >>> found") >>> MCA rmaps: parameter >>> "rmaps_round_robin_debug" (current value: "1") >>> Toggle debug output for Round Robin RMAPS >>> component >>> MCA rmaps: parameter >>> "rmaps_round_robin_priority" (current value: "1") >>> Selection priority for Round Robin RMAPS >>> component >>> MCA rmgr: parameter "rmgr" (current value: <none>) >>> Default selection set of components for >>> the rmgr framework (<none> means "use all components that can be >>> found") >>> MCA rmgr: parameter "rmgr_proxy_priority" (current >>> value: "0") >>> MCA rmgr: parameter "rmgr_urm_priority" (current >>> value: "0") >>> MCA rml: parameter "rml" (current value: <none>) >>> Default selection set of components for >>> the rml framework (<none> means "use all components that can be >>> found") >>> MCA rml: parameter "rml_base_verbose" (current >>> value: "0") >>> Verbosity level for the rml framework (0 = >>> no verbosity) >>> MCA rml: parameter "rml_oob_priority" (current >>> value: "0") >>> MCA pls: parameter >>> "pls_base_reuse_daemons" (current value: "0") >>> If nonzero, reuse daemons to launch >>> dynamically spawned processes. If zero, do not reuse daemons >>> (default) >>> MCA pls: parameter "pls" (current value: <none>) >>> Default selection set of components for >>> the pls framework (<none> means "use all components that can be >>> found") >>> MCA pls: parameter "pls_base_verbose" (current >>> value: "0") >>> Verbosity level for the pls framework (0 = >>> no verbosity) >>> MCA pls: parameter "pls_gridengine_debug" (current >>> value: "0") >>> Enable debugging of gridengine pls >>> component >>> MCA pls: parameter >>> "pls_gridengine_verbose" (current value: "0") >>> Enable verbose output of the gridengine >>> qrsh -inherit command >>> MCA pls: parameter >>> "pls_gridengine_priority" (current value: "100") >>> Priority of the gridengine pls component >>> MCA pls: parameter "pls_gridengine_orted" (current >>> value: "orted") >>> The command name that the gridengine pls >>> component will invoke for the ORTE daemon >>> MCA pls: parameter "pls_proxy_priority" (current >>> value: "0") >>> MCA pls: parameter "pls_rsh_debug" (current value: >>> "0") >>> Whether or not to enable debugging output >>> for the rsh pls component (0 or 1) >>> MCA pls: parameter >>> "pls_rsh_num_concurrent" (current value: "128") >>> How many pls_rsh_agent instances to invoke >>> concurrently (must be > 0) >>> MCA pls: parameter "pls_rsh_force_rsh" (current >>> value: "0") >>> Force the launcher to always use rsh, even >>> for local daemons >>> MCA pls: parameter "pls_rsh_orted" (current value: >>> "orted") >>> The command name that the rsh pls >>> component will invoke for the ORTE daemon >>> MCA pls: parameter "pls_rsh_priority" (current >>> value: "10") >>> Priority of the rsh pls component >>> MCA pls: parameter "pls_rsh_delay" (current value: >>> "1") >>> Delay (in seconds) between invocations of >>> the remote agent, but only used when the "debug" MCA parameter is >>> true, or the top-level MCA debugging is enabled (otherwise this value >>> is ignored) >>> MCA pls: parameter "pls_rsh_reap" (current >>> value: "1") >>> If set to 1, wait for all the processes to >>> complete before exiting. Otherwise, quit immediately -- without >>> waiting for confirmation that all other processes in the job have >>> completed. >>> MCA pls: parameter >>> "pls_rsh_assume_same_shell" (current value: "1") >>> If set to 1, assume that the shell on the >>> remote node is the same as the shell on the local node. Otherwise, >>> probe for what the remote shell. >>> MCA pls: parameter "pls_rsh_agent" (current value: >>> "rsh") >>> The command used to launch executables on >>> remote nodes (typically either "ssh" or "rsh") >>> MCA pls: parameter "pls_slurm_debug" (current >>> value: "0") >>> Enable debugging of slurm pls >>> MCA pls: parameter "pls_slurm_priority" (current >>> value: "75") >>> Default selection priority >>> MCA pls: parameter "pls_slurm_orted" (current >>> value: "orted") >>> Command to use to start proxy orted >>> MCA pls: parameter "pls_slurm_args" (current value: >>> <none>) >>> Custom arguments to srun >>> MCA sds: parameter "sds" (current value: <none>) >>> Default selection set of components for >>> the sds framework (<none> means "use all components that can be >>> found") >>> MCA sds: parameter "sds_base_verbose" (current >>> value: "0") >>> Verbosity level for the sds framework (0 = >>> no verbosity) >>> MCA sds: parameter "sds_env_priority" (current >>> value: "0") >>> MCA sds: parameter "sds_pipe_priority" (current >>> value: "0") >>> MCA sds: parameter "sds_seed_priority" (current >>> value: "0") >>> MCA sds: parameter >>> "sds_singleton_priority" (current value: "0") >>> MCA sds: parameter "sds_slurm_priority" (current >>> value: "0") >>> -bash-3.00$ >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users