It looks to me like the network is losing connections - your error messages all state "no route to host", which implies that the network interface dropped out.
On Sep 3, 2012, at 1:39 PM, Andrea Negri <negri.an...@gmail.com> wrote: > I have asked to my admin and he said that no log messages were present > in /var/log, apart my login on the compute node. > No killed processes, no full stack errors, the memory is ok, 1GB is > used and 2GB are free. > Actually I don't know what kind of problem coud be, does someone have > ideas? Or at least a suspect? > > Please, don't let me alone! > > Sorry for the trouble with the mail > > 2012/9/1 <users-requ...@open-mpi.org>: >> Send users mailing list submissions to >> us...@open-mpi.org >> >> To subscribe or unsubscribe via the World Wide Web, visit >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> or, via email, send a message with subject or body 'help' to >> users-requ...@open-mpi.org >> >> You can reach the person managing the list at >> users-ow...@open-mpi.org >> >> When replying, please edit your Subject line so it is more specific >> than "Re: Contents of users digest..." >> >> >> Today's Topics: >> >> 1. Re: some mpi processes "disappear" on a cluster of servers >> (John Hearns) >> 2. Re: users Digest, Vol 2339, Issue 5 (Andrea Negri) >> >> >> ---------------------------------------------------------------------- >> >> Message: 1 >> Date: Sat, 1 Sep 2012 08:48:56 +0100 >> From: John Hearns <hear...@googlemail.com> >> Subject: Re: [OMPI users] some mpi processes "disappear" on a cluster >> of servers >> To: Open MPI Users <us...@open-mpi.org> >> Message-ID: >> <capqne2wo3bgefjiyfums6yquoduhjj75zjoeeyjrdq60fmz...@mail.gmail.com> >> Content-Type: text/plain; charset=ISO-8859-1 >> >> Apologies, I have not taken the time to read your comprehensive diagnostics! >> >> As Gus says, this sounds like a memory problem. >> My suspicion would be the kernel Out Of Memory (OOM) killer. >> Log into those nodes (or ask your systems manager to do this). Look >> closely at /var/log/messages where there will be notifications when >> the OOM Killer kicks in and - well - kills large memory processes! >> Grep for "killed process" in /var/log/messages >> >> http://linux-mm.org/OOM_Killer >> >> >> ------------------------------ >> >> Message: 2 >> Date: Sat, 1 Sep 2012 11:50:59 +0200 >> From: Andrea Negri <negri.an...@gmail.com> >> Subject: Re: [OMPI users] users Digest, Vol 2339, Issue 5 >> To: us...@open-mpi.org >> Message-ID: >> <CAPUxaiQ4RFqSK1kz7fM7U9XRxjQh8N+=98paqm2ikvr7bv-...@mail.gmail.com> >> Content-Type: text/plain; charset=ISO-8859-1 >> >> Hi, Gus and John, >> >> my code (zeusmp2) is a F77 code ported in F95, and the very first time >> I have launched it, the processed disappears almost immediately. >> I checked the code with valgrind and some unallocated arrays were >> passed wrongly to some subroutines. >> After having corrected this bug, the code runs for a while and after >> occour all the stuff described in my first post. >> However, the code still performs a lot of main temporal cycle before >> "die" (I don't know if thies piece of information is useful). >> >> Now I'm going to check the memory usage, (I also have a lot of unused >> variables in this pretty large code, maybe I shoud comment them). >> >> uname -a returns >> Linux cloud 2.6.9-42.0.3.ELsmp #1 SMP Thu Oct 5 16:29:37 CDT 2006 >> x86_64 x86_64 x86_64 GNU/Linux >> >> ulimit -a returns: >> core file size (blocks, -c) 0 >> data seg size (kbytes, -d) unlimited >> file size (blocks, -f) unlimited >> pending signals (-i) 1024 >> max locked memory (kbytes, -l) 32 >> max memory size (kbytes, -m) unlimited >> open files (-n) 1024 >> pipe size (512 bytes, -p) 8 >> POSIX message queues (bytes, -q) 819200 >> stack size (kbytes, -s) 10240 >> cpu time (seconds, -t) unlimited >> max user processes (-u) 36864 >> virtual memory (kbytes, -v) unlimited >> file locks (-x) unlimited >> >> >> I can log on the logins nodes, but unfortunately the command ls >> /var/log/messages return >> acpid cron.4 messages.3 secure.4 >> anaconda.log cups messages.4 spooler >> anaconda.syslog dmesg mpi_uninstall.log spooler.1 >> anaconda.xlog gdm ppp spooler.2 >> audit httpd prelink.log spooler.3 >> boot.log itac_uninstall.log rpmpkgs spooler.4 >> boot.log.1 lastlog rpmpkgs.1 vbox >> boot.log.2 mail rpmpkgs.2 wtmp >> boot.log.3 maillog rpmpkgs.3 wtmp.1 >> boot.log.4 maillog.1 rpmpkgs.4 Xorg.0.log >> cmkl_install.log maillog.2 samba Xorg.0.log.old >> cmkl_uninstall.log maillog.3 scrollkeeper.log yum.log >> cron maillog.4 secure yum.log.1 >> cron.1 messages secure.1 >> cron.2 messages.1 secure.2 >> cron.3 messages.2 secure.3 >> >> so, the log should be in some of these files (I don't have read >> permission on these files). I'll contact the admin for that. >> >> 2012/9/1 <users-requ...@open-mpi.org>: >>> Send users mailing list submissions to >>> us...@open-mpi.org >>> >>> To subscribe or unsubscribe via the World Wide Web, visit >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> or, via email, send a message with subject or body 'help' to >>> users-requ...@open-mpi.org >>> >>> You can reach the person managing the list at >>> users-ow...@open-mpi.org >>> >>> When replying, please edit your Subject line so it is more specific >>> than "Re: Contents of users digest..." >>> >>> >>> Today's Topics: >>> >>> 1. Re: some mpi processes "disappear" on a cluster of servers >>> (Gus Correa) >>> >>> >>> ---------------------------------------------------------------------- >>> >>> Message: 1 >>> Date: Fri, 31 Aug 2012 20:11:41 -0400 >>> From: Gus Correa <g...@ldeo.columbia.edu> >>> Subject: Re: [OMPI users] some mpi processes "disappear" on a cluster >>> of servers >>> To: Open MPI Users <us...@open-mpi.org> >>> Message-ID: <504152bd.3000...@ldeo.columbia.edu> >>> Content-Type: text/plain; charset=ISO-8859-1; format=flowed >>> >>> Hi Andrea >>> >>> I would guess this is a memory problem. >>> Do you know how much memory each node has? >>> Do you know the memory that >>> each MPI process in the CFD code requires? >>> If the program starts swapping/paging into disk, because of >>> low memory, those interesting things that you described can happen. >>> >>> I would login to the compute nodes and monitor the >>> amount of memory being used with "top" right after the program >>> starts to run. If there is a pattern of which node tends to fail, >>> track login to that fail-prone node and monitor it. >>> >>> Alternatively, if you cluster is running Ganglia, >>> you can see the memory use graphically, >>> in the Ganglia web page in a web browser. >>> >>> If your cluster >>> doesn't allow direct user logins to compute nodes, >>> you can ask the system administrator to do this for you. >>> >>> It may well be that the code has a memory leak, or that >>> it has a memory request spike, which may not be caught by >>> OpenMPI. >>> [Jeff and Ralph will probably correct me soon for >>> saying this, and I know the OpenMPI safeguards against >>> process misbehavior are great, but ...] >>> >>> Anyway, we had several codes here that would make a node go south >>> because of either type of memory problem, and subsequently the >>> program would die, or the other processes in other nodes would >>> continue "running" [i.e. mostly waiting for MPI calls to the >>> dead node that would never return] as you described. >>> >>> If the problem is benign, i.e., if it is just that the >>> memory-per-processor is not large enough to run in 10 processors, >>> you can get around it by running in, say, 20 processors. >>> >>> Yet another issue that you may check is the stacksize in the >>> compute nodes. Many codes require a large stacksize, i.e., >>> they create large arrays in subroutines, etc, and >>> the default stacksize in standard Linux distributions >>> may not be as large as needed. >>> We use ulimited stacksize in our compute nodes. >>> >>> You can ask the system administrator to check this for you, >>> and perhaps change it in /etc/security/limits.conf to make it >>> unlimited or at least larger than the default. >>> The Linux shell command "ulimit -a" [bash] or >>> "limit" [tcsh] will tell what the limits are. >>> >>> I hope this helps, >>> Gus Correa >>> >>> On 08/31/2012 07:15 PM, Andrea Negri wrote: >>>> Hi, I have been in trouble for a year. >>>> >>>> I run a pure MPI (no openMP) Fortran fluid dynamical code on a cluster >>>> of server, and I obtain a strange behaviour by running the code on >>>> multiple nodes. >>>> The cluster is formed by 16 pc (1 pc is a node) with a dual core processor. >>>> Basically, I'm able to run the code from the login node with the command: >>>> mpirun --mca btl_base_verbose 100 --mca backtrace_base_verbose 100 >>>> --mca memory_base_verbose 100 --mca sysinfo_base_verbose 100 -nolocal >>>> -hostfile ./host_file -n 10 ./zeusmp2.x>> zmp_errors 2>&1 >>>> by selecting one process per core (i.e. in this case I use 5 nodes) >>>> >>>> The code starts, and it runs correctely for some time. >>>> After that, an entire node (sometimes two) "disappears" and it cannot >>>> be reached with the ssh command, which returns: No route to host. >>>> Sometimes the node is still reachable, but the two processes that was >>>> running on the node are disappears. >>>> In addition, on the other nodes, the others processes are still running. >>>> >>>> If I have a look on the output and error file of mpirun, the following >>>> error is present: [btl_tcp_frag.c:215:mca_btl_tcp_frag_recv] >>>> mca_btl_tcp_frag_recv: readv failed: No route to host (113) >>>> >>>> PS: I'm not the admin of the cluster, I've installed the gcc and >>>> openmpi on my own because the complier aviable on that machine are 8 >>>> years old. >>>> >>>> >>>> I post here some information, if you want other info, you have only to >>>> tell me which command I have to type on the bash and I will >>>> immediately reply. >>>> >>>> >>>> complier: gcc 4.7 (which was also used to compile openmpi) >>>> openmpi version: 1.6 >>>> >>>> output of "ompi_info --all" from the node where I launch mpirun (which >>>> is also the login node of the cluster) >>>> >>>> Package: Open MPI and...@cloud.bo.astro.it Distribution >>>> Open MPI: 1.6 >>>> Open MPI SVN revision: r26429 >>>> Open MPI release date: May 10, 2012 >>>> Open RTE: 1.6 >>>> Open RTE SVN revision: r26429 >>>> Open RTE release date: May 10, 2012 >>>> OPAL: 1.6 >>>> OPAL SVN revision: r26429 >>>> OPAL release date: May 10, 2012 >>>> MPI API: 2.1 >>>> Ident string: 1.6 >>>> MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6) >>>> MCA memory: linux (MCA v2.0, API v2.0, Component v1.6) >>>> MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6) >>>> MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6) >>>> MCA carto: file (MCA v2.0, API v2.0, Component v1.6) >>>> MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6) >>>> MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6) >>>> MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6) >>>> MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6) >>>> MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6) >>>> MCA timer: linux (MCA v2.0, API v2.0, Component v1.6) >>>> MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6) >>>> MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6) >>>> MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6) >>>> MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6) >>>> MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6) >>>> MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6) >>>> MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6) >>>> MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6) >>>> MCA coll: basic (MCA v2.0, API v2.0, Component v1.6) >>>> MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6) >>>> MCA coll: inter (MCA v2.0, API v2.0, Component v1.6) >>>> MCA coll: self (MCA v2.0, API v2.0, Component v1.6) >>>> MCA coll: sm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA coll: sync (MCA v2.0, API v2.0, Component v1.6) >>>> MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6) >>>> MCA io: romio (MCA v2.0, API v2.0, Component v1.6) >>>> MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6) >>>> MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6) >>>> MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6) >>>> MCA pml: csum (MCA v2.0, API v2.0, Component v1.6) >>>> MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6) >>>> MCA pml: v (MCA v2.0, API v2.0, Component v1.6) >>>> MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6) >>>> MCA btl: self (MCA v2.0, API v2.0, Component v1.6) >>>> MCA btl: sm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6) >>>> MCA topo: unity (MCA v2.0, API v2.0, Component v1.6) >>>> MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6) >>>> MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6) >>>> MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6) >>>> MCA iof: orted (MCA v2.0, API v2.0, Component v1.6) >>>> MCA iof: tool (MCA v2.0, API v2.0, Component v1.6) >>>> MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6) >>>> MCA odls: default (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ras: cm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6) >>>> MCA rml: oob (MCA v2.0, API v2.0, Component v1.6) >>>> MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6) >>>> MCA routed: cm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA routed: direct (MCA v2.0, API v2.0, Component v1.6) >>>> MCA routed: linear (MCA v2.0, API v2.0, Component v1.6) >>>> MCA routed: radix (MCA v2.0, API v2.0, Component v1.6) >>>> MCA routed: slave (MCA v2.0, API v2.0, Component v1.6) >>>> MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6) >>>> MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6) >>>> MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ess: env (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ess: slave (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6) >>>> MCA ess: tool (MCA v2.0, API v2.0, Component v1.6) >>>> MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6) >>>> MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6) >>>> MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6) >>>> MCA notifier: command (MCA v2.0, API v1.0, Component v1.6) >>>> MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6) >>>> Prefix: /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7 >>>> Exec_prefix: /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7 >>>> Bindir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/bin >>>> Sbindir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/sbin >>>> Libdir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/lib >>>> Incdir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/include >>>> Mandir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/share/man >>>> Pkglibdir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/lib/openmpi >>>> Libexecdir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/libexec >>>> Datarootdir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/share >>>> Datadir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/share >>>> Sysconfdir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/etc >>>> Sharedstatedir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/com >>>> Localstatedir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/var >>>> Infodir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/share/info >>>> Pkgdatadir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/share/openmpi >>>> Pkglibdir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/lib/openmpi >>>> Pkgincludedir: >>>> /home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/include/openmpi >>>> Configured architecture: x86_64-unknown-linux-gnu >>>> Configure host: cloud.bo.astro.it >>>> Configured by: andrea >>>> Configured on: Tue Jul 31 10:53:46 CEST 2012 >>>> Configure host: cloud.bo.astro.it >>>> Built by: andrea >>>> Built on: Tue Jul 31 11:08:33 CEST 2012 >>>> Built host: cloud.bo.astro.it >>>> C bindings: yes >>>> C++ bindings: yes >>>> Fortran77 bindings: yes (all) >>>> Fortran90 bindings: yes >>>> Fortran90 bindings size: medium >>>> C compiler: /home/andrea/library/gcc/gcc-objects/bin/gcc >>>> C compiler absolute: >>>> C compiler family name: GNU >>>> C compiler version: 4.7.1 >>>> C char size: 1 >>>> C bool size: 1 >>>> C short size: 2 >>>> C int size: 4 >>>> C long size: 8 >>>> C float size: 4 >>>> C double size: 8 >>>> C pointer size: 8 >>>> C char align: 1 >>>> C bool align: 1 >>>> C int align: 4 >>>> C float align: 4 >>>> C double align: 8 >>>> C++ compiler: /home/andrea/library/gcc/gcc-objects/bin/g++ >>>> C++ compiler absolute: none >>>> Fortran77 compiler: /home/andrea/library/gcc/gcc-objects/bin/gfortran >>>> Fortran77 compiler abs: >>>> Fortran90 compiler: /home/andrea/library/gcc/gcc-objects/bin/gfortran >>>> Fortran90 compiler abs: >>>> Fort integer size: 4 >>>> Fort logical size: 4 >>>> Fort logical value true: 1 >>>> Fort have integer1: yes >>>> Fort have integer2: yes >>>> Fort have integer4: yes >>>> Fort have integer8: yes >>>> Fort have integer16: no >>>> Fort have real4: yes >>>> Fort have real8: yes >>>> Fort have real16: no >>>> Fort have complex8: yes >>>> Fort have complex16: yes >>>> Fort have complex32: no >>>> Fort integer1 size: 1 >>>> Fort integer2 size: 2 >>>> Fort integer4 size: 4 >>>> Fort integer8 size: 8 >>>> Fort integer16 size: -1 >>>> Fort real size: 4 >>>> Fort real4 size: 4 >>>> Fort real8 size: 8 >>>> Fort real16 size: 16 >>>> Fort dbl prec size: 8 >>>> Fort cplx size: 8 >>>> Fort dbl cplx size: 16 >>>> Fort cplx8 size: 8 >>>> Fort cplx16 size: 16 >>>> Fort cplx32 size: 32 >>>> Fort integer align: 4 >>>> Fort integer1 align: 1 >>>> Fort integer2 align: 2 >>>> Fort integer4 align: 4 >>>> Fort integer8 align: 8 >>>> Fort integer16 align: -1 >>>> Fort real align: 4 >>>> Fort real4 align: 4 >>>> Fort real8 align: 8 >>>> Fort real16 align: 16 >>>> Fort dbl prec align: 8 >>>> Fort cplx align: 4 >>>> Fort dbl cplx align: 8 >>>> Fort cplx8 align: 4 >>>> Fort cplx16 align: 8 >>>> Fort cplx32 align: 16 >>>> C profiling: yes >>>> C++ profiling: yes >>>> Fortran77 profiling: yes >>>> Fortran90 profiling: yes >>>> C++ exceptions: no >>>> Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no) >>>> Sparse Groups: no >>>> Build CFLAGS: -DNDEBUG -g -O2 -finline-functions >>>> -fno-strict-aliasing >>>> -pthread >>>> Build CXXFLAGS: -O3 -DNDEBUG -finline-functions -pthread >>>> Build FFLAGS: >>>> Build FCFLAGS: >>>> Build LDFLAGS: -Wl,--rpath >>>> -Wl,/home/andrea/library/gcc/gcc-objects/lib64 >>>> Build LIBS: -lrt -lnsl -lutil -lm >>>> Wrapper extra CFLAGS: -pthread >>>> Wrapper extra CXXFLAGS: -pthread >>>> Wrapper extra FFLAGS: -pthread >>>> Wrapper extra FCFLAGS: -pthread >>>> Wrapper extra LDFLAGS: >>>> Wrapper extra LIBS: -ldl -lm -lnuma -lrt -lnsl -lutil -lm >>>> Internal debug support: no >>>> MPI interface warnings: yes >>>> MPI parameter check: runtime >>>> Memory profiling support: no >>>> Memory debugging support: no >>>> libltdl support: no >>>> Heterogeneous support: no >>>> mpirun default --prefix: yes >>>> MPI I/O support: yes >>>> MPI_WTIME support: gettimeofday >>>> Symbol vis. support: yes >>>> Host topology support: yes >>>> MPI extensions: affinity example >>>> FT Checkpoint support: no (checkpoint thread: no) >>>> VampirTrace support: yes >>>> MPI_MAX_PROCESSOR_NAME: 256 >>>> MPI_MAX_ERROR_STRING: 256 >>>> MPI_MAX_OBJECT_NAME: 64 >>>> MPI_MAX_INFO_KEY: 36 >>>> MPI_MAX_INFO_VAL: 256 >>>> MPI_MAX_PORT_NAME: 1024 >>>> MPI_MAX_DATAREP_STRING: 128 >>>> MCA mca: parameter "mca_param_files" (current value: >>>> >>>> </home/andrea/.openmpi/mca-params.conf:/home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/etc/openmpi-mca-params.conf>, >>>> data source: default value) >>>> Path for MCA configuration files containing >>>> default parameter >>>> values >>>> MCA mca: parameter "mca_base_param_file_prefix" >>>> (current value:<none>, >>>> data source: default value) >>>> Aggregate MCA parameter file sets >>>> MCA mca: parameter "mca_base_param_file_path" (current >>>> value: >>>> >>>> </home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/share/openmpi/amca-param-sets:/home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/bin>, >>>> data source: default value) >>>> Aggregate MCA parameter Search path >>>> MCA mca: parameter "mca_base_param_file_path_force" >>>> (current value: >>>> <none>, data source: default value) >>>> Forced Aggregate MCA parameter Search path >>>> MCA mca: parameter "mca_component_path" (current value: >>>> >>>> </home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/lib/openmpi:/home/andrea/.openmpi/components>, >>>> data source: default value) >>>> Path where to look for Open MPI and ORTE >>>> components >>>> MCA mca: parameter "mca_component_show_load_errors" >>>> (current value:<1>, >>>> data source: default value) >>>> Whether to show errors for components that >>>> failed to load or >>>> not >>>> MCA mca: parameter "mca_component_disable_dlopen" >>>> (current value:<0>, >>>> data source: default value) >>>> Whether to attempt to disable opening >>>> dynamic components or not >>>> MCA mca: parameter "mca_verbose" (current value: >>>> <stderr>, data source: >>>> default value) >>>> Specifies where the default error output >>>> stream goes (this is >>>> separate from distinct help messages). Accepts a >>>> comma-delimited list of: stderr, stdout, syslog, >>>> syslogpri:<notice|info|debug>, >>>> syslogid:<str> (where str is the >>>> prefix string for all syslog notices), >>>> file[:filename] (if >>>> filename is not specified, a default >>>> filename is used), >>>> fileappend (if not specified, the file is opened >>>> for >>>> truncation), level[:N] (if specified, >>>> integer verbose level; >>>> otherwise, 0 is implied) >>>> MCA mpi: parameter "mpi_paffinity_alone" (current >>>> value:<0>, data >>>> source: default value, synonym of: >>>> opal_paffinity_alone) >>>> If nonzero, assume that this job is the only >>>> (set of) >>>> process(es) running on each node and bind >>>> processes to >>>> processors, starting with processor ID 0 >>>> MCA mpi: parameter "mpi_param_check" (current value: >>>> <1>, data source: >>>> default value) >>>> Whether you want MPI API parameters checked >>>> at run-time or not. >>>> Possible values are 0 (no checking) and 1 >>>> (perform checking at >>>> run-time) >>>> MCA mpi: parameter "mpi_yield_when_idle" (current >>>> value:<-1>, data >>>> source: default value) >>>> Yield the processor when waiting for MPI >>>> communication (for MPI >>>> processes, will default to 1 when >>>> oversubscribing nodes) >>>> MCA mpi: parameter "mpi_event_tick_rate" (current >>>> value:<-1>, data >>>> source: default value) >>>> How often to progress TCP communications (0 >>>> = never, otherwise >>>> specified in microseconds) >>>> MCA mpi: parameter "mpi_show_handle_leaks" (current >>>> value:<0>, data >>>> source: default value) >>>> Whether MPI_FINALIZE shows all MPI handles >>>> that were not freed >>>> or not >>>> MCA mpi: parameter "mpi_no_free_handles" (current >>>> value:<0>, data >>>> source: default value) >>>> Whether to actually free MPI objects when >>>> their handles are >>>> freed >>>> MCA mpi: parameter "mpi_show_mpi_alloc_mem_leaks" >>>> (current value:<0>, >>>> data source: default value) >>>> If>0, MPI_FINALIZE will show up to this >>>> many instances of >>>> memory allocated by MPI_ALLOC_MEM that was >>>> not freed by >>>> MPI_FREE_MEM >>>> MCA mpi: parameter "mpi_show_mca_params" (current >>>> value:<none>, data >>>> source: default value) >>>> Whether to show all MCA parameter values >>>> during MPI_INIT or not >>>> (good for reproducability of MPI jobs for >>>> debug purposes). >>>> Accepted values are all, default, file, api, >>>> and enviro - or a >>>> comma delimited combination of them >>>> MCA mpi: parameter "mpi_show_mca_params_file" >>>> (current value:<none>, >>>> data source: default value) >>>> If mpi_show_mca_params is true, setting this >>>> string to a valid >>>> filename tells Open MPI to dump all the MCA >>>> parameter values >>>> into a file suitable for reading via the >>>> mca_param_files >>>> parameter (good for reproducability of MPI jobs) >>>> MCA mpi: parameter "mpi_keep_peer_hostnames" (current >>>> value:<1>, data >>>> source: default value) >>>> If nonzero, save the string hostnames of all >>>> MPI peer processes >>>> (mostly for error / debugging output >>>> messages). This can add >>>> quite a bit of memory usage to each MPI process. >>>> MCA mpi: parameter "mpi_abort_delay" (current value: >>>> <0>, data source: >>>> default value) >>>> If nonzero, print out an identifying message >>>> when MPI_ABORT is >>>> invoked (hostname, PID of the process that >>>> called MPI_ABORT) >>>> and delay for that many seconds before >>>> exiting (a negative >>>> delay value means to never abort). This >>>> allows attaching of a >>>> debugger before quitting the job. >>>> MCA mpi: parameter "mpi_abort_print_stack" (current >>>> value:<0>, data >>>> source: default value) >>>> If nonzero, print out a stack trace when >>>> MPI_ABORT is invoked >>>> MCA mpi: parameter "mpi_preconnect_mpi" (current >>>> value:<0>, data >>>> source: default value, synonyms: >>>> mpi_preconnect_all) >>>> Whether to force MPI processes to fully >>>> wire-up the MPI >>>> connections between MPI processes during >>>> MPI_INIT (vs. making >>>> connections lazily -- upon the first MPI >>>> traffic between each >>>> process peer pair) >>>> MCA mpi: parameter "mpi_preconnect_all" (current >>>> value:<0>, data >>>> source: default value, deprecated, synonym of: >>>> mpi_preconnect_mpi) >>>> Whether to force MPI processes to fully >>>> wire-up the MPI >>>> connections between MPI processes during >>>> MPI_INIT (vs. making >>>> connections lazily -- upon the first MPI >>>> traffic between each >>>> process peer pair) >>>> MCA mpi: parameter "mpi_leave_pinned" (current value: >>>> <-1>, data source: >>>> default value) >>>> Whether to use the "leave pinned" protocol >>>> or not. Enabling >>>> this setting can help bandwidth performance >>>> when repeatedly >>>> sending and receiving large messages with >>>> the same buffers over >>>> RDMA-based networks (0 = do not use "leave >>>> pinned" protocol, 1 >>>> = use "leave pinned" protocol, -1 = allow >>>> network to choose at >>>> runtime). >>>> MCA mpi: parameter "mpi_leave_pinned_pipeline" >>>> (current value:<0>, data >>>> source: default value) >>>> Whether to use the "leave pinned pipeline" >>>> protocol or not. >>>> MCA mpi: parameter "mpi_warn_on_fork" (current value: >>>> <1>, data source: >>>> default value) >>>> If nonzero, issue a warning if program forks >>>> under conditions >>>> that could cause system errors >>>> MCA mpi: information "mpi_have_sparse_group_storage" >>>> (value:<0>, data >>>> source: default value) >>>> Whether this Open MPI installation supports >>>> storing of data in >>>> MPI groups in "sparse" formats (good for >>>> extremely large >>>> process count MPI jobs that create many >>>> communicators/groups) >>>> MCA mpi: parameter "mpi_use_sparse_group_storage" >>>> (current value:<0>, >>>> data source: default value) >>>> Whether to use "sparse" storage formats for >>>> MPI groups (only >>>> relevant if mpi_have_sparse_group_storage is 1) >>>> MCA mpi: parameter "mpi_notify_init_finalize" >>>> (current value:<1>, data >>>> source: default value) >>>> If nonzero, send two notifications during >>>> MPI_INIT: one near >>>> when MPI_INIT starts, and another right >>>> before MPI_INIT >>>> finishes, and send 2 notifications during >>>> MPI_FINALIZE: one >>>> right when MPI_FINALIZE starts, and another near >>>> when >>>> MPI_FINALIZE finishes. >>>> MCA orte: parameter "orte_base_help_aggregate" >>>> (current value:<1>, data >>>> source: default value) >>>> If orte_base_help_aggregate is true, >>>> duplicate help messages >>>> will be aggregated rather than displayed >>>> individually. This >>>> can be helpful for parallel jobs that >>>> experience multiple >>>> identical failures; rather than print out >>>> the same help/failure >>>> message N times, display it once with a >>>> count of how many >>>> processes sent the same message. >>>> MCA orte: parameter "orte_tmpdir_base" (current value: >>>> <none>, data >>>> source: default value) >>>> Base of the session directory tree >>>> MCA orte: parameter "orte_no_session_dirs" (current >>>> value:<none>, data >>>> source: default value) >>>> Prohibited locations for session directories >>>> (multiple >>>> locations separated by ',', default=NULL) >>>> MCA orte: parameter "orte_send_profile" (current >>>> value:<0>, data source: >>>> default value) >>>> Send profile info in launch message >>>> MCA orte: parameter "orte_debug" (current value:<0>, >>>> data source: >>>> default value) >>>> Top-level ORTE debug switch (default verbosity: >>>> 1) >>>> MCA orte: parameter "orte_debug_verbose" (current >>>> value:<-1>, data >>>> source: default value) >>>> Verbosity level for ORTE debug messages >>>> (default: 1) >>>> MCA orte: parameter "orte_debug_daemons" (current >>>> value:<0>, data >>>> source: default value) >>>> Whether to debug the ORTE daemons or not >>>> MCA orte: parameter "orte_debug_daemons_file" (current >>>> value:<0>, data >>>> source: default value) >>>> Whether want stdout/stderr of daemons to go >>>> to a file or not >>>> MCA orte: parameter "orte_daemon_bootstrap" (current >>>> value:<0>, data >>>> source: default value) >>>> Bootstrap the connection to the HNP >>>> MCA orte: parameter "orte_leave_session_attached" >>>> (current value:<0>, >>>> data source: default value) >>>> Whether applications and/or daemons should >>>> leave their sessions >>>> attached so that any output can be received >>>> - this allows X >>>> forwarding without all the attendant debugging >>>> output >>>> MCA orte: parameter "orte_output_debugger_proctable" >>>> (current value:<0>, >>>> data source: default value) >>>> Whether or not to output the debugger >>>> proctable after launch >>>> (default: false) >>>> MCA orte: parameter "orte_debugger_test_daemon" >>>> (current value:<none>, >>>> data source: default value) >>>> Name of the executable to be used to >>>> simulate a debugger >>>> colaunch (relative or absolute path) >>>> MCA orte: parameter "orte_debugger_test_attach" >>>> (current value:<0>, data >>>> source: default value) >>>> Test debugger colaunch after debugger attachment >>>> MCA orte: parameter "orte_debugger_check_rate" >>>> (current value:<0>, data >>>> source: default value) >>>> Set rate (in secs) for auto-detect of >>>> debugger attachment (0 => >>>> do not check) >>>> MCA orte: parameter "orte_do_not_launch" (current >>>> value:<0>, data >>>> source: default value) >>>> Perform all necessary operations to prepare >>>> to launch the >>>> application, but do not actually launch it >>>> MCA orte: parameter "orte_daemon_spin" (current value: >>>> <0>, data source: >>>> default value) >>>> Have any orteds spin until we can connect a >>>> debugger to them >>>> MCA orte: parameter "orte_daemon_fail" (current value: >>>> <-1>, data source: >>>> default value) >>>> Have the specified orted fail after init for >>>> debugging purposes >>>> MCA orte: parameter "orte_daemon_fail_delay" (current >>>> value:<0>, data >>>> source: default value) >>>> Have the specified orted fail after >>>> specified number of seconds >>>> (default: 0 => no delay) >>>> MCA orte: parameter "orte_heartbeat_rate" (current >>>> value:<0>, data >>>> source: default value) >>>> Seconds between checks for daemon >>>> state-of-health (default: 0 >>>> => do not check) >>>> MCA orte: parameter "orte_startup_timeout" (current >>>> value:<0>, data >>>> source: default value) >>>> Milliseconds/daemon to wait for startup >>>> before declaring >>>> failed_to_start (default: 0 => do not check) >>>> MCA orte: parameter "orte_timing" (current value:<0>, >>>> data source: >>>> default value) >>>> Request that critical timing loops be measured >>>> MCA orte: parameter "orte_timing_details" (current >>>> value:<0>, data >>>> source: default value) >>>> Request that detailed timing data by reported >>>> MCA orte: parameter "orte_base_user_debugger" (current >>>> value:<totalview >>>> @mpirun@ -a @mpirun_args@ : ddt -n @np@ >>>> -start @executable@ >>>> @executable_argv@ @single_app@ : fxp @mpirun@ -a >>>> @mpirun_args@>, data source: default value) >>>> Sequence of user-level debuggers to search >>>> for in orterun >>>> MCA orte: parameter "orte_abort_timeout" (current >>>> value:<1>, data >>>> source: default value) >>>> Max time to wait [in secs] before aborting >>>> an ORTE operation >>>> (default: 1sec) >>>> MCA orte: parameter "orte_timeout_step" (current >>>> value:<1000>, data >>>> source: default value) >>>> Time to wait [in usecs/proc] before aborting >>>> an ORTE operation >>>> (default: 1000 usec/proc) >>>> MCA orte: parameter "orte_default_hostfile" (current value: >>>> >>>> </home/andrea/library/openmpi/openmpi-1.6-gnu-4.7/etc/openmpi-default-hostfile>, >>>> data source: default value) >>>> Name of the default hostfile (relative or >>>> absolute path, "none" >>>> to ignore environmental or default MCA param >>>> setting) >>>> MCA orte: parameter "orte_rankfile" (current value: >>>> <none>, data source: >>>> default value, synonyms: rmaps_rank_file_path) >>>> Name of the rankfile to be used for mapping >>>> processes (relative >>>> or absolute path) >>>> MCA orte: parameter "orte_keep_fqdn_hostnames" >>>> (current value:<0>, data >>>> source: default value) >>>> Whether or not to keep FQDN hostnames [default: >>>> no] >>>> MCA orte: parameter "orte_use_regexp" (current value: >>>> <0>, data source: >>>> default value) >>>> Whether or not to use regular expressions >>>> for launch [default: >>>> no] >>>> MCA orte: parameter "orte_tag_output" (current value: >>>> <0>, data source: >>>> default value) >>>> Tag all output with [job,rank] (default: false) >>>> MCA orte: parameter "orte_xml_output" (current value: >>>> <0>, data source: >>>> default value) >>>> Display all output in XML format (default: false) >>>> MCA orte: parameter "orte_xml_file" (current value: >>>> <none>, data source: >>>> default value) >>>> Provide all output in XML format to the >>>> specified file >>>> MCA orte: parameter "orte_timestamp_output" (current >>>> value:<0>, data >>>> source: default value) >>>> Timestamp all application process output >>>> (default: false) >>>> MCA orte: parameter "orte_output_filename" (current >>>> value:<none>, data >>>> source: default value) >>>> Redirect output from application processes >>>> into filename.rank >>>> [default: NULL] >>>> MCA orte: parameter "orte_show_resolved_nodenames" >>>> (current value:<0>, >>>> data source: default value) >>>> Display any node names that are resolved to >>>> a different name >>>> (default: false) >>>> MCA orte: parameter "orte_hetero_apps" (current value: >>>> <0>, data source: >>>> default value) >>>> Indicates that multiple app_contexts are >>>> being provided that >>>> are a mix of 32/64 bit binaries (default: false) >>>> MCA orte: parameter "orte_launch_agent" (current >>>> value:<orted>, data >>>> source: default value) >>>> Command used to start processes on remote >>>> nodes (default: >>>> orted) >>>> MCA orte: parameter "orte_allocation_required" >>>> (current value:<0>, data >>>> source: default value) >>>> Whether or not an allocation by a resource >>>> manager is required >>>> [default: no] >>>> MCA orte: parameter "orte_xterm" (current value: >>>> <none>, data source: >>>> default value) >>>> Create a new xterm window and display output >>>> from the specified >>>> ranks there [default: none] >>>> MCA orte: parameter "orte_forward_job_control" >>>> (current value:<0>, data >>>> source: default value) >>>> Forward SIGTSTP (after converting to >>>> SIGSTOP) and SIGCONT >>>> signals to the application procs [default: no] >>>> MCA orte: parameter "orte_rsh_agent" (current value: >>>> <ssh : rsh>, data >>>> source: default value, synonyms: >>>> pls_rsh_agent, plm_rsh_agent) >>>> The command used to launch executables on remote >>>> nodes >>>> (typically either "ssh" or "rsh") >>>> MCA orte: parameter "orte_assume_same_shell" (current >>>> value:<1>, data >>>> source: default value, synonyms: >>>> plm_rsh_assume_same_shell) >>>> If set to 1, assume that the shell on the >>>> remote node is the >>>> same as the shell on the local node. >>>> Otherwise, probe for what >>>> the remote shell [default: 1] >>>> MCA orte: parameter "orte_report_launch_progress" >>>> (current value:<0>, >>>> data source: default value) >>>> Output a brief periodic report on launch >>>> progress [default: no] >>>> MCA orte: parameter "orte_num_boards" (current value: >>>> <1>, data source: >>>> default value) >>>> Number of processor boards/node (1-256) >>>> [default: 1] >>>> MCA orte: parameter "orte_num_sockets" (current value: >>>> <0>, data source: >>>> default value) >>>> Number of sockets/board (1-256) >>>> MCA orte: parameter "orte_num_cores" (current value: >>>> <0>, data source: >>>> default value) >>>> Number of cores/socket (1-256) >>>> MCA orte: parameter "orte_cpu_set" (current value: >>>> <none>, data source: >>>> default value) >>>> Comma-separated list of ranges specifying >>>> logical cpus >>>> allocated to this job [default: none] >>>> MCA orte: parameter "orte_process_binding" (current >>>> value:<none>, data >>>> source: default value) >>>> Policy for binding processes [none | core | >>>> socket | board] >>>> (supported qualifier: if-avail) >>>> MCA orte: parameter "orte_report_bindings" (current >>>> value:<0>, data >>>> source: default value) >>>> Report bindings >>>> MCA orte: parameter "orte_report_events" (current >>>> value:<none>, data >>>> source: default value) >>>> URI to which events are to be reported >>>> (default: NULL)] >>>> MCA opal: parameter "opal_net_private_ipv4" (current value: >>>> >>>> <10.0.0.0/8;172.16.0.0/12;192.168.0.0/16;169.254.0.0/16>, data >>>> source: default value) >>>> Semicolon-delimited list of CIDR notation >>>> entries specifying >>>> what networks are considered "private" >>>> (default value based on >>>> RFC1918 and RFC3330) >>>> MCA opal: parameter "opal_signal" (current value: >>>> <6,7,8,11>, data >>>> source: default value) >>>> Comma-delimited list of integer signal >>>> numbers to Open MPI to >>>> attempt to intercept. Upon receipt of the >>>> intercepted signal, >>>> Open MPI will display a stack trace and >>>> abort. Open MPI will >>>> *not* replace signals if handlers are >>>> already installed by the >>>> time MPI_INIT is invoked. Optionally append >>>> ":complain" to any >>>> signal number in the comma-delimited list to >>>> make Open MPI >>>> complain if it detects another signal >>>> handler (and therefore >>>> does not insert its own). >>>> MCA opal: parameter "opal_profile" (current value: >>>> <0>, data source: >>>> default value) >>>> Set to non-zero to profile component selections >>>> MCA opal: parameter "opal_profile_file" (current >>>> value:<none>, data >>>> source: default value) >>>> Name of the file containing the cluster >>>> configuration >>>> information >>>> MCA opal: parameter "opal_paffinity_alone" (current >>>> value:<0>, data >>>> source: default value, synonyms: >>>> mpi_paffinity_alone) >>>> If nonzero, assume that this job is the only >>>> (set of) >>>> process(es) running on each node and bind >>>> processes to >>>> processors, starting with processor ID 0 >>>> MCA opal: parameter "opal_set_max_sys_limits" (current >>>> value:<0>, data >>>> source: default value) >>>> Set to non-zero to automatically set any >>>> system-imposed limits >>>> to the maximum allowed >>>> MCA opal: parameter "opal_event_include" (current >>>> value:<poll>, data >>>> source: default value) >>>> Comma-delimited list of libevent subsystems >>>> to use (epoll, >>>> poll, select -- available on your platform) >>>> MCA backtrace: parameter "backtrace" (current value: >>>> <none>, data source: >>>> default value) >>>> Default selection set of components for the >>>> backtrace framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA backtrace: parameter "backtrace_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the backtrace framework >>>> (0 = no verbosity) >>>> MCA backtrace: parameter "backtrace_execinfo_priority" >>>> (current value:<0>, >>>> data source: default value) >>>> MCA memchecker: parameter "memchecker" (current value: >>>> <none>, data source: >>>> default value) >>>> Default selection set of components for the >>>> memchecker >>>> framework (<none> means use all components >>>> that can be found) >>>> MCA memory: parameter "memory" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> memory framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA memory: parameter "memory_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the memory framework (0 >>>> = no verbosity) >>>> MCA memory: information >>>> "memory_linux_ptmalloc2_available" (value:<1>, >>>> data source: default value) >>>> Whether ptmalloc2 support is included in >>>> Open MPI or not (1 = >>>> yes, 0 = no) >>>> MCA memory: information >>>> "memory_linux_ummunotify_available" (value:<0>, >>>> data source: default value) >>>> Whether ummunotify support is included in >>>> Open MPI or not (1 = >>>> yes, 0 = no) >>>> MCA memory: parameter "memory_linux_ptmalloc2_enable" >>>> (current value:<-1>, >>>> data source: default value) >>>> Whether to enable ptmalloc2 support or not >>>> (negative = try to >>>> enable, but continue even if support is not >>>> available, 0 = do >>>> not enable support, positive = try to enable >>>> and fail if >>>> support is not available) >>>> MCA memory: parameter "memory_linux_ummunotify_enable" >>>> (current value: >>>> <-1>, data source: default value) >>>> Whether to enable ummunotify support or not >>>> (negative = try to >>>> enable, but continue even if support is not >>>> available, 0 = do >>>> not enable support, positive = try to enable >>>> and fail if >>>> support is not available) >>>> MCA memory: parameter "memory_linux_disable" (current >>>> value:<0>, data >>>> source: default value) >>>> If this MCA parameter is set to 1 **VIA >>>> ENVIRONMENT VARIABLE >>>> ONLY*** (this MCA parameter *CANNOT* be set >>>> in a file or on the >>>> mpirun command line!), this component will >>>> be disabled and will >>>> not attempt to use either ummunotify or >>>> memory hook support >>>> MCA memory: parameter "memory_linux_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA paffinity: parameter "paffinity_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level of the paffinity framework >>>> MCA paffinity: parameter "paffinity" (current value: >>>> <none>, data source: >>>> default value) >>>> Default selection set of components for the >>>> paffinity framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA paffinity: parameter "paffinity_hwloc_priority" >>>> (current value:<40>, data >>>> source: default value) >>>> Priority of the hwloc paffinity component >>>> MCA carto: parameter "carto_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level of the carto framework >>>> MCA carto: parameter "carto" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> carto framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA carto: parameter "carto_auto_detect_priority" >>>> (current value:<11>, >>>> data source: default value) >>>> Priority of the auto_detect carto component >>>> MCA carto: parameter "carto_file_path" (current value: >>>> <none>, data >>>> source: default value) >>>> The path to the cartography file >>>> MCA carto: parameter "carto_file_priority" (current >>>> value:<10>, data >>>> source: default value) >>>> Priority of the file carto component >>>> MCA shmem: parameter "shmem_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level of the shmem framework >>>> MCA shmem: parameter "shmem" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> shmem framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA shmem: parameter "shmem_mmap_enable_nfs_warning" >>>> (current value:<1>, >>>> data source: default value) >>>> Enable the warning emitted when Open MPI >>>> detects that its >>>> shared memory backing file is located on a >>>> network filesystem >>>> (1 = enabled, 0 = disabled). >>>> MCA shmem: parameter "shmem_mmap_priority" (current >>>> value:<50>, data >>>> source: default value) >>>> Priority of the mmap shmem component >>>> MCA shmem: parameter "shmem_mmap_relocate_backing_file" >>>> (current value: >>>> <0>, data source: default value) >>>> Whether to change the default placement of >>>> backing files or not >>>> (Negative = try to relocate backing files to >>>> an area rooted at >>>> the path specified by >>>> >>>> shmem_mmap_opal_shmem_mmap_backing_file_base_dir, but continue >>>> with the default path if the relocation >>>> fails, 0 = do not >>>> relocate, Positive = same as the negative >>>> option, but will fail >>>> if the relocation fails. >>>> MCA shmem: parameter "shmem_mmap_backing_file_base_dir" >>>> (current value: >>>> </dev/shm>, data source: default value) >>>> Specifies where backing files will be created >>>> when >>>> shmem_mmap_relocate_backing_file is in use. >>>> MCA shmem: parameter "shmem_posix_priority" (current >>>> value:<40>, data >>>> source: default value) >>>> Priority of the posix shmem component >>>> MCA shmem: parameter "shmem_sysv_priority" (current >>>> value:<30>, data >>>> source: default value) >>>> Priority of the sysv shmem component >>>> MCA maffinity: parameter "maffinity_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level of the maffinity framework >>>> MCA maffinity: parameter "maffinity" (current value: >>>> <none>, data source: >>>> default value) >>>> Default selection set of components for the >>>> maffinity framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA maffinity: parameter "maffinity_first_use_priority" >>>> (current value:<10>, >>>> data source: default value) >>>> Priority of the first_use maffinity component >>>> MCA maffinity: parameter "maffinity_hwloc_priority" >>>> (current value:<40>, data >>>> source: default value) >>>> Priority of the hwloc maffinity component >>>> MCA timer: parameter "timer" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> timer framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA timer: parameter "timer_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the timer framework (0 = >>>> no verbosity) >>>> MCA timer: parameter "timer_linux_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA sysinfo: parameter "sysinfo" (current value:<none>, >>>> data source: >>>> default value) >>>> Default selection set of components for the >>>> sysinfo framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA sysinfo: parameter "sysinfo_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the sysinfo framework (0 >>>> = no verbosity) >>>> MCA sysinfo: parameter "sysinfo_linux_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA hwloc: parameter "hwloc_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level of the hwloc framework >>>> MCA hwloc: parameter "hwloc_base_mem_alloc_policy" >>>> (current value:<none>, >>>> data source: default value) >>>> Policy that determines how general memory >>>> allocations are bound >>>> after MPI_INIT. A value of "none" means >>>> that no memory policy >>>> is applied. A value of "local_only" means >>>> that all memory >>>> allocations will be restricted to the local >>>> NUMA node where >>>> each process is placed. Note that operating >>>> system paging >>>> policies are unaffected by this setting. >>>> For example, if >>>> "local_only" is used and local NUMA node >>>> memory is exhausted, a >>>> new memory allocation may cause paging. >>>> MCA hwloc: parameter >>>> "hwloc_base_mem_bind_failure_action" (current value: >>>> <error>, data source: default value) >>>> What Open MPI will do if it explicitly tries >>>> to bind memory to >>>> a specific NUMA location, and fails. Note >>>> that this is a >>>> different case than the general allocation >>>> policy described by >>>> hwloc_base_alloc_policy. A value of "warn" >>>> means that Open MPI >>>> will warn the first time this happens, but >>>> allow the job to >>>> continue (possibly with degraded >>>> performance). A value of >>>> "error" means that Open MPI will abort the >>>> job if this happens. >>>> MCA hwloc: parameter "hwloc" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> hwloc framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA hwloc: parameter "hwloc_hwloc132_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA dpm: parameter "dpm" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> dpm framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA dpm: parameter "dpm_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the dpm framework (0 = >>>> no verbosity) >>>> MCA dpm: parameter "dpm_orte_priority" (current >>>> value:<0>, data source: >>>> default value) >>>> MCA pubsub: parameter "pubsub" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> pubsub framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA pubsub: parameter "pubsub_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the pubsub framework (0 >>>> = no verbosity) >>>> MCA pubsub: parameter "pubsub_orte_priority" (current >>>> value:<50>, data >>>> source: default value) >>>> Priority of the pubsub pmi component >>>> MCA allocator: parameter "allocator" (current value: >>>> <none>, data source: >>>> default value) >>>> Default selection set of components for the >>>> allocator framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA allocator: parameter "allocator_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the allocator framework >>>> (0 = no verbosity) >>>> MCA allocator: parameter "allocator_basic_priority" >>>> (current value:<0>, data >>>> source: default value) >>>> MCA allocator: parameter "allocator_bucket_num_buckets" >>>> (current value:<30>, >>>> data source: default value) >>>> MCA allocator: parameter "allocator_bucket_priority" >>>> (current value:<0>, data >>>> source: default value) >>>> MCA coll: parameter "coll" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> coll framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA coll: parameter "coll_base_verbose" (current >>>> value:<0>, data source: >>>> default value) >>>> Verbosity level for the coll framework (0 = >>>> no verbosity) >>>> MCA coll: parameter "coll_basic_priority" (current >>>> value:<10>, data >>>> source: default value) >>>> Priority of the basic coll component >>>> MCA coll: parameter "coll_basic_crossover" (current >>>> value:<4>, data >>>> source: default value) >>>> Minimum number of processes in a >>>> communicator before using the >>>> logarithmic algorithms >>>> MCA coll: parameter "coll_hierarch_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> Priority of the hierarchical coll component >>>> MCA coll: parameter "coll_hierarch_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Turn verbose message of the hierarchical >>>> coll component on/off >>>> MCA coll: parameter "coll_hierarch_use_rdma" (current >>>> value:<0>, data >>>> source: default value) >>>> Switch from the send btl list used to detect >>>> hierarchies to the >>>> rdma btl list >>>> MCA coll: parameter "coll_hierarch_ignore_sm" (current >>>> value:<0>, data >>>> source: default value) >>>> Ignore sm protocol when detecting >>>> hierarchies. Required to >>>> enable the usage of protocol specific >>>> collective operations >>>> MCA coll: parameter "coll_hierarch_detection_alg" >>>> (current value:<2>, >>>> data source: default value) >>>> Used to specify the algorithm for detecting >>>> Hierarchy.Choose >>>> between all or two levels of hierarchy >>>> MCA coll: parameter "coll_hierarch_bcast_alg" (current >>>> value:<4>, data >>>> source: default value) >>>> Used to specify the algorithm used for bcast >>>> operations. >>>> MCA coll: parameter "coll_hierarch_segment_size" >>>> (current value:<32768>, >>>> data source: default value) >>>> Used to specify the segment size for >>>> segmented algorithms. >>>> MCA coll: parameter "coll_inter_priority" (current >>>> value:<40>, data >>>> source: default value) >>>> Priority of the inter coll component >>>> MCA coll: parameter "coll_inter_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Turn verbose message of the inter coll >>>> component on/off >>>> MCA coll: parameter "coll_self_priority" (current >>>> value:<75>, data >>>> source: default value) >>>> MCA coll: parameter "coll_sm_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> Priority of the sm coll component >>>> MCA coll: parameter "coll_sm_control_size" (current >>>> value:<4096>, data >>>> source: default value) >>>> Length of the control data -- should usually >>>> be either the >>>> length of a cache line on most SMPs, or the >>>> size of a page on >>>> machines that support direct memory affinity >>>> page placement (in >>>> bytes) >>>> MCA coll: parameter "coll_sm_fragment_size" (current >>>> value:<8192>, data >>>> source: default value) >>>> Fragment size (in bytes) used for passing >>>> data through shared >>>> memory (will be rounded up to the nearest >>>> control_size size) >>>> MCA coll: parameter "coll_sm_comm_in_use_flags" >>>> (current value:<2>, data >>>> source: default value) >>>> Number of "in use" flags, used to mark a >>>> message passing area >>>> segment as currently being used or not (must >>>> be>= 2 and<= >>>> comm_num_segments) >>>> MCA coll: parameter "coll_sm_comm_num_segments" >>>> (current value:<8>, data >>>> source: default value) >>>> Number of segments in each communicator's >>>> shared memory message >>>> passing area (must be>= 2, and must be a >>>> multiple of >>>> comm_in_use_flags) >>>> MCA coll: parameter "coll_sm_tree_degree" (current >>>> value:<4>, data >>>> source: default value) >>>> Degree of the tree for tree-based operations >>>> (must be => 1 and >>>> <= min(control_size, 255)) >>>> MCA coll: parameter "coll_sm_info_num_procs" (current >>>> value:<4>, data >>>> source: default value) >>>> Number of processes to use for the calculation >>>> of the >>>> shared_mem_size MCA information parameter >>>> (must be => 2) >>>> MCA coll: information "coll_sm_shared_mem_used_data" >>>> (value:<548864>, >>>> data source: default value) >>>> Amount of shared memory used, per >>>> communicator, in the shared >>>> memory data area for info_num_procs >>>> processes (in bytes) >>>> MCA coll: parameter "coll_sync_priority" (current >>>> value:<50>, data >>>> source: default value) >>>> Priority of the sync coll component; only >>>> relevant if >>>> barrier_before or barrier_after is> 0 >>>> MCA coll: parameter "coll_sync_barrier_before" >>>> (current value:<1000>, >>>> data source: default value) >>>> Do a synchronization before each Nth collective >>>> MCA coll: parameter "coll_sync_barrier_after" (current >>>> value:<0>, data >>>> source: default value) >>>> Do a synchronization after each Nth collective >>>> MCA coll: parameter "coll_tuned_priority" (current >>>> value:<30>, data >>>> source: default value) >>>> Priority of the tuned coll component >>>> MCA coll: parameter >>>> "coll_tuned_pre_allocate_memory_comm_size_limit" >>>> (current value:<32768>, data source: default >>>> value) >>>> Size of communicator were we stop >>>> pre-allocating memory for the >>>> fixed internal buffer used for message >>>> requests etc that is >>>> hung off the communicator data segment. I.e. >>>> if you have a >>>> 100'000 nodes you might not want to >>>> pre-allocate 200'000 >>>> request handle slots per communicator instance! >>>> MCA coll: parameter "coll_tuned_init_tree_fanout" >>>> (current value:<4>, >>>> data source: default value) >>>> Inital fanout used in the tree topologies for >>>> each >>>> communicator. This is only an initial guess, >>>> if a tuned >>>> collective needs a different fanout for an >>>> operation, it build >>>> it dynamically. This parameter is only for >>>> the first guess and >>>> might save a little time >>>> MCA coll: parameter "coll_tuned_init_chain_fanout" >>>> (current value:<4>, >>>> data source: default value) >>>> Inital fanout used in the chain (fanout >>>> followed by pipeline) >>>> topologies for each communicator. This is >>>> only an initial >>>> guess, if a tuned collective needs a >>>> different fanout for an >>>> operation, it build it dynamically. This >>>> parameter is only for >>>> the first guess and might save a little time >>>> MCA coll: parameter "coll_tuned_use_dynamic_rules" >>>> (current value:<0>, >>>> data source: default value) >>>> Switch used to decide if we use static >>>> (compiled/if statements) >>>> or dynamic (built at runtime) decision function >>>> rules >>>> MCA io: parameter "io_base_freelist_initial_size" >>>> (current value:<16>, >>>> data source: default value) >>>> Initial MPI-2 IO request freelist size >>>> MCA io: parameter "io_base_freelist_max_size" >>>> (current value:<64>, >>>> data source: default value) >>>> Max size of the MPI-2 IO request freelist >>>> MCA io: parameter "io_base_freelist_increment" >>>> (current value:<16>, >>>> data source: default value) >>>> Increment size of the MPI-2 IO request freelist >>>> MCA io: parameter "io" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> io framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA io: parameter "io_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the io framework (0 = no >>>> verbosity) >>>> MCA io: parameter "io_romio_priority" (current >>>> value:<10>, data >>>> source: default value) >>>> Priority of the io romio component >>>> MCA io: parameter "io_romio_delete_priority" >>>> (current value:<10>, data >>>> source: default value) >>>> Delete priority of the io romio component >>>> MCA io: information "io_romio_version" (value:<from >>>> MPICH2 v1.3.1 with >>>> an additional patch from >>>> romio-ma...@mcs.anl.gov about an >>>> attribute issue>, data source: default value) >>>> Version of ROMIO >>>> MCA io: information "io_romio_user_configure_params" >>>> (value:<none>, >>>> data source: default value) >>>> User-specified command line parameters >>>> passed to ROMIO's >>>> configure script >>>> MCA io: information >>>> "io_romio_complete_configure_params" (value:< >>>> CFLAGS='-DNDEBUG -g -O2 -finline-functions >>>> -fno-strict-aliasing >>>> -pthread' CPPFLAGS=' >>>> >>>> -I/home/andrea/library/openmpi/openmpi-1.6/opal/mca/hwloc/hwloc132/hwloc/include >>>> -I/usr/include/infiniband >>>> -I/usr/include/infiniband' >>>> FFLAGS='' LDFLAGS='-Wl,--rpath >>>> -Wl,/home/andrea/library/gcc/gcc-objects/lib64 ' >>>> --enable-shared --enable-static >>>> >>>> --prefix=/home/andrea/library/openmpi/openmpi-1.6-gnu-4.7 >>>> --with-mpi=open_mpi --disable-aio>, data >>>> source: default value) >>>> Complete set of command line parameters >>>> passed to ROMIO's >>>> configure script >>>> MCA mpool: parameter "mpool" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> mpool framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA mpool: parameter "mpool_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the mpool framework (0 = >>>> no verbosity) >>>> MCA mpool: parameter "mpool_fake_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA mpool: parameter "mpool_rdma_rcache_name" (current >>>> value:<vma>, data >>>> source: default value) >>>> The name of the registration cache the mpool >>>> should use >>>> MCA mpool: parameter "mpool_rdma_rcache_size_limit" >>>> (current value:<0>, >>>> data source: default value) >>>> the maximum size of registration cache in >>>> bytes. 0 is unlimited >>>> (default 0) >>>> MCA mpool: parameter "mpool_rdma_print_stats" (current >>>> value:<0>, data >>>> source: default value) >>>> print pool usage statistics at the end of the run >>>> MCA mpool: parameter "mpool_rdma_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA mpool: parameter "mpool_sm_allocator" (current >>>> value:<bucket>, data >>>> source: default value) >>>> Name of allocator component to use with sm mpool >>>> MCA mpool: parameter "mpool_sm_min_size" (current >>>> value:<67108864>, data >>>> source: default value) >>>> Minimum size of the sm mpool shared memory file >>>> MCA mpool: parameter "mpool_sm_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Enable verbose output for mpool sm component >>>> MCA mpool: parameter "mpool_sm_priority" (current >>>> value:<0>, data source: >>>> default value) >>>> MCA pml: parameter "pml_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level of the PML framework >>>> MCA pml: parameter "pml" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> pml framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA pml: parameter "pml_bfo_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA pml: parameter "pml_bfo_free_list_num" (current >>>> value:<4>, data >>>> source: default value) >>>> MCA pml: parameter "pml_bfo_free_list_max" (current >>>> value:<-1>, data >>>> source: default value) >>>> MCA pml: parameter "pml_bfo_free_list_inc" (current >>>> value:<64>, data >>>> source: default value) >>>> MCA pml: parameter "pml_bfo_priority" (current value: >>>> <5>, data source: >>>> default value) >>>> MCA pml: parameter "pml_bfo_send_pipeline_depth" >>>> (current value:<3>, >>>> data source: default value) >>>> MCA pml: parameter "pml_bfo_recv_pipeline_depth" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_bfo_rdma_put_retries_limit" >>>> (current value:<5>, >>>> data source: default value) >>>> MCA pml: parameter "pml_bfo_max_rdma_per_request" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_bfo_max_send_per_range" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_bfo_unexpected_limit" >>>> (current value:<128>, >>>> data source: default value) >>>> MCA pml: parameter "pml_bfo_allocator" (current >>>> value:<bucket>, data >>>> source: default value) >>>> Name of allocator component for unexpected >>>> messages >>>> MCA pml: parameter "pml_cm_free_list_num" (current >>>> value:<4>, data >>>> source: default value) >>>> Initial size of request free lists >>>> MCA pml: parameter "pml_cm_free_list_max" (current >>>> value:<-1>, data >>>> source: default value) >>>> Maximum size of request free lists >>>> MCA pml: parameter "pml_cm_free_list_inc" (current >>>> value:<64>, data >>>> source: default value) >>>> Number of elements to add when growing >>>> request free lists >>>> MCA pml: parameter "pml_cm_priority" (current value: >>>> <10>, data source: >>>> default value) >>>> CM PML selection priority >>>> MCA pml: parameter "pml_csum_free_list_num" (current >>>> value:<4>, data >>>> source: default value) >>>> MCA pml: parameter "pml_csum_free_list_max" (current >>>> value:<-1>, data >>>> source: default value) >>>> MCA pml: parameter "pml_csum_free_list_inc" (current >>>> value:<64>, data >>>> source: default value) >>>> MCA pml: parameter "pml_csum_send_pipeline_depth" >>>> (current value:<3>, >>>> data source: default value) >>>> MCA pml: parameter "pml_csum_recv_pipeline_depth" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_csum_rdma_put_retries_limit" >>>> (current value: >>>> <5>, data source: default value) >>>> MCA pml: parameter "pml_csum_max_rdma_per_request" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_csum_max_send_per_range" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_csum_unexpected_limit" >>>> (current value:<128>, >>>> data source: default value) >>>> MCA pml: parameter "pml_csum_allocator" (current >>>> value:<bucket>, data >>>> source: default value) >>>> Name of allocator component for unexpected >>>> messages >>>> MCA pml: parameter "pml_csum_priority" (current >>>> value:<0>, data source: >>>> default value) >>>> MCA pml: parameter "pml_ob1_free_list_num" (current >>>> value:<4>, data >>>> source: default value) >>>> MCA pml: parameter "pml_ob1_free_list_max" (current >>>> value:<-1>, data >>>> source: default value) >>>> MCA pml: parameter "pml_ob1_free_list_inc" (current >>>> value:<64>, data >>>> source: default value) >>>> MCA pml: parameter "pml_ob1_priority" (current value: >>>> <20>, data source: >>>> default value) >>>> MCA pml: parameter "pml_ob1_send_pipeline_depth" >>>> (current value:<3>, >>>> data source: default value) >>>> MCA pml: parameter "pml_ob1_recv_pipeline_depth" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_ob1_rdma_put_retries_limit" >>>> (current value:<5>, >>>> data source: default value) >>>> MCA pml: parameter "pml_ob1_max_rdma_per_request" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_ob1_max_send_per_range" >>>> (current value:<4>, >>>> data source: default value) >>>> MCA pml: parameter "pml_ob1_unexpected_limit" >>>> (current value:<128>, >>>> data source: default value) >>>> MCA pml: parameter "pml_ob1_allocator" (current >>>> value:<bucket>, data >>>> source: default value) >>>> Name of allocator component for unexpected >>>> messages >>>> MCA pml: parameter "pml_v_priority" (current value: >>>> <-1>, data source: >>>> default value) >>>> MCA pml: parameter "pml_v_output" (current value: >>>> <stderr>, data source: >>>> default value) >>>> MCA pml: parameter "pml_v_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA bml: parameter "bml" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> bml framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA bml: parameter "bml_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the bml framework (0 = >>>> no verbosity) >>>> MCA bml: parameter "bml_r2_show_unreach_errors" >>>> (current value:<1>, >>>> data source: default value) >>>> Show error message when procs are unreachable >>>> MCA bml: parameter "bml_r2_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA rcache: parameter "rcache" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> rcache framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA rcache: parameter "rcache_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the rcache framework (0 >>>> = no verbosity) >>>> MCA rcache: parameter "rcache_vma_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA btl: parameter "btl_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level of the BTL framework >>>> MCA btl: parameter "btl" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> btl framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA btl: parameter "btl_self_free_list_num" (current >>>> value:<0>, data >>>> source: default value) >>>> Number of fragments by default >>>> MCA btl: parameter "btl_self_free_list_max" (current >>>> value:<-1>, data >>>> source: default value) >>>> Maximum number of fragments >>>> MCA btl: parameter "btl_self_free_list_inc" (current >>>> value:<32>, data >>>> source: default value) >>>> Increment by this number of fragments >>>> MCA btl: parameter "btl_self_exclusivity" (current >>>> value:<65536>, data >>>> source: default value) >>>> BTL exclusivity (must be>= 0) >>>> MCA btl: parameter "btl_self_flags" (current value: >>>> <10>, data source: >>>> default value) >>>> BTL bit flags (general flags: SEND=1, PUT=2, >>>> GET=4, >>>> SEND_INPLACE=8, RDMA_MATCHED=64, >>>> HETEROGENEOUS_RDMA=256; flags >>>> only used by the "dr" PML (ignored by others): >>>> ACK=16, >>>> CHECKSUM=32, RDMA_COMPLETION=128; flags only >>>> used by the "bfo" >>>> PML (ignored by others): FAILOVER_SUPPORT=512) >>>> MCA btl: parameter "btl_self_rndv_eager_limit" >>>> (current value:<131072>, >>>> data source: default value) >>>> Size (in bytes) of "phase 1" fragment sent >>>> for all large >>>> messages (must be>= 0 and<= eager_limit) >>>> MCA btl: parameter "btl_self_eager_limit" (current >>>> value:<131072>, data >>>> source: default value) >>>> Maximum size (in bytes) of "short" messages >>>> (must be>= 1). >>>> MCA btl: parameter "btl_self_max_send_size" (current >>>> value:<262144>, >>>> data source: default value) >>>> Maximum size (in bytes) of a single "phase >>>> 2" fragment of a >>>> long message when using the pipeline >>>> protocol (must be>= 1) >>>> MCA btl: parameter >>>> "btl_self_rdma_pipeline_send_length" (current value: >>>> <2147483647>, data source: default value) >>>> Length of the "phase 2" portion of a large >>>> message (in bytes) >>>> when using the pipeline protocol. This part >>>> of the message >>>> will be split into fragments of size >>>> max_send_size and sent >>>> using send/receive semantics (must be>= 0; >>>> only relevant when >>>> the PUT flag is set) >>>> MCA btl: parameter "btl_self_rdma_pipeline_frag_size" >>>> (current value: >>>> <2147483647>, data source: default value) >>>> Maximum size (in bytes) of a single "phase >>>> 3" fragment from a >>>> long message when using the pipeline >>>> protocol. These fragments >>>> will be sent using RDMA semantics (must be >>>>> = 1; only relevant >>>> when the PUT flag is set) >>>> MCA btl: parameter "btl_self_min_rdma_pipeline_size" >>>> (current value: >>>> <0>, data source: default value) >>>> Messages smaller than this size (in bytes) >>>> will not use the >>>> RDMA pipeline protocol. Instead, they will >>>> be split into >>>> fragments of max_send_size and sent using >>>> send/receive >>>> semantics (must be>=0, and is automatically >>>> adjusted up to at >>>> least >>>> (eager_limit+btl_rdma_pipeline_send_length); only >>>> relevant when the PUT flag is set) >>>> MCA btl: parameter "btl_self_bandwidth" (current >>>> value:<100>, data >>>> source: default value) >>>> Approximate maximum bandwidth of >>>> interconnect(0 = auto-detect >>>> value at run-time [not supported in all BTL >>>> modules],>= 1 = >>>> bandwidth in Mbps) >>>> MCA btl: parameter "btl_self_latency" (current value: >>>> <0>, data source: >>>> default value) >>>> Approximate latency of interconnect (must be>= 0) >>>> MCA btl: parameter "btl_self_priority" (current >>>> value:<0>, data source: >>>> default value) >>>> MCA btl: information "btl_sm_have_knem_support" >>>> (value:<0>, data >>>> source: default value) >>>> Whether this component supports the knem >>>> Linux kernel module or >>>> not >>>> MCA btl: parameter "btl_sm_use_knem" (current value: >>>> <0>, data source: >>>> default value) >>>> Whether knem support is desired or not >>>> (negative = try to >>>> enable knem support, but continue even if it >>>> is not available, >>>> 0 = do not enable knem support, positive = >>>> try to enable knem >>>> support and fail if it is not available) >>>> MCA btl: parameter "btl_sm_knem_dma_min" (current >>>> value:<0>, data >>>> source: default value) >>>> Minimum message size (in bytes) to use the >>>> knem DMA mode; >>>> ignored if knem does not support DMA mode (0 >>>> = do not use the >>>> knem DMA mode) >>>> MCA btl: parameter "btl_sm_knem_max_simultaneous" >>>> (current value:<0>, >>>> data source: default value) >>>> Max number of simultaneous ongoing knem >>>> operations to support >>>> (0 = do everything synchronously, which >>>> probably gives the best >>>> large message latency;>0 means to do all >>>> operations >>>> asynchronously, which supports better >>>> overlap for simultaneous >>>> large message sends) >>>> MCA btl: parameter "btl_sm_free_list_num" (current >>>> value:<8>, data >>>> source: default value) >>>> MCA btl: parameter "btl_sm_free_list_max" (current >>>> value:<-1>, data >>>> source: default value) >>>> MCA btl: parameter "btl_sm_free_list_inc" (current >>>> value:<64>, data >>>> source: default value) >>>> MCA btl: parameter "btl_sm_max_procs" (current value: >>>> <-1>, data source: >>>> default value) >>>> MCA btl: parameter "btl_sm_mpool" (current value: >>>> <sm>, data source: >>>> default value) >>>> MCA btl: parameter "btl_sm_fifo_size" (current value: >>>> <4096>, data >>>> source: default value) >>>> MCA btl: parameter "btl_sm_num_fifos" (current value: >>>> <1>, data source: >>>> default value) >>>> MCA btl: parameter "btl_sm_fifo_lazy_free" (current >>>> value:<120>, data >>>> source: default value) >>>> MCA btl: parameter "btl_sm_sm_extra_procs" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA btl: parameter "btl_sm_exclusivity" (current >>>> value:<65535>, data >>>> source: default value) >>>> BTL exclusivity (must be>= 0) >>>> MCA btl: parameter "btl_sm_flags" (current value: >>>> <1>, data source: >>>> default value) >>>> BTL bit flags (general flags: SEND=1, PUT=2, >>>> GET=4, >>>> SEND_INPLACE=8, RDMA_MATCHED=64, >>>> HETEROGENEOUS_RDMA=256; flags >>>> only used by the "dr" PML (ignored by others): >>>> ACK=16, >>>> CHECKSUM=32, RDMA_COMPLETION=128; flags only >>>> used by the "bfo" >>>> PML (ignored by others): FAILOVER_SUPPORT=512) >>>> MCA btl: parameter "btl_sm_rndv_eager_limit" (current >>>> value:<4096>, >>>> data source: default value) >>>> Size (in bytes) of "phase 1" fragment sent >>>> for all large >>>> messages (must be>= 0 and<= eager_limit) >>>> MCA btl: parameter "btl_sm_eager_limit" (current >>>> value:<4096>, data >>>> source: default value) >>>> Maximum size (in bytes) of "short" messages >>>> (must be>= 1). >>>> MCA btl: parameter "btl_sm_max_send_size" (current >>>> value:<32768>, data >>>> source: default value) >>>> Maximum size (in bytes) of a single "phase >>>> 2" fragment of a >>>> long message when using the pipeline >>>> protocol (must be>= 1) >>>> MCA btl: parameter "btl_sm_bandwidth" (current value: >>>> <9000>, data >>>> source: default value) >>>> Approximate maximum bandwidth of >>>> interconnect(0 = auto-detect >>>> value at run-time [not supported in all BTL >>>> modules],>= 1 = >>>> bandwidth in Mbps) >>>> MCA btl: parameter "btl_sm_latency" (current value: >>>> <1>, data source: >>>> default value) >>>> Approximate latency of interconnect (must be>= 0) >>>> MCA btl: parameter "btl_sm_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA btl: parameter "btl_tcp_links" (current value: >>>> <1>, data source: >>>> default value) >>>> MCA btl: parameter "btl_tcp_if_include" (current >>>> value:<none>, data >>>> source: default value) >>>> Comma-delimited list of devices or CIDR >>>> notation of networks to >>>> use for MPI communication (e.g., "eth0,eth1" or >>>> "192.168.0.0/16,10.1.4.0/24"). Mutually >>>> exclusive with >>>> btl_tcp_if_exclude. >>>> MCA btl: parameter "btl_tcp_if_exclude" (current >>>> value:<lo,sppp>, data >>>> source: default value) >>>> Comma-delimited list of devices or CIDR >>>> notation of networks to >>>> NOT use for MPI communication -- all devices >>>> not matching these >>>> specifications will be used (e.g., "eth0,eth1" or >>>> "192.168.0.0/16,10.1.4.0/24"). Mutually >>>> exclusive with >>>> btl_tcp_if_include. >>>> MCA btl: parameter "btl_tcp_free_list_num" (current >>>> value:<8>, data >>>> source: default value) >>>> MCA btl: parameter "btl_tcp_free_list_max" (current >>>> value:<-1>, data >>>> source: default value) >>>> MCA btl: parameter "btl_tcp_free_list_inc" (current >>>> value:<32>, data >>>> source: default value) >>>> MCA btl: parameter "btl_tcp_sndbuf" (current value: >>>> <131072>, data >>>> source: default value) >>>> MCA btl: parameter "btl_tcp_rcvbuf" (current value: >>>> <131072>, data >>>> source: default value) >>>> MCA btl: parameter "btl_tcp_endpoint_cache" (current >>>> value:<30720>, >>>> data source: default value) >>>> The size of the internal cache for each TCP >>>> connection. This >>>> cache is used to reduce the number of >>>> syscalls, by replacing >>>> them with memcpy. Every read will read the >>>> expected data plus >>>> the amount of the endpoint_cache >>>> MCA btl: parameter "btl_tcp_use_nagle" (current >>>> value:<0>, data source: >>>> default value) >>>> Whether to use Nagle's algorithm or not (using >>>> Nagle's >>>> algorithm may increase short message latency) >>>> MCA btl: parameter "btl_tcp_port_min_v4" (current >>>> value:<1024>, data >>>> source: default value) >>>> The minimum port where the TCP BTL will try >>>> to bind (default >>>> 1024) >>>> MCA btl: parameter "btl_tcp_port_range_v4" (current >>>> value:<64511>, data >>>> source: default value) >>>> The number of ports where the TCP BTL will >>>> try to bind (default >>>> 64511). This parameter together with the >>>> port min, define a >>>> range of ports where Open MPI will open sockets. >>>> MCA btl: parameter "btl_tcp_exclusivity" (current >>>> value:<100>, data >>>> source: default value) >>>> BTL exclusivity (must be>= 0) >>>> MCA btl: parameter "btl_tcp_flags" (current value: >>>> <314>, data source: >>>> default value) >>>> BTL bit flags (general flags: SEND=1, PUT=2, >>>> GET=4, >>>> SEND_INPLACE=8, RDMA_MATCHED=64, >>>> HETEROGENEOUS_RDMA=256; flags >>>> only used by the "dr" PML (ignored by others): >>>> ACK=16, >>>> CHECKSUM=32, RDMA_COMPLETION=128; flags only >>>> used by the "bfo" >>>> PML (ignored by others): FAILOVER_SUPPORT=512) >>>> MCA btl: parameter "btl_tcp_rndv_eager_limit" >>>> (current value:<65536>, >>>> data source: default value) >>>> Size (in bytes) of "phase 1" fragment sent >>>> for all large >>>> messages (must be>= 0 and<= eager_limit) >>>> MCA btl: parameter "btl_tcp_eager_limit" (current >>>> value:<65536>, data >>>> source: default value) >>>> Maximum size (in bytes) of "short" messages >>>> (must be>= 1). >>>> MCA btl: parameter "btl_tcp_max_send_size" (current >>>> value:<131072>, >>>> data source: default value) >>>> Maximum size (in bytes) of a single "phase >>>> 2" fragment of a >>>> long message when using the pipeline >>>> protocol (must be>= 1) >>>> MCA btl: parameter >>>> "btl_tcp_rdma_pipeline_send_length" (current value: >>>> <131072>, data source: default value) >>>> Length of the "phase 2" portion of a large >>>> message (in bytes) >>>> when using the pipeline protocol. This part >>>> of the message >>>> will be split into fragments of size >>>> max_send_size and sent >>>> using send/receive semantics (must be>= 0; >>>> only relevant when >>>> the PUT flag is set) >>>> MCA btl: parameter "btl_tcp_rdma_pipeline_frag_size" >>>> (current value: >>>> <2147483647>, data source: default value) >>>> Maximum size (in bytes) of a single "phase >>>> 3" fragment from a >>>> long message when using the pipeline >>>> protocol. These fragments >>>> will be sent using RDMA semantics (must be >>>>> = 1; only relevant >>>> when the PUT flag is set) >>>> MCA btl: parameter "btl_tcp_min_rdma_pipeline_size" >>>> (current value:<0>, >>>> data source: default value) >>>> Messages smaller than this size (in bytes) >>>> will not use the >>>> RDMA pipeline protocol. Instead, they will >>>> be split into >>>> fragments of max_send_size and sent using >>>> send/receive >>>> semantics (must be>=0, and is automatically >>>> adjusted up to at >>>> least >>>> (eager_limit+btl_rdma_pipeline_send_length); only >>>> relevant when the PUT flag is set) >>>> MCA btl: parameter "btl_tcp_bandwidth" (current >>>> value:<100>, data >>>> source: default value) >>>> Approximate maximum bandwidth of >>>> interconnect(0 = auto-detect >>>> value at run-time [not supported in all BTL >>>> modules],>= 1 = >>>> bandwidth in Mbps) >>>> MCA btl: parameter "btl_tcp_latency" (current value: >>>> <100>, data source: >>>> default value) >>>> Approximate latency of interconnect (must be>= 0) >>>> MCA btl: parameter "btl_tcp_disable_family" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA btl: parameter "btl_tcp_if_seq" (current value: >>>> <none>, data source: >>>> default value) >>>> If specified, a comma-delimited list of TCP >>>> interfaces. >>>> Interfaces will be assigned, one to each MPI >>>> process, in a >>>> round-robin fashion on each server. For >>>> example, if the list >>>> is "eth0,eth1" and four MPI processes are >>>> run on a single >>>> server, then local ranks 0 and 2 will use >>>> eth0 and local ranks >>>> 1 and 3 will use eth1. >>>> MCA btl: parameter "btl_tcp_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA btl: parameter "btl_base_include" (current value: >>>> <none>, data >>>> source: default value) >>>> MCA btl: parameter "btl_base_exclude" (current value: >>>> <none>, data >>>> source: default value) >>>> MCA btl: parameter "btl_base_warn_component_unused" >>>> (current value:<1>, >>>> data source: default value) >>>> This parameter is used to turn on warning >>>> messages when certain >>>> NICs are not used >>>> MCA mtl: parameter "mtl" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> mtl framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA mtl: parameter "mtl_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the mtl framework (0 = >>>> no verbosity) >>>> MCA topo: parameter "topo" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> topo framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA topo: parameter "topo_base_verbose" (current >>>> value:<0>, data source: >>>> default value) >>>> Verbosity level for the topo framework (0 = >>>> no verbosity) >>>> MCA topo: parameter "topo_unity_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA osc: parameter "osc" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> osc framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA osc: parameter "osc_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the osc framework (0 = >>>> no verbosity) >>>> MCA osc: parameter "osc_pt2pt_no_locks" (current >>>> value:<0>, data >>>> source: default value) >>>> Enable optimizations available only if >>>> MPI_LOCK is not used. >>>> MCA osc: parameter "osc_pt2pt_eager_limit" (current >>>> value:<16384>, data >>>> source: default value) >>>> Max size of eagerly sent data >>>> MCA osc: parameter "osc_pt2pt_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA osc: parameter "osc_rdma_eager_send" (current >>>> value:<1>, data >>>> source: default value) >>>> Attempt to start data movement during >>>> communication call, >>>> instead of at synchrnoization time. Info >>>> key of same name >>>> overrides this value. >>>> MCA osc: parameter "osc_rdma_use_buffers" (current >>>> value:<1>, data >>>> source: default value) >>>> Coalesce messages during an epoch to reduce >>>> network >>>> utilization. Info key of same name >>>> overrides this value. >>>> MCA osc: parameter "osc_rdma_use_rdma" (current >>>> value:<0>, data source: >>>> default value) >>>> Use real RDMA operations to transfer data. >>>> Info key of same >>>> name overrides this value. >>>> MCA osc: parameter "osc_rdma_rdma_completion_wait" >>>> (current value:<1>, >>>> data source: default value) >>>> Wait for all completion of rdma events before >>>> sending >>>> acknowledgment. Info key of same name >>>> overrides this value. >>>> MCA osc: parameter "osc_rdma_no_locks" (current >>>> value:<0>, data source: >>>> default value) >>>> Enable optimizations available only if >>>> MPI_LOCK is not used. >>>> Info key of same name overrides this value. >>>> MCA osc: parameter "osc_rdma_priority" (current >>>> value:<0>, data source: >>>> default value) >>>> MCA op: parameter "op_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level of the op framework >>>> MCA iof: parameter "iof" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> iof framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA iof: parameter "iof_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the iof framework (0 = >>>> no verbosity) >>>> MCA iof: parameter "iof_hnp_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA iof: parameter "iof_orted_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA iof: parameter "iof_tool_priority" (current >>>> value:<0>, data source: >>>> default value) >>>> MCA oob: parameter "oob" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> oob framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA oob: parameter "oob_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the oob framework (0 = >>>> no verbosity) >>>> MCA oob: parameter "oob_tcp_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbose level for the OOB tcp component >>>> MCA oob: parameter "oob_tcp_peer_limit" (current >>>> value:<-1>, data >>>> source: default value) >>>> Maximum number of peer connections to >>>> simultaneously maintain >>>> (-1 = infinite) >>>> MCA oob: parameter "oob_tcp_peer_retries" (current >>>> value:<60>, data >>>> source: default value) >>>> Number of times to try shutting down a >>>> connection before giving >>>> up >>>> MCA oob: parameter "oob_tcp_debug" (current value: >>>> <0>, data source: >>>> default value) >>>> Enable (1) / disable (0) debugging output >>>> for this component >>>> MCA oob: parameter "oob_tcp_sndbuf" (current value: >>>> <131072>, data >>>> source: default value) >>>> TCP socket send buffering size (in bytes) >>>> MCA oob: parameter "oob_tcp_rcvbuf" (current value: >>>> <131072>, data >>>> source: default value) >>>> TCP socket receive buffering size (in bytes) >>>> MCA oob: parameter "oob_tcp_if_include" (current >>>> value:<none>, data >>>> source: default value) >>>> Comma-delimited list of TCP interfaces to use >>>> MCA oob: parameter "oob_tcp_if_exclude" (current >>>> value:<none>, data >>>> source: default value) >>>> Comma-delimited list of TCP interfaces to exclude >>>> MCA oob: parameter "oob_tcp_connect_sleep" (current >>>> value:<1>, data >>>> source: default value) >>>> Enable (1) / disable (0) random sleep for >>>> connection wireup. >>>> MCA oob: parameter "oob_tcp_listen_mode" (current >>>> value:<event>, data >>>> source: default value) >>>> Mode for HNP to accept incoming connections: >>>> event, >>>> listen_thread. >>>> MCA oob: parameter "oob_tcp_listen_thread_max_queue" >>>> (current value: >>>> <10>, data source: default value) >>>> High water mark for queued accepted socket >>>> list size. Used >>>> only when listen_mode is listen_thread. >>>> MCA oob: parameter "oob_tcp_listen_thread_wait_time" >>>> (current value: >>>> <10>, data source: default value) >>>> Time in milliseconds to wait before actively >>>> checking for new >>>> connections when listen_mode is listen_thread. >>>> MCA oob: parameter "oob_tcp_static_ports" (current >>>> value:<none>, data >>>> source: default value) >>>> Static ports for daemons and procs (IPv4) >>>> MCA oob: parameter "oob_tcp_dynamic_ports" (current >>>> value:<none>, data >>>> source: default value) >>>> Range of ports to be dynamically used by >>>> daemons and procs >>>> (IPv4) >>>> MCA oob: parameter "oob_tcp_disable_family" (current >>>> value:<0>, data >>>> source: default value) >>>> Disable IPv4 (4) or IPv6 (6) >>>> MCA oob: parameter "oob_tcp_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA odls: parameter "odls_base_sigkill_timeout" >>>> (current value:<1>, data >>>> source: default value) >>>> Time to wait for a process to die after >>>> issuing a kill signal >>>> to it >>>> MCA odls: parameter "odls" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> odls framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA odls: parameter "odls_base_verbose" (current >>>> value:<0>, data source: >>>> default value) >>>> Verbosity level for the odls framework (0 = >>>> no verbosity) >>>> MCA odls: parameter "odls_default_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA ras: parameter "ras_base_display_alloc" (current >>>> value:<0>, data >>>> source: default value) >>>> Whether to display the allocation after it >>>> is determined >>>> MCA ras: parameter "ras_base_display_devel_alloc" >>>> (current value:<0>, >>>> data source: default value) >>>> Whether to display a developer-detail >>>> allocation after it is >>>> determined >>>> MCA ras: parameter "ras" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> ras framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA ras: parameter "ras_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the ras framework (0 = >>>> no verbosity) >>>> MCA ras: parameter "ras_cm_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA ras: parameter "ras_loadleveler_priority" >>>> (current value:<90>, data >>>> source: default value) >>>> Priority of the loadleveler ras component >>>> MCA ras: parameter "ras_slurm_priority" (current >>>> value:<75>, data >>>> source: default value) >>>> Priority of the slurm ras component >>>> MCA rmaps: parameter "rmaps_rank_file_path" (current >>>> value:<none>, data >>>> source: default value, synonym of: orte_rankfile) >>>> Name of the rankfile to be used for mapping >>>> processes (relative >>>> or absolute path) >>>> MCA rmaps: parameter "rmaps_base_schedule_policy" >>>> (current value:<slot>, >>>> data source: default value) >>>> Scheduling Policy for RMAPS. [slot >>>> (alias:core) | socket | >>>> board | node] >>>> MCA rmaps: parameter "rmaps_base_pernode" (current >>>> value:<0>, data >>>> source: default value) >>>> Launch one ppn as directed >>>> MCA rmaps: parameter "rmaps_base_n_pernode" (current >>>> value:<-1>, data >>>> source: default value) >>>> Launch n procs/node >>>> MCA rmaps: parameter "rmaps_base_n_perboard" (current >>>> value:<-1>, data >>>> source: default value) >>>> Launch n procs/board >>>> MCA rmaps: parameter "rmaps_base_n_persocket" (current >>>> value:<-1>, data >>>> source: default value) >>>> Launch n procs/socket >>>> MCA rmaps: parameter "rmaps_base_loadbalance" (current >>>> value:<0>, data >>>> source: default value) >>>> Balance total number of procs across all >>>> allocated nodes >>>> MCA rmaps: parameter "rmaps_base_cpus_per_proc" >>>> (current value:<1>, data >>>> source: default value, synonyms: >>>> rmaps_base_cpus_per_rank) >>>> Number of cpus to use for each rank [1-2**15 >>>> (default=1)] >>>> MCA rmaps: parameter "rmaps_base_cpus_per_rank" >>>> (current value:<1>, data >>>> source: default value, synonym of: >>>> rmaps_base_cpus_per_proc) >>>> Number of cpus to use for each rank [1-2**15 >>>> (default=1)] >>>> MCA rmaps: parameter "rmaps_base_stride" (current >>>> value:<1>, data source: >>>> default value) >>>> When binding multiple cores to a rank, the >>>> step size to use >>>> between cores [1-2**15 (default: 1)] >>>> MCA rmaps: parameter "rmaps_base_slot_list" (current >>>> value:<none>, data >>>> source: default value) >>>> List of processor IDs to bind MPI processes >>>> to (e.g., used in >>>> conjunction with rank files) [default=NULL] >>>> MCA rmaps: parameter "rmaps_base_no_schedule_local" >>>> (current value:<0>, >>>> data source: default value) >>>> If false, allow scheduling MPI applications >>>> on the same node as >>>> mpirun (default). If true, do not schedule any >>>> MPI >>>> applications on the same node as mpirun >>>> MCA rmaps: parameter "rmaps_base_no_oversubscribe" >>>> (current value:<0>, >>>> data source: default value) >>>> If true, then do not allow oversubscription >>>> of nodes - mpirun >>>> will return an error if there aren't enough >>>> nodes to launch all >>>> processes without oversubscribing >>>> MCA rmaps: parameter "rmaps_base_display_map" (current >>>> value:<0>, data >>>> source: default value) >>>> Whether to display the process map after it >>>> is computed >>>> MCA rmaps: parameter "rmaps_base_display_devel_map" >>>> (current value:<0>, >>>> data source: default value) >>>> Whether to display a developer-detail >>>> process map after it is >>>> computed >>>> MCA rmaps: parameter "rmaps" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> rmaps framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA rmaps: parameter "rmaps_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the rmaps framework (0 = >>>> no verbosity) >>>> MCA rmaps: parameter "rmaps_load_balance_priority" >>>> (current value:<0>, >>>> data source: default value) >>>> MCA rmaps: parameter "rmaps_rank_file_priority" >>>> (current value:<0>, data >>>> source: default value) >>>> MCA rmaps: parameter "rmaps_resilient_fault_grp_file" >>>> (current value: >>>> <none>, data source: default value) >>>> Filename that contains a description of >>>> fault groups for this >>>> system >>>> MCA rmaps: parameter "rmaps_resilient_priority" >>>> (current value:<0>, data >>>> source: default value) >>>> MCA rmaps: parameter "rmaps_round_robin_priority" >>>> (current value:<0>, >>>> data source: default value) >>>> MCA rmaps: parameter "rmaps_seq_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA rmaps: parameter "rmaps_topo_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA rml: parameter "rml_wrapper" (current value: >>>> <none>, data source: >>>> default value) >>>> Use a Wrapper component around the selected >>>> RML component >>>> MCA rml: parameter "rml" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> rml framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA rml: parameter "rml_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the rml framework (0 = >>>> no verbosity) >>>> MCA rml: parameter "rml_oob_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA routed: parameter "routed" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> routed framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA routed: parameter "routed_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the routed framework (0 >>>> = no verbosity) >>>> MCA routed: parameter "routed_binomial_priority" >>>> (current value:<0>, data >>>> source: default value) >>>> MCA routed: parameter "routed_cm_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA routed: parameter "routed_direct_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA routed: parameter "routed_linear_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA routed: parameter "routed_radix_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA routed: parameter "routed_slave_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA plm: parameter "plm_rsh_agent" (current value: >>>> <ssh : rsh>, data >>>> source: default value, deprecated, synonym >>>> of: orte_rsh_agent) >>>> The command used to launch executables on remote >>>> nodes >>>> (typically either "ssh" or "rsh") >>>> MCA plm: parameter "plm_rsh_assume_same_shell" >>>> (current value:<1>, data >>>> source: default value, deprecated, synonym of: >>>> orte_assume_same_shell) >>>> If set to 1, assume that the shell on the >>>> remote node is the >>>> same as the shell on the local node. >>>> Otherwise, probe for what >>>> the remote shell [default: 1] >>>> MCA plm: parameter "plm" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> plm framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA plm: parameter "plm_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the plm framework (0 = >>>> no verbosity) >>>> MCA plm: parameter "plm_rsh_num_concurrent" (current >>>> value:<128>, data >>>> source: default value) >>>> How many plm_rsh_agent instances to invoke >>>> concurrently (must >>>> be> 0) >>>> MCA plm: parameter "plm_rsh_force_rsh" (current >>>> value:<0>, data source: >>>> default value) >>>> Force the launcher to always use rsh >>>> MCA plm: parameter "plm_rsh_disable_qrsh" (current >>>> value:<0>, data >>>> source: default value) >>>> Disable the launcher to use qrsh when under >>>> the SGE parallel >>>> environment >>>> MCA plm: parameter "plm_rsh_daemonize_qrsh" (current >>>> value:<0>, data >>>> source: default value) >>>> Daemonize the orted under the SGE parallel >>>> environment >>>> MCA plm: parameter "plm_rsh_disable_llspawn" (current >>>> value:<0>, data >>>> source: default value) >>>> Disable the use of llspawn when under the >>>> LoadLeveler >>>> environment >>>> MCA plm: parameter "plm_rsh_daemonize_llspawn" >>>> (current value:<0>, data >>>> source: default value) >>>> Daemonize the orted when under the >>>> LoadLeveler environment >>>> MCA plm: parameter "plm_rsh_priority" (current value: >>>> <10>, data source: >>>> default value) >>>> Priority of the rsh plm component >>>> MCA plm: parameter "plm_rsh_delay" (current value: >>>> <1>, data source: >>>> default value) >>>> Delay (in seconds) between invocations of >>>> the remote agent, but >>>> only used when the "debug" MCA parameter is >>>> true, or the >>>> top-level MCA debugging is enabled >>>> (otherwise this value is >>>> ignored) >>>> MCA plm: parameter "plm_rsh_tree_spawn" (current >>>> value:<0>, data >>>> source: default value) >>>> If set to 1, launch via a tree-based topology >>>> MCA plm: parameter "plm_slurm_args" (current value: >>>> <none>, data source: >>>> default value) >>>> Custom arguments to srun >>>> MCA plm: parameter "plm_slurm_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA filem: parameter "filem" (current value:<none>, >>>> data source: default >>>> value) >>>> Which Filem component to use (empty = >>>> auto-select) >>>> MCA filem: parameter "filem_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the filem framework (0 = >>>> no verbosity) >>>> MCA filem: parameter "filem_rsh_priority" (current >>>> value:<20>, data >>>> source: default value) >>>> Priority of the FILEM rsh component >>>> MCA filem: parameter "filem_rsh_verbose" (current >>>> value:<0>, data source: >>>> default value) >>>> Verbose level for the FILEM rsh component >>>> MCA filem: parameter "filem_rsh_rcp" (current value: >>>> <scp>, data source: >>>> default value) >>>> The rsh cp command for the FILEM rsh component >>>> MCA filem: parameter "filem_rsh_cp" (current value: >>>> <cp>, data source: >>>> default value) >>>> The Unix cp command for the FILEM rsh component >>>> MCA filem: parameter "filem_rsh_rsh" (current value: >>>> <ssh>, data source: >>>> default value) >>>> The remote shell command for the FILEM rsh >>>> component >>>> MCA filem: parameter "filem_rsh_max_incomming" (current >>>> value:<10>, data >>>> source: default value) >>>> Maximum number of incomming connections (0 = any) >>>> MCA filem: parameter "filem_rsh_max_outgoing" (current >>>> value:<10>, data >>>> source: default value) >>>> Maximum number of out going connections (0 = any) >>>> MCA errmgr: parameter "errmgr" (current value:<none>, >>>> data source: default >>>> value) >>>> Default selection set of components for the >>>> errmgr framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA errmgr: parameter "errmgr_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the errmgr framework (0 >>>> = no verbosity) >>>> MCA errmgr: parameter "errmgr_default_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA ess: parameter "ess" (current value:<none>, data >>>> source: default >>>> value) >>>> Default selection set of components for the >>>> ess framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA ess: parameter "ess_base_verbose" (current value: >>>> <0>, data source: >>>> default value) >>>> Verbosity level for the ess framework (0 = >>>> no verbosity) >>>> MCA ess: parameter "ess_env_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA ess: parameter "ess_hnp_priority" (current value: >>>> <0>, data source: >>>> default value) >>>> MCA ess: parameter "ess_singleton_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA ess: parameter "ess_slave_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA ess: parameter "ess_slurm_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA ess: parameter "ess_slurmd_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA ess: parameter "ess_tool_priority" (current >>>> value:<0>, data source: >>>> default value) >>>> MCA grpcomm: parameter "grpcomm" (current value:<none>, >>>> data source: >>>> default value) >>>> Default selection set of components for the >>>> grpcomm framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA grpcomm: parameter "grpcomm_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the grpcomm framework (0 >>>> = no verbosity) >>>> MCA grpcomm: parameter "grpcomm_bad_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA grpcomm: parameter "grpcomm_basic_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA grpcomm: parameter "grpcomm_hier_priority" (current >>>> value:<0>, data >>>> source: default value) >>>> MCA notifier: parameter "notifier_threshold_severity" >>>> (current value: >>>> <critical>, data source: default value) >>>> Report all events at or above this severity >>>> [default: critical] >>>> MCA notifier: parameter "notifier" (current value:<none>, >>>> data source: >>>> default value) >>>> Default selection set of components for the >>>> notifier framework >>>> (<none> means use all components that can be >>>> found) >>>> MCA notifier: parameter "notifier_base_verbose" (current >>>> value:<0>, data >>>> source: default value) >>>> Verbosity level for the notifier framework >>>> (0 = no verbosity) >>>> MCA notifier: parameter "notifier_command_cmd" (current >>>> value:</sbin/initlog >>>> -f $s -n "Open MPI" -s "$S: $m (errorcode: >>>> $e)">, data source: >>>> default value) >>>> Command to execute, with substitution. $s = >>>> integer severity; >>>> $S = string severity; $e = integer error >>>> code; $m = string >>>> message >>>> MCA notifier: parameter "notifier_command_timeout" >>>> (current value:<30>, data >>>> source: default value) >>>> Timeout (in seconds) of the command >>>> MCA notifier: parameter "notifier_command_priority" >>>> (current value:<10>, >>>> data source: default value) >>>> Priority of this component >>>> MCA notifier: parameter "notifier_syslog_priority" >>>> (current value:<0>, data >>>> source: default value) >>>> >>>> ==================================================================================================== >>>> >>>> output of cat /proc/cpuinfo >>>> processor : 0 >>>> vendor_id : AuthenticAMD >>>> cpu family : 15 >>>> model : 75 >>>> model name : AMD Athlon(tm) 64 X2 Dual Core Processor 3800+ >>>> stepping : 2 >>>> cpu MHz : 1002.094 >>>> cache size : 512 KB >>>> physical id : 0 >>>> siblings : 2 >>>> core id : 0 >>>> cpu cores : 2 >>>> fpu : yes >>>> fpu_exception : yes >>>> cpuid level : 1 >>>> wp : yes >>>> flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca >>>> cmov >>>> pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext lm 3dnowext >>>> 3dnow pni cx16 >>>> bogomips : 2003.90 >>>> TLB size : 1088 4K pages >>>> clflush size : 64 >>>> cache_alignment : 64 >>>> address sizes : 40 bits physical, 48 bits virtual >>>> power management: ts fid vid ttp [4] [5] >>>> >>>> processor : 1 >>>> vendor_id : AuthenticAMD >>>> cpu family : 15 >>>> model : 75 >>>> model name : AMD Athlon(tm) 64 X2 Dual Core Processor 3800+ >>>> stepping : 2 >>>> cpu MHz : 1002.094 >>>> cache size : 512 KB >>>> physical id : 0 >>>> siblings : 2 >>>> core id : 1 >>>> cpu cores : 2 >>>> fpu : yes >>>> fpu_exception : yes >>>> cpuid level : 1 >>>> wp : yes >>>> flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca >>>> cmov >>>> pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext lm 3dnowext >>>> 3dnow pni cx16 >>>> bogomips : 2003.90 >>>> TLB size : 1088 4K pages >>>> clflush size : 64 >>>> cache_alignment : 64 >>>> address sizes : 40 bits physical, 48 bits virtual >>>> power management: ts fid vid ttp [4] [5] >>>> >>>> >>>> ==================================================================================================== >>>> output of ifconfig -a from a compute node >>>> eth0 Link encap:Ethernet HWaddr 00:18:F3:3F:84:A1 >>>> inet addr:192.168.0.2 Bcast:192.168.0.255 Mask:255.255.255.0 >>>> inet6 addr: fe80::218:f3ff:fe3f:84a1/64 Scope:Link >>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>> RX packets:2006 errors:0 dropped:0 overruns:0 frame:0 >>>> TX packets:2064 errors:0 dropped:0 overruns:0 carrier:0 >>>> collisions:0 txqueuelen:1000 >>>> RX bytes:242685 (236.9 KiB) TX bytes:0 (0.0 b) >>>> Interrupt:11 Base address:0x8000 >>>> >>>> lo Link encap:Local Loopback >>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>> inet6 addr: ::1/128 Scope:Host >>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>> RX packets:60 errors:0 dropped:0 overruns:0 frame:0 >>>> TX packets:60 errors:0 dropped:0 overruns:0 carrier:0 >>>> collisions:0 txqueuelen:0 >>>> RX bytes:4440 (4.3 KiB) TX bytes:4440 (4.3 KiB) >>>> >>>> sit0 Link encap:IPv6-in-IPv4 >>>> NOARP MTU:1480 Metric:1 >>>> RX packets:0 errors:0 dropped:0 overruns:0 frame:0 >>>> TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 >>>> collisions:0 txqueuelen:0 >>>> RX bytes:0 (0.0 b) TX bytes:0 (0.0 b) >>>> >>>> >>>> ==================================================================================================== >>>> output of ifconfig -a from the login node, where I run mpirun >>>> >>>> eth0 Link encap:Ethernet HWaddr 00:18:F3:51:B3:6E >>>> inet addr:192.168.0.1 Bcast:192.168.0.255 Mask:255.255.255.0 >>>> inet6 addr: fe80::218:f3ff:fe51:b36e/64 Scope:Link >>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>> RX packets:7180758 errors:0 dropped:0 overruns:0 frame:0 >>>> TX packets:4989496 errors:0 dropped:0 overruns:0 carrier:0 >>>> collisions:0 txqueuelen:1000 >>>> RX bytes:6045614452 (5.6 GiB) TX bytes:0 (0.0 b) >>>> Interrupt:201 Base address:0xe000 >>>> >>>> eth1 Link encap:Ethernet HWaddr 00:01:02:13:AA:3C >>>> inet addr:137.204.66.188 Bcast:137.204.66.255 >>>> Mask:255.255.255.0 >>>> inet6 addr: fe80::201:2ff:fe13:aa3c/64 Scope:Link >>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>> RX packets:4750212 errors:0 dropped:0 overruns:0 frame:0 >>>> TX packets:405027 errors:0 dropped:0 overruns:0 carrier:0 >>>> collisions:0 txqueuelen:1000 >>>> RX bytes:629146679 (600.0 MiB) TX bytes:332118265 (316.7 MiB) >>>> Interrupt:177 Base address:0x9c00 >>>> >>>> lo Link encap:Local Loopback >>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>> inet6 addr: ::1/128 Scope:Host >>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>> RX packets:288455 errors:0 dropped:0 overruns:0 frame:0 >>>> TX packets:288455 errors:0 dropped:0 overruns:0 carrier:0 >>>> collisions:0 txqueuelen:0 >>>> RX bytes:35908038 (34.2 MiB) TX bytes:35908038 (34.2 MiB) >>>> >>>> sit0 Link encap:IPv6-in-IPv4 >>>> NOARP MTU:1480 Metric:1 >>>> RX packets:0 errors:0 dropped:0 overruns:0 frame:0 >>>> TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 >>>> collisions:0 txqueuelen:0 >>>> RX bytes:0 (0.0 b) TX bytes:0 (0.0 b) >>>> >>>> ==================================================================================================== >>>> output of mpirun --bynode --hostfile my_hostfile.txt --tag-output >>>> ompi_info -v ompi full --parsable >>>> >>>> [1,0]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,0]<stdout>:ompi:version:full:1.6 >>>> [1,0]<stdout>:ompi:version:svn:r26429 >>>> [1,0]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,0]<stdout>:orte:version:full:1.6 >>>> [1,0]<stdout>:orte:version:svn:r26429 >>>> [1,0]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,0]<stdout>:opal:version:full:1.6 >>>> [1,0]<stdout>:opal:version:svn:r26429 >>>> [1,0]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,0]<stdout>:mpi-api:version:full:2.1 >>>> [1,0]<stdout>:ident:1.6 >>>> [1,6]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,6]<stdout>:ompi:version:full:1.6 >>>> [1,6]<stdout>:ompi:version:svn:r26429 >>>> [1,6]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,6]<stdout>:orte:version:full:1.6 >>>> [1,6]<stdout>:orte:version:svn:r26429 >>>> [1,6]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,6]<stdout>:opal:version:full:1.6 >>>> [1,6]<stdout>:opal:version:svn:r26429 >>>> [1,6]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,6]<stdout>:mpi-api:version:full:2.1 >>>> [1,6]<stdout>:ident:1.6 >>>> [1,9]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,10]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,3]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,3]<stdout>:ompi:version:full:1.6 >>>> [1,3]<stdout>:ompi:version:svn:r26429 >>>> [1,3]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,3]<stdout>:orte:version:full:1.6 >>>> [1,3]<stdout>:orte:version:svn:r26429 >>>> [1,3]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,3]<stdout>:opal:version:full:1.6 >>>> [1,3]<stdout>:opal:version:svn:r26429 >>>> [1,3]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,3]<stdout>:mpi-api:version:full:2.1 >>>> [1,3]<stdout>:ident:1.6 >>>> [1,4]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,4]<stdout>:ompi:version:full:1.6 >>>> [1,4]<stdout>:ompi:version:svn:r26429 >>>> [1,4]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,4]<stdout>:orte:version:full:1.6 >>>> [1,4]<stdout>:orte:version:svn:r26429 >>>> [1,4]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,4]<stdout>:opal:version:full:1.6 >>>> [1,9]<stdout>:ompi:version:full:1.6 >>>> [1,4]<stdout>:opal:version:svn:r26429 >>>> [1,4]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,4]<stdout>:mpi-api:version:full:2.1 >>>> [1,4]<stdout>:ident:1.6 >>>> [1,9]<stdout>:ompi:version:svn:r26429 >>>> [1,10]<stdout>:ompi:version:full:1.6 >>>> [1,9]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,10]<stdout>:ompi:version:svn:r26429 >>>> [1,9]<stdout>:orte:version:full:1.6 >>>> [1,10]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,9]<stdout>:orte:version:svn:r26429 >>>> [1,10]<stdout>:orte:version:full:1.6 >>>> [1,10]<stdout>:orte:version:svn:r26429 >>>> [1,9]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,10]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,9]<stdout>:opal:version:full:1.6 >>>> [1,10]<stdout>:opal:version:full:1.6 >>>> [1,9]<stdout>:opal:version:svn:r26429 >>>> [1,10]<stdout>:opal:version:svn:r26429 >>>> [1,9]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,10]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,9]<stdout>:mpi-api:version:full:2.1 >>>> [1,9]<stdout>:ident:1.6 >>>> [1,10]<stdout>:mpi-api:version:full:2.1 >>>> [1,10]<stdout>:ident:1.6 >>>> [1,2]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,2]<stdout>:ompi:version:full:1.6 >>>> [1,2]<stdout>:ompi:version:svn:r26429 >>>> [1,2]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,2]<stdout>:orte:version:full:1.6 >>>> [1,2]<stdout>:orte:version:svn:r26429 >>>> [1,2]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,2]<stdout>:opal:version:full:1.6 >>>> [1,2]<stdout>:opal:version:svn:r26429 >>>> [1,2]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,2]<stdout>:mpi-api:version:full:2.1 >>>> [1,2]<stdout>:ident:1.6 >>>> [1,8]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,8]<stdout>:ompi:version:full:1.6 >>>> [1,8]<stdout>:ompi:version:svn:r26429 >>>> [1,8]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,8]<stdout>:orte:version:full:1.6 >>>> [1,8]<stdout>:orte:version:svn:r26429 >>>> [1,8]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,8]<stdout>:opal:version:full:1.6 >>>> [1,8]<stdout>:opal:version:svn:r26429 >>>> [1,8]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,8]<stdout>:mpi-api:version:full:2.1 >>>> [1,8]<stdout>:ident:1.6 >>>> [1,11]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,11]<stdout>:ompi:version:full:1.6 >>>> [1,11]<stdout>:ompi:version:svn:r26429 >>>> [1,11]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,11]<stdout>:orte:version:full:1.6 >>>> [1,11]<stdout>:orte:version:svn:r26429 >>>> [1,11]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,11]<stdout>:opal:version:full:1.6 >>>> [1,11]<stdout>:opal:version:svn:r26429 >>>> [1,11]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,11]<stdout>:mpi-api:version:full:2.1 >>>> [1,11]<stdout>:ident:1.6 >>>> [1,5]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,5]<stdout>:ompi:version:full:1.6 >>>> [1,5]<stdout>:ompi:version:svn:r26429 >>>> [1,5]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,5]<stdout>:orte:version:full:1.6 >>>> [1,5]<stdout>:orte:version:svn:r26429 >>>> [1,5]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,5]<stdout>:opal:version:full:1.6 >>>> [1,5]<stdout>:opal:version:svn:r26429 >>>> [1,5]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,5]<stdout>:mpi-api:version:full:2.1 >>>> [1,5]<stdout>:ident:1.6 >>>> [1,1]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,7]<stdout>:package:Open MPI and...@cloud.bo.astro.it Distribution >>>> [1,7]<stdout>:ompi:version:full:1.6 >>>> [1,7]<stdout>:ompi:version:svn:r26429 >>>> [1,7]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,7]<stdout>:orte:version:full:1.6 >>>> [1,7]<stdout>:orte:version:svn:r26429 >>>> [1,7]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,7]<stdout>:opal:version:full:1.6 >>>> [1,7]<stdout>:opal:version:svn:r26429 >>>> [1,7]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,7]<stdout>:mpi-api:version:full:2.1 >>>> [1,7]<stdout>:ident:1.6 >>>> [1,1]<stdout>:ompi:version:full:1.6 >>>> [1,1]<stdout>:ompi:version:svn:r26429 >>>> [1,1]<stdout>:ompi:version:release_date:May 10, 2012 >>>> [1,1]<stdout>:orte:version:full:1.6 >>>> [1,1]<stdout>:orte:version:svn:r26429 >>>> [1,1]<stdout>:orte:version:release_date:May 10, 2012 >>>> [1,1]<stdout>:opal:version:full:1.6 >>>> [1,1]<stdout>:opal:version:svn:r26429 >>>> [1,1]<stdout>:opal:version:release_date:May 10, 2012 >>>> [1,1]<stdout>:mpi-api:version:full:2.1 >>>> [1,1]<stdout>:ident:1.6 >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> >>> >>> ------------------------------ >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> End of users Digest, Vol 2339, Issue 5 >>> ************************************** >> >> >> ------------------------------ >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> End of users Digest, Vol 2340, Issue 1 >> ************************************** > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users