Okay - so effectively you have no hostfile, and no allocation. So this is running just on the one node where mpirun exists?
Add “-mca ras_base_verbose 10 --display-allocation” to your cmd line and let’s see what it found > On Apr 25, 2017, at 12:56 PM, Eric Chamberland > <eric.chamberl...@giref.ulaval.ca> wrote: > > Hi, > > the host file has been constructed automatically by the > configuration+installation process and seems to contain only comments and a > blank line: > > (15:53:50) [zorg]:~> cat /opt/openmpi-3.x_debug/etc/openmpi-default-hostfile > # > # Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana > # University Research and Technology > # Corporation. All rights reserved. > # Copyright (c) 2004-2005 The University of Tennessee and The University > # of Tennessee Research Foundation. All rights > # reserved. > # Copyright (c) 2004-2005 High Performance Computing Center Stuttgart, > # University of Stuttgart. All rights reserved. > # Copyright (c) 2004-2005 The Regents of the University of California. > # All rights reserved. > # $COPYRIGHT$ > # > # Additional copyrights may follow > # > # $HEADER$ > # > # This is the default hostfile for Open MPI. Notice that it does not > # contain any hosts (not even localhost). This file should only > # contain hosts if a system administrator wants users to always have > # the same set of default hosts, and is not using a batch scheduler > # (such as SLURM, PBS, etc.). > # > # Note that this file is *not* used when running in "managed" > # environments (e.g., running in a job under a job scheduler, such as > # SLURM or PBS / Torque). > # > # If you are primarily interested in running Open MPI on one node, you > # should *not* simply list "localhost" in here (contrary to prior MPI > # implementations, such as LAM/MPI). A localhost-only node list is > # created by the RAS component named "localhost" if no other RAS > # components were able to find any hosts to run on (this behavior can > # be disabled by excluding the localhost RAS component by specifying > # the value "^localhost" [without the quotes] to the "ras" MCA > # parameter). > > (15:53:52) [zorg]:~> > > Thanks! > > Eric > > > On 25/04/17 03:52 PM, r...@open-mpi.org wrote: >> What is in your hostfile? >> >> >>> On Apr 25, 2017, at 11:39 AM, Eric Chamberland >>> <eric.chamberl...@giref.ulaval.ca> wrote: >>> >>> Hi, >>> >>> just testing the 3.x branch... I launch: >>> >>> mpirun -n 8 echo "hello" >>> >>> and I get: >>> >>> -------------------------------------------------------------------------- >>> There are not enough slots available in the system to satisfy the 8 slots >>> that were requested by the application: >>> echo >>> >>> Either request fewer slots for your application, or make more slots >>> available >>> for use. >>> -------------------------------------------------------------------------- >>> >>> I have to oversubscribe, so what do I have to do to bypass this >>> "limitation"? >>> >>> Thanks, >>> >>> Eric >>> >>> configure log: >>> >>> http://www.giref.ulaval.ca/~cmpgiref/ompi_3.x/2017.04.25.10h46m08s_config.log >>> http://www.giref.ulaval.ca/~cmpgiref/ompi_3.x/2017.04.25.10h46m08s_ompi_info_all.txt >>> >>> >>> here is the complete message: >>> >>> [zorg:30036] [[INVALID],INVALID] plm:rsh_lookup on agent ssh : rsh path NULL >>> [zorg:30036] plm:base:set_hnp_name: initial bias 30036 nodename hash >>> 810220270 >>> [zorg:30036] plm:base:set_hnp_name: final jobfam 49136 >>> [zorg:30036] [[49136,0],0] plm:rsh_setup on agent ssh : rsh path NULL >>> [zorg:30036] [[49136,0],0] plm:base:receive start comm >>> [zorg:30036] [[49136,0],0] plm:base:setup_job >>> [zorg:30036] [[49136,0],0] plm:base:setup_vm >>> [zorg:30036] [[49136,0],0] plm:base:setup_vm creating map >>> [zorg:30036] [[49136,0],0] setup:vm: working unmanaged allocation >>> [zorg:30036] [[49136,0],0] using default hostfile >>> /opt/openmpi-3.x_debug/etc/openmpi-default-hostfile >>> [zorg:30036] [[49136,0],0] plm:base:setup_vm only HNP in allocation >>> [zorg:30036] [[49136,0],0] plm:base:setting slots for node zorg by cores >>> [zorg:30036] [[49136,0],0] complete_setup on job [49136,1] >>> [zorg:30036] [[49136,0],0] plm:base:launch_apps for job [49136,1] >>> -------------------------------------------------------------------------- >>> There are not enough slots available in the system to satisfy the 8 slots >>> that were requested by the application: >>> echo >>> >>> Either request fewer slots for your application, or make more slots >>> available >>> for use. >>> -------------------------------------------------------------------------- >>> [zorg:30036] [[49136,0],0] plm:base:orted_cmd sending orted_exit commands >>> [zorg:30036] [[49136,0],0] plm:base:receive stop comm >>> >>> _______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >> > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users