Okay - so effectively you have no hostfile, and no allocation. So this is 
running just on the one node where mpirun exists?

Add “-mca ras_base_verbose 10 --display-allocation” to your cmd line and let’s 
see what it found

> On Apr 25, 2017, at 12:56 PM, Eric Chamberland 
> <eric.chamberl...@giref.ulaval.ca> wrote:
> 
> Hi,
> 
> the host file has been constructed automatically by the 
> configuration+installation process and seems to contain only comments and a 
> blank line:
> 
> (15:53:50) [zorg]:~> cat /opt/openmpi-3.x_debug/etc/openmpi-default-hostfile
> #
> # Copyright (c) 2004-2005 The Trustees of Indiana University and Indiana
> #                         University Research and Technology
> #                         Corporation.  All rights reserved.
> # Copyright (c) 2004-2005 The University of Tennessee and The University
> #                         of Tennessee Research Foundation.  All rights
> #                         reserved.
> # Copyright (c) 2004-2005 High Performance Computing Center Stuttgart,
> #                         University of Stuttgart.  All rights reserved.
> # Copyright (c) 2004-2005 The Regents of the University of California.
> #                         All rights reserved.
> # $COPYRIGHT$
> #
> # Additional copyrights may follow
> #
> # $HEADER$
> #
> # This is the default hostfile for Open MPI.  Notice that it does not
> # contain any hosts (not even localhost).  This file should only
> # contain hosts if a system administrator wants users to always have
> # the same set of default hosts, and is not using a batch scheduler
> # (such as SLURM, PBS, etc.).
> #
> # Note that this file is *not* used when running in "managed"
> # environments (e.g., running in a job under a job scheduler, such as
> # SLURM or PBS / Torque).
> #
> # If you are primarily interested in running Open MPI on one node, you
> # should *not* simply list "localhost" in here (contrary to prior MPI
> # implementations, such as LAM/MPI).  A localhost-only node list is
> # created by the RAS component named "localhost" if no other RAS
> # components were able to find any hosts to run on (this behavior can
> # be disabled by excluding the localhost RAS component by specifying
> # the value "^localhost" [without the quotes] to the "ras" MCA
> # parameter).
> 
> (15:53:52) [zorg]:~>
> 
> Thanks!
> 
> Eric
> 
> 
> On 25/04/17 03:52 PM, r...@open-mpi.org wrote:
>> What is in your hostfile?
>> 
>> 
>>> On Apr 25, 2017, at 11:39 AM, Eric Chamberland 
>>> <eric.chamberl...@giref.ulaval.ca> wrote:
>>> 
>>> Hi,
>>> 
>>> just testing the 3.x branch... I launch:
>>> 
>>> mpirun -n 8 echo "hello"
>>> 
>>> and I get:
>>> 
>>> --------------------------------------------------------------------------
>>> There are not enough slots available in the system to satisfy the 8 slots
>>> that were requested by the application:
>>> echo
>>> 
>>> Either request fewer slots for your application, or make more slots 
>>> available
>>> for use.
>>> --------------------------------------------------------------------------
>>> 
>>> I have to oversubscribe, so what do I have to do to bypass this 
>>> "limitation"?
>>> 
>>> Thanks,
>>> 
>>> Eric
>>> 
>>> configure log:
>>> 
>>> http://www.giref.ulaval.ca/~cmpgiref/ompi_3.x/2017.04.25.10h46m08s_config.log
>>> http://www.giref.ulaval.ca/~cmpgiref/ompi_3.x/2017.04.25.10h46m08s_ompi_info_all.txt
>>> 
>>> 
>>> here is the complete message:
>>> 
>>> [zorg:30036] [[INVALID],INVALID] plm:rsh_lookup on agent ssh : rsh path NULL
>>> [zorg:30036] plm:base:set_hnp_name: initial bias 30036 nodename hash 
>>> 810220270
>>> [zorg:30036] plm:base:set_hnp_name: final jobfam 49136
>>> [zorg:30036] [[49136,0],0] plm:rsh_setup on agent ssh : rsh path NULL
>>> [zorg:30036] [[49136,0],0] plm:base:receive start comm
>>> [zorg:30036] [[49136,0],0] plm:base:setup_job
>>> [zorg:30036] [[49136,0],0] plm:base:setup_vm
>>> [zorg:30036] [[49136,0],0] plm:base:setup_vm creating map
>>> [zorg:30036] [[49136,0],0] setup:vm: working unmanaged allocation
>>> [zorg:30036] [[49136,0],0] using default hostfile 
>>> /opt/openmpi-3.x_debug/etc/openmpi-default-hostfile
>>> [zorg:30036] [[49136,0],0] plm:base:setup_vm only HNP in allocation
>>> [zorg:30036] [[49136,0],0] plm:base:setting slots for node zorg by cores
>>> [zorg:30036] [[49136,0],0] complete_setup on job [49136,1]
>>> [zorg:30036] [[49136,0],0] plm:base:launch_apps for job [49136,1]
>>> --------------------------------------------------------------------------
>>> There are not enough slots available in the system to satisfy the 8 slots
>>> that were requested by the application:
>>> echo
>>> 
>>> Either request fewer slots for your application, or make more slots 
>>> available
>>> for use.
>>> --------------------------------------------------------------------------
>>> [zorg:30036] [[49136,0],0] plm:base:orted_cmd sending orted_exit commands
>>> [zorg:30036] [[49136,0],0] plm:base:receive stop comm
>>> 
>>> _______________________________________________
>>> users mailing list
>>> users@lists.open-mpi.org
>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>> 
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>> 
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to