On Nov 13, 2013, at 4:43 PM, tmish...@jcity.maeda.co.jp wrote:

> 
> 
> Yes, the node08 has 8 slots but the process I run is also 8.
> 
> #PBS -l nodes=node08:ppn=8
> 
> Therefore, I think it should allow this allocation. Is that right?

Correct

> 
> My question is why scritp1 works and script2 does not. They are
> almost same.
> 
> #PBS -l nodes=node08:ppn=8
> export OMP_NUM_THREADS=1
> cd $PBS_O_WORKDIR
> cp $PBS_NODEFILE pbs_hosts
> NPROCS=`wc -l < pbs_hosts`
> 
> #SCRITP1
> mpirun -report-bindings -bind-to core Myprog
> 
> #SCRIPT2
> mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings -bind-to core

This version is not only reading the PBS allocation, but also invoking the 
hostfile filter on top of it. Different code path. I'll take a look - it should 
still match up assuming NPROCS=8. Any possibility that it is a different 
number? I don't recall, but isn't there some extra lines in the nodefile - 
e.g., comments?


> Myprog
> 
> tmishima
> 
>> I guess here's my confusion. If you are using only one node, and that
> node has 8 allocated slots, then we will not allow you to run more than 8
> processes on that node unless you specifically provide
>> the --oversubscribe flag. This is because you are operating in a managed
> environment (in this case, under Torque), and so we treat the allocation as
> "mandatory" by default.
>> 
>> I suspect that is the issue here, in which case the system is behaving as
> it should.
>> 
>> Is the above accurate?
>> 
>> 
>> On Nov 13, 2013, at 4:11 PM, Ralph Castain <r...@open-mpi.org> wrote:
>> 
>>> It has nothing to do with LAMA as you aren't using that mapper.
>>> 
>>> How many nodes are in this allocation?
>>> 
>>> On Nov 13, 2013, at 4:06 PM, tmish...@jcity.maeda.co.jp wrote:
>>> 
>>>> 
>>>> 
>>>> Hi Ralph, this is an additional information.
>>>> 
>>>> Here is the main part of output by adding "-mca rmaps_base_verbose
> 50".
>>>> 
>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm
>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm creating map
>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm only HNP in
>>>> allocation
>>>> [node08.cluster:26952] mca:rmaps: mapping job [56581,1]
>>>> [node08.cluster:26952] mca:rmaps: creating new map for job [56581,1]
>>>> [node08.cluster:26952] mca:rmaps:ppr: job [56581,1] not using ppr
> mapper
>>>> [node08.cluster:26952] [[56581,0],0] rmaps:seq mapping job [56581,1]
>>>> [node08.cluster:26952] mca:rmaps:seq: job [56581,1] not using seq
> mapper
>>>> [node08.cluster:26952] mca:rmaps:resilient: cannot perform initial map
> of
>>>> job [56581,1] - no fault groups
>>>> [node08.cluster:26952] mca:rmaps:mindist: job [56581,1] not using
> mindist
>>>> mapper
>>>> [node08.cluster:26952] mca:rmaps:rr: mapping job [56581,1]
>>>> [node08.cluster:26952] [[56581,0],0] Starting with 1 nodes in list
>>>> [node08.cluster:26952] [[56581,0],0] Filtering thru apps
>>>> [node08.cluster:26952] [[56581,0],0] Retained 1 nodes in list
>>>> [node08.cluster:26952] [[56581,0],0] Removing node node08 slots 0
> inuse 0
>>>> 
>>>> From this result, I guess it's related to oversubscribe.
>>>> So I added "-oversubscribe" and rerun, then it worked well as show
> below:
>>>> 
>>>> [node08.cluster:27019] [[56774,0],0] Starting with 1 nodes in list
>>>> [node08.cluster:27019] [[56774,0],0] Filtering thru apps
>>>> [node08.cluster:27019] [[56774,0],0] Retained 1 nodes in list
>>>> [node08.cluster:27019] AVAILABLE NODES FOR MAPPING:
>>>> [node08.cluster:27019]     node: node08 daemon: 0
>>>> [node08.cluster:27019] [[56774,0],0] Starting bookmark at node node08
>>>> [node08.cluster:27019] [[56774,0],0] Starting at node node08
>>>> [node08.cluster:27019] mca:rmaps:rr: mapping by slot for job [56774,1]
>>>> slots 1 num_procs 8
>>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08
>>>> [node08.cluster:27019] mca:rmaps:rr:slot node node08 is full -
> skipping
>>>> [node08.cluster:27019] mca:rmaps:rr:slot job [56774,1] is
> oversubscribed -
>>>> performing second pass
>>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08
>>>> [node08.cluster:27019] mca:rmaps:rr:slot adding up to 8 procs to node
>>>> node08
>>>> [node08.cluster:27019] mca:rmaps:base: computing vpids by slot for job
>>>> [56774,1]
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 0 to node node08
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 1 to node node08
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 2 to node node08
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 3 to node node08
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 4 to node node08
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 5 to node node08
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 6 to node node08
>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 7 to node node08
>>>> 
>>>> I think something is wrong with treatment of oversubscription, which
> might
>>>> be
>>>> related to "#3893: LAMA mapper has problems"
>>>> 
>>>> tmishima
>>>> 
>>>>> Hmmm...looks like we aren't getting your allocation. Can you rerun
> and
>>>> add -mca ras_base_verbose 50?
>>>>> 
>>>>> On Nov 12, 2013, at 11:30 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>> 
>>>>>> 
>>>>>> 
>>>>>> Hi Ralph,
>>>>>> 
>>>>>> Here is the output of "-mca plm_base_verbose 5".
>>>>>> 
>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying component
> [rsh]
>>>>>> [node08.cluster:23573] [[INVALID],INVALID] plm:rsh_lookup on
>>>>>> agent /usr/bin/rsh path NULL
>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Query of component
> [rsh]
>>>> set
>>>>>> priority to 10
>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying component
>>>> [slurm]
>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Skipping component
>>>> [slurm].
>>>>>> Query failed to return a module
>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying component
> [tm]
>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Query of component
> [tm]
>>>> set
>>>>>> priority to 75
>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Selected component
> [tm]
>>>>>> [node08.cluster:23573] plm:base:set_hnp_name: initial bias 23573
>>>> nodename
>>>>>> hash 85176670
>>>>>> [node08.cluster:23573] plm:base:set_hnp_name: final jobfam 59480
>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:receive start comm
>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_job
>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm
>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm creating map
>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm only HNP in
>>>>>> allocation
>>>>>> 
>>>> 
> --------------------------------------------------------------------------
>>>>>> All nodes which are allocated for this job are already filled.
>>>>>> 
>>>> 
> --------------------------------------------------------------------------
>>>>>> 
>>>>>> Here, openmpi's configuration is as follows:
>>>>>> 
>>>>>> ./configure \
>>>>>> --prefix=/home/mishima/opt/mpi/openmpi-1.7.4a1-pgi13.10 \
>>>>>> --with-tm \
>>>>>> --with-verbs \
>>>>>> --disable-ipv6 \
>>>>>> --disable-vt \
>>>>>> --enable-debug \
>>>>>> CC=pgcc CFLAGS="-tp k8-64e" \
>>>>>> CXX=pgCC CXXFLAGS="-tp k8-64e" \
>>>>>> F77=pgfortran FFLAGS="-tp k8-64e" \
>>>>>> FC=pgfortran FCFLAGS="-tp k8-64e"
>>>>>> 
>>>>>>> Hi Ralph,
>>>>>>> 
>>>>>>> Okey, I can help you. Please give me some time to report the
> output.
>>>>>>> 
>>>>>>> Tetsuya Mishima
>>>>>>> 
>>>>>>>> I can try, but I have no way of testing Torque any more - so all I
>>>> can
>>>>>> do
>>>>>>> is a code review. If you can build --enable-debug and add -mca
>>>>>>> plm_base_verbose 5 to your cmd line, I'd appreciate seeing the
>>>>>>>> output.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Nov 12, 2013, at 9:58 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Hi Ralph,
>>>>>>>>> 
>>>>>>>>> Thank you for your quick response.
>>>>>>>>> 
>>>>>>>>> I'd like to report one more regressive issue about Torque support
> of
>>>>>>>>> openmpi-1.7.4a1r29646, which might be related to "#3893: LAMA
> mapper
>>>>>>>>> has problems" I reported a few days ago.
>>>>>>>>> 
>>>>>>>>> The script below does not work with openmpi-1.7.4a1r29646,
>>>>>>>>> although it worked with openmpi-1.7.3 as I told you before.
>>>>>>>>> 
>>>>>>>>> #!/bin/sh
>>>>>>>>> #PBS -l nodes=node08:ppn=8
>>>>>>>>> export OMP_NUM_THREADS=1
>>>>>>>>> cd $PBS_O_WORKDIR
>>>>>>>>> cp $PBS_NODEFILE pbs_hosts
>>>>>>>>> NPROCS=`wc -l < pbs_hosts`
>>>>>>>>> mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings
>>>> -bind-to
>>>>>>> core
>>>>>>>>> Myprog
>>>>>>>>> 
>>>>>>>>> If I drop "-machinefile pbs_hosts -np ${NPROCS} ", then it works
>>>>>> fine.
>>>>>>>>> Since this happens without lama request, I guess it's not the
>>>> problem
>>>>>>>>> in lama itself. Anyway, please look into this issue as well.
>>>>>>>>> 
>>>>>>>>> Regards,
>>>>>>>>> Tetsuya Mishima
>>>>>>>>> 
>>>>>>>>>> Done - thanks!
>>>>>>>>>> 
>>>>>>>>>> On Nov 12, 2013, at 7:35 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> Dear openmpi developers,
>>>>>>>>>>> 
>>>>>>>>>>> I got a segmentation fault in traial use of
> openmpi-1.7.4a1r29646
>>>>>>> built
>>>>>>>>> by
>>>>>>>>>>> PGI13.10 as shown below:
>>>>>>>>>>> 
>>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ mpirun -np 4
>>>> -cpus-per-proc
>>>>>> 2
>>>>>>>>>>> -report-bindings mPre
>>>>>>>>>>> [manage.cluster:23082] MCW rank 2 bound to socket 0[core 4[hwt
>>>> 0]],
>>>>>>>>> socket
>>>>>>>>>>> 0[core 5[hwt 0]]: [././././B/B][./././././.]
>>>>>>>>>>> [manage.cluster:23082] MCW rank 3 bound to socket 1[core 6[hwt
>>>> 0]],
>>>>>>>>> socket
>>>>>>>>>>> 1[core 7[hwt 0]]: [./././././.][B/B/./././.]
>>>>>>>>>>> [manage.cluster:23082] MCW rank 0 bound to socket 0[core 0[hwt
>>>> 0]],
>>>>>>>>> socket
>>>>>>>>>>> 0[core 1[hwt 0]]: [B/B/./././.][./././././.]
>>>>>>>>>>> [manage.cluster:23082] MCW rank 1 bound to socket 0[core 2[hwt
>>>> 0]],
>>>>>>>>> socket
>>>>>>>>>>> 0[core 3[hwt 0]]: [././B/B/./.][./././././.]
>>>>>>>>>>> [manage:23082] *** Process received signal ***
>>>>>>>>>>> [manage:23082] Signal: Segmentation fault (11)
>>>>>>>>>>> [manage:23082] Signal code: Address not mapped (1)
>>>>>>>>>>> [manage:23082] Failing at address: 0x34
>>>>>>>>>>> [manage:23082] *** End of error message ***
>>>>>>>>>>> Segmentation fault (core dumped)
>>>>>>>>>>> 
>>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ gdb mpirun core.23082
>>>>>>>>>>> GNU gdb (GDB) CentOS (7.0.1-42.el5.centos.1)
>>>>>>>>>>> Copyright (C) 2009 Free Software Foundation, Inc.
>>>>>>>>>>> ...
>>>>>>>>>>> Core was generated by `mpirun -np 4 -cpus-per-proc 2
>>>>>> -report-bindings
>>>>>>>>>>> mPre'.
>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>> #0  0x00002b5f861c9c4f in recv_connect (mod=0x5f861ca20b00007f,
>>>>>>>>> sd=32767,
>>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631
>>>>>>>>>>> 631             peer = OBJ_NEW(mca_oob_tcp_peer_t);
>>>>>>>>>>> (gdb) where
>>>>>>>>>>> #0  0x00002b5f861c9c4f in recv_connect (mod=0x5f861ca20b00007f,
>>>>>>>>> sd=32767,
>>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631
>>>>>>>>>>> #1  0x00002b5f861ca20b in recv_handler (sd=1778385023,
>>>> flags=32767,
>>>>>>>>>>> cbdata=0x8eb06a00007fff25) at ./oob_tcp.c:760
>>>>>>>>>>> #2  0x00002b5f848eb06a in event_process_active_single_queue
>>>>>>>>>>> (base=0x5f848eb27000007f, activeq=0x848eb27000007fff)
>>>>>>>>>>> at ./event.c:1366
>>>>>>>>>>> #3  0x00002b5f848eb270 in event_process_active
>>>>>>>>> (base=0x5f848eb84900007f)
>>>>>>>>>>> at ./event.c:1435
>>>>>>>>>>> #4  0x00002b5f848eb849 in opal_libevent2021_event_base_loop
>>>>>>>>>>> (base=0x4077a000007f, flags=32767) at ./event.c:1645
>>>>>>>>>>> #5  0x00000000004077a0 in orterun (argc=7, argv=0x7fff25bbd4a8)
>>>>>>>>>>> at ./orterun.c:1030
>>>>>>>>>>> #6  0x00000000004067fb in main (argc=7, argv=0x7fff25bbd4a8)
>>>>>>>>> at ./main.c:13
>>>>>>>>>>> (gdb) quit
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> The line 627 in orte/mca/oob/tcp/oob_tcp.c is apparently
>>>>>> unnecessary,
>>>>>>>>> which
>>>>>>>>>>> causes the segfault.
>>>>>>>>>>> 
>>>>>>>>>>> 624      /* lookup the corresponding process */
>>>>>>>>>>> 625      peer = mca_oob_tcp_peer_lookup(mod, &hdr->origin);
>>>>>>>>>>> 626      if (NULL == peer) {
>>>>>>>>>>> 627          ui64 = (uint64_t*)(&peer->name);
>>>>>>>>>>> 628          opal_output_verbose(OOB_TCP_DEBUG_CONNECT,
>>>>>>>>>>> orte_oob_base_framework.framework_output,
>>>>>>>>>>> 629                              "%s mca_oob_tcp_recv_connect:
>>>>>>>>>>> connection from new peer",
>>>>>>>>>>> 630                              ORTE_NAME_PRINT
>>>>>>> (ORTE_PROC_MY_NAME));
>>>>>>>>>>> 631          peer = OBJ_NEW(mca_oob_tcp_peer_t);
>>>>>>>>>>> 632          peer->mod = mod;
>>>>>>>>>>> 633          peer->name = hdr->origin;
>>>>>>>>>>> 634          peer->state = MCA_OOB_TCP_ACCEPTING;
>>>>>>>>>>> 635          ui64 = (uint64_t*)(&peer->name);
>>>>>>>>>>> 636          if (OPAL_SUCCESS !=
> opal_hash_table_set_value_uint64
>>>>>>>>> (&mod->
>>>>>>>>>>> peers, (*ui64), peer)) {
>>>>>>>>>>> 637              OBJ_RELEASE(peer);
>>>>>>>>>>> 638              return;
>>>>>>>>>>> 639          }
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> Please fix this mistake in the next release.
>>>>>>>>>>> 
>>>>>>>>>>> Regards,
>>>>>>>>>>> Tetsuya Mishima
>>>>>>>>>>> 
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> users mailing list
>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>> 
>>>>>>>>>> _______________________________________________
>>>>>>>>>> users mailing list
>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>> 
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> us...@open-mpi.org
>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to