Actually, I should say that I discovered that if you put --prefix on each line 
of the app context file, then the first case (running the app context file) 
works fine; it adheres to the --prefix behavior.

Ralph: is this intended behavior?  (I don't know if I have an opinion either 
way)


On Mar 1, 2012, at 4:51 PM, Jeffrey Squyres wrote:

> I see the problem.
> 
> It looks like the use of the app context file is triggering different 
> behavior, and that behavior is erasing the use of --prefix.  If I replace the 
> app context file with a complete command line, it works and the --prefix 
> behavior is observed.
> 
> Specifically:
> 
> $mpirunfile $mcaparams --app addmpw-hostname
> 
> ^^ This one seems to ignore --prefix behavior.
> 
> $mpirunfile $mcaparams --host svbu-mpi,svbu-mpi001 -np 2 hostname
> $mpirunfile $mcaparams --host svbu-mpi -np 1 hostname : --host svbu-mpi001 
> -np 1 hostname
> 
> ^^ These two seem to adhere to the proper --prefix behavior.
> 
> Ralph -- can you have a look?
> 
> 
> 
> 
> On Mar 1, 2012, at 2:59 PM, Yiguang Yan wrote:
> 
>> Hi Ralph,
>> 
>> Thanks, here is what I did as suggested by Jeff:
>> 
>>> What did this command line look like? Can you provide the configure line as 
>>> well? 
>> 
>> As in my previous post, the script as following:
>> 
>> (1) debug messages:
>>>>> 
>> yiguang@gulftown testdmp]$ ./test.bash
>> [gulftown:28340] mca: base: components_open: Looking for plm components
>> [gulftown:28340] mca: base: components_open: opening plm components
>> [gulftown:28340] mca: base: components_open: found loaded component rsh
>> [gulftown:28340] mca: base: components_open: component rsh has no register 
>> function
>> [gulftown:28340] mca: base: components_open: component rsh open function 
>> successful
>> [gulftown:28340] mca: base: components_open: found loaded component slurm
>> [gulftown:28340] mca: base: components_open: component slurm has no register 
>> function
>> [gulftown:28340] mca: base: components_open: component slurm open function 
>> successful
>> [gulftown:28340] mca: base: components_open: found loaded component tm
>> [gulftown:28340] mca: base: components_open: component tm has no register 
>> function
>> [gulftown:28340] mca: base: components_open: component tm open function 
>> successful
>> [gulftown:28340] mca:base:select: Auto-selecting plm components
>> [gulftown:28340] mca:base:select:(  plm) Querying component [rsh]
>> [gulftown:28340] mca:base:select:(  plm) Query of component [rsh] set 
>> priority to 10
>> [gulftown:28340] mca:base:select:(  plm) Querying component [slurm]
>> [gulftown:28340] mca:base:select:(  plm) Skipping component [slurm]. Query 
>> failed to return a module
>> [gulftown:28340] mca:base:select:(  plm) Querying component [tm]
>> [gulftown:28340] mca:base:select:(  plm) Skipping component [tm]. Query 
>> failed to return a module
>> [gulftown:28340] mca:base:select:(  plm) Selected component [rsh]
>> [gulftown:28340] mca: base: close: component slurm closed
>> [gulftown:28340] mca: base: close: unloading component slurm
>> [gulftown:28340] mca: base: close: component tm closed
>> [gulftown:28340] mca: base: close: unloading component tm
>> [gulftown:28340] plm:base:set_hnp_name: initial bias 28340 nodename hash 
>> 3546479048
>> [gulftown:28340] plm:base:set_hnp_name: final jobfam 17438
>> [gulftown:28340] [[17438,0],0] plm:base:receive start comm
>> [gulftown:28340] [[17438,0],0] plm:rsh: setting up job [17438,1]
>> [gulftown:28340] [[17438,0],0] plm:base:setup_job for job [17438,1]
>> [gulftown:28340] [[17438,0],0] plm:rsh: local shell: 0 (bash)
>> [gulftown:28340] [[17438,0],0] plm:rsh: assuming same remote shell as local 
>> shell
>> [gulftown:28340] [[17438,0],0] plm:rsh: remote shell: 0 (bash)
>> [gulftown:28340] [[17438,0],0] plm:rsh: final template argv:
>>       /usr/bin/rsh <template>  orted --daemonize -mca ess env -mca 
>> orte_ess_jobid 1142816768 -mca 
>> orte_ess_vpid <template> -mca orte_ess_num_procs 4 --hnp-uri 
>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>  -
>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>> btl openib,sm,self --mca 
>> orte_tmpdir_base /tmp --mca plm_base_verbose 100
>> [gulftown:28340] [[17438,0],0] plm:rsh:launch daemon already exists on node 
>> gulftown
>> [gulftown:28340] [[17438,0],0] plm:rsh: launching on node ibnode001
>> [gulftown:28340] [[17438,0],0] plm:rsh: recording launch of daemon 
>> [[17438,0],1]
>> [gulftown:28340] [[17438,0],0] plm:rsh: executing: (//usr/bin/rsh) 
>> [/usr/bin/rsh ibnode001  orted --daemonize -mca 
>> ess env -mca orte_ess_jobid 1142816768 -mca orte_ess_vpid 1 -mca 
>> orte_ess_num_procs 4 --hnp-uri 
>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>  -
>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>> btl openib,sm,self --mca 
>> orte_tmpdir_base /tmp --mca plm_base_verbose 100]
>> bash: orted: command not found
>> [gulftown:28340] [[17438,0],0] plm:rsh: launching on node ibnode002
>> [gulftown:28340] [[17438,0],0] plm:rsh: recording launch of daemon 
>> [[17438,0],2]
>> [gulftown:28340] [[17438,0],0] plm:rsh: executing: (//usr/bin/rsh) 
>> [/usr/bin/rsh ibnode002  orted --daemonize -mca 
>> ess env -mca orte_ess_jobid 1142816768 -mca orte_ess_vpid 2 -mca 
>> orte_ess_num_procs 4 --hnp-uri 
>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>  -
>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>> btl openib,sm,self --mca 
>> orte_tmpdir_base /tmp --mca plm_base_verbose 100]
>> bash: orted: command not found
>> [gulftown:28340] [[17438,0],0] plm:rsh: launching on node ibnode003
>> [gulftown:28340] [[17438,0],0] plm:rsh: executing: (//usr/bin/rsh) 
>> [/usr/bin/rsh ibnode003  orted --daemonize -mca 
>> ess env -mca orte_ess_jobid 1142816768 -mca orte_ess_vpid 3 -mca 
>> orte_ess_num_procs 4 --hnp-uri 
>> "1142816768.0;tcp://198.177.146.70:43159;tcp://10.10.10.4:43159;tcp://172.23.10.1:43159;tcp://172.33.10.1:43159"
>>  -
>> -mca plm_rsh_agent rsh:ssh --mca btl_openib_warn_default_gid_prefix 0 --mca 
>> btl openib,sm,self --mca 
>> orte_tmpdir_base /tmp --mca plm_base_verbose 100]
>> [gulftown:28340] [[17438,0],0] plm:rsh: recording launch of daemon 
>> [[17438,0],3]
>> bash: orted: command not found
>> [gulftown:28340] [[17438,0],0] plm:base:daemon_callback
>> <<<
>> 
>> (2) test.bash script:
>>>>> 
>> #!/bin/sh -f
>> #nohup
>> #
>> # 
>> >-------------------------------------------------------------------------------------------<
>> adinahome=/usr/adina/system8.8dmp
>> mpirunfile=$adinahome/bin/mpirun
>> #
>> # Set envars for mpirun and orted
>> #
>> export PATH=$adinahome/bin:$adinahome/tools:$PATH
>> export LD_LIBRARY_PATH=$adinahome/lib:$LD_LIBRARY_PATH
>> #
>> #
>> # run DMP problem
>> #
>> mcaprefix="--prefix $adinahome"
>> mcarshagent="--mca plm_rsh_agent rsh:ssh"
>> mcatmpdir="--mca orte_tmpdir_base /tmp"
>> mcaopenibmsg="--mca btl_openib_warn_default_gid_prefix 0"
>> mcaenvars="-x PATH -x LD_LIBRARY_PATH"
>> mcabtlconn="--mca btl openib,sm,self"
>> mcaplmbase="--mca plm_base_verbose 100"
>> 
>> mcaparams="$mcaprefix $mcaenvars $mcarshagent $mcaopenibmsg $mcabtlconn 
>> $mcatmpdir $mcaplmbase"
>> 
>> $mpirunfile $mcaparams --app addmpw-hostname
>> <<<
>> 
>> (3) the contend of app file addmpw-hostname:
>>>>> 
>> -n 1 -host gulftown hostname
>> -n 1 -host ibnode001 hostname
>> -n 1 -host ibnode002 hostname
>> -n 1 -host ibnode003 thostname
>> <<<
>> 
>> Any comments?
>> 
>> Thanks,
>> Yiguang
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> -- 
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to: 
> http://www.cisco.com/web/about/doing_business/legal/cri/
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to