I imagine you must have a system admin down there - perhaps you should ask for 
their help in finding the OMPI installation. Such systems often use "modules" 
to assist in setting up paths. I'm afraid I can't really help you find it, all 
I can do is tell you that you are using the MPICH version.

On Sep 25, 2014, at 4:33 AM, XingFENG <xingf...@cse.unsw.edu.au> wrote:

> It returns /usr/bin/mpiexec.
> 
> On Thu, Sep 25, 2014 at 8:57 PM, Ralph Castain <r...@open-mpi.org> wrote:
> Do "which mpiexec" and look at the path. The options you show are from MPICH, 
> not OMPI.
> 
> On Sep 25, 2014, at 12:15 AM, XingFENG <xingf...@cse.unsw.edu.au> wrote:
> 
>> Hi Ralph,
>> 
>> Thanks for your reply.
>> 
>> I am not pretty sure about the version of mpiexec. The documentation claims 
>> that two mpi are installed, namely, OpenMPI and MPICH2.
>> 
>> On Thu, Sep 25, 2014 at 11:45 AM, Ralph Castain <r...@open-mpi.org> wrote:
>> No, it doesn't matter at all for OMPI - any order is fine. The issue I see 
>> is that your mpiexec isn't the OMPI one, but is from someone else. I have no 
>> idea whose mpiexec you are using
>> 
>> 
>> On Sep 24, 2014, at 6:38 PM, XingFENG <xingf...@cse.unsw.edu.au> wrote:
>> 
>>> I have found the solution. The command mpirun -machinefile ./my_hosts -n 3 
>>> ./testMPI works. I think the order of arguments matters here.
>>> 
>>> On Thu, Sep 25, 2014 at 11:02 AM, XingFENG <xingf...@cse.unsw.edu.au> wrote:
>>> Hi all,
>>> 
>>> I got problem with running program on a cluster.
>>> I used the following command. my_hosts is a file containing 3 hosts while 
>>> testMPI is a very simple MPI program.
>>> ==========================================
>>> mpirun -np 2 --hostfile ./my_hosts ./testMPI
>>> mpirun -np 2 --machinefile ./my_hosts ./testMPI
>>> mpirun -np 2 --f ./my_hosts ./testMPI
>>> ==========================================
>>> 
>>> And the output is like this.
>>> ==========================================
>>> invalid "local" arg: --hostfile
>>> 
>>> usage:
>>> mpiexec [-h or -help or --help]    # get this message
>>> mpiexec -file filename             # (or -f) filename contains XML job 
>>> description
>>> mpiexec [global args] [local args] executable [args]
>>>    where global args may be
>>>       -l                           # line labels by MPI rank
>>>       -bnr                         # MPICH1 compatibility mode
>>>       -machinefile                 # file mapping procs to machines
>>>       -s <spec>                    # direct stdin to "all" or 1,2 or 2-4,6 
>>>       -1                           # override default of trying 1st proc 
>>> locally
>>>       -ifhn                        # network interface to use locally
>>>       -tv                          # run procs under totalview (must be 
>>> installed)
>>>       -tvsu                        # totalview startup only
>>>       -gdb                         # run procs under gdb
>>>       -m                           # merge output lines (default with gdb)
>>>       -a                           # means assign this alias to the job
>>>       -ecfn                        # output_xml_exit_codes_filename
>>>       -recvtimeout <integer_val>   # timeout for recvs to fail (e.g. from 
>>> mpd daemon)
>>>       -g<local arg name>           # global version of local arg (below)
>>>     and local args may be
>>>       -n <n> or -np <n>            # number of processes to start
>>>       -wdir <dirname>              # working directory to start in
>>>       -umask <umask>               # umask for remote process
>>>       -path <dirname>              # place to look for executables
>>>       -host <hostname>             # host to start on
>>>       -soft <spec>                 # modifier of -n value
>>>       -arch <arch>                 # arch type to start on (not implemented)
>>>       -envall                      # pass all env vars in current 
>>> environment
>>>       -envnone                     # pass no env vars
>>>       -envlist <list of env var names> # pass current values of these vars
>>>       -env <name> <value>          # pass this value of this env var
>>> mpiexec [global args] [local args] executable args : [local args] 
>>> executable...
>>> mpiexec -gdba jobid                # gdb-attach to existing jobid
>>> mpiexec -configfile filename       # filename contains cmd line segs as 
>>> lines
>>>   (See User Guide for more details)
>>> 
>>> Examples:
>>>    mpiexec -l -n 10 cpi 100
>>>    mpiexec -genv QPL_LICENSE 4705 -n 3 a.out
>>> 
>>>    mpiexec -n 1 -host foo master : -n 4 -host mysmp slave
>>> 
>>> ==========================================
>>> 
>>> 
>>> Another problem is that I cannot get the version of MPI. With command 
>>> mpirun --version I got
>>> 
>>> ==========================================
>>> invalid "local" arg: --version
>>> 
>>> usage:
>>> mpiexec [-h or -help or --help]    # get this message
>>> mpiexec -file filename             # (or -f) filename contains XML job 
>>> description
>>> mpiexec [global args] [local args] executable [args]
>>>    where global args may be
>>>       -l                           # line labels by MPI rank
>>>       -bnr                         # MPICH1 compatibility mode
>>>       -machinefile                 # file mapping procs to machines
>>>       -s <spec>                    # direct stdin to "all" or 1,2 or 2-4,6 
>>>       -1                           # override default of trying 1st proc 
>>> locally
>>>       -ifhn                        # network interface to use locally
>>>       -tv                          # run procs under totalview (must be 
>>> installed)
>>>       -tvsu                        # totalview startup only
>>>       -gdb                         # run procs under gdb
>>>       -m                           # merge output lines (default with gdb)
>>>       -a                           # means assign this alias to the job
>>>       -ecfn                        # output_xml_exit_codes_filename
>>>       -recvtimeout <integer_val>   # timeout for recvs to fail (e.g. from 
>>> mpd daemon)
>>>       -g<local arg name>           # global version of local arg (below)
>>>     and local args may be
>>>       -n <n> or -np <n>            # number of processes to start
>>>       -wdir <dirname>              # working directory to start in
>>>       -umask <umask>               # umask for remote process
>>>       -path <dirname>              # place to look for executables
>>>       -host <hostname>             # host to start on
>>>       -soft <spec>                 # modifier of -n value
>>>       -arch <arch>                 # arch type to start on (not implemented)
>>>       -envall                      # pass all env vars in current 
>>> environment
>>>       -envnone                     # pass no env vars
>>>       -envlist <list of env var names> # pass current values of these vars
>>>       -env <name> <value>          # pass this value of this env var
>>> mpiexec [global args] [local args] executable args : [local args] 
>>> executable...
>>> mpiexec -gdba jobid                # gdb-attach to existing jobid
>>> mpiexec -configfile filename       # filename contains cmd line segs as 
>>> lines
>>>   (See User Guide for more details)
>>> 
>>> Examples:
>>>    mpiexec -l -n 10 cpi 100
>>>    mpiexec -genv QPL_LICENSE 4705 -n 3 a.out
>>> 
>>>    mpiexec -n 1 -host foo master : -n 4 -host mysmp slave
>>> 
>>> 
>>> ==========================================
>>> 
>>> Any help would be greatly appreciated!
>>> 
>>> 
>>> -- 
>>> Best Regards.
>>> ---
>>> Xing FENG
>>> PhD Candidate
>>> Database Research Group
>>> 
>>> School of Computer Science and Engineering
>>> University of New South Wales
>>> NSW 2052, Sydney
>>> 
>>> Phone: (+61) 413 857 288
>>> 
>>> 
>>> 
>>> -- 
>>> Best Regards.
>>> ---
>>> Xing FENG
>>> PhD Candidate
>>> Database Research Group
>>> 
>>> School of Computer Science and Engineering
>>> University of New South Wales
>>> NSW 2052, Sydney
>>> 
>>> Phone: (+61) 413 857 288
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2014/09/25386.php
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2014/09/25387.php
>> 
>> 
>> 
>> -- 
>> Best Regards.
>> ---
>> Xing FENG
>> PhD Candidate
>> Database Research Group
>> 
>> School of Computer Science and Engineering
>> University of New South Wales
>> NSW 2052, Sydney
>> 
>> Phone: (+61) 413 857 288
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2014/09/25388.php
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/09/25390.php
> 
> 
> 
> -- 
> Best Regards.
> ---
> Xing FENG
> PhD Candidate
> Database Research Group
> 
> School of Computer Science and Engineering
> University of New South Wales
> NSW 2052, Sydney
> 
> Phone: (+61) 413 857 288
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/09/25391.php

Reply via email to