Is ompi installed on the other node and at the same location ?
did you configure ompi with --enable-mpirun-prefix-by-default ?
(note that should not be necessary if you invoke mpirun with full path )

you can also try
/.../bin/mpirun --mca plm_base_verbose 100 ...

and see if there is something wrong

last but not least, can you try to use full path for both mpirun and nwchem
?


Cheers,

Gilles

On Sunday, August 2, 2015, abhisek Mondal <abhisek.m...@gmail.com> wrote:

> Yes, I have tried this and got following error:
>
> *mpirun was unable to launch the specified application as it could not
> find an executable:*
>
> *Executable: nwchem*
> *Node: cx934*
>
> *while attempting to start process rank 16.*
>
> Given that: I have to run my code with "nwchem filename.nw" command.
> While I run the same thing on 1 node with 16 processor, it works fine
> (mpirun -np 16 nwchem filename.nw).
> Can't understand why am I having problem while trying to go for multinode
> operation.
>
> Thanks.
>
> On Sun, Aug 2, 2015 at 3:41 PM, Gilles Gouaillardet <
> gilles.gouaillar...@gmail.com
> <javascript:_e(%7B%7D,'cvml','gilles.gouaillar...@gmail.com');>> wrote:
>
>> Can you try running invoking mpirun with its full path instead ?
>> e.g. /usr/local/bin/mpirun instead of mpirun
>>
>> Cheers,
>>
>> Gilles
>>
>> On Sunday, August 2, 2015, abhisek Mondal <abhisek.m...@gmail.com
>> <javascript:_e(%7B%7D,'cvml','abhisek.m...@gmail.com');>> wrote:
>>
>>> Here is the other details,
>>>
>>> a. The Openmpi version is 1.6.4
>>>
>>> b. The error as being generated is :
>>> *Warning: Permanently added 'cx0937,10.1.4.1' (RSA) to the list of known
>>> hosts.*
>>> *Warning: Permanently added 'cx0934,10.1.3.255' (RSA) to the list of
>>> known hosts.*
>>> *orted: Command not found.*
>>> *orted: Command not found.*
>>>
>>> *--------------------------------------------------------------------------*
>>> *A daemon (pid 53580) died unexpectedly with status 1 while attempting*
>>> *to launch so we are aborting.*
>>>
>>> *There may be more information reported by the environment (see above).*
>>>
>>> *This may be because the daemon was unable to find all the needed shared*
>>> *libraries on the remote node. You may set your LD_LIBRARY_PATH to have
>>> the*
>>> *location of the shared libraries on the remote nodes and this will*
>>> *automatically be forwarded to the remote nodes.*
>>>
>>> *--------------------------------------------------------------------------*
>>>
>>> *--------------------------------------------------------------------------*
>>> *mpirun noticed that the job aborted, but has no info as to the process*
>>> *that caused that situation.*
>>>
>>> *--------------------------------------------------------------------------*
>>>
>>>
>>> I'm not being able to understand why "command not found" error is being
>>> raised.
>>> Thank you.
>>>
>>> On Sun, Aug 2, 2015 at 1:43 AM, Ralph Castain <r...@open-mpi.org> wrote:
>>>
>>>> Would you please tell us:
>>>>
>>>> (a) what version of OMPI you are using
>>>>
>>>> (b) what error message you are getting when the job terminates
>>>>
>>>>
>>>> On Aug 1, 2015, at 12:22 PM, abhisek Mondal <abhisek.m...@gmail.com>
>>>> wrote:
>>>>
>>>> I'm working on an openmpi enabled cluster. I'm trying to run a job with
>>>> 2 different nodes and 16 processors per nodes.
>>>> Using this command:
>>>>
>>>> *mpirun -np 32 --hostfile myhostfile -loadbalance exe*
>>>>
>>>> The contents of myhostfile:
>>>>
>>>> *cx0937 slots=16    *
>>>> *cx0934 slots=16*
>>>>
>>>>
>>>> But the job is getting terminated each time before job allocation
>>>> happens as per desired way.
>>>>
>>>> So, it'll very nice if I get some suggestions regarding the facts I'm
>>>> missing.
>>>>
>>>> Thank you
>>>>
>>>> --
>>>> Abhisek Mondal
>>>>
>>>> *Research Fellow*
>>>>
>>>> *Structural Biology and Bioinformatics*
>>>> *Indian Institute of Chemical Biology*
>>>>
>>>> *Kolkata 700032*
>>>>
>>>> *INDIA*
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Searchable archives:
>>>> http://www.open-mpi.org/community/lists/users/2015/08/27367.php
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Link to this post:
>>>> http://www.open-mpi.org/community/lists/users/2015/08/27367.php
>>>>
>>>
>>>
>>>
>>> --
>>> Abhisek Mondal
>>>
>>> *Research Fellow*
>>>
>>> *Structural Biology and Bioinformatics*
>>> *Indian Institute of Chemical Biology*
>>>
>>> *Kolkata 700032*
>>>
>>> *INDIA*
>>>
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org <javascript:_e(%7B%7D,'cvml','us...@open-mpi.org');>
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post:
>> http://www.open-mpi.org/community/lists/users/2015/08/27369.php
>>
>
>
>
> --
> Abhisek Mondal
>
> *Research Fellow*
>
> *Structural Biology and Bioinformatics*
> *Indian Institute of Chemical Biology*
>
> *Kolkata 700032*
>
> *INDIA*
>

Reply via email to