HI,
I have tried using full paths for both of them. But stuck in the same issue.

On Sun, Aug 2, 2015 at 4:39 PM, Gilles Gouaillardet <
gilles.gouaillar...@gmail.com> wrote:

> Is ompi installed on the other node and at the same location ?
> did you configure ompi with --enable-mpirun-prefix-by-default ?
> (note that should not be necessary if you invoke mpirun with full path )
>
> you can also try
> /.../bin/mpirun --mca plm_base_verbose 100 ...
>
> and see if there is something wrong
>
> last but not least, can you try to use full path for both mpirun and
> nwchem ?
>
>
> Cheers,
>
> Gilles
>
> On Sunday, August 2, 2015, abhisek Mondal <abhisek.m...@gmail.com> wrote:
>
>> Yes, I have tried this and got following error:
>>
>> *mpirun was unable to launch the specified application as it could not
>> find an executable:*
>>
>> *Executable: nwchem*
>> *Node: cx934*
>>
>> *while attempting to start process rank 16.*
>>
>> Given that: I have to run my code with "nwchem filename.nw" command.
>> While I run the same thing on 1 node with 16 processor, it works fine
>> (mpirun -np 16 nwchem filename.nw).
>> Can't understand why am I having problem while trying to go for multinode
>> operation.
>>
>> Thanks.
>>
>> On Sun, Aug 2, 2015 at 3:41 PM, Gilles Gouaillardet <
>> gilles.gouaillar...@gmail.com> wrote:
>>
>>> Can you try running invoking mpirun with its full path instead ?
>>> e.g. /usr/local/bin/mpirun instead of mpirun
>>>
>>> Cheers,
>>>
>>> Gilles
>>>
>>> On Sunday, August 2, 2015, abhisek Mondal <abhisek.m...@gmail.com>
>>> wrote:
>>>
>>>> Here is the other details,
>>>>
>>>> a. The Openmpi version is 1.6.4
>>>>
>>>> b. The error as being generated is :
>>>> *Warning: Permanently added 'cx0937,10.1.4.1' (RSA) to the list of
>>>> known hosts.*
>>>> *Warning: Permanently added 'cx0934,10.1.3.255' (RSA) to the list of
>>>> known hosts.*
>>>> *orted: Command not found.*
>>>> *orted: Command not found.*
>>>>
>>>> *--------------------------------------------------------------------------*
>>>> *A daemon (pid 53580) died unexpectedly with status 1 while attempting*
>>>> *to launch so we are aborting.*
>>>>
>>>> *There may be more information reported by the environment (see above).*
>>>>
>>>> *This may be because the daemon was unable to find all the needed
>>>> shared*
>>>> *libraries on the remote node. You may set your LD_LIBRARY_PATH to have
>>>> the*
>>>> *location of the shared libraries on the remote nodes and this will*
>>>> *automatically be forwarded to the remote nodes.*
>>>>
>>>> *--------------------------------------------------------------------------*
>>>>
>>>> *--------------------------------------------------------------------------*
>>>> *mpirun noticed that the job aborted, but has no info as to the process*
>>>> *that caused that situation.*
>>>>
>>>> *--------------------------------------------------------------------------*
>>>>
>>>>
>>>> I'm not being able to understand why "command not found" error is being
>>>> raised.
>>>> Thank you.
>>>>
>>>> On Sun, Aug 2, 2015 at 1:43 AM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>
>>>>> Would you please tell us:
>>>>>
>>>>> (a) what version of OMPI you are using
>>>>>
>>>>> (b) what error message you are getting when the job terminates
>>>>>
>>>>>
>>>>> On Aug 1, 2015, at 12:22 PM, abhisek Mondal <abhisek.m...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> I'm working on an openmpi enabled cluster. I'm trying to run a job
>>>>> with 2 different nodes and 16 processors per nodes.
>>>>> Using this command:
>>>>>
>>>>> *mpirun -np 32 --hostfile myhostfile -loadbalance exe*
>>>>>
>>>>> The contents of myhostfile:
>>>>>
>>>>> *cx0937 slots=16    *
>>>>> *cx0934 slots=16*
>>>>>
>>>>>
>>>>> But the job is getting terminated each time before job allocation
>>>>> happens as per desired way.
>>>>>
>>>>> So, it'll very nice if I get some suggestions regarding the facts I'm
>>>>> missing.
>>>>>
>>>>> Thank you
>>>>>
>>>>> --
>>>>> Abhisek Mondal
>>>>>
>>>>> *Research Fellow*
>>>>>
>>>>> *Structural Biology and Bioinformatics*
>>>>> *Indian Institute of Chemical Biology*
>>>>>
>>>>> *Kolkata 700032*
>>>>>
>>>>> *INDIA*
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Searchable archives:
>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27367.php
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27367.php
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Abhisek Mondal
>>>>
>>>> *Research Fellow*
>>>>
>>>> *Structural Biology and Bioinformatics*
>>>> *Indian Institute of Chemical Biology*
>>>>
>>>> *Kolkata 700032*
>>>>
>>>> *INDIA*
>>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/users/2015/08/27369.php
>>>
>>
>>
>>
>> --
>> Abhisek Mondal
>>
>> *Research Fellow*
>>
>> *Structural Biology and Bioinformatics*
>> *Indian Institute of Chemical Biology*
>>
>> *Kolkata 700032*
>>
>> *INDIA*
>>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2015/08/27371.php
>



-- 
Abhisek Mondal

*Research Fellow*

*Structural Biology and Bioinformatics*
*Indian Institute of Chemical Biology*

*Kolkata 700032*

*INDIA*

Reply via email to