Hi
I think I'am not understanding what you said , here is the hello.py and next 
the command mpirun…

Thanks!

#!/usr/bin/env python
"""
Parallel Hello World
"""

from mpi4py import MPI
import sys

size = MPI.COMM_WORLD.Get_size()
rank = MPI.COMM_WORLD.Get_rank()
name = MPI.Get_processor_name()

sys.stdout.write(
    "Hello, World! I am process %d of %d on %s.\n"
    % (rank, size, name))

 ~/bin/mpirun -np 70 python2.7 helloworld.py 
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.
Hello, World! I am process 0 of 1 on ferrari.

On Sep 25, 2012, at 4:46 PM, Ralph Castain <r...@open-mpi.org> wrote:

> The usual reason for this is that you aren't launching these processes 
> correctly. How are you starting your job? Are you using mpirun?
> 
> 
> On Sep 25, 2012, at 1:43 PM, mariana Vargas <mmaria...@yahoo.com.mx> wrote:
> 
>> Hi
>> 
>> I fact I found what is the origin of this problem and it is because all 
>> processes have rank 0, so I tested and in effect even when I send the 
>> clasical Hello.py give the same, how can I solved this?? Do I  re installed 
>> every again???
>> 
>> Help please...
>> 
>> Mariana
>> 
>> 
>> 
>> On Sep 24, 2012, at 9:13 PM, Mariana Vargas Magana wrote:
>> 
>>> 
>>> 
>>> Yes you are right this is what it says but if fact the weird thing is that 
>>> not all times the error message appears….I send to 20 nodes and only one 
>>> gives this message, is this normal…
>>> 
>>> 
>>> 
>>> 
>>> On Sep 24, 2012, at 8:00 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>> 
>>>> Well, as it says, your processes called MPI_Init, but at least one of them 
>>>> exited without calling MPI_Finalize. That violates the MPI rules and we 
>>>> therefore terminate the remaining processes.
>>>> 
>>>> Check your code and see how/why you are doing that - you probably have a 
>>>> code path whereby a process exits without calling finalize.
>>>> 
>>>> 
>>>> On Sep 24, 2012, at 4:37 PM, mariana Vargas <mmaria...@yahoo.com.mx> wrote:
>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> Hi all
>>>>> 
>>>>> I get this error when I run a paralelized python code in a cluster, could 
>>>>> anyone give me an idea of what is happening? I'am new in this Thanks...
>>>>> 
>>>>> mpirun has exited due to process rank 2 with PID 10259 on
>>>>> node f01 exiting improperly. There are two reasons this could occur:
>>>>> 
>>>>> 1. this process did not call "init" before exiting, but others in
>>>>> the job did. This can cause a job to hang indefinitely while it waits
>>>>> for all processes to call "init". By rule, if one process calls "init",
>>>>> then ALL processes must call "init" prior to termination.
>>>>> 
>>>>> 2. this process called "init", but exited without calling "finalize".
>>>>> By rule, all processes that call "init" MUST call "finalize" prior to
>>>>> exiting or it will be considered an "abnormal termination"
>>>>> 
>>>>> This may have caused other processes in the application to be
>>>>> terminated by signals sent by mpirun (as reported here).
>>>>> 
>>>>> Thanks!!
>>>>> 
>>>>> 
>>>>> 
>>>>>> 
>>>>>> Dr. Mariana Vargas Magana
>>>>>> Astroparticule et Cosmologie - Bureau 409B
>>>>>> PHD student- Université Denis Diderot-Paris 7
>>>>>> 10, rue Alice Domon et Léonie Duquet
>>>>>> 75205 Paris Cedex - France
>>>>>> Tel. +33 (0)1 57 27 70 32
>>>>>> Fax. +33 (0)1 57 27 60 71
>>>>>> mari...@apc.univ-paris7.fr
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to