You forgot to call MPI_Init at the beginning of your program.

On Sep 25, 2012, at 2:08 PM, Mariana Vargas Magana <mmaria...@yahoo.com.mx> 
wrote:

> Hi
> I think I'am not understanding what you said , here is the hello.py and next 
> the command mpirun…
> 
> Thanks!
> 
> #!/usr/bin/env python
> """
> Parallel Hello World
> """
> 
> from mpi4py import MPI
> import sys
> 
> size = MPI.COMM_WORLD.Get_size()
> rank = MPI.COMM_WORLD.Get_rank()
> name = MPI.Get_processor_name()
> 
> sys.stdout.write(
>     "Hello, World! I am process %d of %d on %s.\n"
>     % (rank, size, name))
> 
>  ~/bin/mpirun -np 70 python2.7 helloworld.py 
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> Hello, World! I am process 0 of 1 on ferrari.
> 
> On Sep 25, 2012, at 4:46 PM, Ralph Castain <r...@open-mpi.org> wrote:
> 
>> The usual reason for this is that you aren't launching these processes 
>> correctly. How are you starting your job? Are you using mpirun?
>> 
>> 
>> On Sep 25, 2012, at 1:43 PM, mariana Vargas <mmaria...@yahoo.com.mx> wrote:
>> 
>>> Hi
>>> 
>>> I fact I found what is the origin of this problem and it is because all 
>>> processes have rank 0, so I tested and in effect even when I send the 
>>> clasical Hello.py give the same, how can I solved this?? Do I  re installed 
>>> every again???
>>> 
>>> Help please...
>>> 
>>> Mariana
>>> 
>>> 
>>> 
>>> On Sep 24, 2012, at 9:13 PM, Mariana Vargas Magana wrote:
>>> 
>>>> 
>>>> 
>>>> Yes you are right this is what it says but if fact the weird thing is that 
>>>> not all times the error message appears….I send to 20 nodes and only one 
>>>> gives this message, is this normal…
>>>> 
>>>> 
>>>> 
>>>> 
>>>> On Sep 24, 2012, at 8:00 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>> 
>>>>> Well, as it says, your processes called MPI_Init, but at least one of 
>>>>> them exited without calling MPI_Finalize. That violates the MPI rules and 
>>>>> we therefore terminate the remaining processes.
>>>>> 
>>>>> Check your code and see how/why you are doing that - you probably have a 
>>>>> code path whereby a process exits without calling finalize.
>>>>> 
>>>>> 
>>>>> On Sep 24, 2012, at 4:37 PM, mariana Vargas <mmaria...@yahoo.com.mx> 
>>>>> wrote:
>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> Hi all
>>>>>> 
>>>>>> I get this error when I run a paralelized python code in a cluster, 
>>>>>> could anyone give me an idea of what is happening? I'am new in this 
>>>>>> Thanks...
>>>>>> 
>>>>>> mpirun has exited due to process rank 2 with PID 10259 on
>>>>>> node f01 exiting improperly. There are two reasons this could occur:
>>>>>> 
>>>>>> 1. this process did not call "init" before exiting, but others in
>>>>>> the job did. This can cause a job to hang indefinitely while it waits
>>>>>> for all processes to call "init". By rule, if one process calls "init",
>>>>>> then ALL processes must call "init" prior to termination.
>>>>>> 
>>>>>> 2. this process called "init", but exited without calling "finalize".
>>>>>> By rule, all processes that call "init" MUST call "finalize" prior to
>>>>>> exiting or it will be considered an "abnormal termination"
>>>>>> 
>>>>>> This may have caused other processes in the application to be
>>>>>> terminated by signals sent by mpirun (as reported here).
>>>>>> 
>>>>>> Thanks!!
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>>> 
>>>>>>> Dr. Mariana Vargas Magana
>>>>>>> Astroparticule et Cosmologie - Bureau 409B
>>>>>>> PHD student- Université Denis Diderot-Paris 7
>>>>>>> 10, rue Alice Domon et Léonie Duquet
>>>>>>> 75205 Paris Cedex - France
>>>>>>> Tel. +33 (0)1 57 27 70 32
>>>>>>> Fax. +33 (0)1 57 27 60 71
>>>>>>> mari...@apc.univ-paris7.fr
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to