I don't think that is true, but I suggest you check the mpi4py examples. I believe all import does is import function definitions - it doesn't execute anything.
Sent from my iPad On Sep 25, 2012, at 2:41 PM, mariana Vargas <mmaria...@yahoo.com.mx> wrote: > MPI_init() is actually called when import MPI module from MPi package... > > > On Sep 25, 2012, at 5:17 PM, Ralph Castain wrote: > >> You forgot to call MPI_Init at the beginning of your program. >> >> On Sep 25, 2012, at 2:08 PM, Mariana Vargas Magana <mmaria...@yahoo.com.mx> >> wrote: >> >>> Hi >>> I think I'am not understanding what you said , here is the hello.py and >>> next the command mpirun… >>> >>> Thanks! >>> >>> #!/usr/bin/env python >>> """ >>> Parallel Hello World >>> """ >>> >>> from mpi4py import MPI >>> import sys >>> >>> size = MPI.COMM_WORLD.Get_size() >>> rank = MPI.COMM_WORLD.Get_rank() >>> name = MPI.Get_processor_name() >>> >>> sys.stdout.write( >>> "Hello, World! I am process %d of %d on %s.\n" >>> % (rank, size, name)) >>> >>> ~/bin/mpirun -np 70 python2.7 helloworld.py >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> Hello, World! I am process 0 of 1 on ferrari. >>> >>> On Sep 25, 2012, at 4:46 PM, Ralph Castain <r...@open-mpi.org> wrote: >>> >>>> The usual reason for this is that you aren't launching these processes >>>> correctly. How are you starting your job? Are you using mpirun? >>>> >>>> >>>> On Sep 25, 2012, at 1:43 PM, mariana Vargas <mmaria...@yahoo.com.mx> wrote: >>>> >>>>> Hi >>>>> >>>>> I fact I found what is the origin of this problem and it is because all >>>>> processes have rank 0, so I tested and in effect even when I send the >>>>> clasical Hello.py give the same, how can I solved this?? Do I re >>>>> installed every again??? >>>>> >>>>> Help please... >>>>> >>>>> Mariana >>>>> >>>>> >>>>> >>>>> On Sep 24, 2012, at 9:13 PM, Mariana Vargas Magana wrote: >>>>> >>>>>> >>>>>> >>>>>> Yes you are right this is what it says but if fact the weird thing is >>>>>> that not all times the error message appears….I send to 20 nodes and >>>>>> only one gives this message, is this normal… >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Sep 24, 2012, at 8:00 PM, Ralph Castain <r...@open-mpi.org> wrote: >>>>>> >>>>>>> Well, as it says, your processes called MPI_Init, but at least one of >>>>>>> them exited without calling MPI_Finalize. That violates the MPI rules >>>>>>> and we therefore terminate the remaining processes. >>>>>>> >>>>>>> Check your code and see how/why you are doing that - you probably have >>>>>>> a code path whereby a process exits without calling finalize. >>>>>>> >>>>>>> >>>>>>> On Sep 24, 2012, at 4:37 PM, mariana Vargas <mmaria...@yahoo.com.mx> >>>>>>> wrote: >>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Hi all >>>>>>>> >>>>>>>> I get this error when I run a paralelized python code in a cluster, >>>>>>>> could anyone give me an idea of what is happening? I'am new in this >>>>>>>> Thanks... >>>>>>>> >>>>>>>> mpirun has exited due to process rank 2 with PID 10259 on >>>>>>>> node f01 exiting improperly. There are two reasons this could occur: >>>>>>>> >>>>>>>> 1. this process did not call "init" before exiting, but others in >>>>>>>> the job did. This can cause a job to hang indefinitely while it waits >>>>>>>> for all processes to call "init". By rule, if one process calls "init", >>>>>>>> then ALL processes must call "init" prior to termination. >>>>>>>> >>>>>>>> 2. this process called "init", but exited without calling "finalize". >>>>>>>> By rule, all processes that call "init" MUST call "finalize" prior to >>>>>>>> exiting or it will be considered an "abnormal termination" >>>>>>>> >>>>>>>> This may have caused other processes in the application to be >>>>>>>> terminated by signals sent by mpirun (as reported here). >>>>>>>> >>>>>>>> Thanks!! >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> Dr. Mariana Vargas Magana >>>>>>>>> Astroparticule et Cosmologie - Bureau 409B >>>>>>>>> PHD student- Université Denis Diderot-Paris 7 >>>>>>>>> 10, rue Alice Domon et Léonie Duquet >>>>>>>>> 75205 Paris Cedex - France >>>>>>>>> Tel. +33 (0)1 57 27 70 32 >>>>>>>>> Fax. +33 (0)1 57 27 60 71 >>>>>>>>> mari...@apc.univ-paris7.fr >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> us...@open-mpi.org >>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> us...@open-mpi.org >>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> us...@open-mpi.org >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users