Well, not sure what I can advise. Check to ensure that your LD_LIBRARY_PATH is pointing to the same installation where your mpirun is located. For whatever reason, the processes think they are singletons - i.e., that they were not actually started by mpirun.
You might also want to ask the mpi4py folks - we aren't very familiar with that package over here. It could be that you need to configure it for OpenMPI as opposed to mpich. On Tue, Sep 25, 2012 at 7:08 PM, Mariana Vargas Magana < mmaria...@yahoo.com.mx> wrote: > > Yes I am sure I read from a mpi4py guide I already check the examples if > fact this an example extracted from a guide…!! Evenmore this example if I > use with mpich2 it runs very nicely, even though for the other code I need > openmpi working =s > > Mariana > > > > On Sep 25, 2012, at 8:00 PM, Ralph Castain <rhc.open...@gmail.com> wrote: > > I don't think that is true, but I suggest you check the mpi4py examples. I > believe all import does is import function definitions - it doesn't execute > anything. > > Sent from my iPad > > On Sep 25, 2012, at 2:41 PM, mariana Vargas <mmaria...@yahoo.com.mx> > wrote: > > MPI_init() is actually called when import MPI module from MPi package... > > > On Sep 25, 2012, at 5:17 PM, Ralph Castain wrote: > > You forgot to call MPI_Init at the beginning of your program. > > On Sep 25, 2012, at 2:08 PM, Mariana Vargas Magana <mmaria...@yahoo.com.mx> > wrote: > > Hi > I think I'am not understanding what you said , here is the hello.py and > next the command mpirun… > > Thanks! > > #!/usr/bin/env python > """ > Parallel Hello World > """ > > from mpi4py import MPI > import sys > > size = MPI.COMM_WORLD.Get_size() > rank = MPI.COMM_WORLD.Get_rank() > name = MPI.Get_processor_name() > > sys.stdout.write( > "Hello, World! I am process %d of %d on %s.\n" > % (rank, size, name)) > > ~/bin/mpirun -np 70 python2.7 helloworld.py > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > Hello, World! I am process 0 of 1 on ferrari. > > On Sep 25, 2012, at 4:46 PM, Ralph Castain <r...@open-mpi.org> wrote: > > The usual reason for this is that you aren't launching these processes > correctly. How are you starting your job? Are you using mpirun? > > > On Sep 25, 2012, at 1:43 PM, mariana Vargas <mmaria...@yahoo.com.mx> > wrote: > > Hi > > I fact I found what is the origin of this problem and it is because all > processes have rank 0, so I tested and in effect even when I send the > clasical Hello.py give the same, how can I solved this?? Do I re installed > every again??? > > Help please... > > Mariana > > > > On Sep 24, 2012, at 9:13 PM, Mariana Vargas Magana wrote: > > > > Yes you are right this is what it says but if fact the weird thing is that > not all times the error message appears….I send to 20 nodes and only one > gives this message, is this normal… > > > > > On Sep 24, 2012, at 8:00 PM, Ralph Castain <r...@open-mpi.org> wrote: > > Well, as it says, your processes called MPI_Init, but at least one of them > exited without calling MPI_Finalize. That violates the MPI rules and we > therefore terminate the remaining processes. > > Check your code and see how/why you are doing that - you probably have a > code path whereby a process exits without calling finalize. > > > On Sep 24, 2012, at 4:37 PM, mariana Vargas <mmaria...@yahoo.com.mx> > wrote: > > > > > Hi all > > > I get this error when I run a paralelized python code in a cluster, could > anyone give me an idea of what is happening? I'am new in this Thanks... > > > mpirun has exited due to process rank 2 with PID 10259 on > > node f01 exiting improperly. There are two reasons this could occur: > > > 1. this process did not call "init" before exiting, but others in > > the job did. This can cause a job to hang indefinitely while it waits > > for all processes to call "init". By rule, if one process calls "init", > > then ALL processes must call "init" prior to termination. > > > 2. this process called "init", but exited without calling "finalize". > > By rule, all processes that call "init" MUST call "finalize" prior to > > exiting or it will be considered an "abnormal termination" > > > This may have caused other processes in the application to be > > terminated by signals sent by mpirun (as reported here). > > > Thanks!! > > > > > > Dr. Mariana Vargas Magana > Astroparticule et Cosmologie - Bureau 409B > PHD student- Université Denis Diderot-Paris 7 > 10, rue Alice Domon et Léonie Duquet > 75205 Paris Cedex - France > Tel. +33 (0)1 57 27 70 32 > Fax. +33 (0)1 57 27 60 71 > mari...@apc.univ-paris7.fr > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >