On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote: > I am new to mpi, so I'm sorry for any silly questions. > > My idea was to try to use dual core machine as two nodes. I have a limited > access to a cluster, so this was just for "testing" purposes. > My default hostfile contains usual comments and this two nodes: > >> node0 >> node1 > I thought that each processor is a node for MPI purpose.
I'm afraid not - it is just another processor on that node. So you only have one node as far as OMPI is concerned. > Im not sure what do you mean with "mpirun cmd line"? How are you starting your job? The usual way is with "mpirun -n N ...". That is what we mean by the "mpirun cmd line" - i.e., what command are you using to start your job? It sounds like things are actually working correctly. You might look at "mpirun -h" for possible options of interest. > > Regards, > > Mario > > On Sat, Apr 17, 2010 at 1:54 AM, Ralph Castain <r...@open-mpi.org> wrote: > > On Apr 16, 2010, at 5:08 PM, Mario Ogrizek wrote: > >> I checked the default MCA param file, and found it was there that was >> (automatically) specified as a relative path, so i changed it. >> So now, it works, altho, still something is not right. >> Seems like its creating 4 times only 1 process. >> Not sure if it has to do something with my hostfile, it contains: >> >> node0 >> node1 >> >> I am running this on a simple dualcore machine, so i specified it as a >> localhost with two nodes. > > I don't understand this comment - a dual core machine would still be a single > node. Just happens to have two processors in it. > > Could you send the contents of your hostfile and your mpirun cmd line? > >> >> Regards, >> >> Mario >> >> On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek <mario.guard...@gmail.com> >> wrote: >> I understand, so, its looking for a >> working_dir/usr/local/etc/openmpi-default-hostfile >> I managed to run a hello world program from the console, while my wd was >> just "/" and it worked, altho strangely... >> example for 4 procs: >> >> Hello MPI World From process 0: Num processes: 1 >> Hello MPI World From process 0: Num processes: 1 >> Hello MPI World From process 0: Num processes: 1 >> Hello MPI World From process 0: Num processes: 1 >> >> So, you are saying i allways have to be in "/" to run mpi programs, or there >> is a way for mpi to search absolute path? >> It seems pretty inconvinient this way. >> I think v 1.2 didnt have this limitation. >> >> Does this have to do anything with LD_LIBRARY_PATH? >> >> Regards, >> >> Mario >> >> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain <r...@open-mpi.org> wrote: >> How did you specify it? Command line? Default MCA param file? >> >> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote: >> >>> Any idea how to solve this? >>> >>> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov >>> <timur.magome...@developonbox.ru> wrote: >>> Hello. >>> It looks that you hostfile path should >>> be /usr/local/etc/openmpi-default-hostfile not >>> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the >>> second path. >>> >>> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет: >>> > Well, im not sure why should i name it /openmpi-default-hostfile >>> > Especially, because mpirun v1.2 executes without any errors. >>> > But, i made a copy named /openmpi-default-hostfile, and still, the >>> > same result. >>> > >>> > This is the whole error message for a simple hello world program: >>> > >>> > >>> > Open RTE was unable to open the hostfile: >>> > usr/local/etc/openmpi-default-hostfile >>> > Check to make sure the path and filename are correct. >>> > -------------------------------------------------------------------------- >>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file >>> > base/ras_base_allocate.c at line 186 >>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file >>> > base/plm_base_launch_support.c at line 72 >>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file >>> > plm_rsh_module.c at line 990 >>> > -------------------------------------------------------------------------- >>> > A daemon (pid unknown) died unexpectedly on signal 1 while attempting >>> > to >>> > launch so we are aborting. >>> > >>> > >>> > There may be more information reported by the environment (see above). >>> > >>> > >>> > This may be because the daemon was unable to find all the needed >>> > shared >>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have >>> > the >>> > location of the shared libraries on the remote nodes and this will >>> > automatically be forwarded to the remote nodes. >>> > -------------------------------------------------------------------------- >>> > -------------------------------------------------------------------------- >>> > mpirun noticed that the job aborted, but has no info as to the process >>> > that caused that situation. >>> > -------------------------------------------------------------------------- >>> > mpirun: clean termination accomplished >>> > >>> > >>> > >>> > >>> > ps. PTP is a parallel tools platform plugin for eclipse >>> > >>> > >>> > Regards, >>> > >>> > >>> > Mario >>> > >>> > _______________________________________________ >>> > users mailing list >>> > us...@open-mpi.org >>> > http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> >>> -- >>> Kind regards, >>> Timur Magomedov >>> Senior C++ Developer >>> DevelopOnBox LLC / Zodiac Interactive >>> http://www.zodiac.tv/ >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users