Again, what is PTP?

I can't replicate this on any system we can access, so it may be something 
about this PTP thing.

On Apr 18, 2010, at 1:37 AM, Mario Ogrizek wrote:

> Ofcourse i checked that, i have all of this things, 
> I simplified the program, and its the same.
> Nothing gave me clue, except the more detailed writeout of the PTP.
> Here is the critical part of it:
> (1.2 one, this is correct)
> [Mario.local:05548]  Map for job: 1   Generated by mapping mode: byslot
>       Starting vpid: 0        Vpid range: 4   Num app_contexts: 1
> ...
> ...
> 
> (1.4 one)
> [Mario.local:05542]  Map for job: 0   Generated by mapping mode: byslot
>       Starting vpid: 0        Vpid range: 1   Num app_contexts: 1
> ...
> ...
> 
> Seems the 1.4 mapps the wrong job, Im not sure to what is it referred to, but 
> hope it will give you some clues.
>  
> On Sun, Apr 18, 2010 at 4:07 AM, Ralph Castain <r...@open-mpi.org> wrote:
> Just to check what is going on, why don't you remove that message passing 
> code and just
> 
> printf("Hello MPI World from process %d!", my_rank
> 
>  in each process? Much more direct - avoids any ambiguity.
> 
> Also, be certain that you compile this program for the specific OMPI version 
> you are running it under. OMPI is NOT binary compatible across releases - you 
> have to recompile the program for the specific release you are going to use.
> 
> 
> On Apr 17, 2010, at 4:52 PM, Mario Ogrizek wrote:
> 
>> Ofcourse, its the same program, wasnt recompiled for a week.
>> 
>> 
>> #include <stdio.h>
>> #include <string.h>
>> #include "mpi.h"
>> 
>> int main(int argc, char* argv[]){
>>      int  my_rank; /* rank of process */
>>      int  p;       /* number of processes */
>>      int source;   /* rank of sender */
>>      int dest;     /* rank of receiver */
>>      int tag=0;    /* tag for messages */
>>      char message[100];        /* storage for message */
>>      MPI_Status status ;   /* return status for receive */
>>      
>>      /* start up MPI */
>>      
>>      MPI_Init(&argc, &argv);
>> 
>>      /* find out process rank */
>>      MPI_Comm_rank(MPI_COMM_WORLD, &my_rank); 
>>      
>> 
>>      /* find out number of processes */
>>      MPI_Comm_size(MPI_COMM_WORLD, &p);
>> 
>>      
>>      if (my_rank !=0){
>>              /* create message */
>>              sprintf(message, "Hello MPI World from process %d!", my_rank);
>>              dest = 0;
>>              /* use strlen+1 so that '\0' get transmitted */
>>              MPI_Send(message, strlen(message)+1, MPI_CHAR,
>>                 dest, tag, MPI_COMM_WORLD);
>>      }
>>      else{
>>              printf("Hello MPI World From process 0: Num processes: %d\n",p);
>>              for (source = 1; source < p; source++) {
>>                      MPI_Recv(message, 100, MPI_CHAR, source, tag,
>>                            MPI_COMM_WORLD, &status);
>>                      printf("%s\n",message);
>>              }
>>      }
>>      /* shut down MPI */
>>      MPI_Finalize(); 
>>      
>>      
>>      return 0;
>> }
>> 
>> I triplechecked:
>> v1.2 output
>> Hello MPI World From process 0: Num processes: 4
>> Hello MPI World from process 1!
>> Hello MPI World from process 2!
>> Hello MPI World from process 3!
>> 
>> v1.4 output:
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> On Sat, Apr 17, 2010 at 9:13 PM, Ralph Castain <r...@open-mpi.org> wrote:
>> 
>> On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:
>> 
>>> Hahaha, ok then that WAS silly! :D
>>> So there is no way to utilize both cores with mpi?
>> 
>> We are using both cores - it is just that they are on the same node. Unless 
>> told otherwise, the processes will use shared memory for communication.
>> 
>>> 
>>> Ah well, I'll correct that.
>>> 
>>> From console, im starting a job like this: mpirun -np 4 Program, where i 
>>> want to run a Program on 4 processors.
>>> I was just stumbled when i got same output 4 times, like there are 4 
>>> processes ranked 0.
>>> While with the old version of mpi (1.2) same execution would give 4 
>>> processes ranked 0..3.
>> 
>> And so you should - if not, then there is something wrong. No way mpirun 
>> would start 4 processes ranked 0. How are you printing the rank? Are you 
>> sure you are getting it correctly?
>> 
>> 
>>> 
>>> Hope you see my question.
>>> 
>>> On Sat, Apr 17, 2010 at 6:29 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>> 
>>> On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:
>>> 
>>>> I am new to mpi, so I'm sorry for any silly questions.
>>>> 
>>>> My idea was to try to use dual core machine as two nodes. I have a limited 
>>>> access to a cluster, so this was just for "testing" purposes.
>>>> My default hostfile contains usual comments and this two nodes:
>>>> 
>>>>> node0
>>>>> node1
>>>> I thought that each processor is a node for MPI purpose.
>>> 
>>> I'm afraid not - it is just another processor on that node. So you only 
>>> have one node as far as OMPI is concerned.
>>> 
>>>> Im not sure what do you mean with "mpirun cmd line"?
>>> 
>>> How are you starting your job? The usual way is with "mpirun -n N ...". 
>>> That is what we mean by the "mpirun cmd line" - i.e., what command are you 
>>> using to start your job?
>>> 
>>> It sounds like things are actually working correctly. You might look at 
>>> "mpirun -h" for possible options of interest.
>>> 
>>> 
>>>> 
>>>> Regards,
>>>> 
>>>> Mario
>>>> 
>>>> On Sat, Apr 17, 2010 at 1:54 AM, Ralph Castain <r...@open-mpi.org> wrote:
>>>> 
>>>> On Apr 16, 2010, at 5:08 PM, Mario Ogrizek wrote:
>>>> 
>>>>> I checked the default MCA param file, and found it was there that was 
>>>>> (automatically) specified as a relative path, so i changed it.
>>>>> So now, it works, altho, still something is not right.
>>>>> Seems like its creating 4 times only 1 process.
>>>>> Not sure if it has to do something with my hostfile, it contains:
>>>>> 
>>>>> node0
>>>>> node1
>>>>> 
>>>>> I am running this on a simple dualcore machine, so i specified it as a 
>>>>> localhost with two nodes.
>>>> 
>>>> I don't understand this comment - a dual core machine would still be a 
>>>> single node. Just happens to have two processors in it.
>>>> 
>>>> Could you send the contents of your hostfile and your mpirun cmd line?
>>>> 
>>>>> 
>>>>> Regards,
>>>>> 
>>>>> Mario
>>>>> 
>>>>> On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek 
>>>>> <mario.guard...@gmail.com> wrote:
>>>>> I understand, so, its looking for a 
>>>>> working_dir/usr/local/etc/openmpi-default-hostfile
>>>>> I managed to run a hello world program from the console, while my wd was 
>>>>> just "/" and it worked, altho strangely...
>>>>> example for 4 procs:
>>>>> 
>>>>> Hello MPI World From process 0: Num processes: 1
>>>>> Hello MPI World From process 0: Num processes: 1
>>>>> Hello MPI World From process 0: Num processes: 1
>>>>> Hello MPI World From process 0: Num processes: 1
>>>>> 
>>>>> So, you are saying i allways have to be in "/" to run mpi programs, or 
>>>>> there is a way for mpi to search absolute path?
>>>>> It seems pretty inconvinient this way.
>>>>> I think v 1.2 didnt have this limitation.
>>>>> 
>>>>> Does this have to do anything with LD_LIBRARY_PATH?
>>>>> 
>>>>> Regards,
>>>>> 
>>>>> Mario
>>>>> 
>>>>> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>> How did you specify it? Command line? Default MCA param file?
>>>>> 
>>>>> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:
>>>>> 
>>>>>> Any idea how to solve this?
>>>>>> 
>>>>>> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov 
>>>>>> <timur.magome...@developonbox.ru> wrote:
>>>>>> Hello.
>>>>>> It looks that you hostfile path should
>>>>>> be /usr/local/etc/openmpi-default-hostfile not
>>>>>> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
>>>>>> second path.
>>>>>> 
>>>>>> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
>>>>>> > Well, im not sure why should i name it /openmpi-default-hostfile
>>>>>> > Especially, because mpirun v1.2 executes without any errors.
>>>>>> > But, i made a copy named /openmpi-default-hostfile, and still, the
>>>>>> > same result.
>>>>>> >
>>>>>> > This is the whole error message for a simple hello world program:
>>>>>> >
>>>>>> >
>>>>>> > Open RTE was unable to open the hostfile:
>>>>>> >     usr/local/etc/openmpi-default-hostfile
>>>>>> > Check to make sure the path and filename are correct.
>>>>>> > --------------------------------------------------------------------------
>>>>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>>>>> > base/ras_base_allocate.c at line 186
>>>>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>>>>> > base/plm_base_launch_support.c at line 72
>>>>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>>>>> > plm_rsh_module.c at line 990
>>>>>> > --------------------------------------------------------------------------
>>>>>> > A daemon (pid unknown) died unexpectedly on signal 1  while attempting
>>>>>> > to
>>>>>> > launch so we are aborting.
>>>>>> >
>>>>>> >
>>>>>> > There may be more information reported by the environment (see above).
>>>>>> >
>>>>>> >
>>>>>> > This may be because the daemon was unable to find all the needed
>>>>>> > shared
>>>>>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
>>>>>> > the
>>>>>> > location of the shared libraries on the remote nodes and this will
>>>>>> > automatically be forwarded to the remote nodes.
>>>>>> > --------------------------------------------------------------------------
>>>>>> > --------------------------------------------------------------------------
>>>>>> > mpirun noticed that the job aborted, but has no info as to the process
>>>>>> > that caused that situation.
>>>>>> > --------------------------------------------------------------------------
>>>>>> > mpirun: clean termination accomplished
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > ps. PTP is a parallel tools platform plugin for eclipse
>>>>>> >
>>>>>> >
>>>>>> > Regards,
>>>>>> >
>>>>>> >
>>>>>> > Mario
>>>>>> >
>>>>>> > _______________________________________________
>>>>>> > users mailing list
>>>>>> > us...@open-mpi.org
>>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> 
>>>>>> 
>>>>>> --
>>>>>> Kind regards,
>>>>>> Timur Magomedov
>>>>>> Senior C++ Developer
>>>>>> DevelopOnBox LLC / Zodiac Interactive
>>>>>> http://www.zodiac.tv/
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to