Whenever I have seen this problem, it has been because of a mismatch between 
mpirun and the back-end libraries that are linked against the executable.  For 
example, the app was compiled/linked against Open MPI a.b.c and either some 
other mpirun was used (e.g., from MPICH) or the mpirun from Open MPI x.y.z was 
used.



On Apr 18, 2010, at 2:28 PM, Ralph Castain wrote:

> Afraid I can't help you - I've never seen that behavior on any system, can't 
> replicate it anywhere, and have no idea what might cause it.
> 
> On Apr 18, 2010, at 9:15 AM, Mario Ogrizek wrote:
> 
>> It is a parallel tools platform for eclipse IDE, a plugin.
>> I dont think it is a source of problem.
>> 
>> The same thing is happening running it from shell. It has something to do 
>> with mapping or something else. Since it allways maps for job 0, what ever 
>> that means.
>> 
>> On Sun, Apr 18, 2010 at 4:50 PM, Ralph Castain <r...@open-mpi.org> wrote:
>> Again, what is PTP?
>> 
>> I can't replicate this on any system we can access, so it may be something 
>> about this PTP thing.
>> 
>> On Apr 18, 2010, at 1:37 AM, Mario Ogrizek wrote:
>> 
>>> Ofcourse i checked that, i have all of this things, 
>>> I simplified the program, and its the same.
>>> Nothing gave me clue, except the more detailed writeout of the PTP.
>>> Here is the critical part of it:
>>> (1.2 one, this is correct)
>>> [Mario.local:05548]  Map for job: 1 Generated by mapping mode: byslot
>>>     Starting vpid: 0        Vpid range: 4   Num app_contexts: 1
>>> ...
>>> ...
>>> 
>>> (1.4 one)
>>> [Mario.local:05542]  Map for job: 0 Generated by mapping mode: byslot
>>>     Starting vpid: 0        Vpid range: 1   Num app_contexts: 1
>>> ...
>>> ...
>>> 
>>> Seems the 1.4 mapps the wrong job, Im not sure to what is it referred to, 
>>> but hope it will give you some clues.
>>>  
>>> On Sun, Apr 18, 2010 at 4:07 AM, Ralph Castain <r...@open-mpi.org> wrote:
>>> Just to check what is going on, why don't you remove that message passing 
>>> code and just
>>> 
>>> printf("Hello MPI World from process %d!", my_rank
>>> 
>>>  in each process? Much more direct - avoids any ambiguity.
>>> 
>>> Also, be certain that you compile this program for the specific OMPI 
>>> version you are running it under. OMPI is NOT binary compatible across 
>>> releases - you have to recompile the program for the specific release you 
>>> are going to use.
>>> 
>>> 
>>> On Apr 17, 2010, at 4:52 PM, Mario Ogrizek wrote:
>>> 
>>>> Ofcourse, its the same program, wasnt recompiled for a week.
>>>> 
>>>> 
>>>> #include <stdio.h>
>>>> #include <string.h>
>>>> #include "mpi.h"
>>>> 
>>>> int main(int argc, char* argv[]){
>>>>    int  my_rank; /* rank of process */
>>>>    int  p;       /* number of processes */
>>>>    int source;   /* rank of sender */
>>>>    int dest;     /* rank of receiver */
>>>>    int tag=0;    /* tag for messages */
>>>>    char message[100];        /* storage for message */
>>>>    MPI_Status status ;   /* return status for receive */
>>>>    
>>>>    /* start up MPI */
>>>>    
>>>>    MPI_Init(&argc, &argv);
>>>> 
>>>>    /* find out process rank */
>>>>    MPI_Comm_rank(MPI_COMM_WORLD, &my_rank); 
>>>>    
>>>> 
>>>>    /* find out number of processes */
>>>>    MPI_Comm_size(MPI_COMM_WORLD, &p);
>>>> 
>>>>    
>>>>    if (my_rank !=0){
>>>>            /* create message */
>>>>            sprintf(message, "Hello MPI World from process %d!", my_rank);
>>>>            dest = 0;
>>>>            /* use strlen+1 so that '\0' get transmitted */
>>>>            MPI_Send(message, strlen(message)+1, MPI_CHAR,
>>>>               dest, tag, MPI_COMM_WORLD);
>>>>    }
>>>>    else{
>>>>            printf("Hello MPI World From process 0: Num processes: %d\n",p);
>>>>            for (source = 1; source < p; source++) {
>>>>                    MPI_Recv(message, 100, MPI_CHAR, source, tag,
>>>>                          MPI_COMM_WORLD, &status);
>>>>                    printf("%s\n",message);
>>>>            }
>>>>    }
>>>>    /* shut down MPI */
>>>>    MPI_Finalize(); 
>>>>    
>>>>    
>>>>    return 0;
>>>> }
>>>> 
>>>> I triplechecked:
>>>> v1.2 output
>>>> Hello MPI World From process 0: Num processes: 4
>>>> Hello MPI World from process 1!
>>>> Hello MPI World from process 2!
>>>> Hello MPI World from process 3!
>>>> 
>>>> v1.4 output:
>>>> Hello MPI World From process 0: Num processes: 1
>>>> Hello MPI World From process 0: Num processes: 1
>>>> Hello MPI World From process 0: Num processes: 1
>>>> Hello MPI World From process 0: Num processes: 1
>>>> 
>>>> 
>>>> 
>>>> 
>>>> 
>>>> 
>>>> 
>>>> On Sat, Apr 17, 2010 at 9:13 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>> 
>>>> On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:
>>>> 
>>>>> Hahaha, ok then that WAS silly! :D
>>>>> So there is no way to utilize both cores with mpi?
>>>> 
>>>> We are using both cores - it is just that they are on the same node. 
>>>> Unless told otherwise, the processes will use shared memory for 
>>>> communication.
>>>> 
>>>>> 
>>>>> Ah well, I'll correct that.
>>>>> 
>>>>> From console, im starting a job like this: mpirun -np 4 Program, where i 
>>>>> want to run a Program on 4 processors.
>>>>> I was just stumbled when i got same output 4 times, like there are 4 
>>>>> processes ranked 0.
>>>>> While with the old version of mpi (1.2) same execution would give 4 
>>>>> processes ranked 0..3.
>>>> 
>>>> And so you should - if not, then there is something wrong. No way mpirun 
>>>> would start 4 processes ranked 0. How are you printing the rank? Are you 
>>>> sure you are getting it correctly?
>>>> 
>>>> 
>>>>> 
>>>>> Hope you see my question.
>>>>> 
>>>>> On Sat, Apr 17, 2010 at 6:29 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>> 
>>>>> On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:
>>>>> 
>>>>>> I am new to mpi, so I'm sorry for any silly questions.
>>>>>> 
>>>>>> My idea was to try to use dual core machine as two nodes. I have a 
>>>>>> limited access to a cluster, so this was just for "testing" purposes.
>>>>>> My default hostfile contains usual comments and this two nodes:
>>>>>> 
>>>>>>> node0
>>>>>>> node1
>>>>>> I thought that each processor is a node for MPI purpose.
>>>>> 
>>>>> I'm afraid not - it is just another processor on that node. So you only 
>>>>> have one node as far as OMPI is concerned.
>>>>> 
>>>>>> Im not sure what do you mean with "mpirun cmd line"?
>>>>> 
>>>>> How are you starting your job? The usual way is with "mpirun -n N ...". 
>>>>> That is what we mean by the "mpirun cmd line" - i.e., what command are 
>>>>> you using to start your job?
>>>>> 
>>>>> It sounds like things are actually working correctly. You might look at 
>>>>> "mpirun -h" for possible options of interest.
>>>>> 
>>>>> 
>>>>>> 
>>>>>> Regards,
>>>>>> 
>>>>>> Mario
>>>>>> 
>>>>>> On Sat, Apr 17, 2010 at 1:54 AM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>>> 
>>>>>> On Apr 16, 2010, at 5:08 PM, Mario Ogrizek wrote:
>>>>>> 
>>>>>>> I checked the default MCA param file, and found it was there that was 
>>>>>>> (automatically) specified as a relative path, so i changed it.
>>>>>>> So now, it works, altho, still something is not right.
>>>>>>> Seems like its creating 4 times only 1 process.
>>>>>>> Not sure if it has to do something with my hostfile, it contains:
>>>>>>> 
>>>>>>> node0
>>>>>>> node1
>>>>>>> 
>>>>>>> I am running this on a simple dualcore machine, so i specified it as a 
>>>>>>> localhost with two nodes.
>>>>>> 
>>>>>> I don't understand this comment - a dual core machine would still be a 
>>>>>> single node. Just happens to have two processors in it.
>>>>>> 
>>>>>> Could you send the contents of your hostfile and your mpirun cmd line?
>>>>>> 
>>>>>>> 
>>>>>>> Regards,
>>>>>>> 
>>>>>>> Mario
>>>>>>> 
>>>>>>> On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek 
>>>>>>> <mario.guard...@gmail.com> wrote:
>>>>>>> I understand, so, its looking for a 
>>>>>>> working_dir/usr/local/etc/openmpi-default-hostfile
>>>>>>> I managed to run a hello world program from the console, while my wd 
>>>>>>> was just "/" and it worked, altho strangely...
>>>>>>> example for 4 procs:
>>>>>>> 
>>>>>>> Hello MPI World From process 0: Num processes: 1
>>>>>>> Hello MPI World From process 0: Num processes: 1
>>>>>>> Hello MPI World From process 0: Num processes: 1
>>>>>>> Hello MPI World From process 0: Num processes: 1
>>>>>>> 
>>>>>>> So, you are saying i allways have to be in "/" to run mpi programs, or 
>>>>>>> there is a way for mpi to search absolute path?
>>>>>>> It seems pretty inconvinient this way.
>>>>>>> I think v 1.2 didnt have this limitation.
>>>>>>> 
>>>>>>> Does this have to do anything with LD_LIBRARY_PATH?
>>>>>>> 
>>>>>>> Regards,
>>>>>>> 
>>>>>>> Mario
>>>>>>> 
>>>>>>> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain <r...@open-mpi.org> 
>>>>>>> wrote:
>>>>>>> How did you specify it? Command line? Default MCA param file?
>>>>>>> 
>>>>>>> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:
>>>>>>> 
>>>>>>>> Any idea how to solve this?
>>>>>>>> 
>>>>>>>> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov 
>>>>>>>> <timur.magome...@developonbox.ru> wrote:
>>>>>>>> Hello.
>>>>>>>> It looks that you hostfile path should
>>>>>>>> be /usr/local/etc/openmpi-default-hostfile not
>>>>>>>> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
>>>>>>>> second path.
>>>>>>>> 
>>>>>>>> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
>>>>>>>> > Well, im not sure why should i name it /openmpi-default-hostfile
>>>>>>>> > Especially, because mpirun v1.2 executes without any errors.
>>>>>>>> > But, i made a copy named /openmpi-default-hostfile, and still, the
>>>>>>>> > same result.
>>>>>>>> >
>>>>>>>> > This is the whole error message for a simple hello world program:
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > Open RTE was unable to open the hostfile:
>>>>>>>> >     usr/local/etc/openmpi-default-hostfile
>>>>>>>> > Check to make sure the path and filename are correct.
>>>>>>>> > --------------------------------------------------------------------------
>>>>>>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>>>>>>> > base/ras_base_allocate.c at line 186
>>>>>>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>>>>>>> > base/plm_base_launch_support.c at line 72
>>>>>>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>>>>>>> > plm_rsh_module.c at line 990
>>>>>>>> > --------------------------------------------------------------------------
>>>>>>>> > A daemon (pid unknown) died unexpectedly on signal 1  while 
>>>>>>>> > attempting
>>>>>>>> > to
>>>>>>>> > launch so we are aborting.
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > There may be more information reported by the environment (see 
>>>>>>>> > above).
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > This may be because the daemon was unable to find all the needed
>>>>>>>> > shared
>>>>>>>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to 
>>>>>>>> > have
>>>>>>>> > the
>>>>>>>> > location of the shared libraries on the remote nodes and this will
>>>>>>>> > automatically be forwarded to the remote nodes.
>>>>>>>> > --------------------------------------------------------------------------
>>>>>>>> > --------------------------------------------------------------------------
>>>>>>>> > mpirun noticed that the job aborted, but has no info as to the 
>>>>>>>> > process
>>>>>>>> > that caused that situation.
>>>>>>>> > --------------------------------------------------------------------------
>>>>>>>> > mpirun: clean termination accomplished
>>>>>>>> >
>>>>>>>> >
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > ps. PTP is a parallel tools platform plugin for eclipse
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > Regards,
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > Mario
>>>>>>>> >
>>>>>>>> > _______________________________________________
>>>>>>>> > users mailing list
>>>>>>>> > us...@open-mpi.org
>>>>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> 
>>>>>>>> 
>>>>>>>> --
>>>>>>>> Kind regards,
>>>>>>>> Timur Magomedov
>>>>>>>> Senior C++ Developer
>>>>>>>> DevelopOnBox LLC / Zodiac Interactive
>>>>>>>> http://www.zodiac.tv/
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> 
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> 
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to