[OMPI users] Gathering results of MPI_SCAN

2012-05-29 Thread Rohan Deshpande
Hi, I have this array [1 2 3 4 5 6 7 8 9] and i am performing scan operation on that. I have 3 mpi tasks and each task gets 3 elements then each task calculates its scan and returns result to master task task 0 - [1 2 3] => [1 3 6] task 1 - [4 5 6 ] => [4 9 15] task 2 - [7 8 9] => [7 15 24] Now

Re: [OMPI users] MPI_SCAN

2012-05-19 Thread Rohan Deshpande
;http://www.open-mpi.org/doc/v1.6/man3/MPI_Reduce.3.php > > > On May 15, 2012, at 11:27 PM, Rohan Deshpande wrote: > > > I am performing Prefix scan operation on cluster > > > > I have 3 MPI tasks and master task is responsible for distributing the > data > &

Re: [OMPI users] MPI - memory sharing?

2012-05-17 Thread Rohan Deshpande
MCW rank 0 will be totally different than the malloc() that you do in > MCW rank 1. > > Make sense? > > On May 16, 2012, at 8:08 AM, Rohan Deshpande wrote: > > > I have following structure of MPI code - > > > > int main(){ > > > > MPI_I

[OMPI users] MPI - memory sharing?

2012-05-16 Thread Rohan Deshpande
I have following structure of MPI code - int main(){ MPI_INIT. //initialize MPI data = malloc(sizeof(int)*200); //initialize data /*Master-*/ if(taskid == 0){ //send data to each slave MPI_SEND } /*Slaves---*/ if(taskid > 0){ //accep

[OMPI users] MPI_SCAN

2012-05-15 Thread Rohan Deshpande
I am performing Prefix scan operation on cluster I have 3 MPI tasks and master task is responsible for distributing the data Now, each task calculates sum of its own part of array using GPUs and returns the results to master task. Master task also calculates its own part of array using GPU. Whe

[OMPI users] MPI books and resources

2012-05-12 Thread Rohan Deshpande
Hi, Can anyone point me to good resources and books to understand the detailed architecture and working of MPI. I would like to know all the minute details. Thanks, --

Re: [OMPI users] GPU and CPU timing - OpenMPI and Thrust

2012-05-08 Thread Rohan Deshpande
come out better. > > ** ** > > ** ** > > *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On > Behalf Of *Rohan Deshpande > *Sent:* Monday, May 07, 2012 9:38 PM > *To:* Open MPI Users > *Subject:* [OMPI users] GPU and CPU timing - OpenMPI and Thrust > &g

[OMPI users] GPU and CPU timing - OpenMPI and Thrust

2012-05-07 Thread Rohan Deshpande
I am running MPI and Thrust code on a cluster and measuring time for calculations. My MPI code - #include "mpi.h" #include #include #include #include #include #include #define MASTER 0 #define ARRAYSIZE 2000 int *masterarray,*onearray,*twoarray,*threearray,*fourarray,*fivearray,*s

Re: [OMPI users] Segmentation fault on master task

2012-05-03 Thread Rohan Deshpande
uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au> >> >ecm-c-l-207-004.**uniwa.uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>exited >> > on signal 11 (Segmentation fault). >> >> --**--** >> -- >> >> Sadly i cant install memory checker such as valgrind on my machine due to >> some restrictions. I could not spot any error by looking in code. >> >> Can anyone help me ?what is wrong in above code. >> >> Thanks >> > > > -- Best Regards, ROHAN DESHPANDE

[OMPI users] Segmentation fault on master task

2012-05-02 Thread Rohan Deshpande
Hi, I am trying to execute following code on cluster. run_kernel is a cuda call with signature int run_kernel(int array[],int nelements, int taskid, char hostname[]); My MPI code is #include "mpi.h" #include #include #include #define MASTER0 int *masterarray; int *onearray; int

Re: [OMPI users] OpenMPI and CUDA on cluster

2012-05-01 Thread Rohan Deshpande
> > http://forums.nvidia.com/index.php?showtopic=227854 > > ** ** > > Maybe that could be your issue too. > > ** ** > > ** ** > > ** ** > > *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On > Behalf Of *Rohan Deshpande > *Se

[OMPI users] OpenMPI and CUDA on cluster

2012-05-01 Thread Rohan Deshpande
Hi, I am trying to execute OpenMPI and CUDA code on a cluster. The code works fine on single machine but when I try to execute it on cluster I get error: error while loading shared libraries: libcudart.so.4: cannot open shared object file: No such file or directory I checked my PATH and LD_PATH

[OMPI users] MPI and CUDA

2012-04-24 Thread Rohan Deshpande
I am combining mpi and cuda. Trying to find out sum of array elements using cuda and using mpi to distribute the array. my cuda code #include __global__ void add(int *devarray, int *devsum) { int index = blockIdx.x * blockDim.x + threadIdx.x; *devsum = *devsum + devarray[index]

Re: [OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-19 Thread Rohan Deshpande
/* printf("task %d has elements:",myid); for(j = myoffset; j wrote: > Send the most recent version of your code. > > Have you tried running it through a memory-checking debugger, like > valgrind? > > > On Apr 19, 2012, at 4:24 AM, Rohan Deshpande wrote: >

Re: [OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-19 Thread Rohan Deshpande
d be the problem. > > Pascal > > users-boun...@open-mpi.org a écrit sur 19/04/2012 09:18:31 : > > > De : Rohan Deshpande > > A : Open MPI Users > > Date : 19/04/2012 09:18 > > Objet : Re: [OMPI users] machine exited on signal 11 (Segmentation > fault). &g

Re: [OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-19 Thread Rohan Deshpande
, Jeffrey Squyres wrote: > As a guess, you're passing in a bad address. > > Double check the buffers that you're sending to MPI_SEND/MPI_RECV/etc. > > > On Apr 17, 2012, at 10:43 PM, Rohan Deshpande wrote: > > > After using malloc i am getting following error

Re: [OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-17 Thread Rohan Deshpande
cally on the stack. > Something like: > > int *data; > > int main(..) { > { >data = malloc(ARRAYSIZE * sizeof(int)); >if (NULL == data) { >perror("malloc"); >exit(1); >} > // ... > } > > > On Apr 17, 2012, at 5:

[OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-17 Thread Rohan Deshpande
Hi, I am trying to distribute large amount of data using MPI. When I exceed the certain data size the segmentation fault occurs. Here is my code, #include "mpi.h" #include #include #include #define ARRAYSIZE200 #define MASTER0 int data[ARRAYSIZE]; int main(int argc, ch

[OMPI users] Open MPI 1.4.5 and CUDA support

2012-04-17 Thread Rohan Deshpande
Hi, I am using Open MPI 1.4.5 and I have CUDA 3.2 installed. Anyone knows whether CUDA 3.2 is supported by OpenMPI? Thanks

Re: [OMPI users] Cluster : received unexpected process identifier

2012-04-05 Thread Rohan Deshpande
I actually had too many Ethernet ports open. When specify which Ethernet to use by btl_tcp_if_include the program works like a charm :) Thanks for inputs. Cheers On Thu, Apr 5, 2012 at 8:11 AM, Jeffrey Squyres wrote: > On Apr 4, 2012, at 8:04 PM, Rohan Deshpande wrote: > > > Yes

Re: [OMPI users] Cluster : received unexpected process identifier

2012-04-04 Thread Rohan Deshpande
e.g., one IP interface > 192.168.1.x/24 and another on 192.168.1.y/24) > > > On Apr 4, 2012, at 7:55 PM, Rohan Deshpande wrote: > > > Hi, > > > > I am not quite sure what does that mean. Can you please explain more. > > > > Thanks > > > > On

Re: [OMPI users] Cluster : received unexpected process identifier

2012-04-04 Thread Rohan Deshpande
your IP addresses in a different > IP subnet. > > > On Apr 3, 2012, at 10:30 PM, Rohan Deshpande wrote: > > > Hi, > > > > I am running a MPI program using cluster and tcp communication. > > > > To run I am using: mpirun --prefix /usr/local/ --mca btl tcp,s

[OMPI users] Cluster : received unexpected process identifier

2012-04-03 Thread Rohan Deshpande
Hi, I am running a MPI program using cluster and tcp communication. To run I am using: *mpirun --prefix /usr/local/ --mca btl tcp,self --hostfile slaves -np 6 scatter* I am getting following output: Process 0 on host host1 has elements 0 1 2 3 Process 1 on host host1 has elements 4 5 6 7 Proc

Re: [OMPI users] Error while loading shared libraries

2012-04-02 Thread Rohan Deshpande
Thanks guys. Using absolute path of mpirun fixes my problem. Cheers On Mon, Apr 2, 2012 at 6:24 PM, Reuti wrote: > Am 02.04.2012 um 09:56 schrieb Rohan Deshpande: > > > Yes, I am trying to run the program using multiple hosts. > > > > The program executes fine but does

Re: [OMPI users] Error while loading shared libraries

2012-04-02 Thread Rohan Deshpande
at 2:52 PM, Rayson Ho wrote: > On Sun, Apr 1, 2012 at 11:27 PM, Rohan Deshpande > wrote: > > error while loading shared libraries: libmpi.so.0: cannot open shared > > object file no such object file: No such file or directory. > > Were you trying to run the MPI program

[OMPI users] Error while loading shared libraries

2012-04-01 Thread Rohan Deshpande
Hi , I have installed mpi successfully and I am able to compile the programs using mpicc But when I run using mpirun, I get following error * error while loading shared libraries: libmpi.so.0: cannot open shared object file no such object file: No such file or directory. * When I run using - m

Re: [OMPI users] mpicc command not found - Fedora

2012-03-30 Thread Rohan Deshpande
ackage. >> >> The rationale is that you don't need mpicc (and friends) to *run* Open >> MPI applications -- you only need mpicc (etc.) to *develop* Open MPI >> applications. >> >> Poke around and see if you can find a devel-like Open MPI package in

[OMPI users] mpicc command not found - Fedora

2012-03-29 Thread Rohan Deshpande
Hi, I have installed mpi successfully on fedora using *yum install openmpi openmpi-devel openmpi-libs* * * I have also added */usr/lib/openmpi/bin* to *PATH *and* LD_LIBRARY_PATH*variable. But when I try to complie my program using *mpicc hello.c* or*/usr/lib/openmpi/bin/mpicc hello.c * I get err

[OMPI users] Data distribution on different machines

2012-03-25 Thread Rohan Deshpande
Hi, I want to distribute the data on different machines using open mpi. I am a new user. Can some one point me to the resources or atleast functions i would have to use to complete the task? I am using red hat linux. Thanks, --