Hi,
I have this array [1 2 3 4 5 6 7 8 9] and i am performing scan operation on
that.
I have 3 mpi tasks and each task gets 3 elements then each task calculates
its scan and returns result to master task
task 0 - [1 2 3] => [1 3 6]
task 1 - [4 5 6 ] => [4 9 15]
task 2 - [7 8 9] => [7 15 24]
Now
;http://www.open-mpi.org/doc/v1.6/man3/MPI_Reduce.3.php
>
>
> On May 15, 2012, at 11:27 PM, Rohan Deshpande wrote:
>
> > I am performing Prefix scan operation on cluster
> >
> > I have 3 MPI tasks and master task is responsible for distributing the
> data
> &
MCW rank 0 will be totally different than the malloc() that you do in
> MCW rank 1.
>
> Make sense?
>
> On May 16, 2012, at 8:08 AM, Rohan Deshpande wrote:
>
> > I have following structure of MPI code -
> >
> > int main(){
> >
> > MPI_I
I have following structure of MPI code -
int main(){
MPI_INIT.
//initialize MPI
data = malloc(sizeof(int)*200);
//initialize data
/*Master-*/
if(taskid == 0){
//send data to each slave
MPI_SEND
}
/*Slaves---*/
if(taskid > 0){
//accep
I am performing Prefix scan operation on cluster
I have 3 MPI tasks and master task is responsible for distributing the data
Now, each task calculates sum of its own part of array using GPUs and
returns the results to master task.
Master task also calculates its own part of array using GPU.
Whe
Hi,
Can anyone point me to good resources and books to understand the detailed
architecture and working of MPI.
I would like to know all the minute details.
Thanks,
--
come out better.
>
> ** **
>
> ** **
>
> *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On
> Behalf Of *Rohan Deshpande
> *Sent:* Monday, May 07, 2012 9:38 PM
> *To:* Open MPI Users
> *Subject:* [OMPI users] GPU and CPU timing - OpenMPI and Thrust
>
&g
I am running MPI and Thrust code on a cluster and measuring time for
calculations.
My MPI code -
#include "mpi.h"
#include
#include
#include
#include
#include
#include
#define MASTER 0
#define ARRAYSIZE 2000
int
*masterarray,*onearray,*twoarray,*threearray,*fourarray,*fivearray,*s
uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>
>> >ecm-c-l-207-004.**uniwa.uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>exited
>> > on signal 11 (Segmentation fault).
>>
>> --**--**
>> --
>>
>> Sadly i cant install memory checker such as valgrind on my machine due to
>> some restrictions. I could not spot any error by looking in code.
>>
>> Can anyone help me ?what is wrong in above code.
>>
>> Thanks
>>
>
>
>
--
Best Regards,
ROHAN DESHPANDE
Hi,
I am trying to execute following code on cluster.
run_kernel is a cuda call with signature int run_kernel(int array[],int
nelements, int taskid, char hostname[]);
My MPI code is
#include "mpi.h"
#include
#include
#include
#define MASTER0
int *masterarray;
int *onearray;
int
>
> http://forums.nvidia.com/index.php?showtopic=227854
>
> ** **
>
> Maybe that could be your issue too.
>
> ** **
>
> ** **
>
> ** **
>
> *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On
> Behalf Of *Rohan Deshpande
> *Se
Hi,
I am trying to execute OpenMPI and CUDA code on a cluster. The code works
fine on single machine but when I try to execute it on cluster I get error:
error while loading shared libraries: libcudart.so.4: cannot open shared
object file: No such file or directory
I checked my PATH and LD_PATH
I am combining mpi and cuda. Trying to find out sum of array elements
using cuda and using mpi to distribute the array.
my cuda code
#include
__global__ void add(int *devarray, int *devsum)
{
int index = blockIdx.x * blockDim.x + threadIdx.x;
*devsum = *devsum + devarray[index]
/* printf("task %d has elements:",myid);
for(j = myoffset; j wrote:
> Send the most recent version of your code.
>
> Have you tried running it through a memory-checking debugger, like
> valgrind?
>
>
> On Apr 19, 2012, at 4:24 AM, Rohan Deshpande wrote:
>
d be the problem.
>
> Pascal
>
> users-boun...@open-mpi.org a écrit sur 19/04/2012 09:18:31 :
>
> > De : Rohan Deshpande
> > A : Open MPI Users
> > Date : 19/04/2012 09:18
> > Objet : Re: [OMPI users] machine exited on signal 11 (Segmentation
> fault).
&g
, Jeffrey Squyres wrote:
> As a guess, you're passing in a bad address.
>
> Double check the buffers that you're sending to MPI_SEND/MPI_RECV/etc.
>
>
> On Apr 17, 2012, at 10:43 PM, Rohan Deshpande wrote:
>
> > After using malloc i am getting following error
cally on the stack.
> Something like:
>
> int *data;
>
> int main(..) {
> {
>data = malloc(ARRAYSIZE * sizeof(int));
>if (NULL == data) {
>perror("malloc");
>exit(1);
>}
> // ...
> }
>
>
> On Apr 17, 2012, at 5:
Hi,
I am trying to distribute large amount of data using MPI.
When I exceed the certain data size the segmentation fault occurs.
Here is my code,
#include "mpi.h"
#include
#include
#include
#define ARRAYSIZE200
#define MASTER0
int data[ARRAYSIZE];
int main(int argc, ch
Hi,
I am using Open MPI 1.4.5 and I have CUDA 3.2 installed.
Anyone knows whether CUDA 3.2 is supported by OpenMPI?
Thanks
I actually had too many Ethernet ports open.
When specify which Ethernet to use by btl_tcp_if_include the program works
like a charm :)
Thanks for inputs.
Cheers
On Thu, Apr 5, 2012 at 8:11 AM, Jeffrey Squyres wrote:
> On Apr 4, 2012, at 8:04 PM, Rohan Deshpande wrote:
>
> > Yes
e.g., one IP interface
> 192.168.1.x/24 and another on 192.168.1.y/24)
>
>
> On Apr 4, 2012, at 7:55 PM, Rohan Deshpande wrote:
>
> > Hi,
> >
> > I am not quite sure what does that mean. Can you please explain more.
> >
> > Thanks
> >
> > On
your IP addresses in a different
> IP subnet.
>
>
> On Apr 3, 2012, at 10:30 PM, Rohan Deshpande wrote:
>
> > Hi,
> >
> > I am running a MPI program using cluster and tcp communication.
> >
> > To run I am using: mpirun --prefix /usr/local/ --mca btl tcp,s
Hi,
I am running a MPI program using cluster and tcp communication.
To run I am using: *mpirun --prefix /usr/local/ --mca btl tcp,self
--hostfile slaves -np 6 scatter*
I am getting following output:
Process 0 on host host1 has elements 0 1 2 3
Process 1 on host host1 has elements 4 5 6 7
Proc
Thanks guys.
Using absolute path of mpirun fixes my problem.
Cheers
On Mon, Apr 2, 2012 at 6:24 PM, Reuti wrote:
> Am 02.04.2012 um 09:56 schrieb Rohan Deshpande:
>
> > Yes, I am trying to run the program using multiple hosts.
> >
> > The program executes fine but does
at 2:52 PM, Rayson Ho wrote:
> On Sun, Apr 1, 2012 at 11:27 PM, Rohan Deshpande
> wrote:
> > error while loading shared libraries: libmpi.so.0: cannot open shared
> > object file no such object file: No such file or directory.
>
> Were you trying to run the MPI program
Hi ,
I have installed mpi successfully and I am able to compile the programs
using mpicc
But when I run using mpirun, I get following error
* error while loading shared libraries: libmpi.so.0: cannot open shared
object file no such object file: No such file or directory. *
When I run using - m
ackage.
>>
>> The rationale is that you don't need mpicc (and friends) to *run* Open
>> MPI applications -- you only need mpicc (etc.) to *develop* Open MPI
>> applications.
>>
>> Poke around and see if you can find a devel-like Open MPI package in
Hi,
I have installed mpi successfully on fedora using *yum install openmpi
openmpi-devel openmpi-libs*
*
*
I have also added */usr/lib/openmpi/bin* to *PATH *and*
LD_LIBRARY_PATH*variable.
But when I try to complie my program using *mpicc hello.c*
or*/usr/lib/openmpi/bin/mpicc hello.c
* I get err
Hi,
I want to distribute the data on different machines using open mpi.
I am a new user. Can some one point me to the resources or atleast
functions i would have to use to complete the task?
I am using red hat linux.
Thanks,
--
29 matches
Mail list logo