Hi jeffy,

Thanks for your reply.

I am not understanding how MPI_Reduce would be useful.

Say I have 3 processes and each process has array [1,2,3,4]

When each process calculates the prefix sum using cuda each process will
have array as [1,3,6,10]

so if I use MPI_Reduce to gather results it returns me sum as 30.

but the original array is [1,2,3,4,1,2,3,4,1,2,3,4] and the prefix sum of
this array should be

[1,3,6,10,11,13,16,20,21,23,26,30]

Is my understanding wrong somewhere?

On Fri, May 18, 2012 at 7:05 AM, Jeff Squyres <jsquy...@cisco.com> wrote:

> You probably want MPI_Reduce, instead.
>
>    http://www.open-mpi.org/doc/v1.6/man3/MPI_Reduce.3.php
>
>
> On May 15, 2012, at 11:27 PM, Rohan Deshpande wrote:
>
> > I am performing Prefix scan operation on cluster
> >
> > I have 3 MPI tasks and master task is responsible for distributing the
> data
> >
> > Now, each task calculates sum of its own part of array using GPUs and
> returns the results to master task.
> >
> > Master task also calculates its own part of array using GPU.
> >
> > When each task returns its result (which would be array) master task
> needs to combine all the results to get the final result.
> >
> > Can I use MPI_SCAN to combine the results?
> >
> >
> >
> >
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>



-- 

Best Regards,

ROHAN DESHPANDE

Reply via email to