On Apr 8, 2014, at 8:05 AM, Hamid Saeed <e.hamidsa...@gmail.com> wrote:

> Yes i meant Parallel file system.
> 
> And can you kindly explain what exactly happens if 
> the RANK0 want to send to RANK0?

It goes thru the "self" BTL, which is pretty fast but does require a little 
time You also have the collective operation overhead in the scatterv algorithm.

> 
> Why does MPIO is different in time consumption than RANK0 to RANK0 
> communication?

Again, it depends on how you did the test. MPIO on a single node is just a file 
read operation - depending on the parallel file system, that can be heavily 
optimized with pre-fetch and memory caching.


> 
> 
> On Tue, Apr 8, 2014 at 4:45 PM, Ralph Castain <r...@open-mpi.org> wrote:
> I suspect it all depends on when you start the clock. If the data is sitting 
> in the file at time=0, then the file I/O method will likely be faster as 
> every proc just reads its data in parallel - no comm required as it is all 
> handled by the parallel file system.
> 
> I confess I don't quite understand your reference to "shared memory" in the 
> MPIO case, but I suspect what you really meant was just "parallel file 
> system"?
> 
> 
> On Apr 8, 2014, at 6:12 AM, Hamid Saeed <e.hamidsa...@gmail.com> wrote:
> 
>> Can someone kindly reply?
>> 
>> 
>> 
>> On Tue, Apr 8, 2014 at 1:01 PM, Hamid Saeed <e.hamidsa...@gmail.com> wrote:
>> Hello,
>> I think that the MPI open its sockets even though the number of processor is 
>> only 1 on the same machine?
>> regards.
>> 
>> 
>> On Tue, Apr 8, 2014 at 9:43 AM, Hamid Saeed <e.hamidsa...@gmail.com> wrote:
>> Hello all,
>> 
>> I have a very basic question regarding MPI communication.
>> 
>> In my Task, what i am doing is..
>> Comparing Scatterv and MPIO.
>> 1) In scatterv, I scatter all the data to the other ranks and SCAN for the 
>> specific characters.
>> MPI_Scatterv (chunk, send_counts, displacements, MPI_CHAR, copychunk, 
>> smallchunk_size, MPI_CHAR, 0,  MPI_COMM_WORLD);  
>> &
>> 2) On the other hand using MPIO, I have the data available in the shared 
>> memory and every rank searches in the specific chunk.
>> MPI_File_open(MPI_COMM_WORLD, "170mb.txt", MPI_MODE_RDONLY, MPI_INFO_NULL, 
>> &in);
>> here i assign every processor to search in a specific "chunk"
>> 
>> 
>> My question is..
>> 
>> Why MPI_Scatterv using 1 processor takes more time then MPI_File_open?
>> How does MPI sending and receiving takes place?
>> 
>> I think using 1 processor does not include physical sending and receiving. 
>> Then why it consumes more clock?
>> 
>> In the attachment you can observe the plots in which i performed some tests 
>> using both algorithms.
>> 
>> Kindly explain briefly the mpi communication using 1 processor and multiple 
>> processors.
>> 
>> 
>> Thanks in advance.
>> 
>> Regards
>> Hamid
>> 
>> 
>> 
>> -- 
>> Hamid
>> 
>> 
>> 
>> 
>> -- 
>> _______________________________________________
>> Hamid Saeed
>> CoSynth GmbH & Co. KG
>> Escherweg 2 - 26121 Oldenburg - Germany
>> Tel +49 441 9722 738 | Fax -278
>> http://www.cosynth.com
>> _______________________________________________
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> 
> -- 
> _______________________________________________
> Hamid Saeed
> CoSynth GmbH & Co. KG
> Escherweg 2 - 26121 Oldenburg - Germany
> Tel +49 441 9722 738 | Fax -278
> http://www.cosynth.com
> _______________________________________________
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to