I'll give it a go.  Attached is the code.

Thanks!

-Mike

Tim S. Woodall wrote:

Hello Mike,

Mike Houston wrote:
When only sending a few messages, we get reasonably good IB performance, ~500MB/s (MVAPICH is 850MB/s). However, if I crank the number of messages up, we drop to 3MB/s(!!!). This is with the OSU NBCL mpi_bandwidth test. We are running Mellanox IB Gold 1.8 with 3.3.3 firmware on PCI-X (Couger) boards. Everything works with MVAPICH, but we really need the thread support in OpenMPI.

Ideas? I noticed there are a plethora of runtime options configurable for mvapi. Do I need to tweak these to get performacne up?


You might try running w/ the:

mpirun -mca mpi_leave_pinned 1

Which will cause mvapi port to maintain an mru cache of registrations,
rather than dynamically pinning/unpinning memory.

If this does not resolve the BW problems, try increasing the
resources allocated to each connection:

-mca btl_mvapi_rd_min 128
-mca btl_mvapi_rd_max 256

Also can you forward me a copy of the test code or a reference to it?

Thanks,
Tim
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

/*
 * Copyright (C) 2002-2003 the Network-Based Computing Laboratory
 * (NBCL), The Ohio State University.  
 */

#include "mpi.h"
#include <stdio.h>
#include <unistd.h>
#include <stdlib.h>
#include <math.h>
#include <assert.h>


#define MYBUFSIZE (4*1024*1028)
#define MAX_REQ_NUM 100000

char s_buf1[MYBUFSIZE];
char r_buf1[MYBUFSIZE];


MPI_Request request[MAX_REQ_NUM];
MPI_Status stat[MAX_REQ_NUM];

int main(int argc,char *argv[])
{
    int  myid, numprocs, i;
    int size, loop, page_size;
    char *s_buf, *r_buf;
    double t_start=0.0, t_end=0.0, t=0.0;


    MPI_Init(&argc,&argv);
    MPI_Comm_size(MPI_COMM_WORLD,&numprocs);
    MPI_Comm_rank(MPI_COMM_WORLD,&myid);

    if ( argc < 3 ) {
       fprintf(stderr, "Usage: bw loop msg_size\n");
       MPI_Finalize();
       return 0;
    }    
    size=atoi(argv[2]);
    loop = atoi(argv[1]);

    if(size > MYBUFSIZE){
             fprintf(stderr, "Maximum message size is %d\n",MYBUFSIZE);
             MPI_Finalize();
             return 0;
    }

    if(loop > MAX_REQ_NUM){
             fprintf(stderr, "Maximum number of iterations is 
%d\n",MAX_REQ_NUM);
             MPI_Finalize();
             return 0;
    }

    page_size = getpagesize();

    s_buf = (char*)(((unsigned long)s_buf1 + (page_size -1))/page_size * 
page_size);
    r_buf = (char*)(((unsigned long)r_buf1 + (page_size -1))/page_size * 
page_size);

    assert( (s_buf != NULL) && (r_buf != NULL) );

    for ( i=0; i<size; i++ ){
           s_buf[i]='a';
           r_buf[i]='b';
    }

    /*warmup */
    if (myid == 0)
    {
        for ( i=0; i< loop; i++ ) {
            MPI_Isend(s_buf, size, MPI_CHAR, 1, 100, MPI_COMM_WORLD, request+i);
        }

        MPI_Waitall(loop, request, stat);
        MPI_Recv(r_buf, 4, MPI_CHAR, 1, 101, MPI_COMM_WORLD, &stat[0]);

    }else{
        for ( i=0; i< loop; i++ ) {
            MPI_Irecv(r_buf, size, MPI_CHAR, 0, 100, MPI_COMM_WORLD, request+i);
        }
        MPI_Waitall(loop, request, stat);
        MPI_Send(s_buf, 4, MPI_CHAR, 0, 101, MPI_COMM_WORLD);
    }

    MPI_Barrier(MPI_COMM_WORLD);

    if (myid == 0)
    {
        t_start=MPI_Wtime();
        for ( i=0; i< loop; i++ ) {
            MPI_Isend(s_buf, size, MPI_CHAR, 1, 100, MPI_COMM_WORLD, request+i);
        }

        MPI_Waitall(loop, request, stat);
        MPI_Recv(r_buf, 4, MPI_CHAR, 1, 101, MPI_COMM_WORLD, &stat[0]);

        t_end=MPI_Wtime();
        t = t_end - t_start;

    }else{
        for ( i=0; i< loop; i++ ) {
            MPI_Irecv(r_buf, size, MPI_CHAR, 0, 100, MPI_COMM_WORLD, request+i);
        }
        MPI_Waitall(loop, request, stat);
        MPI_Send(s_buf, 4, MPI_CHAR, 0, 101, MPI_COMM_WORLD);
    }

    if ( myid == 0 ) {
        double tmp;
        tmp = ((size*1.0)/1.0e6)*loop; 
        fprintf(stdout,"%d\t%f (MillionBytes/sec)\t%f(MegaBytes/sec)\n", size, 
tmp/t, tmp/t/1.048576);
    }

    MPI_Barrier(MPI_COMM_WORLD);
    MPI_Finalize();
    return 0;
}

Reply via email to