Thanks Brian, Thanks Michael
I wanted to benchmark the communcation throughput and latency using multiple using gigabit eithernet controller.
So here are the results which i want share with you all
I used .
OpenMPI version 1.0.2a10r9275
Hpcbench
Two Dell Precision 650 workstation.
The Dell Precision 650 workstation has three separate PCI bus segments.
Segment 1 -> PCI Slot1,2 -> 32 bit, 33MHz, Shared with integrated 1394
Segment 2 -> PCI SLot3,4 -> 64 bit, 100MHz, Shared with the Gb Ethernet connection
Segment 3 -> PCI Slot 5 -> Shared with Integrated Ultra 320 controller
The workstation has Integrated PCI-X 64-bit Intel 10/100/1000 Gigabit Ethernet.
I added three D-Link DGE-530T 1000 Mbps Ethernet Card in Slot2, Slot4 and Slot5 respectively.
As i expected, the Card in slot5 performed better than the cards in other slots. Hereare the results.
(Using Slot2)-# MPI communication latency (roundtrip time) test -- Wed Mar 15 09:19:10 2006# Hosts: DELL <> DELL2# Blocking Communication (MPI_Send/MPI_Recv)# Message size (Bytes) : 40960# Iteration: 7# Test time (Seconds): 0.20
# RTT-time # Microseconds1 25953.5652 25569.4393 22392.0004 20876.5785 21327.1216 19597.1567 21264.0088 24109.5689 23877.85910 24064.575
# MPI RTT min/avg/max = 19597.156/22903.187/25953.565 usec
--
# MPI communication test -- Wed Mar 15 10:16:22 2006# Test mode: Fixed-size stream (unidirectional) test# Hosts: DELL <> DELL2# Blocking communication (MPI_Send/MPI_Recv)# Total data size of each test (Bytes): 524288000# Message size (Bytes): 104857600# Iteration : 5# Test time: 5.00# Test repetition: 10## Overall Master-node M-process M-process Slave-node S-process S-process# Throughput Elapsed-time User-mode Sys-mode Elapsed-time User-mode Sys-mode# Mbps Seconds Seconds Seconds Seconds
Seconds Seconds1 521.9423 8.04 1.42 6.62 8.04 0.93 7.102 551.5377 7.60 1.20 6.41 7.60 0.77 6.873 552.5600 7.59 1.27 6.32
7.59 0.82 6.814 552.6328 7.59 1.28 6.31 7.59 0.80 6.835 552.6334 7.59 1.24 6.35 7.59 0.86 6.776 552.7048 7.59
1.26 6.33 7.59 0.77 6.867 563.6736 7.44 1.22 6.22 7.44 0.78 6.708 552.2710 7.59 1.22 6.37 7.59 0.83 6.809
520.9938 8.05 1.37 6.68 8.05 0.93 7.1610 535.0131 7.84 1.36 6.48 7.84 0.84 7.04
==
(Using Slot3)-# MPI communication latency (roundtrip time) test -- Thu Mar 16 10:15:58 2006# Hosts: DELL <> DELL2# Blocking Communication (MPI_Send/MPI_Recv)# Message size (Bytes) : 40960# Iteration: 10# Test time (Seconds): 0.20
# RTT-time # Microseconds1 20094.2042 14773.5123 14846.0154 17756.8205 18419.2906 23394.7997 21840.5968 17727.4949 21822.09510 17659.688
# MPI RTT min/avg/max = 14773.512/18833.451/23394.799 usec
--
# MPI communication test -- Wed Mar 15 09:17:54 2006# Test mode: Fixed-size stream (unidirectional) test# Hosts: DELL <> DELL2# Blocking communication (MPI_Send/MPI_Recv)# Total data size of each test (Bytes): 524288000# Message size (Bytes): 104857600# Iteration : 5# Test time: 5.00# Test repetition: 10## Overall Master-node M-process M-process Slave-node S-process S-process# Throughput Elapsed-time User-mode Sys-mode Elapsed-time User-mode Sys-mode# Mbps Seconds Seconds Seconds Seconds
Seconds Seconds1 794.9650 5.28 1.04 4.24 5.28 0.47 4.812 838.1621 5.00 0.91 4.09 5.00 0.39 4.653 898.3811 4.67 0.84 3.82
4.67 0.34 4.374 798.9575 5.25 1.03 4.22 5.25 0.40 4.895 829.7181 5.06 0.94 4.11 5.05 0.40 4.696 881.5526 4.76
0.86 3.90 4.76 0.28 4.527 827.9215 5.07 0.96 4.11 5.07 0.41 4.708 845.6428 4.96 0.87 4.09 4.96 0.38 4.629
845.6903 4.96 0.90 4.06 4.96