> Try to separate the two things:> > (1) Try /dev/zero -> mbuffer --- network 
> ---> mbuffer > /dev/null
> That should give you wirespeed
I tried that already.  It still gets just 10-11MB/s from this server.
I can get zfs send / receive and mbuffer working at 30MB/s though from a couple 
of test servers (with much lower specs).
 
> (2) Try zfs send | mbuffer > /dev/null> That should give you an idea how fast 
> zfs send really is locally.
Hmm, that's better than 10MB/s, but the average is still only around 20MB/s:
summary:  942 MByte in 47.4 sec - average of 19.9 MB/s
 
I think that points to another problem though as the send mbuffer is 100% full. 
 Certainly the pool itself doesn't appear under any strain at all while this is 
going on:
 
               capacity     operations    bandwidthpool         used  avail   
read  write   read  write----------  -----  -----  -----  -----  -----  
-----rc-pool      732G  1.55T    171     85  21.3M  1.01M  mirror     144G   
320G     38      0  4.78M      0    c1t1d0      -      -      6      0   779K   
   0    c1t2d0      -      -     17      0  2.17M      0    c2t1d0      -      
-     14      0  1.85M      0  mirror     146G   318G     39      0  4.89M      
0    c1t3d0      -      -     20      0  2.50M      0    c2t2d0      -      -   
  13      0  1.63M      0    c2t0d0      -      -      6      0   779K      0  
mirror     146G   318G     34      0  4.35M      0    c2t3d0      -      -     
19      0  2.39M      0    c1t5d0      -      -      7      0  1002K      0    
c1t4d0      -      -      7      0  1002K      0  mirror     148G   316G     23 
     0  2.93M      0    c2t4d0      -      -      8      0  1.09M      0    
c2t5d0      -      -      6      0   890K      0    c1t6d0      -      -      7 
     0  1002K      0  mirror     148G   316G     35      0  4.35M      0    
c1t7d0      -      -      6      0   779K      0    c2t6d0      -      -     12 
     0  1.52M      0    c2t7d0      -      -     17      0  2.07M      0  
c3d1p0      12K   504M      0     85      0  1.01M----------  -----  -----  
-----  -----  -----  -----
Especially when compared to the zfs send stats on my backup server which 
managed 30MB/s via mbuffer (Being received on a single virtual SATA disk):
               capacity     operations    bandwidthpool         used  avail   
read  write   read  write----------  -----  -----  -----  -----  -----  
-----rpool       5.12G  42.6G      0      5      0  27.1K  c4t0d0s0  5.12G  
42.6G      0      5      0  27.1K----------  -----  -----  -----  -----  -----  
-----zfspool      431G  4.11T    261      0  31.4M      0  raidz2     431G  
4.11T    261      0  31.4M      0    c4t1d0      -      -    155      0  6.28M  
    0    c4t2d0      -      -    155      0  6.27M      0    c4t3d0      -      
-    155      0  6.27M      0    c4t4d0      -      -    155      0  6.27M      
0    c4t5d0      -      -    155      0  6.27M      0----------  -----  -----  
-----  -----  -----  -----
The really ironic thing is that the 30MB/s send / receive was sending to a 
virtual SATA disk which is stored (via sync NFS) on the server I'm having 
problems with...
 
Ross
 
 
_________________________________________________________________
Win New York holidays with Kellogg’s & Live Search
http://clk.atdmt.com/UKM/go/111354033/direct/01/
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to