Hi again, Thomas Maier-Komor wrote: > Carsten Aulbert schrieb: >> Hi Thomas, > I don't know socat or what benefit it gives you, but have you tried > using mbuffer to send and receive directly (options -I and -O)?
I thought we tried that in the past and with socat it seemed faster, but I just made a brief test and I got (/dev/zero -> remote /dev/null) 330 MB/s with mbuffer+socat and 430MB/s with mbuffer alone. > Additionally, try to set the block size of mbuffer to the recordsize of > zfs (usually 128k): > receiver$ mbuffer -I sender:10000 -s 128k -m 2048M | zfs receive > sender$ zfs send blabla | mbuffer -s 128k -m 2048M -O receiver:10000 We are using 32k since many of our user use tiny files (and then I need to reduce the buffer size because of this 'funny' error): mbuffer: fatal: Cannot address so much memory (32768*65536=2147483648>1544040742911). Does this qualify for a bug report? Thanks for the hint of looking into this again! Cheers Carsten _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss