Hi Robert,

Could you try 35 concurrent dd each issuing 128K I/O ?
That would be closer to how ZFS would behave.

-r

Robert Milkowski writes:
 > Well I have just tested UFS on the same disk.
 > 
 > bash-3.00# newfs -v /dev/rdsk/c5t500000E0119495A0d0s0
 > newfs: construct a new file system /dev/rdsk/c5t500000E0119495A0d0s0: (y/n)? 
 > y
 > mkfs -F ufs /dev/rdsk/c5t500000E0119495A0d0s0 143358287 128 48 8192 1024 16 
 > 1 1 8192 t 0 -1 1 1024 n
 > Warning: 5810 sector(s) in last cylinder unallocated
 > /dev/rdsk/c5t500000E0119495A0d0s0:      143358286 sectors in 23334 cylinders 
 > of 48 tracks, 128 sectors
 >         69999.2MB in 1459 cyl groups (16 c/g, 48.00MB/g, 5824 i/g)
 > super-block backups (for fsck -F ufs -o b=#) at:
 >  32, 98464, 196896, 295328, 393760, 492192, 590624, 689056, 787488, 885920,
 > Initializing cylinder groups:
 > ............................
 > super-block backups for last 10 cylinder groups at:
 >  142447776, 142546208, 142644640, 142743072, 142841504, 142939936, 143038368,
 >  143136800, 143235232, 143333664
 > bash-3.00# mkdir /mnt/1
 > bash-3.00# mount -o noatime /dev/dsk/c5t500000E0119495A0d0s0 /mnt/1
 > 
 > bash-3.00# dd if=/dev/zero of=/mnt/1/q1 bs=8192k
 > ^C110+0 records in
 > 110+0 records out
 > bash-3.00#
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     5.0   25.0   35.0 82408.8  0.0  3.6    0.0  120.3   0  99 c5
 >     5.0   25.0   35.0 82409.7  0.0  3.6    0.0  120.3   0  99 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     4.0   25.0   28.0 79832.1  0.0  3.9    0.0  133.4   0  97 c5
 >     4.0   25.0   28.0 79831.5  0.0  3.9    0.0  133.4   0  97 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     6.0   25.0   42.0 81921.3  0.0  4.7    0.0  151.6   0 100 c5
 >     6.0   25.0   42.0 81921.4  0.0  4.7    0.0  151.6   0 100 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     4.0   21.0   28.0 73555.6  0.0  3.5    0.0  138.7   0  97 c5
 >     4.0   21.0   28.0 73555.7  0.0  3.5    0.0  138.7   0  97 
 > c5t500000E0119495A0d0
 > 
 > 
 > bash-3.00# tunefs -a 2048 /mnt/1
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0   22.0    0.0 83240.1  0.0  3.5    0.0  157.1   0  97 c5
 >     0.0   22.0    0.0 83240.5  0.0  3.5    0.0  157.1   0  97 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0   19.0    0.0 81837.1  0.0  3.4    0.0  180.1   0  98 c5
 >     0.0   19.0    0.0 81837.2  0.0  3.4    0.0  180.1   0  98 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0   21.0    0.0 94004.1  0.0  4.6    0.0  218.1   0 100 c5
 >     0.0   21.0    0.0 94002.6  0.0  4.6    0.0  218.1   0 100 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0   20.0    0.0 70116.6  0.0  4.3    0.0  216.5   0 100 c5
 >     0.0   20.0    0.0 70116.7  0.0  4.3    0.0  216.5   0 100 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0   21.0    0.0 82140.7  0.0  3.3    0.0  158.0   0  95 c5
 >     0.0   21.0    0.0 82140.8  0.0  3.3    0.0  158.0   0  95 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0   72.0    0.0 82279.7  0.0  5.0    0.0   69.9   0  98 c5
 >     0.0   72.0    0.0 82279.6  0.0  5.0    0.0   69.9   0  98 
 > c5t500000E0119495A0d0
 > 
 > So sometimes it still can even push more from the disk. 
 > 
 > So even UFS in this case is much faster than ZFS. And UFS issued something 
 > like 3,5MB block sizes.
 > 
 > 
 > bash-3.00# tunefs -a 16 /mnt/1
 > maximum contiguous block count changes from 2048 to 16
 > 
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0  350.9    0.0 44533.6  0.0 118.1    0.0  336.6   0 100 c5
 >     0.0  350.9    0.0 44531.0  0.0 118.1    0.0  336.6   0 100 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0  381.0    0.0 48466.4  0.0 112.9    0.0  296.4   0 100 c5
 >     0.0  381.0    0.0 48468.7  0.0 112.9    0.0  296.4   0 100 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0  369.9    0.0 47057.3  0.0 110.8    0.0  299.6   0 100 c5
 >     0.0  369.9    0.0 47057.3  0.0 110.8    0.0  299.6   0 100 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0  399.1    0.0 50566.4  0.0 108.8    0.0  272.7   0 100 c5
 >     0.0  399.1    0.0 50566.5  0.0 108.8    0.0  272.7   0 100 
 > c5t500000E0119495A0d0
 >                     extended device statistics
 >     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
 >     0.0  345.0    0.0 44171.3  0.0 87.7    0.0  254.3   0 100 c5
 >     0.0  345.0    0.0 44171.4  0.0 87.7    0.0  254.3   0 100 
 > c5t500000E0119495A0d0
 > 
 > So now UFS was issuing 128KB IOs and now with UFS I get similar
 > performance to ZFS. 
 > 
 > So I would say that larger IOs greatly could help ZFS performance
 > while writing large sequential files (with large writes). 
 >  
 >  
 > This message posted from opensolaris.org
 > _______________________________________________
 > zfs-discuss mailing list
 > zfs-discuss@opensolaris.org
 > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to