I suppose that the write operations may use wal more when block size is small.
[image: image.png] rainning <tweety...@qq.com> 于2020年7月16日周四 上午10:39写道: > I tested osd bench with different block size: 1MB, 512KB, 256KB, 128KB, > 64KB, and 32KB. osd.2 is one from the cluster where osds have better 4KB > osd bench, and osd.30 is from the cluster where osds have lower 4KB osd > bench. Before 32KB, osd.30 was better than osd.2, however, there was a big > drop on osd.30 with 32KB block size. > > root@cmn01:~# ceph tell osd.2 bench 1073741824 1048576 > { > "bytes_written": 1073741824, > "blocksize": 1048576, > "bytes_per_sec": 188747963 > } > root@cmn01:~# ceph tell osd.2 bench 1073741824 524288 > { > "bytes_written": 1073741824, > "blocksize": 524288, > "bytes_per_sec": 181071543 > } > root@cmn01:~# ceph tell osd.2 bench 786432000 262144 > { > "bytes_written": 786432000, > "blocksize": 262144, > "bytes_per_sec": 159007035 > } > root@cmn01:~# ceph tell osd.2 bench 393216000 131072 > { > "bytes_written": 393216000, > "blocksize": 131072, > "bytes_per_sec": 127179122 > } > root@cmn01:~# ceph tell osd.2 bench 196608000 65536 > { > "bytes_written": 196608000, > "blocksize": 65536, > "bytes_per_sec": 83365482 > } > root@cmn01:~# ceph tell osd.2 bench 98304000 32768 > { > "bytes_written": 98304000, > "blocksize": 32768, > "bytes_per_sec": 48351258 > } > root@cmn01:~# ceph tell osd.2 bench 49152000 16384 > { > "bytes_written": 49152000, > "blocksize": 16384, > "bytes_per_sec": 31725841 > } > > > ------------------------------------------------------------------------------------------------------------------ > > root@stor-mgt01:~# ceph tell osd.30 bench 1073741824 1048576 > { > "bytes_written": 1073741824, > "blocksize": 1048576, > "elapsed_sec": 5.344805, > "bytes_per_sec": 200894474.890259, > "iops": 191.587901 > } > root@stor-mgt01:~# ceph tell osd.30 bench 1073741824 524288 > { > "bytes_written": 1073741824, > "blocksize": 524288, > "elapsed_sec": 5.303052, > "bytes_per_sec": 202476205.680661, > "iops": 386.192714 > } > root@stor-mgt01:~# ceph tell osd.30 bench 786432000 262144 > { > "bytes_written": 786432000, > "blocksize": 262144, > "elapsed_sec": 3.878248, > "bytes_per_sec": 202780204.655892, > "iops": 773.545092 > } > root@stor-mgt01:~# ceph tell osd.30 bench 393216000 131072 > { > "bytes_written": 393216000, > "blocksize": 131072, > "elapsed_sec": 1.939532, > "bytes_per_sec": 202737591.242988, > "iops": 1546.765070 > } > root@stor-mgt01:~# ceph tell osd.30 bench 196608000 65536 > { > "bytes_written": 196608000, > "blocksize": 65536, > "elapsed_sec": 1.081617, > "bytes_per_sec": 181772360.338257, > "iops": 2773.626104 > } > root@stor-mgt01:~# ceph tell osd.30 bench 98304000 32768 > { > "bytes_written": 98304000, > "blocksize": 32768, > "elapsed_sec": 2.908703, > "bytes_per_sec": 33796507.598640, > "iops": 1031.387561 > } > root@stor-mgt01:~# ceph tell osd.30 bench 49152000 16384 > { > "bytes_written": 49152000, > "blocksize": 16384, > "elapsed_sec": 3.907744, > "bytes_per_sec": 12578102.861185, > "iops": 767.706473 > } > > > ------------------ 原始邮件 ------------------ > *发件人:* "rainning" <tweety...@qq.com>; > *发送时间:* 2020年7月16日(星期四) 上午9:42 > *收件人:* "Zhenshi Zhou"<deader...@gmail.com>; > *抄送:* "ceph-users"<ceph-users@ceph.io>; > *主题:* 回复:[ceph-users] Re: osd bench with or without a separate WAL device > deployed > > Hi Zhenshi, > > I did try with bigger block size. Interestingly, the one whose 4KB osd > bench was lower performed slightly better in 4MB osd bench. > > Let me try some other bigger block sizes, e.g. 16K, 64K, 128K, 1M etc, to > see if there is any pattern. > > Moreover, I did compare two SSDs, they respectively are INTEL > SSDSC2KB480G8 and INTEL SSDSC2KB960G8. Performance wise, there is no much > difference. > > Thanks, > Ning > > > ------------------ 原始邮件 ------------------ > *发件人:* "Zhenshi Zhou" <deader...@gmail.com>; > *发送时间:* 2020年7月16日(星期四) 上午9:24 > *收件人:* "rainning"<tweety...@qq.com>; > *抄送:* "ceph-users"<ceph-users@ceph.io>; > *主题:* [ceph-users] Re: osd bench with or without a separate WAL device > deployed > > Maybe you can try writing with bigger block size and compare the results. > For bluestore, the write operations contain two modes. One is COW, the > other is RMW. AFAIK only RMW uses wal in order to prevent data from > being interrupted. > > rainning <tweety...@qq.com> 于2020年7月15日周三 下午11:04写道: > > > Hi Zhenshi, thanks very much for the reply. > > > > Yes I know it is ood that the bluestore is deployed only with a separate > > db device but no a WAL device. The cluster was deployed in k8s using > rook. > > I was told it was because the rook we used didn't support that. > > > > Moreover, the comparison was made on osd bench, so the network should not > > be the case. As far as the storage node hardware, although two clusters > are > > indeed different, their CPUs and HDDs do have almost same performance > > numbers. I haven't compared SSDs that are used as db/WAL devices, it > might > > cause difference, but I am not sure if it can make two times difference. > > > > ---Original--- > > *From:* "Zhenshi Zhou"<deader...@gmail.com> > > *Date:* Wed, Jul 15, 2020 18:39 PM > > *To:* "rainning"<tweety...@qq.com>; > > *Cc:* "ceph-users"<ceph-users@ceph.io>; > > *Subject:* [ceph-users] Re: osd bench with or without a separate WAL > > device deployed > > > > I deployed the cluster either with separate db/wal or put db/wal/data > > together. Never tried to have only a seperate db. > > AFAIK wal does have an effect on writing but I'm not sure if it could be > > two times of the bench value. Hardware and > > network environment are also important factors. > > > > rainning <tweety...@qq.com> 于2020年7月15日周三 下午4:35写道: > > > > > Hi all, > > > > > > > > > I am wondering if there is any performance comparison done on osd bench > > > with and without a separate WAL device deployed given that there is > > always > > > a separate db device deployed on SSD in both cases. > > > > > > > > > The reason I am asking this question is that we have two clusters and > > osds > > > in one have separate db and WAL device deployed on SSD but osds in > > another > > > only have a separate db device deployed. And we found 4KB osd bench > (i.e. > > > ceph tell osd.X bench 12288000 4096) for the ones having a separate WAL > > > device was two times of the ones without a separate WAL device. Is the > > > performance difference caused by the separate WAL device? > > > > > > > > > Thanks, > > > Ning > > > _______________________________________________ > > > ceph-users mailing list -- ceph-users@ceph.io > > > To unsubscribe send an email to ceph-users-le...@ceph.io > > > > > _______________________________________________ > > ceph-users mailing list -- ceph-users@ceph.io > > To unsubscribe send an email to ceph-users-le...@ceph.io > > > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io >
_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io