I suppose that the write operations may use wal more when block size is
small.

[image: image.png]

rainning <tweety...@qq.com> 于2020年7月16日周四 上午10:39写道:

> I tested osd bench with different block size: 1MB, 512KB, 256KB, 128KB,
> 64KB, and 32KB. osd.2 is one from the cluster where osds have better 4KB
> osd bench, and osd.30 is from the cluster where osds have lower 4KB osd
> bench.  Before 32KB, osd.30 was better than osd.2, however, there was a big
> drop on osd.30 with 32KB block size.
>
> root@cmn01:~# ceph tell osd.2 bench 1073741824 1048576
> {
>     "bytes_written": 1073741824,
>     "blocksize": 1048576,
>     "bytes_per_sec": 188747963
> }
> root@cmn01:~# ceph tell osd.2 bench 1073741824 524288
> {
>     "bytes_written": 1073741824,
>     "blocksize": 524288,
>     "bytes_per_sec": 181071543
> }
> root@cmn01:~# ceph tell osd.2 bench 786432000 262144
> {
>     "bytes_written": 786432000,
>     "blocksize": 262144,
>     "bytes_per_sec": 159007035
> }
> root@cmn01:~# ceph tell osd.2 bench 393216000 131072
> {
>     "bytes_written": 393216000,
>     "blocksize": 131072,
>     "bytes_per_sec": 127179122
> }
> root@cmn01:~# ceph tell osd.2 bench 196608000 65536
> {
>     "bytes_written": 196608000,
>     "blocksize": 65536,
>     "bytes_per_sec": 83365482
> }
> root@cmn01:~# ceph tell osd.2 bench 98304000 32768
> {
>     "bytes_written": 98304000,
>     "blocksize": 32768,
>     "bytes_per_sec": 48351258
> }
> root@cmn01:~# ceph tell osd.2 bench 49152000 16384
> {
>     "bytes_written": 49152000,
>     "blocksize": 16384,
>     "bytes_per_sec": 31725841
> }
>
>
> ------------------------------------------------------------------------------------------------------------------
>
> root@stor-mgt01:~# ceph tell osd.30 bench 1073741824 1048576
> {
>     "bytes_written": 1073741824,
>     "blocksize": 1048576,
>     "elapsed_sec": 5.344805,
>     "bytes_per_sec": 200894474.890259,
>     "iops": 191.587901
> }
> root@stor-mgt01:~# ceph tell osd.30 bench 1073741824 524288
> {
>     "bytes_written": 1073741824,
>     "blocksize": 524288,
>     "elapsed_sec": 5.303052,
>     "bytes_per_sec": 202476205.680661,
>     "iops": 386.192714
> }
> root@stor-mgt01:~# ceph tell osd.30 bench 786432000 262144
> {
>     "bytes_written": 786432000,
>     "blocksize": 262144,
>     "elapsed_sec": 3.878248,
>     "bytes_per_sec": 202780204.655892,
>     "iops": 773.545092
> }
> root@stor-mgt01:~# ceph tell osd.30 bench 393216000 131072
> {
>     "bytes_written": 393216000,
>     "blocksize": 131072,
>     "elapsed_sec": 1.939532,
>     "bytes_per_sec": 202737591.242988,
>     "iops": 1546.765070
> }
> root@stor-mgt01:~# ceph tell osd.30 bench 196608000 65536
> {
>     "bytes_written": 196608000,
>     "blocksize": 65536,
>     "elapsed_sec": 1.081617,
>     "bytes_per_sec": 181772360.338257,
>     "iops": 2773.626104
> }
> root@stor-mgt01:~# ceph tell osd.30 bench 98304000 32768
> {
>     "bytes_written": 98304000,
>     "blocksize": 32768,
>     "elapsed_sec": 2.908703,
>     "bytes_per_sec": 33796507.598640,
>     "iops": 1031.387561
> }
> root@stor-mgt01:~# ceph tell osd.30 bench 49152000 16384
> {
>     "bytes_written": 49152000,
>     "blocksize": 16384,
>     "elapsed_sec": 3.907744,
>     "bytes_per_sec": 12578102.861185,
>     "iops": 767.706473
> }
>
>
> ------------------ 原始邮件 ------------------
> *发件人:* "rainning" <tweety...@qq.com>;
> *发送时间:* 2020年7月16日(星期四) 上午9:42
> *收件人:* "Zhenshi Zhou"<deader...@gmail.com>;
> *抄送:* "ceph-users"<ceph-users@ceph.io>;
> *主题:* 回复:[ceph-users] Re: osd bench with or without a separate WAL device
> deployed
>
> Hi Zhenshi,
>
> I did try with bigger block size. Interestingly, the one whose 4KB osd
> bench was lower performed slightly better in 4MB osd bench.
>
> Let me try some other bigger block sizes, e.g. 16K, 64K, 128K, 1M etc, to
> see if there is any pattern.
>
> Moreover, I did compare two SSDs, they respectively are INTEL
> SSDSC2KB480G8 and INTEL SSDSC2KB960G8. Performance wise, there is no much
> difference.
>
> Thanks,
> Ning
>
>
> ------------------ 原始邮件 ------------------
> *发件人:* "Zhenshi Zhou" <deader...@gmail.com>;
> *发送时间:* 2020年7月16日(星期四) 上午9:24
> *收件人:* "rainning"<tweety...@qq.com>;
> *抄送:* "ceph-users"<ceph-users@ceph.io>;
> *主题:* [ceph-users] Re: osd bench with or without a separate WAL device
> deployed
>
> Maybe you can try writing with bigger block size and compare the results.
> For bluestore, the write operations contain two modes. One is COW, the
> other is RMW. AFAIK only RMW uses wal in order to prevent data from
> being interrupted.
>
> rainning <tweety...@qq.com> 于2020年7月15日周三 下午11:04写道:
>
> > Hi Zhenshi, thanks very much for the reply.
> >
> > Yes I know it is ood that the bluestore is deployed only with a separate
> > db device  but no a WAL device. The cluster was deployed in k8s using
> rook.
> > I was told it was because the rook we used didn't support that.
> >
> > Moreover, the comparison was made on osd bench, so the network should not
> > be the case. As far as the storage node hardware, although two clusters
> are
> > indeed different, their CPUs and HDDs do have almost same performance
> > numbers. I haven't compared SSDs that are used as db/WAL devices, it
> might
> > cause difference, but I am not sure if it can make two times difference.
> >
> > ---Original---
> > *From:* "Zhenshi Zhou"<deader...@gmail.com>
> > *Date:* Wed, Jul 15, 2020 18:39 PM
> > *To:* "rainning"<tweety...@qq.com>;
> > *Cc:* "ceph-users"<ceph-users@ceph.io>;
> > *Subject:* [ceph-users] Re: osd bench with or without a separate WAL
> > device deployed
> >
> > I deployed the cluster either with separate db/wal or put db/wal/data
> > together. Never tried to have only a seperate db.
> > AFAIK wal does have an effect on writing but I'm not sure if it could be
> > two times of the bench value. Hardware and
> > network environment are also important factors.
> >
> > rainning <tweety...@qq.com> 于2020年7月15日周三 下午4:35写道:
> >
> > > Hi all,
> > >
> > >
> > > I am wondering if there is any performance comparison done on osd bench
> > > with and without a separate WAL device deployed given that there is
> > always
> > > a separate db device deployed on SSD in both cases.
> > >
> > >
> > > The reason I am asking this question is that we have two clusters and
> > osds
> > > in one have separate db and WAL device deployed on SSD but osds in
> > another
> > > only have a separate db device deployed. And we found 4KB osd bench
> (i.e.
> > > ceph tell osd.X bench 12288000 4096) for the ones having a separate WAL
> > > device was two times of the ones without a separate WAL device. Is the
> > > performance difference caused by the separate WAL device?
> > >
> > >
> > > Thanks,
> > > Ning
> > > _______________________________________________
> > > ceph-users mailing list -- ceph-users@ceph.io
> > > To unsubscribe send an email to ceph-users-le...@ceph.io
> > >
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
> >
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to