If you are mapping the RBD with the kernel driver then you're not using
librbd so these settings will have no effect I believe.  The kernel driver
does its own caching but I don't believe there are any settings to change
its default behavior.


On Mon, Feb 29, 2016 at 9:36 PM, Shinobu Kinjo <ski...@redhat.com> wrote:

> You may want to set "ioengine=rbd", I guess.
>
> Cheers,
>
> ----- Original Message -----
> From: "min fang" <louisfang2...@gmail.com>
> To: "ceph-users" <ceph-users@lists.ceph.com>
> Sent: Tuesday, March 1, 2016 1:28:54 PM
> Subject: [ceph-users]  rbd cache did not help improve performance
>
> Hi, I set the following parameters in ceph.conf
>
> [client]
> rbd cache=true
> rbd cache size= 25769803776
> rbd readahead disable after byte=0
>
>
> map a rbd image to a rbd device then run fio testing on 4k read as the
> command
> ./fio -filename=/dev/rbd4 -direct=1 -iodepth 64 -thread -rw=read
> -ioengine=aio -bs=4K -size=500G -numjobs=32 -runtime=300 -group_reporting
> -name=mytest2
>
> Compared the result with setting rbd cache=false and enable cache model, I
> did not see performance improved by librbd cache.
>
> Is my setting not right, or it is true that ceph librbd cache will not
> have benefit on 4k seq read?
>
> thanks.
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to