Hi Irek, Good day to you, and thank you for your e-mail.
Is there a better way other than patching the kernel? I would like to avoid having to compile a custom kernel for my OS. I read that I can disable write-caching on the drive using hdparm: hdparm -W0 /dev/sdf hdparm -W0 /dev/sdg I tested on one of my test servers and it seems I can disable it using the command. Current setup, write-caching is on: ==== root@ceph-osd-09:/home/indra# hdparm -W /dev/sdg /dev/sdg: write-caching = 1 (on) ==== I tried to disable write-caching and it's successful: ==== root@ceph-osd-09:/home/indra# hdparm -W0 /dev/sdg /dev/sdg: setting drive write-caching to 0 (off) write-caching = 0 (off) ==== I check again, and now write-caching is disabled. ==== root@ceph-osd-09:/home/indra# hdparm -W /dev/sdg /dev/sdg: write-caching = 0 (off) ==== Would the above give the same result? If yes, I will try to do that on our running cluster tonight. May I also know how I can confirm if my SSD comes with "volatile cache" as mentioned on your article? I tried to check my SSD's data sheet and there's no information on whether it comes with volatile cache or not. I also read that disabling write-caching will also increase the risk of data-loss. Can you comment on that? Looking forward to your reply, thank you. Cheers. On Mon, Apr 28, 2014 at 7:49 PM, Irek Fasikhov <malm...@gmail.com> wrote: > This is my article :). > To patch to the kernel ( > http://www.theirek.com/downloads/code/CMD_FLUSH.diff). > After rebooting, run the following commands: > echo temporary write through > /sys/class/scsi_disk/<disk>/cache_type > > > 2014-04-28 15:44 GMT+04:00 Indra Pramana <in...@sg.or.id>: > > Hi Irek, >> >> Thanks for the article. Do you have any other web sources pertaining to >> the same issue, which is in English? >> >> Looking forward to your reply, thank you. >> >> Cheers. >> >> >> On Mon, Apr 28, 2014 at 7:40 PM, Irek Fasikhov <malm...@gmail.com> wrote: >> >>> Most likely you need to apply a patch to the kernel. >>> >>> >>> http://www.theirek.com/blog/2014/02/16/patch-dlia-raboty-s-enierghoniezavisimym-keshiem-ssd-diskov >>> >>> >>> 2014-04-28 15:20 GMT+04:00 Indra Pramana <in...@sg.or.id>: >>> >>> Hi Udo and Irek, >>>> >>>> Good day to you, and thank you for your emails. >>>> >>>> >>>> >perhaps due IOs from the journal? >>>> >You can test with iostat (like "iostat -dm 5 sdg"). >>>> >>>> Yes, I have shared the iostat result earlier on this same thread. At >>>> times the utilisation of the 2 journal drives will hit 100%, especially >>>> when I simulate writing data using rados bench command. Any suggestions >>>> what could be the cause of the I/O issue? >>>> >>>> >>>> ==== >>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>> 1.85 0.00 1.65 3.14 0.00 93.36 >>>> >>>> >>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>> sdg 0.00 0.00 0.00 55.00 0.00 25365.33 >>>> 922.38 34.22 568.90 0.00 568.90 17.82 98.00 >>>> sdf 0.00 0.00 0.00 55.67 0.00 25022.67 >>>> 899.02 29.76 500.57 0.00 500.57 17.60 98.00 >>>> >>>> >>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>> 2.10 0.00 1.37 2.07 0.00 94.46 >>>> >>>> >>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>> sdg 0.00 0.00 0.00 56.67 0.00 25220.00 >>>> 890.12 23.60 412.14 0.00 412.14 17.62 99.87 >>>> sdf 0.00 0.00 0.00 52.00 0.00 24637.33 >>>> 947.59 33.65 587.41 0.00 587.41 19.23 100.00 >>>> >>>> >>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>> 2.21 0.00 1.77 6.75 0.00 89.27 >>>> >>>> >>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>> sdg 0.00 0.00 0.00 54.33 0.00 24802.67 >>>> 912.98 25.75 486.36 0.00 486.36 18.40 100.00 >>>> sdf 0.00 0.00 0.00 53.00 0.00 24716.00 >>>> 932.68 35.26 669.89 0.00 669.89 18.87 100.00 >>>> >>>> >>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>> 1.87 0.00 1.67 5.25 0.00 91.21 >>>> >>>> >>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>> sdg 0.00 0.00 0.00 94.33 0.00 26257.33 >>>> 556.69 18.29 208.44 0.00 208.44 10.50 99.07 >>>> sdf 0.00 0.00 0.00 51.33 0.00 24470.67 >>>> 953.40 32.75 684.62 0.00 684.62 19.51 100.13 >>>> >>>> >>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>> 1.51 0.00 1.34 7.25 0.00 89.89 >>>> >>>> >>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>> sdg 0.00 0.00 0.00 52.00 0.00 22565.33 >>>> 867.90 24.73 446.51 0.00 446.51 19.10 99.33 >>>> sdf 0.00 0.00 0.00 64.67 0.00 24892.00 >>>> 769.86 19.50 330.02 0.00 330.02 15.32 99.07 >>>> ==== >>>> >>>> >You what model SSD? >>>> >>>> For this one, I am using Seagate 100GB SSD, model: HDS-2TM-ST100FM0012 >>>> >>>> >Which version of the kernel? >>>> >>>> Ubuntu 13.04, Linux kernel version: 3.8.0-19-generic #30-Ubuntu SMP Wed >>>> May 1 16:35:23 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux >>>> >>>> Looking forward to your reply, thank you. >>>> >>>> Cheers. >>>> >>>> >>>> >>>> On Mon, Apr 28, 2014 at 4:45 PM, Irek Fasikhov <malm...@gmail.com>wrote: >>>> >>>>> You what model SSD? >>>>> Which version of the kernel? >>>>> >>>>> >>>>> >>>>> 2014-04-28 12:35 GMT+04:00 Udo Lembke <ulem...@polarzone.de>: >>>>> >>>>>> Hi, >>>>>> perhaps due IOs from the journal? >>>>>> You can test with iostat (like "iostat -dm 5 sdg"). >>>>>> >>>>>> on debian iostat is in the package sysstat. >>>>>> >>>>>> Udo >>>>>> >>>>>> Am 28.04.2014 07:38, schrieb Indra Pramana: >>>>>> > Hi Craig, >>>>>> > >>>>>> > Good day to you, and thank you for your enquiry. >>>>>> > >>>>>> > As per your suggestion, I have created a 3rd partition on the SSDs >>>>>> and did >>>>>> > the dd test directly into the device, and the result is very slow. >>>>>> > >>>>>> > ==== >>>>>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero of=/dev/sdg3 >>>>>> > conv=fdatasync oflag=direct >>>>>> > 128+0 records in >>>>>> > 128+0 records out >>>>>> > 134217728 bytes (134 MB) copied, 19.5223 s, 6.9 MB/s >>>>>> > >>>>>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero of=/dev/sdf3 >>>>>> > conv=fdatasync oflag=direct >>>>>> > 128+0 records in >>>>>> > 128+0 records out >>>>>> > 134217728 bytes (134 MB) copied, 5.34405 s, 25.1 MB/s >>>>>> > ==== >>>>>> > >>>>>> > I did a test onto another server with exactly similar specification >>>>>> and >>>>>> > similar SSD drive (Seagate SSD 100 GB) but not added into the >>>>>> cluster yet >>>>>> > (thus no load), and the result is fast: >>>>>> > >>>>>> > ==== >>>>>> > root@ceph-osd-09:/home/indra# dd bs=1M count=128 if=/dev/zero >>>>>> of=/dev/sdf1 >>>>>> > conv=fdatasync oflag=direct >>>>>> > 128+0 records in >>>>>> > 128+0 records out >>>>>> > 134217728 bytes (134 MB) copied, 0.742077 s, 181 MB/s >>>>>> > ==== >>>>>> > >>>>>> > Is the Ceph journal load really takes up a lot of the SSD >>>>>> resources? I >>>>>> > don't understand how come the performance can drop significantly. >>>>>> > Especially since the two Ceph journals are only taking the first 20 >>>>>> GB out >>>>>> > of the 100 GB of the SSD total capacity. >>>>>> > >>>>>> > Any advice is greatly appreciated. >>>>>> > >>>>>> > Looking forward to your reply, thank you. >>>>>> > >>>>>> > Cheers. >>>>>> > >>>>>> > >>>>>> > >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> С уважением, Фасихов Ирек Нургаязович >>>>> Моб.: +79229045757 >>>>> >>>> >>>> >>> >>> >>> -- >>> С уважением, Фасихов Ирек Нургаязович >>> Моб.: +79229045757 >>> >> >> > > > -- > С уважением, Фасихов Ирек Нургаязович > Моб.: +79229045757 >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com