Hi Irek, Good day to you.
Any updates/comments on below? Looking forward to your reply, thank you. Cheers. On Tue, Apr 29, 2014 at 12:47 PM, Indra Pramana <in...@sg.or.id> wrote: > Hi Irek, > > Good day to you, and thank you for your e-mail. > > Is there a better way other than patching the kernel? I would like to > avoid having to compile a custom kernel for my OS. I read that I can > disable write-caching on the drive using hdparm: > > hdparm -W0 /dev/sdf > hdparm -W0 /dev/sdg > > I tested on one of my test servers and it seems I can disable it using the > command. > > Current setup, write-caching is on: > > ==== > root@ceph-osd-09:/home/indra# hdparm -W /dev/sdg > > /dev/sdg: > write-caching = 1 (on) > ==== > > I tried to disable write-caching and it's successful: > > ==== > root@ceph-osd-09:/home/indra# hdparm -W0 /dev/sdg > > /dev/sdg: > setting drive write-caching to 0 (off) > write-caching = 0 (off) > ==== > > I check again, and now write-caching is disabled. > > ==== > root@ceph-osd-09:/home/indra# hdparm -W /dev/sdg > > /dev/sdg: > write-caching = 0 (off) > ==== > > Would the above give the same result? If yes, I will try to do that on our > running cluster tonight. > > May I also know how I can confirm if my SSD comes with "volatile cache" as > mentioned on your article? I tried to check my SSD's data sheet and there's > no information on whether it comes with volatile cache or not. I also read > that disabling write-caching will also increase the risk of data-loss. Can > you comment on that? > > Looking forward to your reply, thank you. > > Cheers. > > > > On Mon, Apr 28, 2014 at 7:49 PM, Irek Fasikhov <malm...@gmail.com> wrote: > >> This is my article :). >> To patch to the kernel ( >> http://www.theirek.com/downloads/code/CMD_FLUSH.diff). >> After rebooting, run the following commands: >> echo temporary write through > /sys/class/scsi_disk/<disk>/cache_type >> >> >> 2014-04-28 15:44 GMT+04:00 Indra Pramana <in...@sg.or.id>: >> >> Hi Irek, >>> >>> Thanks for the article. Do you have any other web sources pertaining to >>> the same issue, which is in English? >>> >>> Looking forward to your reply, thank you. >>> >>> Cheers. >>> >>> >>> On Mon, Apr 28, 2014 at 7:40 PM, Irek Fasikhov <malm...@gmail.com>wrote: >>> >>>> Most likely you need to apply a patch to the kernel. >>>> >>>> >>>> http://www.theirek.com/blog/2014/02/16/patch-dlia-raboty-s-enierghoniezavisimym-keshiem-ssd-diskov >>>> >>>> >>>> 2014-04-28 15:20 GMT+04:00 Indra Pramana <in...@sg.or.id>: >>>> >>>> Hi Udo and Irek, >>>>> >>>>> Good day to you, and thank you for your emails. >>>>> >>>>> >>>>> >perhaps due IOs from the journal? >>>>> >You can test with iostat (like "iostat -dm 5 sdg"). >>>>> >>>>> Yes, I have shared the iostat result earlier on this same thread. At >>>>> times the utilisation of the 2 journal drives will hit 100%, especially >>>>> when I simulate writing data using rados bench command. Any suggestions >>>>> what could be the cause of the I/O issue? >>>>> >>>>> >>>>> ==== >>>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>>> 1.85 0.00 1.65 3.14 0.00 93.36 >>>>> >>>>> >>>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>>> sdg 0.00 0.00 0.00 55.00 0.00 25365.33 >>>>> 922.38 34.22 568.90 0.00 568.90 17.82 98.00 >>>>> sdf 0.00 0.00 0.00 55.67 0.00 25022.67 >>>>> 899.02 29.76 500.57 0.00 500.57 17.60 98.00 >>>>> >>>>> >>>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>>> 2.10 0.00 1.37 2.07 0.00 94.46 >>>>> >>>>> >>>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>>> sdg 0.00 0.00 0.00 56.67 0.00 25220.00 >>>>> 890.12 23.60 412.14 0.00 412.14 17.62 99.87 >>>>> sdf 0.00 0.00 0.00 52.00 0.00 24637.33 >>>>> 947.59 33.65 587.41 0.00 587.41 19.23 100.00 >>>>> >>>>> >>>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>>> 2.21 0.00 1.77 6.75 0.00 89.27 >>>>> >>>>> >>>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>>> sdg 0.00 0.00 0.00 54.33 0.00 24802.67 >>>>> 912.98 25.75 486.36 0.00 486.36 18.40 100.00 >>>>> sdf 0.00 0.00 0.00 53.00 0.00 24716.00 >>>>> 932.68 35.26 669.89 0.00 669.89 18.87 100.00 >>>>> >>>>> >>>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>>> 1.87 0.00 1.67 5.25 0.00 91.21 >>>>> >>>>> >>>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>>> sdg 0.00 0.00 0.00 94.33 0.00 26257.33 >>>>> 556.69 18.29 208.44 0.00 208.44 10.50 99.07 >>>>> sdf 0.00 0.00 0.00 51.33 0.00 24470.67 >>>>> 953.40 32.75 684.62 0.00 684.62 19.51 100.13 >>>>> >>>>> >>>>> avg-cpu: %user %nice %system %iowait %steal %idle >>>>> 1.51 0.00 1.34 7.25 0.00 89.89 >>>>> >>>>> >>>>> Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s >>>>> avgrq-sz avgqu-sz await r_await w_await svctm %util >>>>> sdg 0.00 0.00 0.00 52.00 0.00 22565.33 >>>>> 867.90 24.73 446.51 0.00 446.51 19.10 99.33 >>>>> sdf 0.00 0.00 0.00 64.67 0.00 24892.00 >>>>> 769.86 19.50 330.02 0.00 330.02 15.32 99.07 >>>>> ==== >>>>> >>>>> >You what model SSD? >>>>> >>>>> For this one, I am using Seagate 100GB SSD, model: HDS-2TM-ST100FM0012 >>>>> >>>>> >Which version of the kernel? >>>>> >>>>> Ubuntu 13.04, Linux kernel version: 3.8.0-19-generic #30-Ubuntu SMP >>>>> Wed May 1 16:35:23 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux >>>>> >>>>> Looking forward to your reply, thank you. >>>>> >>>>> Cheers. >>>>> >>>>> >>>>> >>>>> On Mon, Apr 28, 2014 at 4:45 PM, Irek Fasikhov <malm...@gmail.com>wrote: >>>>> >>>>>> You what model SSD? >>>>>> Which version of the kernel? >>>>>> >>>>>> >>>>>> >>>>>> 2014-04-28 12:35 GMT+04:00 Udo Lembke <ulem...@polarzone.de>: >>>>>> >>>>>>> Hi, >>>>>>> perhaps due IOs from the journal? >>>>>>> You can test with iostat (like "iostat -dm 5 sdg"). >>>>>>> >>>>>>> on debian iostat is in the package sysstat. >>>>>>> >>>>>>> Udo >>>>>>> >>>>>>> Am 28.04.2014 07:38, schrieb Indra Pramana: >>>>>>> > Hi Craig, >>>>>>> > >>>>>>> > Good day to you, and thank you for your enquiry. >>>>>>> > >>>>>>> > As per your suggestion, I have created a 3rd partition on the SSDs >>>>>>> and did >>>>>>> > the dd test directly into the device, and the result is very slow. >>>>>>> > >>>>>>> > ==== >>>>>>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero >>>>>>> of=/dev/sdg3 >>>>>>> > conv=fdatasync oflag=direct >>>>>>> > 128+0 records in >>>>>>> > 128+0 records out >>>>>>> > 134217728 bytes (134 MB) copied, 19.5223 s, 6.9 MB/s >>>>>>> > >>>>>>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero >>>>>>> of=/dev/sdf3 >>>>>>> > conv=fdatasync oflag=direct >>>>>>> > 128+0 records in >>>>>>> > 128+0 records out >>>>>>> > 134217728 bytes (134 MB) copied, 5.34405 s, 25.1 MB/s >>>>>>> > ==== >>>>>>> > >>>>>>> > I did a test onto another server with exactly similar >>>>>>> specification and >>>>>>> > similar SSD drive (Seagate SSD 100 GB) but not added into the >>>>>>> cluster yet >>>>>>> > (thus no load), and the result is fast: >>>>>>> > >>>>>>> > ==== >>>>>>> > root@ceph-osd-09:/home/indra# dd bs=1M count=128 if=/dev/zero >>>>>>> of=/dev/sdf1 >>>>>>> > conv=fdatasync oflag=direct >>>>>>> > 128+0 records in >>>>>>> > 128+0 records out >>>>>>> > 134217728 bytes (134 MB) copied, 0.742077 s, 181 MB/s >>>>>>> > ==== >>>>>>> > >>>>>>> > Is the Ceph journal load really takes up a lot of the SSD >>>>>>> resources? I >>>>>>> > don't understand how come the performance can drop significantly. >>>>>>> > Especially since the two Ceph journals are only taking the first >>>>>>> 20 GB out >>>>>>> > of the 100 GB of the SSD total capacity. >>>>>>> > >>>>>>> > Any advice is greatly appreciated. >>>>>>> > >>>>>>> > Looking forward to your reply, thank you. >>>>>>> > >>>>>>> > Cheers. >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> _______________________________________________ >>>>>>> ceph-users mailing list >>>>>>> ceph-users@lists.ceph.com >>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> С уважением, Фасихов Ирек Нургаязович >>>>>> Моб.: +79229045757 >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> С уважением, Фасихов Ирек Нургаязович >>>> Моб.: +79229045757 >>>> >>> >>> >> >> >> -- >> С уважением, Фасихов Ирек Нургаязович >> Моб.: +79229045757 >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com