Most likely you need to apply a patch to the kernel. http://www.theirek.com/blog/2014/02/16/patch-dlia-raboty-s-enierghoniezavisimym-keshiem-ssd-diskov
2014-04-28 15:20 GMT+04:00 Indra Pramana <in...@sg.or.id>: > Hi Udo and Irek, > > Good day to you, and thank you for your emails. > > > >perhaps due IOs from the journal? > >You can test with iostat (like "iostat -dm 5 sdg"). > > Yes, I have shared the iostat result earlier on this same thread. At times > the utilisation of the 2 journal drives will hit 100%, especially when I > simulate writing data using rados bench command. Any suggestions what could > be the cause of the I/O issue? > > > ==== > avg-cpu: %user %nice %system %iowait %steal %idle > 1.85 0.00 1.65 3.14 0.00 93.36 > > > Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz > avgqu-sz await r_await w_await svctm %util > sdg 0.00 0.00 0.00 55.00 0.00 25365.33 > 922.38 34.22 568.90 0.00 568.90 17.82 98.00 > sdf 0.00 0.00 0.00 55.67 0.00 25022.67 > 899.02 29.76 500.57 0.00 500.57 17.60 98.00 > > > avg-cpu: %user %nice %system %iowait %steal %idle > 2.10 0.00 1.37 2.07 0.00 94.46 > > > Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz > avgqu-sz await r_await w_await svctm %util > sdg 0.00 0.00 0.00 56.67 0.00 25220.00 > 890.12 23.60 412.14 0.00 412.14 17.62 99.87 > sdf 0.00 0.00 0.00 52.00 0.00 24637.33 > 947.59 33.65 587.41 0.00 587.41 19.23 100.00 > > > avg-cpu: %user %nice %system %iowait %steal %idle > 2.21 0.00 1.77 6.75 0.00 89.27 > > > Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz > avgqu-sz await r_await w_await svctm %util > sdg 0.00 0.00 0.00 54.33 0.00 24802.67 > 912.98 25.75 486.36 0.00 486.36 18.40 100.00 > sdf 0.00 0.00 0.00 53.00 0.00 24716.00 > 932.68 35.26 669.89 0.00 669.89 18.87 100.00 > > > avg-cpu: %user %nice %system %iowait %steal %idle > 1.87 0.00 1.67 5.25 0.00 91.21 > > > Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz > avgqu-sz await r_await w_await svctm %util > sdg 0.00 0.00 0.00 94.33 0.00 26257.33 > 556.69 18.29 208.44 0.00 208.44 10.50 99.07 > sdf 0.00 0.00 0.00 51.33 0.00 24470.67 > 953.40 32.75 684.62 0.00 684.62 19.51 100.13 > > > avg-cpu: %user %nice %system %iowait %steal %idle > 1.51 0.00 1.34 7.25 0.00 89.89 > > > Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz > avgqu-sz await r_await w_await svctm %util > sdg 0.00 0.00 0.00 52.00 0.00 22565.33 > 867.90 24.73 446.51 0.00 446.51 19.10 99.33 > sdf 0.00 0.00 0.00 64.67 0.00 24892.00 > 769.86 19.50 330.02 0.00 330.02 15.32 99.07 > ==== > > >You what model SSD? > > For this one, I am using Seagate 100GB SSD, model: HDS-2TM-ST100FM0012 > > >Which version of the kernel? > > Ubuntu 13.04, Linux kernel version: 3.8.0-19-generic #30-Ubuntu SMP Wed > May 1 16:35:23 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux > > Looking forward to your reply, thank you. > > Cheers. > > > > On Mon, Apr 28, 2014 at 4:45 PM, Irek Fasikhov <malm...@gmail.com> wrote: > >> You what model SSD? >> Which version of the kernel? >> >> >> >> 2014-04-28 12:35 GMT+04:00 Udo Lembke <ulem...@polarzone.de>: >> >>> Hi, >>> perhaps due IOs from the journal? >>> You can test with iostat (like "iostat -dm 5 sdg"). >>> >>> on debian iostat is in the package sysstat. >>> >>> Udo >>> >>> Am 28.04.2014 07:38, schrieb Indra Pramana: >>> > Hi Craig, >>> > >>> > Good day to you, and thank you for your enquiry. >>> > >>> > As per your suggestion, I have created a 3rd partition on the SSDs and >>> did >>> > the dd test directly into the device, and the result is very slow. >>> > >>> > ==== >>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero of=/dev/sdg3 >>> > conv=fdatasync oflag=direct >>> > 128+0 records in >>> > 128+0 records out >>> > 134217728 bytes (134 MB) copied, 19.5223 s, 6.9 MB/s >>> > >>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero of=/dev/sdf3 >>> > conv=fdatasync oflag=direct >>> > 128+0 records in >>> > 128+0 records out >>> > 134217728 bytes (134 MB) copied, 5.34405 s, 25.1 MB/s >>> > ==== >>> > >>> > I did a test onto another server with exactly similar specification and >>> > similar SSD drive (Seagate SSD 100 GB) but not added into the cluster >>> yet >>> > (thus no load), and the result is fast: >>> > >>> > ==== >>> > root@ceph-osd-09:/home/indra# dd bs=1M count=128 if=/dev/zero >>> of=/dev/sdf1 >>> > conv=fdatasync oflag=direct >>> > 128+0 records in >>> > 128+0 records out >>> > 134217728 bytes (134 MB) copied, 0.742077 s, 181 MB/s >>> > ==== >>> > >>> > Is the Ceph journal load really takes up a lot of the SSD resources? I >>> > don't understand how come the performance can drop significantly. >>> > Especially since the two Ceph journals are only taking the first 20 GB >>> out >>> > of the 100 GB of the SSD total capacity. >>> > >>> > Any advice is greatly appreciated. >>> > >>> > Looking forward to your reply, thank you. >>> > >>> > Cheers. >>> > >>> > >>> > >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >> >> >> >> -- >> С уважением, Фасихов Ирек Нургаязович >> Моб.: +79229045757 >> > > -- С уважением, Фасихов Ирек Нургаязович Моб.: +79229045757
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com