So hdparam -W 0 /dev/sdx doesn't work or it makes no difference?  Also I am
not sure I understand why it should happen before OSD have been started.
At least in my experience hdparam does it to hardware regardless.

On Mon, Jan 20, 2020, 2:25 AM Frank Schilder <fr...@dtu.dk> wrote:

> We are using Micron 5200 PRO, 1.92TB for RBD images on KVM and are very
> happy with the performance. We are using EC 6+2 pools, which really eat up
> IOPs. Still, we get enough performance out to run 20-50 VMs per disk, which
> results in good space utilisation as well since our default image size is
> 50GB and we take rolling snapshots. I was thinking about 4TB disks also,
> but am concerned that their IOPs/TB performance is too low for images on EC
> pools.
>
> We found the raw throughput in fio benchmarks to be very different for
> write-cache enabled and disabled, exactly as explained in the performance
> article. Changing write cache settings is a boot-time operation.
> Unfortunately, I couldn't find a reliable way to disable write cache at
> boot time (I was looking for tuned configs) and ended up adding this to a
> container startup script:
>
>   if [[ "$1" == "osd_ceph_disk_activate" && -n "${OSD_DEVICE}" ]] ; then
>     echo "Disabling write cache on ${OSD_DEVICE}"
>     /usr/sbin/smartctl -s wcache=off "${OSD_DEVICE}"
>   fi
>
> This works for both, SAS and SATA drives and ensures that write cache is
> disabled before an OSD daemon starts.
>
> Best regards,
>
> =================
> Frank Schilder
> AIT Risø Campus
> Bygning 109, rum S14
>
> ________________________________________
> From: ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Eric K.
> Miller <emil...@genesishosting.com>
> Sent: 19 January 2020 04:24:33
> To: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] low io with enterprise SSDs ceph luminous - can
> we expect more? [klartext]
>
> Hi Vitaliy,
>
> Similar to Stefan, we have a bunch of Micron 5200's (3.84TB ECO SATA
> version) in a Ceph cluster (Nautilus) and performance seems less than
> optimal.  I have followed all instructions on your site (thank you for your
> wonderful article btw!!), but I haven't seen much change.
>
> The only thing I could think of is that "maybe" disabling the write cache
> only takes place upon a reboot or power cycle?  Is that necessary?  Or is
> it a "live" change?
>
> I have tested with the cache disabled as well as enabled on all drives.
> We're using fio running in a QEMU/KVM VM in an OpenStack cluster, so not
> "raw" access to the Micron 5200's.  OSD (Bluestore) nodes run CentOS 7
> using a 4.18.x kernel.  Testing doesn't show any, or much, difference,
> enough that the variations could be considered "noise" in the results.
> Certainly no change that anyone could tell.
>
> Thought I'd check to see if you, or anyone else, might have any
> suggestions specific to the Micron 5200.
>
> We have some Micron 5300's inbound, but probably won't have them here for
> another few weeks due to Micron's manufacturing delays, so will be able to
> test these raw drives soon.  I will report back after, but if you know
> anything about these, I'm all ears. :)
>
> Thank you!
>
> Eric
>
>
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
> Stefan Bauer
> Sent: Tuesday, January 14, 2020 10:28 AM
> To: undisclosed-recipients
> Cc: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] low io with enterprise SSDs ceph luminous - can
> we expect more? [klartext]
>
>
> Thank you all,
>
>
>
> performance is indeed better now. Can now go back to sleep ;)
>
>
>
> KR
>
>
>
> Stefan
>
>
> -----Ursprüngliche Nachricht-----
> Von: Виталий Филиппов <vita...@yourcmc.ru>
> Gesendet: Dienstag 14 Januar 2020 10:28
> An: Wido den Hollander <w...@42on.com>; Stefan Bauer <
> stefan.ba...@cubewerk.de>
> CC: ceph-users@lists.ceph.com
> Betreff: Re: [ceph-users] low io with enterprise SSDs ceph luminous - can
> we expect more? [klartext]
>
> ...disable signatures and rbd cache. I didn't mention it in the email to
> not repeat myself. But I have it in the article :-)
> --
> With best regards,
> Vitaliy Filippov
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to