spinners are slow anyway, but on top of that SAS disks often default to writecache=off. In use as a single disk with no risk of raid write-holes, you can turn on writecache. On SAS, I would assume the firmware does not lie about writes reaching stable storage (flushes).
# turn on temporarily: sdparm --set=WCE /dev/sdX # turn on persistently: sdparm --set=WCE --save /dev/sdX To check current state: sdparm --get=WCE /dev/sdf /dev/sdf: SEAGATE ST2000NM0045 DS03 WCE 0 [cha: y, def: 0, sav: 0] "WCE 0" means: off "sav: 0" means: off next time the disk is powered on Matthias On Thu, Apr 06, 2023 at 09:26:27AM -0400, Anthony D'Atri wrote: > How bizarre, I haven’t dealt with this specific SKU before. Some Dell / LSI > HBAs call this passthrough mode, some “personality”, some “jbod mode”, dunno > why they can’t be consistent. > > > > We are testing an experimental Ceph cluster with server and controller at > > subject. > > > > The controller have not an HBA mode, but only a 'NonRAID' mode, come sort of > > 'auto RAID0' configuration. > > Dell’s CLI guide describes setting individual drives in Non-RAID, which > *smells* like passthrough, not the more-complex RAID0 workaround we had to do > before passthrough. > > https://www.dell.com/support/manuals/en-nz/perc-h750-sas/perc_cli_rg/set-drive-state-commands?guid=guid-d4750845-1f57-434c-b4a9-935876ee1a8e&lang=en-us > > > > We are using SSD SATA disks (MICRON MTFDDAK480TDT) that perform very well, > > and SAS HDD disks (SEAGATE ST8000NM014A) that instead perform very bad > > (particulary, very low IOPS). > > Spinners are slow, this is news? > > That said, how slow is slow? Testing commands and results or it didn’t > happen. > > Also, firmware matters. Run Dell’s DSU. > > > There's some hint for disk/controller configuration/optimization? > > Give us details, perccli /c0 show, test results etc. > > Use a different HBA if you have to use an HBA, one that doesn’t suffer an > RoC. Better yet, take an expansive look at TCO and don’t write off NVMe as > infeasible. If your cluster is experimental hopefully you aren’t stuck with > a lot of these. Add up the cost of an RoC HBA, optionally with cache RAM and > BBU/supercap, add in the cost delta for SAS HDDs over SATA. Add in the > operational hassle of managing WAL+DB on those boot SSDs. Add in the extra > HDDs you’ll need to provision because of IOPS. > > > > > > > Thanks. > > > > -- > > Io credo nella chimica tanto quanto Giulio Cesare credeva nel caso... > > mi va bene fino a quando non riguarda me :) (Emanuele Pucciarelli) > > > > _______________________________________________ > > ceph-users mailing list -- ceph-users@ceph.io > > To unsubscribe send an email to ceph-users-le...@ceph.io > > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io