Ah. I think the FIO tests on page 3 are intended to be run *before* deploying 
OSDs, and the rados bench tests on page 6 against the cluster after deployment.

When running rados bench, be sure to clean up afterwards.  It’s all to easy to 
clutter up a RADOS pool with orphaned benchmark objects.



> On Jul 23, 2025, at 9:24 AM, GLE, Vivien <vivien....@inist.fr> wrote:
> 
> I followed this doc from Proxmox to test IO on my OSD
> https://www.proxmox.com/images/download/pve/docs/Proxmox-VE-Ceph-Benchmark-202312-rev0.pdf
> 
> Did it on 3 OSD to bench disk on a testing ceph cluster and one went down
> 
> Wanted to see the difference between a rados bench and fio
> 
> Vivien
> De : Anthony D'Atri <a...@dreamsnake.net <mailto:a...@dreamsnake.net>>
> Envoyé : mercredi 23 juillet 2025 14:06:55
> À : GLE, Vivien
> Cc : Sinan Polat; ceph-users@ceph.io <mailto:ceph-users@ceph.io>
> Objet : Re: [ceph-users] Ceph OSD down (unable to mount object store)
>  
> You ran a write fio job on the underlying device.  That would make any 
> software unhappy.
> 
> Did you mean to run a read test?  Or to test on an RBD volume or a filesystem 
> built within an RBD volume?
> 
> > 
> > 
> > Thanks for your answer !
> > 
> > 
> > Do you know by any chance why fio does this ? Are fio and ceph incompatible 
> > ?
> > 
> > 
> > Let's destroy this OSD then
> > 
> > 
> > Vivien
> > 
> > ________________________________
> > De : Sinan Polat <sinan86po...@gmail.com <mailto:sinan86po...@gmail.com>>
> > Envoyé : mercredi 23 juillet 2025 13:57:07
> > À : GLE, Vivien
> > Cc : ceph-users@ceph.io <mailto:ceph-users@ceph.io>
> > Objet : Re: [ceph-users] Ceph OSD down (unable to mount object store)
> > 
> > Hi Vivien,
> > 
> > Your fio test has very likely destroyed the Ceph OSD block device and the 
> > problem is not just the symlink, it's data corruption on the underlying 
> > device.
> > 
> > Zap the drive, recreate the OSD and let your cluster rebalance.
> > 
> > Sinan
> > 
> > Op wo 23 jul 2025 om 14:10 schreef GLE, Vivien <vivien....@inist.fr 
> > <mailto:vivien....@inist.fr><mailto:vivien....@inist.fr>>:
> > Hi,
> > 
> > 
> > I did a fio bench and believe that it destroy one of my OSD, this is the 
> > command used  :
> > 
> > 
> > fio --ioengine=libaio --filename=/dev/sda --direct=1 --sync=1 --rw=write 
> > --bs=4k --numjobs=1 --iodepth=1 --runtime=60 --time_based --name=fio
> > fio --ioengine=libaio --filename=/dev/sda --direct=1 --sync=1 --rw=write 
> > --bs=4M --numjobs=1 --iodepth=1 --runtime=60 --time_based --name=fio
> > 
> > 
> > Log file after the command :
> > 
> > 2025-07-23T08:41:44.448+0000 734bf7f6e680  1 bdev(0x59ba86d4ee00 
> > /var/lib/ceph/osd/ceph-2/block) close
> > 2025-07-23T08:41:44.719+0000 734bf7f6e680  1 bdev(0x59ba86d4ee00 
> > /var/lib/ceph/osd/ceph-2/block) open path /var/lib/ceph/osd/ceph-2/block
> > 2025-07-23T08:41:44.719+0000 734bf7f6e680  0 bdev(0x59ba86d4ee00 
> > /var/lib/ceph/osd/ceph-2/block) ioctl(F_SET_FILE_RW_HINT) on 
> > /var/lib/ceph/osd/ceph-2/block failed: (22) Invalid argument
> > 2025-07-23T08:41:44.720+0000 734bf7f6e680  1 bdev(0x59ba86d4ee00 
> > /var/lib/ceph/osd/ceph-2/block) open size 1000203091968 (0xe8e0c00000, 932 
> > GiB) block_size 4096 (4 KiB) rotational device, discard supported
> > 2025-07-23T08:41:44.723+0000 734bf7f6e680 -1 
> > bluestore(/var/lib/ceph/osd/ceph-2/block) _read_bdev_label unable to decode 
> > label /var/lib/ceph/osd/ceph-2/block at offset 66: void 
> > bluestore_bdev_label_t::decode(ceph::buffer::v15_2_0::list::const_iterator&)
> >  decode past end of struct encoding: Malformed input [buffer:3]
> > 2025-07-23T08:41:44.724+0000 734bf7f6e680 -1 
> > bluestore(/var/lib/ceph/osd/ceph-2/block) _read_bdev_label unable to decode 
> > label /var/lib/ceph/osd/ceph-2/block at offset 4096: End of buffer 
> > [buffer:2]
> > 2025-07-23T08:41:44.724+0000 734bf7f6e680 -1 
> > bluestore(/var/lib/ceph/osd/ceph-2) _check_main_bdev_label not all labels 
> > read properly
> > 2025-07-23T08:41:44.724+0000 734bf7f6e680  1 bdev(0x59ba86d4ee00 
> > /var/lib/ceph/osd/ceph-2/block) close
> > 2025-07-23T08:41:44.983+0000 734bf7f6e680 -1 osd.2 0 OSD:init: unable to 
> > mount object store
> > 2025-07-23T08:41:44.983+0000 734bf7f6e680 -1  ** ERROR: osd init failed: 
> > (5) Input/output error
> > 
> > 
> > 
> > After checking I saw that the block soft link might be wrong
> > 
> > On a healthy osd :
> > 
> > # ll /var/lib/ceph/cluster-id/osd.5/
> > total 72
> > drwx------  2 167 167 4096 Jul 22 10:36 ./
> > drwx------ 12 167 167 4096 Jul 22 10:30 ../
> > lrwxrwxrwx  1 167 167   93 Jul 15 14:39 block -> 
> > /dev/ceph-c31f0e16-0460-4bc5-9470-468270b4c68a/osd-block-72aa3074-e2f9-45f8-a468-03c02d36f1de
> > 
> > 
> > On my breaking osd :
> > 
> > /var/lib/ceph/cluster-id/osd.2# ll
> > total 72
> > drwx------  2 167 167 4096 Jul 23 10:41 ./
> > drwx------ 11 167 167 4096 Jul 22 10:35 ../
> > lrwxrwxrwx  1 167 167  111 Jul 23 10:41 block -> 
> > /dev/mapper/ceph--bd0ca671--2b89--4530--be2b--f41443822a91-osd--block--342578c3--4603--4023--a564--8fca6dcc1040
> > 
> > Is there a correct way of changing it or I'm going in the wrong direction ?
> > 
> > Thanks
> > 
> > Vivien
> > 
> > 
> > 
> > 
> > 
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io 
> > <mailto:ceph-users@ceph.io><mailto:ceph-users@ceph.io>
> > To unsubscribe send an email to ceph-users-le...@ceph.io 
> > <mailto:ceph-users-le...@ceph.io><mailto:ceph-users-le...@ceph.io>
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io <mailto:ceph-users@ceph.io>
> > To unsubscribe send an email to ceph-users-le...@ceph.io 
> > <mailto:ceph-users-le...@ceph.io>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to