Yikes, something is off here. Is your entire cluster on one host? If so, beware that this is a very risky proposition.
ceph config set global osd_crush_chooseleaf_type 0 https://docs.ceph.com/en/latest/rados/troubleshooting/troubleshooting-pg/#one-node-cluster I suspect that will at least distribute your data better. If not, please send `ceph osd crush dump` > On Mar 28, 2025, at 6:59 AM, Mihai Ciubancan <mihai.ciuban...@eli-np.ro> > wrote: > > Hi Anthony, > Thanks for the answer: > > The output of 'ceph osd df' is: > > ceph osd df tree > ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META > AVAIL %USE VAR PGS STATUS TYPE NAME > -1 167.64825 - 168 TiB 12 TiB 12 TiB 100 MiB 27 GiB > 156 TiB 7.13 1.00 - root default > -3 167.64825 - 168 TiB 12 TiB 12 TiB 100 MiB 27 GiB > 156 TiB 7.13 1.00 - host sto-core-hpc01 > 0 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.0 > 1 ssd 13.97069 1.00000 14 TiB 12 TiB 12 TiB 6 KiB 26 GiB > 2.0 TiB 85.53 12.00 129 up osd.1 > 2 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.2 > 3 ssd 13.97069 1.00000 14 TiB 1.7 GiB 258 MiB 100 MiB 1.3 GiB > 14 TiB 0.01 0.00 16 up osd.3 > 4 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.4 > 5 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.5 > 6 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.6 > 7 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.7 > 8 ssd 13.97069 1.00000 14 TiB 68 MiB 4.8 MiB 12 KiB 63 MiB > 14 TiB 0 0 1 up osd.8 > 9 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.9 > 10 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB 28 MiB > 14 TiB 0 0 0 up osd.10 > 11 ssd 13.97069 1.00000 14 TiB 68 MiB 4.8 MiB 12 KiB 63 MiB > 14 TiB 0 0 1 up osd.11 > TOTAL 168 TiB 12 TiB 12 TiB 100 MiB 27 GiB > 156 TiB 7.13 > > So all the date is on osd.1 > > But I have checked the balancer and seems active: > ceph balancer status > { > "active": true, > "last_optimize_duration": "0:00:00.000368", > "last_optimize_started": "Fri Mar 28 10:55:06 2025", > "mode": "upmap", > "no_optimization_needed": false, > "optimize_result": "Some objects (0.500000) are degraded; try again later", > "plans": [] > } > > But the output of the commnad 'ceph config dump|grep balancer' gives me > nothing. > > Best, > Mihai > > On 2025-03-27 23:06, Anthony D'Atri wrote: >> Look at `ceph osd df`. Is the balancer enabled? >>> On Mar 27, 2025, at 8:50 AM, Mihai Ciubancan <mihai.ciuban...@eli-np.ro> >>> wrote: >>> Hello, >>> My name is Mihai, and I have started using CEPH this mount for a HPC >>> cluster. >>> When was lunch in the production the available space shown was 80TB now is >>> 16TB and I didn't do anything, while I'm having 12 OSD (SSD of 14TB): >>> sudo ceph osd tree >>> ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF >>> -1 167.64825 root default >>> -3 167.64825 host sto-core-hpc01 >>> 0 ssd 13.97069 osd.0 up 1.00000 1.00000 >>> 1 ssd 13.97069 osd.1 up 1.00000 1.00000 >>> 2 ssd 13.97069 osd.2 up 1.00000 1.00000 >>> 3 ssd 13.97069 osd.3 up 1.00000 1.00000 >>> 4 ssd 13.97069 osd.4 up 1.00000 1.00000 >>> 5 ssd 13.97069 osd.5 up 1.00000 1.00000 >>> 6 ssd 13.97069 osd.6 up 1.00000 1.00000 >>> 7 ssd 13.97069 osd.7 up 1.00000 1.00000 >>> 8 ssd 13.97069 osd.8 up 1.00000 1.00000 >>> 9 ssd 13.97069 osd.9 up 1.00000 1.00000 >>> 10 ssd 13.97069 osd.10 up 1.00000 1.00000 >>> 11 ssd 13.97069 osd.11 up 1.00000 1.00000 >>> sudo ceph df detail >>> --- RAW STORAGE --- >>> CLASS SIZE AVAIL USED RAW USED %RAW USED >>> ssd 168 TiB 156 TiB 12 TiB 12 TiB 7.12 >>> TOTAL 168 TiB 156 TiB 12 TiB 12 TiB 7.12 >>> --- POOLS --- >>> POOL ID PGS STORED (DATA) (OMAP) OBJECTS USED >>> (DATA) (OMAP) %USED MAX AVAIL QUOTA OBJECTS QUOTA BYTES DIRTY USED >>> COMPR UNDER COMPR >>> .mgr 1 1 705 KiB 705 KiB 0 B 2 1.4 MiB >>> 1.4 MiB 0 B 0 8.1 TiB N/A N/A N/A >>> 0 B 0 B >>> cephfs.cephfs.meta 2 16 270 MiB 270 MiB 0 B 85.96k 270 MiB >>> 270 MiB 0 B 0 16 TiB N/A N/A N/A >>> 0 B 0 B >>> cephfs.cephfs.data 3 129 12 TiB 12 TiB 0 B 3.73M 12 TiB >>> 12 TiB 0 B 42.49 16 TiB N/A N/A N/A >>> 0 B 0 B >>> While on the client side I have this: >>> $ df -h >>> 10.18.31.1:6789:/ 21T 13T 8.1T 61% /data >>> I don't know where it's gone all the space that was at the beginning. >>> Someone has any hint? >>> Best regards, >>> Mihai >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@ceph.io >>> To unsubscribe send an email to ceph-users-le...@ceph.io > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io