Hi Eugen,
Thanks for your answer. Please find below the output of the command:
ceph osd pool ls detail
pool 1 '.mgr' replicated size 2 min_size 1 crush_rule 0 object_hash
rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 323 flags
hashpspool,nearfull stripe_width 0 pg_num_max 32 pg_num_min 1
application mgr read_balance_score 12.50
pool 2 'cephfs.cephfs.meta' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 16 pgp_num 1 pgp_num_target 16
autoscale_mode on last_change 295 lfor 0/0/54 flags hashpspool,nearfull
stripe_width 0 pg_autoscale_bias 4 pg_num_min 16 recovery_priority 5
application cephfs read_balance_score 12.50
pool 3 'cephfs.cephfs.data' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 129 pgp_num 1 pg_num_target 512
pgp_num_target 512 autoscale_mode on last_change 326 lfor 0/0/54 flags
hashpspool,nearfull,bulk stripe_width 0 application cephfs
read_balance_score 12.50
How can I decrese the number of pg_num?
Best,
Mihai
On 2025-03-28 13:19, Eugen Block wrote:
Do you use size 1 for your data? You know that's bad, right? Please
share 'ceph osd pool ls detail'.
Also, it's recommended to use a power of 2 for pg_num, so you should
decrease the pg_num for the pool cephfs.cephfs.data.
Zitat von Mihai Ciubancan <mihai.ciuban...@eli-np.ro>:
Hi Anthony,
Thanks for the answer:
The output of 'ceph osd df' is:
ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP
META AVAIL %USE VAR PGS STATUS TYPE NAME
-1 167.64825 - 168 TiB 12 TiB 12 TiB 100 MiB
27 GiB 156 TiB 7.13 1.00 - root default
-3 167.64825 - 168 TiB 12 TiB 12 TiB 100 MiB
27 GiB 156 TiB 7.13 1.00 - host sto-core-hpc01
0 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.0
1 ssd 13.97069 1.00000 14 TiB 12 TiB 12 TiB 6 KiB
26 GiB 2.0 TiB 85.53 12.00 129 up osd.1
2 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.2
3 ssd 13.97069 1.00000 14 TiB 1.7 GiB 258 MiB 100 MiB
1.3 GiB 14 TiB 0.01 0.00 16 up osd.3
4 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.4
5 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.5
6 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.6
7 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.7
8 ssd 13.97069 1.00000 14 TiB 68 MiB 4.8 MiB 12 KiB
63 MiB 14 TiB 0 0 1 up osd.8
9 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.9
10 ssd 13.97069 1.00000 14 TiB 32 MiB 4.1 MiB 12 KiB
28 MiB 14 TiB 0 0 0 up osd.10
11 ssd 13.97069 1.00000 14 TiB 68 MiB 4.8 MiB 12 KiB
63 MiB 14 TiB 0 0 1 up osd.11
TOTAL 168 TiB 12 TiB 12 TiB 100 MiB
27 GiB 156 TiB 7.13
So all the date is on osd.1
But I have checked the balancer and seems active:
ceph balancer status
{
"active": true,
"last_optimize_duration": "0:00:00.000368",
"last_optimize_started": "Fri Mar 28 10:55:06 2025",
"mode": "upmap",
"no_optimization_needed": false,
"optimize_result": "Some objects (0.500000) are degraded; try
again later",
"plans": []
}
But the output of the commnad 'ceph config dump|grep balancer' gives
me nothing.
Best,
Mihai
On 2025-03-27 23:06, Anthony D'Atri wrote:
Look at `ceph osd df`. Is the balancer enabled?
On Mar 27, 2025, at 8:50 AM, Mihai Ciubancan
<mihai.ciuban...@eli-np.ro> wrote:
Hello,
My name is Mihai, and I have started using CEPH this mount for a
HPC cluster.
When was lunch in the production the available space shown was 80TB
now is 16TB and I didn't do anything, while I'm having 12 OSD (SSD
of 14TB):
sudo ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT
PRI-AFF
-1 167.64825 root default
-3 167.64825 host sto-core-hpc01
0 ssd 13.97069 osd.0 up 1.00000
1.00000
1 ssd 13.97069 osd.1 up 1.00000
1.00000
2 ssd 13.97069 osd.2 up 1.00000
1.00000
3 ssd 13.97069 osd.3 up 1.00000
1.00000
4 ssd 13.97069 osd.4 up 1.00000
1.00000
5 ssd 13.97069 osd.5 up 1.00000
1.00000
6 ssd 13.97069 osd.6 up 1.00000
1.00000
7 ssd 13.97069 osd.7 up 1.00000
1.00000
8 ssd 13.97069 osd.8 up 1.00000
1.00000
9 ssd 13.97069 osd.9 up 1.00000
1.00000
10 ssd 13.97069 osd.10 up 1.00000
1.00000
11 ssd 13.97069 osd.11 up 1.00000
1.00000
sudo ceph df detail
--- RAW STORAGE ---
CLASS SIZE AVAIL USED RAW USED %RAW USED
ssd 168 TiB 156 TiB 12 TiB 12 TiB 7.12
TOTAL 168 TiB 156 TiB 12 TiB 12 TiB 7.12
--- POOLS ---
POOL ID PGS STORED (DATA) (OMAP) OBJECTS
USED (DATA) (OMAP) %USED MAX AVAIL QUOTA OBJECTS QUOTA BYTES
DIRTY USED COMPR UNDER COMPR
.mgr 1 1 705 KiB 705 KiB 0 B 2 1.4
MiB 1.4 MiB 0 B 0 8.1 TiB N/A N/A
N/A 0 B 0 B
cephfs.cephfs.meta 2 16 270 MiB 270 MiB 0 B 85.96k 270
MiB 270 MiB 0 B 0 16 TiB N/A N/A
N/A 0 B 0 B
cephfs.cephfs.data 3 129 12 TiB 12 TiB 0 B 3.73M 12
TiB 12 TiB 0 B 42.49 16 TiB N/A N/A
N/A 0 B 0 B
While on the client side I have this:
$ df -h
10.18.31.1:6789:/ 21T 13T 8.1T 61% /data
I don't know where it's gone all the space that was at the
beginning.
Someone has any hint?
Best regards,
Mihai
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io