I set the target_size_ratio of pools by mistake as multiple pools sharing
the same raw capacity. After I adjust it, a large number of pgs are in the
backfill state, but the usage rate of osds is still growing, How do I need
to adjust it?

[root@node01 smd]# ceph osd pool autoscale-statusPOOL
                    SIZE  TARGET SIZE                RATE  RAW
CAPACITY   RATIO  TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW
PG_NUM  AUTOSCALE  BULK   device_health_metrics
291.9M                              3.0        48289G  0.0000
                        1.0       1              on         False
deeproute-replica-hdd-pool              12613M
     3.0        785.8T  0.0010        0.1000           0.0010   1.0
  32              on         False  deeproute-replica-ssd-pool
     12352G                              3.0        48289G  0.9901
  50.0000           0.9901   1.0    1024              on         False
 .rgw.root                                5831
      3.0        48289G  0.0099        0.5000           0.0099   1.0
    8              on         False  default.rgw.log
        182                               3.0        48289G  0.0000
                              1.0      32              on
False  default.rgw.control                         0
            3.0        48289G  0.0000
1.0      32              on         False  default.rgw.meta
                0                               3.0        48289G
0.0000                                  4.0       8              on
     False  os-dsglczutvqsgowpz.rgw.control             0
                 3.0        16096G  0.1667        0.5000
0.1667   1.0      64              on         False
os-dsglczutvqsgowpz.rgw.meta            104.2k
     3.0        16096G  0.1667        0.5000           0.1667   1.0
  64              on         False
os-dsglczutvqsgowpz.rgw.buckets.index   57769M
     3.0        16096G  0.1667        0.5000           0.1667   1.0
  32              on         False
os-dsglczutvqsgowpz.rgw.buckets.non-ec  496.5M
     3.0        16096G  0.1667        0.5000           0.1667   1.0
  32              on         False  os-dsglczutvqsgowpz.rgw.log
     309.0M                              3.0        16096G  0.1667
   0.5000           0.1667   1.0      32              on         False
 os-dsglczutvqsgowpz.rgw.buckets.data    147.7T
1.3333333730697632        785.8T  0.7992       80.0000
0.7992   1.0    1024              on         False  cephfs-metadata
                      3231M                              3.0
16096G  0.0006                                  4.0      32
  on         False  cephfs-replicated-pool                  23137G
                         3.0        785.8T  0.1998       20.0000
    0.1998   1.0     128              on         False  .nfs
                         100.1k                              3.0
 48289G  0.0000                                  1.0      32
   on         False  os-dsglczutvqsgowpz.rgw.otp                 0
                          3.0        16096G  0.1667        0.5000
     0.1667   1.0       8              on         False

[root@node01 pg]# ceph osd df | egrep -i "name|hdd"
  1    hdd  10.91399   1.00000   11 TiB  5.3 TiB  5.2 TiB   17 KiB
55 GiB  5.6 TiB  48.26  1.33  195      up
 22    hdd  10.91399   1.00000   11 TiB  8.6 TiB  8.6 TiB    8 KiB
77 GiB  2.3 TiB  79.11  2.17  212      up
 31    hdd  10.91399   1.00000   11 TiB  3.1 TiB  3.0 TiB   12 KiB
29 GiB  7.8 TiB  28.11  0.77  197      up
 51    hdd  10.91399   1.00000   11 TiB  3.9 TiB  3.8 TiB   10 KiB
38 GiB  7.1 TiB  35.28  0.97  186      up
 60    hdd  10.91399   1.00000   11 TiB  927 GiB  916 GiB   14 KiB
11 GiB   10 TiB   8.29  0.23  167      up
 70    hdd  10.91399   1.00000   11 TiB  2.3 TiB  2.2 TiB   13 KiB
5.2 GiB  8.7 TiB  20.63  0.57  177      up
 78    hdd  10.91399   1.00000   11 TiB  3.2 TiB  3.2 TiB   17 KiB
31 GiB  7.7 TiB  29.56  0.81  185      up
 96    hdd  10.91399   1.00000   11 TiB  3.9 TiB  3.8 TiB   11 KiB
38 GiB  7.1 TiB  35.31  0.97  195      up
  9    hdd  10.91399   1.00000   11 TiB  3.0 TiB  2.9 TiB   17 KiB
14 GiB  8.0 TiB  27.11  0.75  183      up
 19    hdd  10.91399   1.00000   11 TiB  4.0 TiB  4.0 TiB   11 KiB
47 GiB  6.9 TiB  36.66  1.01  192      up
 29    hdd  10.91399   1.00000   11 TiB  5.3 TiB  5.2 TiB   14 KiB
40 GiB  5.6 TiB  48.40  1.33  202      up
 47    hdd  10.91399   1.00000   11 TiB  736 GiB  734 GiB   10 KiB
2.1 GiB   10 TiB   6.59  0.18  172      up
 56    hdd  10.91399   1.00000   11 TiB  3.9 TiB  3.8 TiB   10 KiB
38 GiB  7.0 TiB  35.56  0.98  184      up
 65    hdd  10.91399   1.00000   11 TiB  6.1 TiB  6.1 TiB    9 KiB
57 GiB  4.8 TiB  56.22  1.55  214      up
 88    hdd  10.91399   1.00000   11 TiB  4.6 TiB  4.6 TiB   13 KiB
47 GiB  6.3 TiB  42.50  1.17  194      up
 98    hdd  10.91399   1.00000   11 TiB  7.6 TiB  7.5 TiB   14 KiB
60 GiB  3.3 TiB  69.31  1.90  210      up
  2    hdd  10.91399   1.00000   11 TiB  1.6 TiB  1.6 TiB   16 KiB
19 GiB  9.3 TiB  14.38  0.40  182      up
 30    hdd  10.91399   1.00000   11 TiB  4.6 TiB  4.5 TiB   13 KiB
39 GiB  6.3 TiB  41.87  1.15  204      up
 40    hdd  10.91399   1.00000   11 TiB  1.7 TiB  1.7 TiB   10 KiB
13 GiB  9.2 TiB  15.43  0.42  168      up
 48    hdd  10.91399   1.00000   11 TiB  1.5 TiB  1.5 TiB   14 KiB
11 GiB  9.4 TiB  13.78  0.38  168      up
 58    hdd  10.91399   1.00000   11 TiB  7.7 TiB  7.7 TiB   15 KiB
75 GiB  3.2 TiB  70.85  1.95  216      up
 74    hdd  10.91399   1.00000   11 TiB  6.2 TiB  6.2 TiB   13 KiB
65 GiB  4.7 TiB  57.22  1.57  215      up
 83    hdd  10.91399   1.00000   11 TiB  2.4 TiB  2.4 TiB   18 KiB
21 GiB  8.5 TiB  21.81  0.60  183      up
 92    hdd  10.91399   1.00000   11 TiB  3.1 TiB  3.0 TiB   15 KiB
29 GiB  7.8 TiB  28.13  0.77  183      up
 12    hdd  10.91399   1.00000   11 TiB  5.3 TiB  5.2 TiB    9 KiB
40 GiB  5.6 TiB  48.40  1.33  203      up
 23    hdd  10.91399   1.00000   11 TiB  4.7 TiB  4.7 TiB   19 KiB
41 GiB  6.2 TiB  43.50  1.20  188      up
 32    hdd  10.91399   1.00000   11 TiB  2.4 TiB  2.3 TiB    6 KiB
21 GiB  8.6 TiB  21.66  0.60  186      up
 43    hdd  10.91399   1.00000   11 TiB  2.3 TiB  2.2 TiB   12 KiB
5.0 GiB  8.7 TiB  20.62  0.57  177      up
 52    hdd  10.91399   1.00000   11 TiB  3.8 TiB  3.8 TiB   14 KiB
30 GiB  7.1 TiB  34.67  0.95  187      up
 71    hdd  10.91399   1.00000   11 TiB  3.1 TiB  3.1 TiB   14 KiB
29 GiB  7.8 TiB  28.35  0.78  174      up
 95    hdd  10.91399   1.00000   11 TiB  5.5 TiB  5.4 TiB   10 KiB
63 GiB  5.4 TiB  50.29  1.38  197      up
104    hdd  10.91399   1.00000   11 TiB  3.8 TiB  3.8 TiB   11 KiB
38 GiB  7.1 TiB  34.93  0.96  180      up
  3    hdd  10.91399   1.00000   11 TiB  2.4 TiB  2.3 TiB   12 KiB
21 GiB  8.6 TiB  21.65  0.60  179      up
 24    hdd  10.91399   1.00000   11 TiB  2.3 TiB  2.3 TiB   10 KiB
20 GiB  8.6 TiB  21.16  0.58  170      up
 33    hdd  10.91399   1.00000   11 TiB  7.0 TiB  6.9 TiB   13 KiB
73 GiB  3.9 TiB  64.03  1.76  214      up
 45    hdd  10.91399   1.00000   11 TiB  3.2 TiB  3.1 TiB   13 KiB
23 GiB  7.8 TiB  28.92  0.79  188      up

 69    hdd  10.91399   1.00000   11 TiB  3.0 TiB  3.0 TiB   16 KiB
21 GiB  7.9 TiB  27.48  0.76  180      up
 79    hdd  10.91399   1.00000   11 TiB  6.2 TiB  6.1 TiB   23 KiB
57 GiB  4.8 TiB  56.44  1.55  206      up
 89    hdd  10.91399   1.00000   11 TiB  2.2 TiB  2.2 TiB   16 KiB
12 GiB  8.7 TiB  20.30  0.56  180      up
 99    hdd  10.91399   1.00000   11 TiB  4.6 TiB  4.6 TiB   15 KiB
47 GiB  6.3 TiB  42.46  1.17  194      up
  5    hdd  10.91399   1.00000   11 TiB  2.3 TiB  2.3 TiB   18 KiB
20 GiB  8.6 TiB  20.91  0.57  183      up
 15    hdd  10.91399   1.00000   11 TiB  7.6 TiB  7.5 TiB    8 KiB
59 GiB  3.3 TiB  69.31  1.90  215      up
 25    hdd  10.91399   1.00000   11 TiB  1.6 TiB  1.6 TiB   11 KiB
12 GiB  9.3 TiB  14.73  0.40  176      up
 44    hdd  10.91399   1.00000   11 TiB  3.1 TiB  3.1 TiB    9 KiB
36 GiB  7.8 TiB  28.74  0.79  189      up
 63    hdd  10.91399   1.00000   11 TiB  3.0 TiB  3.0 TiB    9 KiB
21 GiB  7.9 TiB  27.49  0.76  189      up
 72    hdd  10.91399   1.00000   11 TiB  4.0 TiB  4.0 TiB   14 KiB
40 GiB  6.9 TiB  36.71  1.01  195      up
 81    hdd  10.91399   1.00000   11 TiB  6.9 TiB  6.9 TiB   14 KiB
66 GiB  4.0 TiB  63.43  1.74  209      up
 91    hdd  10.91399   1.00000   11 TiB  2.3 TiB  2.3 TiB   15 KiB
20 GiB  8.6 TiB  20.91  0.57  179      up
  7    hdd  10.91399   1.00000   11 TiB  4.7 TiB  4.7 TiB   16 KiB
55 GiB  6.2 TiB  43.10  1.18  198      up
 37    hdd  10.91399   1.00000   11 TiB  5.4 TiB  5.4 TiB   14 KiB
56 GiB  5.5 TiB  49.68  1.37  212      up
 46    hdd  10.91399   1.00000   11 TiB  832 GiB  822 GiB   15 KiB
10 GiB   10 TiB   7.45  0.20  175      up
 55    hdd  10.91399   1.00000   11 TiB  6.1 TiB  6.1 TiB   20 KiB
56 GiB  4.8 TiB  56.20  1.54  198      up
 64    hdd  10.91399   1.00000   11 TiB  3.8 TiB  3.8 TiB   25 KiB
30 GiB  7.1 TiB  34.63  0.95  190      up
 84    hdd  10.91399   1.00000   11 TiB  830 GiB  820 GiB   11 KiB
10 GiB   10 TiB   7.43  0.20  177      up
 94    hdd  10.91399   1.00000   11 TiB  2.2 TiB  2.2 TiB    9 KiB
12 GiB  8.7 TiB  20.33  0.56  176      up
102    hdd  10.91399   1.00000   11 TiB  3.2 TiB  3.2 TiB   11 KiB
31 GiB  7.7 TiB  29.52  0.81  173      up
  6    hdd  10.91399   1.00000   11 TiB  4.6 TiB  4.6 TiB   21 KiB
47 GiB  6.3 TiB  42.12  1.16  200      up
 18    hdd  10.91399   1.00000   11 TiB  6.2 TiB  6.1 TiB   11 KiB
64 GiB  4.7 TiB  56.83  1.56  210      up
 41    hdd  10.91399   1.00000   11 TiB  3.1 TiB  3.0 TiB   20 KiB
28 GiB  7.8 TiB  28.09  0.77  188      up
 49    hdd  10.91399   1.00000   11 TiB  904 GiB  893 GiB   21 KiB
11 GiB   10 TiB   8.09  0.22  182      up
 59    hdd  10.91399   1.00000   11 TiB  2.4 TiB  2.4 TiB    8 KiB
21 GiB  8.5 TiB  22.33  0.61  188      up
 68    hdd  10.91399   1.00000   11 TiB  3.9 TiB  3.8 TiB   17 KiB
38 GiB  7.1 TiB  35.31  0.97  190      up
 97    hdd  10.91399   1.00000   11 TiB  3.1 TiB  3.0 TiB   17 KiB
29 GiB  7.8 TiB  28.12  0.77  195      up
105    hdd  10.91399   1.00000   11 TiB  4.0 TiB  3.9 TiB   13 KiB
39 GiB  7.0 TiB  36.25  1.00  193      up
  8    hdd  10.91399   1.00000   11 TiB  6.9 TiB  6.8 TiB   16 KiB
58 GiB  4.1 TiB  62.80  1.73  209      up
 17    hdd  10.91399   1.00000   11 TiB  4.7 TiB  4.7 TiB   13 KiB
54 GiB  6.2 TiB  43.10  1.18  196      up
 27    hdd  10.91399   1.00000   11 TiB  803 GiB  793 GiB   10 KiB
9.8 GiB   10 TiB   7.19  0.20  178      up
 38    hdd  10.91399   1.00000   11 TiB  2.3 TiB  2.3 TiB   14 KiB
20 GiB  8.6 TiB  20.90  0.57  177      up
 57    hdd  10.91399   1.00000   11 TiB  2.4 TiB  2.4 TiB   13 KiB
28 GiB  8.5 TiB  22.05  0.61  181      up
 67    hdd  10.91399   1.00000   11 TiB  4.8 TiB  4.7 TiB    9 KiB
48 GiB  6.1 TiB  43.90  1.21  195      up
 93    hdd  10.91399   1.00000   11 TiB  3.2 TiB  3.2 TiB    8 KiB
37 GiB  7.7 TiB  29.45  0.81  186      up
103    hdd  10.91399   1.00000   11 TiB  4.4 TiB  4.4 TiB   15 KiB
23 GiB  6.5 TiB  40.62  1.12  191      up

[root@node01 smd]# ceph -s  cluster:    id:
f7f1c8ba-f793-436b-bb73-0964108a30c1    health: HEALTH_OK   services:
  mon: 3 daemons, quorum a,b,c (age 8w)    mgr: a(active, since 3w),
standbys: b    mds: 1/1 daemons up, 1 hot standby    osd: 108 osds:
108 up (since 19h), 108 in (since 4M); 104 remapped pgs    rgw: 9
daemons active (9 hosts, 1 zones)   data:    volumes: 1/1 healthy
pools:   17 pools, 5787 pgs    objects: 160.99M objects, 181 TiB
usage:   304 TiB used, 545 TiB / 849 TiB avail    pgs:
119776974/1218292354 objects misplaced (9.832%)             5683
active+clean             102  active+remapped+backfill_wait
 2    active+remapped+backfilling   io:    client:   9.4 MiB/s rd, 87
MiB/s wr, 1.79k op/s rd, 2.34k op/s wr    recovery: 11 MiB/s, 7
objects/s   progress:    Global Recovery Event (0s)
[............................]


郑亮 <zhengliang0...@gmail.com> 于2023年3月16日周四 17:04写道:

> Hi all,
>
> I have a 9 node cluster running *Pacific 16.2.10*. OSDs live on 9 of the
> nodes with each one having 4 x 1.8T ssd and 8 x 10.9T hdd for a total of
> 108 OSDs. We create three crush roots as belows.
>
> 1. The hdds (8x9=72) of all nodes form a large crush root, which is used
> as a data pool, and object storage and cephfs share this crush root.
> 2. Take 3 ssds from the 4 ssds of each node as rbd block storage.
> 3. An ssd on each remaining node is used as an index pool for cephfs and
> object storage.
>
> [root@node01 smd]# ceph osd treeID    CLASS  WEIGHT     TYPE NAME             
>                                 STATUS  REWEIGHT  PRI-AFF                     
>                   -92          15.71910  root 
> root-1c31624a-ad18-445e-8e42-86b71c1fd76f                                     
>                                   -112           1.74657      host 
> node01-fa2cdf3e-7212-4b5f-b62a-3ab1e803547f                                   
>                                 13    ssd    1.74657          osd.13          
>                                   up   1.00000  1.00000                       
>                -103           1.74657      host 
> node02-4e232f27-fe4b-4d0e-bd2a-67d5006a0cdd                                   
>                                 34    ssd    1.74657          osd.34          
>                                   up   1.00000  1.00000                       
>                -109           1.74657      host 
> node03-3ae63d7a-9f65-4bea-b2ba-ff3fe342753d                                   
>                                 28    ssd    1.74657          osd.28          
>                                   up   1.00000  1.00000                       
>                -118           1.74657      host 
> node04-37a3f92a-f6d8-41f9-a774-3069fc2f50b8                                   
>                                 54    ssd    1.74657          osd.54          
>                                   up   1.00000  1.00000                       
>                -106           1.74657      host 
> node05-f667fa27-cc13-4b93-ad56-5dc4c31ffd77                                   
>                                 53    ssd    1.74657          osd.53          
>                                   up   1.00000  1.00000                       
>                 -91           1.74657      host 
> node06-3808c8f6-8e10-47c7-8456-62c1e0e800ed                                   
>                                 61    ssd    1.74657          osd.61          
>                                   up   1.00000  1.00000                       
>                 -97           1.74657      host 
> node07-78216b0d-0999-44e8-905d-8737a5f6f51f                                   
>                                 50    ssd    1.74657          osd.50          
>                                   up   1.00000  1.00000                       
>                -115           1.74657      host 
> node08-947bd556-fb06-497d-8f2c-c4a679d2b06f                                   
>                                 86    ssd    1.74657          osd.86          
>                                   up   1.00000  1.00000                       
>                -100           1.74657      host 
> node09-d9ae9046-0716-454f-ba0c-b03cf9986ba8                             85    
> ssd    1.74657          osd.85                                            up  
>  1.00000  1.00000                                                             
>                                          -38         785.80701  root 
> root-6041a4dc-7c9a-44ed-999c-a847cca81012                                   
> -85          87.31189      host node01-e5646053-2cf8-4ba5-90d5-bb1a63b1234c   
>                            1    hdd   10.91399          osd.1                 
>                             up   1.00000  1.00000  22    hdd   10.91399       
>    osd.22                                            up   0.90002  1.00000  
> 31    hdd   10.91399          osd.31                                          
>   up   1.00000  1.00000  51    hdd   10.91399          osd.51                 
>                            up   1.00000  1.00000  60    hdd   10.91399        
>   osd.60                                            up   1.00000  1.00000  70 
>    hdd   10.91399          osd.70                                            
> up   1.00000  1.00000  78    hdd   10.91399          osd.78                   
>                          up   1.00000  1.00000                                
>         96    hdd   10.91399          osd.96                                  
>           up   1.00000  1.00000                                       -37     
>      87.31189      host node02-be9925fd-60de-4147-81eb-720d7145715f           
>                                                          9    hdd   10.91399  
>         osd.9                                             up   1.00000  
> 1.00000                                        19    hdd   10.91399          
> osd.19                                            up   1.00000  1.00000       
>                                  29    hdd   10.91399          osd.29         
>                                    up   1.00000  1.00000                      
>                   47    hdd   10.91399          osd.47                        
>                     up   1.00000  1.00000                                     
>    56    hdd   10.91399          osd.56                                       
>      up   1.00000  1.00000                                        65    hdd   
> 10.91399          osd.65                                            up   
> 1.00000  1.00000                                        88    hdd   10.91399  
>         osd.88                                            up   1.00000  
> 1.00000                                        98    hdd   10.91399          
> osd.98                                            up   1.00000  1.00000       
>                                 -52          87.31189      host 
> node03-7828653d-6033-4e88-92b0-d8709b0ab218                                   
>                                  2    hdd   10.91399          osd.2           
>                                   up   1.00000  1.00000                       
>                  30    hdd   10.91399          osd.30                         
>                    up   1.00000  1.00000                                      
>   40    hdd   10.91399          osd.40                                        
>     up   1.00000  1.00000                                        48    hdd   
> 10.91399          osd.48                                            up   
> 1.00000  1.00000                                        58    hdd   10.91399  
>         osd.58                                            up   1.00000  
> 1.00000                                        74    hdd   10.91399          
> osd.74                                            up   1.00000  1.00000  83   
>  hdd   10.91399          osd.83                                            up 
>   1.00000  1.00000  92    hdd   10.91399          osd.92                      
>                       up   1.00000  1.00000 -46          87.31189      host 
> node04-e986c3fc-a21b-44ff-9b02-b60b82ee63d7                              12   
>  hdd   10.91399          osd.12                                            up 
>   1.00000  1.00000   23    hdd   10.91399          osd.23                     
>                        up   1.00000  1.00000  32    hdd   10.91399          
> osd.32                                            up   1.00000  1.00000  43   
>  hdd   10.91399          osd.43                                            up 
>   1.00000  1.00000  52    hdd   10.91399          osd.52                      
>                       up   1.00000  1.00000  71    hdd   10.91399          
> osd.71                                            up   1.00000  1.00000       
>                                  95    hdd   10.91399          osd.95         
>                                    up   1.00000  1.00000  104    hdd   
> 10.91399          osd.104                                           up   
> 1.00000  1.00000 -88          87.31189      host 
> node05-fe31d85f-b3b9-4393-b24b-030dbcdfacea                              3    
> hdd   10.91399          osd.3                                             up  
>  1.00000  1.00000  24    hdd   10.91399          osd.24                       
>                      up   1.00000  1.00000  33    hdd   10.91399          
> osd.33                                            up   1.00000  1.00000  45   
>  hdd   10.91399          osd.45                                            up 
>   1.00000  1.00000  69    hdd   10.91399          osd.69                      
>                       up   1.00000  1.00000  79    hdd   10.91399          
> osd.79                                            up   1.00000  1.00000       
>                                  89    hdd   10.91399          osd.89         
>                                    up   1.00000  1.00000                      
>                   99    hdd   10.91399          osd.99                        
>                     up   1.00000  1.00000                                     
>   -55          87.31189      host node06-6f16ba4b-0082-472a-b243-b1a058070918 
>                                                                    5    hdd   
> 10.91399          osd.5                                             up   
> 1.00000  1.00000                                        15    hdd   10.91399  
>         osd.15                                            up   1.00000  
> 1.00000                                        25    hdd   10.91399          
> osd.25                                            up   1.00000  1.00000       
>                                  44    hdd   10.91399          osd.44         
>                                    up   1.00000  1.00000                      
>                   63    hdd   10.91399          osd.63                        
>                     up   1.00000  1.00000                                     
>    72    hdd   10.91399          osd.72                                       
>      up   1.00000  1.00000                                        81    hdd   
> 10.91399          osd.81                                            up   
> 1.00000  1.00000                                        91    hdd   10.91399  
>         osd.91                                            up   1.00000  
> 1.00000                                       -43          87.31189      host 
> node07-5dee846a-2814-4e04-bcfd-ff689d49795c                                   
>                                  7    hdd   10.91399          osd.7           
>                                   up   1.00000  1.00000                       
>                  37    hdd   10.91399          osd.37                         
>                    up   1.00000  1.00000                                      
>   46    hdd   10.91399          osd.46                                        
>     up   1.00000  1.00000                                        55    hdd   
> 10.91399          osd.55                                            up   
> 1.00000  1.00000                                        64    hdd   10.91399  
>         osd.64                                            up   1.00000  
> 1.00000  84    hdd   10.91399          osd.84                                 
>            up   1.00000  1.00000  94    hdd   10.91399          osd.94        
>                                     up   1.00000  1.00000 102    hdd   
> 10.91399          osd.102                                           up   
> 1.00000  1.00000  -58          87.31189      host 
> node08-2d6b7ab3-2067-4e94-b77c-24d6e626e396                               6   
>  hdd   10.91399          osd.6                                             up 
>   1.00000  1.00000  18    hdd   10.91399          osd.18                      
>                       up   1.00000  1.00000  41    hdd   10.91399          
> osd.41                                            up   1.00000  1.00000  49   
>  hdd   10.91399          osd.49                                            up 
>   1.00000  1.00000  59    hdd   10.91399          osd.59                      
>                       up   1.00000  1.00000                                   
>      68    hdd   10.91399          osd.68                                     
>        up   1.00000  1.00000   97    hdd   10.91399          osd.97           
>                                  up   1.00000  1.00000 105    hdd   10.91399  
>         osd.105                                           up   1.00000  
> 1.00000 -49          87.31189      host 
> node09-e6bf0642-e3b4-48f0-9d0e-3d87ceacced8                              8    
> hdd   10.91399          osd.8                                             up  
>  1.00000  1.00000  17    hdd   10.91399          osd.17                       
>                      up   1.00000  1.00000  27    hdd   10.91399          
> osd.27                                            up   1.00000  1.00000  38   
>  hdd   10.91399          osd.38                                            up 
>   1.00000  1.00000  57    hdd   10.91399          osd.57                      
>                       up   1.00000  1.00000  67    hdd   10.91399          
> osd.67                                            up   1.00000  1.00000  93   
>  hdd   10.91399          osd.93                                            up 
>   1.00000  1.00000 103    hdd   10.91399          osd.103                     
>                       up   1.00000  1.00000  -8          47.16115  root 
> root-ea7a1878-722e-49d5-8a91-c618a6aefe29                                  
> -13           5.24013      host node01-4c465825-6bd1-42a5-b087-51a82cb2865c   
>                            0    ssd    1.74657          osd.0                 
>                             up   0.95001  1.00000  11    ssd    1.74699       
>    osd.11                                            up   1.00000  1.00000  
> 16    ssd    1.74657          osd.16                                          
>   up   1.00000  1.00000 -22           5.24013      host 
> node02-0e3e418d-9129-44ee-8453-53680171270e                              4    
> ssd    1.74657          osd.4                                             up  
>  1.00000  1.00000  14    ssd    1.74657          osd.14                       
>                      up   0.95001  1.00000  82    ssd    1.74699          
> osd.82                                            up   1.00000  1.00000 -28   
>         5.24013      host node03-274397b4-8559-4a17-a9eb-1c16920ad432         
>                     10    ssd    1.74657          osd.10                      
>                       up   1.00000  1.00000  76    ssd    1.74657          
> osd.76                                            up   1.00000  1.00000 100   
>  ssd    1.74699          osd.100                                           up 
>   0.95001  1.00000 -31           5.24013      host 
> node04-df05c3cc-08ef-426e-85b8-088cb8c1b4e2                             20    
> ssd    1.74657          osd.20                                            up  
>  1.00000  1.00000  39    ssd    1.74657          osd.39                       
>                      up   0.95001  1.00000  90    ssd    1.74699          
> osd.90                                            up   0.95001  1.00000 -25   
>         5.24013      host node05-d07e7e92-8290-49ba-b09c-ad34bddb1eae         
>                     26    ssd    1.74657          osd.26                      
>                       up   0.95001  1.00000   62    ssd    1.74699          
> osd.62                                            up   1.00000  1.00000  73   
>  ssd    1.74657          osd.73                                            up 
>   1.00000  1.00000 -16           5.24013      host 
> node06-a7a40371-a128-4ab1-90a2-62c99e040036                             42    
> ssd    1.74657          osd.42                                            up  
>  0.95001  1.00000  75    ssd    1.74657          osd.75                       
>                      up   0.95001  1.00000 107    ssd    1.74699          
> osd.107                                           up   1.00000  1.00000  -7   
>         5.24013      host node07-971dee6e-dec4-4c0a-86d1-54d0b23832bd         
>                     21    ssd    1.74699          osd.21                      
>                       up   1.00000  1.00000  66    ssd    1.74657          
> osd.66                                            up   0.95001  1.00000  77   
>  ssd    1.74657          osd.77                                            up 
>   1.00000  1.00000 -19           5.24013      host 
> node08-9a47282b-9530-4cc7-9e29-7f6c0b2f5184                             35    
> ssd    1.74699          osd.35                                            up  
>  1.00000  1.00000  80    ssd    1.74657          osd.80                       
>                      up   1.00000  1.00000 101    ssd    1.74657          
> osd.101                                           up   1.00000  1.00000 -34   
>         5.24013      host node09-f09a8013-2426-4ced-b9c7-02c06ca9d6fc         
>                      36    ssd    1.74699          osd.36                     
>                        up   0.95001  1.00000  87    ssd    1.74657          
> osd.87                                            up   1.00000  1.00000 106   
>  ssd    1.74657          osd.106                                           up 
>   1.00000  1.00000
>
> After ceph has been running normally for a period of time, I found that
> the distribution of osd data is seriously uneven, and autoscale PGs were
> turned on in my envs. My hdds are identical, but utilization ranges from
> 6.14% to 74.35%
>
> root@node09:/# ceph osd df | grep hdd ID   CLASS  WEIGHT    REWEIGHT  SIZE    
>  RAW USE  DATA      OMAP     META     AVAIL    %USE   VAR   PGS  STATUS  1    
> hdd  10.91399   1.00000   11 TiB  5.3 TiB   5.2 TiB   15 KiB   53 GiB  5.7 
> TiB  48.15  1.38  174      up 22    hdd  10.91399   0.90002   11 TiB  8.1 TiB 
>   8.0 TiB   12 KiB   71 GiB  2.8 TiB  74.35  2.13  160      up 31    hdd  
> 10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   13 KiB   27 GiB  8.0 TiB  
> 27.14  0.78  179      up 51    hdd  10.91399   1.00000   11 TiB  3.7 TiB   
> 3.7 TiB    9 KiB   36 GiB  7.2 TiB  34.12  0.98  173      up 60    hdd  
> 10.91399   1.00000   11 TiB  973 GiB   962 GiB   14 KiB   11 GiB   10 TiB   
> 8.71  0.25  167      up 70    hdd  10.91399   1.00000   11 TiB  2.0 TiB   2.0 
> TiB   11 KiB  3.9 GiB  8.9 TiB  18.42  0.53  173      up 78    hdd  10.91399  
>  1.00000   11 TiB  3.0 TiB   2.9 TiB   12 KiB   28 GiB  7.9 TiB  27.18  0.78  
> 173      up 96    hdd  10.91399   1.00000   11 TiB  3.7 TiB   3.7 TiB   10 
> KiB   36 GiB  7.2 TiB  34.15  0.98  176      up  9    hdd  10.91399   1.00000 
>   11 TiB  2.8 TiB   2.8 TiB   16 KiB   13 GiB  8.1 TiB  25.38  0.73  178      
> up 19    hdd  10.91399   1.00000   11 TiB  3.8 TiB   3.8 TiB   13 KiB   43 
> GiB  7.1 TiB  35.04  1.00  177      up 29    hdd  10.91399   1.00000   11 TiB 
>  5.1 TiB   5.0 TiB   14 KiB   38 GiB  5.9 TiB  46.38  1.33  182      up 47    
> hdd  10.91399   1.00000   11 TiB  686 GiB   684 GiB    9 KiB  1.4 GiB   10 
> TiB   6.14  0.18  172      up 56    hdd  10.91399   1.00000   11 TiB  3.7 TiB 
>   3.7 TiB   12 KiB   36 GiB  7.2 TiB  34.16  0.98  174      up 65    hdd  
> 10.91399   1.00000   11 TiB  5.9 TiB   5.9 TiB    7 KiB   54 GiB  5.0 TiB  
> 54.27  1.55  183      up 88    hdd  10.91399   1.00000   11 TiB  4.5 TiB   
> 4.4 TiB   12 KiB   44 GiB  6.4 TiB  41.17  1.18  173      up 98    hdd  
> 10.91399   1.00000   11 TiB  7.3 TiB   7.2 TiB   13 KiB   56 GiB  3.7 TiB  
> 66.49  1.90  179      up  2    hdd  10.91399   1.00000   11 TiB  1.5 TiB   
> 1.5 TiB   12 KiB   18 GiB  9.4 TiB  14.02  0.40  177      up 30    hdd  
> 10.91399   1.00000   11 TiB  4.4 TiB   4.4 TiB   10 KiB   37 GiB  6.5 TiB  
> 40.29  1.15  180      up 40    hdd  10.91399   1.00000   11 TiB  1.4 TiB   
> 1.4 TiB   10 KiB   11 GiB  9.5 TiB  13.14  0.38  172      up 48    hdd  
> 10.91399   1.00000   11 TiB  1.4 TiB   1.4 TiB   12 KiB   11 GiB  9.5 TiB  
> 13.17  0.38  168      up 58    hdd  10.91399   1.00000   11 TiB  7.5 TiB   
> 7.4 TiB   14 KiB   70 GiB  3.5 TiB  68.31  1.96  182      up 74    hdd  
> 10.91399   1.00000   11 TiB  6.0 TiB   6.0 TiB   10 KiB   61 GiB  4.9 TiB  
> 55.15  1.58  181      up 83    hdd  10.91399   1.00000   11 TiB  2.2 TiB   
> 2.2 TiB   22 KiB   19 GiB  8.7 TiB  20.13  0.58  174      up 92    hdd  
> 10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   11 KiB   27 GiB  8.0 TiB  
> 27.15  0.78  171      up 12    hdd  10.91399   1.00000   11 TiB  5.1 TiB   
> 5.0 TiB    9 KiB   38 GiB  5.9 TiB  46.37  1.33  176      up 23    hdd  
> 10.91399   1.00000   11 TiB  4.4 TiB   4.4 TiB   22 KiB   37 GiB  6.5 TiB  
> 40.24  1.15  173      up 32    hdd  10.91399   1.00000   11 TiB  2.2 TiB   
> 2.2 TiB   10 KiB   19 GiB  8.7 TiB  20.13  0.58  176      up 43    hdd  
> 10.91399   1.00000   11 TiB  2.0 TiB   2.0 TiB   10 KiB  4.4 GiB  8.9 TiB  
> 18.41  0.53  176      up 52    hdd  10.91399   1.00000   11 TiB  3.6 TiB   
> 3.6 TiB   12 KiB   29 GiB  7.3 TiB  33.27  0.95  171      up 71    hdd  
> 10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   15 KiB   27 GiB  8.0 TiB  
> 27.13  0.78  170      up 95    hdd  10.91399   1.00000   11 TiB  5.4 TiB   
> 5.3 TiB   10 KiB   60 GiB  5.6 TiB  49.04  1.40  173      up104    hdd  
> 10.91399   1.00000   11 TiB  3.7 TiB   3.7 TiB    8 KiB   36 GiB  7.2 TiB  
> 34.13  0.98  171      up  3    hdd  10.91399   1.00000   11 TiB  2.2 TiB   
> 2.2 TiB    9 KiB   19 GiB  8.7 TiB  20.14  0.58  173      up 24    hdd  
> 10.91399   1.00000   11 TiB  2.2 TiB   2.2 TiB    9 KiB   19 GiB  8.7 TiB  
> 20.12  0.58  170      up 33    hdd  10.91399   1.00000   11 TiB  6.8 TiB   
> 6.7 TiB   14 KiB   69 GiB  4.1 TiB  62.15  1.78  181      up 45    hdd  
> 10.91399   1.00000   11 TiB  2.9 TiB   2.8 TiB   10 KiB   20 GiB  8.0 TiB  
> 26.27  0.75  178      up 69    hdd  10.91399   1.00000   11 TiB  2.9 TiB   
> 2.8 TiB   15 KiB   20 GiB  8.0 TiB  26.25  0.75  173      up 79    hdd  
> 10.91399   1.00000   11 TiB  5.9 TiB   5.9 TiB   28 KiB   54 GiB  5.0 TiB  
> 54.25  1.55  178      up 89    hdd  10.91399   1.00000   11 TiB  2.1 TiB   
> 2.1 TiB   16 KiB   11 GiB  8.8 TiB  19.25  0.55  173      up 99    hdd  
> 10.91399   1.00000   11 TiB  4.5 TiB   4.4 TiB   13 KiB   44 GiB  6.4 TiB  
> 41.14  1.18  174      up  5    hdd  10.91399   1.00000   11 TiB  2.2 TiB   
> 2.2 TiB   14 KiB   19 GiB  8.7 TiB  20.12  0.58  172      up 15    hdd  
> 10.91399   1.00000   11 TiB  7.3 TiB   7.2 TiB    7 KiB   56 GiB  3.7 TiB  
> 66.48  1.90  177      up 25    hdd  10.91399   1.00000   11 TiB  1.4 TiB   
> 1.4 TiB   11 KiB   10 GiB  9.5 TiB  13.13  0.38  176      up 44    hdd  
> 10.91399   1.00000   11 TiB  3.1 TiB   3.0 TiB    9 KiB   34 GiB  7.9 TiB  
> 28.03  0.80  179      up 63    hdd  10.91399   1.00000   11 TiB  2.9 TiB   
> 2.8 TiB   10 KiB   20 GiB  8.0 TiB  26.26  0.75  178      up 72    hdd  
> 10.91399   1.00000   11 TiB  3.7 TiB   3.7 TiB   12 KiB   36 GiB  7.2 TiB  
> 34.13  0.98  180      up 81    hdd  10.91399   1.00000   11 TiB  6.7 TiB   
> 6.6 TiB   11 KiB   62 GiB  4.2 TiB  61.30  1.76  181      up 91    hdd  
> 10.91399   1.00000   11 TiB  2.2 TiB   2.2 TiB   16 KiB   19 GiB  8.7 TiB  
> 20.11  0.58  176      up  7    hdd  10.91399   1.00000   11 TiB  4.6 TiB   
> 4.5 TiB   13 KiB   52 GiB  6.3 TiB  42.03  1.20  180      up 37    hdd  
> 10.91399   1.00000   11 TiB  5.3 TiB   5.2 TiB   14 KiB   52 GiB  5.7 TiB  
> 48.16  1.38  182      up 46    hdd  10.91399   1.00000   11 TiB  784 GiB   
> 775 GiB   14 KiB  8.8 GiB   10 TiB   7.01  0.20  173      up 55    hdd  
> 10.91399   1.00000   11 TiB  5.9 TiB   5.9 TiB   20 KiB   54 GiB  5.0 TiB  
> 54.25  1.55  177      up 64    hdd  10.91399   1.00000   11 TiB  3.6 TiB   
> 3.6 TiB   18 KiB   28 GiB  7.3 TiB  33.23  0.95  176      up 84    hdd  
> 10.91399   1.00000   11 TiB  784 GiB   775 GiB   10 KiB  9.0 GiB   10 TiB   
> 7.01  0.20  176      up 94    hdd  10.91399   1.00000   11 TiB  2.1 TiB   2.1 
> TiB    8 KiB   12 GiB  8.8 TiB  19.28  0.55  169      up102    hdd  10.91399  
>  1.00000   11 TiB  3.0 TiB   2.9 TiB   10 KiB   28 GiB  8.0 TiB  27.13  0.78  
> 170      up  6    hdd  10.91399   1.00000   11 TiB  4.5 TiB   4.4 TiB   23 
> KiB   44 GiB  6.4 TiB  41.14  1.18  181      up 18    hdd  10.91399   1.00000 
>   11 TiB  6.0 TiB   6.0 TiB    9 KiB   61 GiB  4.9 TiB  55.15  1.58  183      
> up 41    hdd  10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   15 KiB   27 
> GiB  8.0 TiB  27.12  0.78  179      up 49    hdd  10.91399   1.00000   11 TiB 
>  785 GiB   775 GiB   19 KiB  9.5 GiB   10 TiB   7.02  0.20  176      up 59    
> hdd  10.91399   1.00000   11 TiB  2.2 TiB   2.2 TiB    8 KiB   19 GiB  8.7 
> TiB  20.12  0.58  178      up 68    hdd  10.91399   1.00000   11 TiB  3.7 TiB 
>   3.7 TiB   13 KiB   36 GiB  7.2 TiB  34.15  0.98  172      up 97    hdd  
> 10.91399   1.00000   11 TiB  3.0 TiB   2.9 TiB   15 KiB   27 GiB  8.0 TiB  
> 27.13  0.78  178      up105    hdd  10.91399   1.00000   11 TiB  3.7 TiB   
> 3.7 TiB   13 KiB   36 GiB  7.2 TiB  34.14  0.98  179      up  8    hdd  
> 10.91399   1.00000   11 TiB  6.6 TiB   6.5 TiB   13 KiB   55 GiB  4.3 TiB  
> 60.40  1.73  181      up 17    hdd  10.91399   1.00000   11 TiB  4.6 TiB   
> 4.5 TiB   14 KiB   52 GiB  6.3 TiB  42.03  1.20  177      up 27    hdd  
> 10.91399   1.00000   11 TiB  783 GiB   774 GiB    8 KiB  9.4 GiB   10 TiB   
> 7.01  0.20  176      up 38    hdd  10.91399   1.00000   11 TiB  2.2 TiB   2.2 
> TiB   15 KiB   19 GiB  8.7 TiB  20.11  0.58  174      up 57    hdd  10.91399  
>  1.00000   11 TiB  2.3 TiB   2.3 TiB   12 KiB   26 GiB  8.6 TiB  21.01  0.60  
> 177      up 67    hdd  10.91399   1.00000   11 TiB  4.5 TiB   4.4 TiB    9 
> KiB   44 GiB  6.4 TiB  41.14  1.18  176      up 93    hdd  10.91399   1.00000 
>   11 TiB  3.1 TiB   3.0 TiB    7 KiB   35 GiB  7.9 TiB  28.01  0.80  171      
> up103    hdd  10.91399   1.00000   11 TiB  4.2 TiB   4.2 TiB   10 KiB   22 
> GiB  6.7 TiB  38.52  1.10  176      up
>
> three pools (deeproute-replica-hdd-pool,
> os-dsglczutvqsgowpz.rgw.buckets.data, and cephfs-replicated-pool ) are
> set to use the same device class hdd. I noticed that the *effective ratio* of
> the three pools is very different, not sure if it will affect the data
> rebalancing.
>
> [root@node01 smd]# ceph osd pool autoscale-statusPOOL                         
>              SIZE  TARGET SIZE                RATE  RAW CAPACITY   RATIO  
> TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE  BULK   
> device_health_metrics                   278.5M                              
> 3.0        48289G  0.0000                                  1.0       1        
>       on         False  deeproute-replica-hdd-pool              12231M        
>                       3.0        785.8T  0.9901       50.0000           
> 0.9901   1.0    4096              on         False  
> deeproute-replica-ssd-pool              12333G                              
> 3.0        48289G  0.9901       50.0000           0.9901   1.0    1024        
>       on         False  .rgw.root                                5831         
>                       3.0        48289G  0.0099        0.5000           
> 0.0099   1.0       8              on         False  default.rgw.log           
>                 182                               3.0        48289G  0.0000   
>                                1.0      32              on         False  
> default.rgw.control                         0                               
> 3.0        48289G  0.0000                                  1.0      32        
>       on         False  default.rgw.meta                            0         
>                       3.0        48289G  0.0000                               
>    4.0       8              on         False  os-dsglczutvqsgowpz.rgw.control 
>             0                               3.0        16096G  0.1667        
> 0.5000           0.1667   1.0      64              on         False  
> os-dsglczutvqsgowpz.rgw.meta            99.81k                              
> 3.0        16096G  0.1667        0.5000           0.1667   1.0      64        
>       on         False  os-dsglczutvqsgowpz.rgw.buckets.index   59437M        
>                       3.0        16096G  0.1667        0.5000           
> 0.1667   1.0      32              on         False  
> os-dsglczutvqsgowpz.rgw.buckets.non-ec  498.7M                              
> 3.0        16096G  0.1667        0.5000           0.1667   1.0      32        
>       on         False  os-dsglczutvqsgowpz.rgw.log              1331M        
>                       3.0        16096G  0.1667        0.5000           
> 0.1667   1.0      32              on         False  
> os-dsglczutvqsgowpz.rgw.buckets.data    143.1T               
> 1.3333333730697632        785.8T  0.2429        0.5000           0.0099   1.0 
>      32              on         False  cephfs-metadata                        
>   3161M                              3.0        16096G  0.0006                
>                   4.0      32              on         False  
> cephfs-replicated-pool                  21862G                              
> 3.0        785.8T  0.0815                                  1.0      32        
>       on         False  .nfs                                    84599         
>                       3.0        48289G  0.0000                               
>    1.0      32              on         False  os-dsglczutvqsgowpz.rgw.otp     
>             0                               3.0        16096G  0.1667        
> 0.5000
>
> [root@node01 smd]# ceph df--- RAW STORAGE ---CLASS     SIZE    AVAIL     USED 
>  RAW USED  %RAW USEDhdd    786 TiB  545 TiB  240 TiB   240 TiB      30.60ssd  
>    63 TiB   25 TiB   38 TiB    38 TiB      59.95TOTAL  849 TiB  571 TiB  278 
> TiB   278 TiB      32.77 --- POOLS ---POOL                                    
> ID   PGS   STORED  OBJECTS     USED  %USED  MAX AVAILdevice_health_metrics    
>                 1     1  280 MiB      135  841 MiB   0.02    1.8 
> TiBdeeproute-replica-hdd-pool              11  4096   12 GiB    9.44k   36 
> GiB   0.02     69 TiBdeeproute-replica-ssd-pool              12  1024   12 
> TiB    3.43M   37 TiB  87.62    1.8 TiB.rgw.root                              
>  25     8  5.7 KiB       20  228 KiB      0    1.8 TiBdefault.rgw.log         
>                 26    32    182 B        2   24 KiB      0    1.8 
> TiBdefault.rgw.control                     27    32      0 B        8      0 
> B      0    1.8 TiBdefault.rgw.meta                        28     8      0 B  
>       0      0 B      0    1.8 TiBos-dsglczutvqsgowpz.rgw.control         29  
>   64      0 B        8      0 B      0    4.9 TiBos-dsglczutvqsgowpz.rgw.meta 
>            30    64  100 KiB      456  4.3 MiB      0    4.9 
> TiBos-dsglczutvqsgowpz.rgw.buckets.index   31    32   58 GiB   10.40k  174 
> GiB   1.15    4.9 TiBos-dsglczutvqsgowpz.rgw.buckets.non-ec  32    32  500 
> MiB  247.78k  4.3 GiB   0.03    4.9 TiBos-dsglczutvqsgowpz.rgw.log            
>  33    32  689 MiB      432  2.0 GiB   0.01    4.9 
> TiBos-dsglczutvqsgowpz.rgw.buckets.data    34    32  130 TiB  134.96M  174 
> TiB  45.69    155 TiBcephfs-metadata                         35    32  3.1 
> GiB    1.67M  9.3 GiB   0.06    4.9 TiBcephfs-replicated-pool                 
>  36    32   21 TiB    8.25M   64 TiB  23.79     69 TiB.nfs                    
>                 37    32   89 KiB        9  341 KiB      0    1.8 
> TiBos-dsglczutvqsgowpz.rgw.otp             38     8      0 B        0      0 
> B      0    4.9 TiB
>
> [root@node01 smd]# ceph -s
>   cluster:
>     id:     f7f1c8ba-f793-436b-bb73-0964108a30c1
>     health: HEALTH_OK
>
>   services:
>     mon: 3 daemons, quorum a,b,c (age 7w)
>     mgr: a(active, since 2w), standbys: b
>     mds: 1/1 daemons up, 1 hot standby
>     osd: 108 osds: 108 up (since 5h), 108 in (since 4M); 1 remapped pgs
>     rgw: 9 daemons active (9 hosts, 1 zones)
>
>   data:
>     volumes: 1/1 healthy
>     pools:   17 pools, 5561 pgs
>     objects: 148.37M objects, 162 TiB
>     usage:   276 TiB used, 572 TiB / 849 TiB avail
>     pgs:     2786556/1119434807 objects misplaced (0.249%)
>              5559 active+clean
>              1    active+remapped+backfilling
>              1    active+clean+scrubbing+deep
>
>   io:
>     client:   212 MiB/s rd, 27 MiB/s wr, 4.04k op/s rd, 2.01k op/s wr
>     recovery: 4.0 MiB/s, 4 objects/s
>
>   progress:
>     Global Recovery Event (2d)
>       [===========================.] (remaining: 39m)
>
>
> So my question is, how can I adjust to make the osd data more evenly
> distributed? Thanks!
>
> Best Regards,
> Liang Zheng
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to