Hi David,

root@or1010051251044:~# ceph df
GLOBAL:
    SIZE       AVAIL      RAW USED     %RAW USED
    79793G     56832G       22860G         28.65
POOLS:
    NAME        ID     USED      %USED     MAX AVAIL     OBJECTS
    rbd         0          0         0        14395G           0
    compute     1          0         0        14395G           0
    volumes     2      7605G     28.60        14395G     1947372
    images      4          0         0        14395G           0
root@or1010051251044:~#



pool : 4 0 1 2 | SUM
------------------------------------------------
osd.10 8 10 44 96 | 158
osd.11 14 8 58 100 | 180
osd.12 12 6 50 95 | 163
osd.13 14 4 49 121 | 188
osd.14 9 8 54 86 | 157
osd.15 12 5 55 103 | 175
osd.16 23 5 56 99 | 183
osd.30 6 4 31 47 | 88
osd.17 8 8 50 114 | 180
osd.31 7 1 23 35 | 66
osd.18 15 5 42 94 | 156
osd.32 12 6 24 54 | 96
osd.19 13 5 54 116 | 188
osd.33 4 2 28 49 | 83
osd.34 7 5 18 62 | 92
osd.35 10 2 21 56 | 89
osd.36 5 1 34 35 | 75
osd.37 4 4 24 45 | 77
osd.39 14 8 48 106 | 176
osd.0 12 3 27 67 | 109
osd.1 8 3 27 43 | 81
osd.2 4 5 27 45 | 81
osd.3 4 3 19 50 | 76
osd.4 4 1 23 54 | 82
osd.5 4 2 23 56 | 85
osd.6 1 5 32 50 | 88
osd.7 9 1 32 66 | 108
osd.8 7 4 27 49 | 87
osd.9 6 4 24 55 | 89
osd.20 7 4 43 122 | 176
osd.21 14 5 46 95 | 160
osd.22 13 8 51 107 | 179
osd.23 11 7 54 105 | 177
osd.24 11 6 52 112 | 181
osd.25 16 6 36 98 | 156
osd.26 15 7 59 101 | 182
osd.27 7 9 58 101 | 175
osd.28 16 5 60 89 | 170
osd.29 18 7 53 94 | 172
------------------------------------------------
SUM : 384 192 1536 3072



root@or1010051251044:~# for i in `rados lspools`; do echo
"================="; echo Working on pool: $i; ceph osd pool get $i pg_num;
ceph osd pool get $i pgp_num; done ================= Working on pool: rbd
pg_num: 64 pgp_num: 64 ================= Working on pool: compute pg_num:
512 pgp_num: 512 ================= Working on pool: volumes pg_num: 1024
pgp_num: 1024 ================= Working on pool: images pg_num: 128
pgp_num: 128 root@or1010051251044:~#



Thanks,
Pardhiv Karri

On Tue, May 22, 2018 at 9:16 AM, David Turner <drakonst...@gmail.com> wrote:

> This is all weird. Maybe it just doesn't have any PGs with data on them.
> `ceph df`, how many PGs you have in each pool, and which PGs are on osd 38.
>
>
> On Tue, May 22, 2018, 11:19 AM Pardhiv Karri <meher4in...@gmail.com>
> wrote:
>
>> Hi David,
>>
>>
>>
>> root@or1010051251044:~# ceph osd tree
>> ID  WEIGHT   TYPE NAME                    UP/DOWN REWEIGHT
>> PRIMARY-AFFINITY
>>  -1 80.00000 root default
>>
>>  -2 40.00000     rack rack_A1
>>
>>  -3 20.00000         host or1010051251040
>>
>>   0  2.00000             osd.0                 up  1.00000
>>  1.00000
>>   1  2.00000             osd.1                 up  1.00000
>>  1.00000
>>   2  2.00000             osd.2                 up  1.00000
>>  1.00000
>>   3  2.00000             osd.3                 up  1.00000
>>  1.00000
>>   4  2.00000             osd.4                 up  1.00000
>>  1.00000
>>   5  2.00000             osd.5                 up  1.00000
>>  1.00000
>>   6  2.00000             osd.6                 up  1.00000
>>  1.00000
>>   7  2.00000             osd.7                 up  1.00000
>>  1.00000
>>   8  2.00000             osd.8                 up  1.00000
>>  1.00000
>>   9  2.00000             osd.9                 up  1.00000
>>  1.00000
>>  -8 20.00000         host or1010051251044
>>
>>  30  2.00000             osd.30                up  1.00000
>>  1.00000
>>  31  2.00000             osd.31                up  1.00000
>>  1.00000
>>  32  2.00000             osd.32                up  1.00000
>>  1.00000
>>  33  2.00000             osd.33                up  1.00000
>>  1.00000
>>  34  2.00000             osd.34                up  1.00000
>>  1.00000
>>  35  2.00000             osd.35                up  1.00000
>>  1.00000
>>  36  2.00000             osd.36                up  1.00000
>>  1.00000
>>  37  2.00000             osd.37                up  1.00000
>>  1.00000
>>  38  2.00000             osd.38                up  1.00000
>>  1.00000
>>  39  2.00000             osd.39                up  1.00000
>>  1.00000
>>  -4 20.00000     rack rack_B1
>>
>>  -5 20.00000         host or1010051251041
>>
>>  10  2.00000             osd.10                up  1.00000
>>  1.00000
>>  11  2.00000             osd.11                up  1.00000
>>  1.00000
>>  12  2.00000             osd.12                up  1.00000
>>  1.00000
>>  13  2.00000             osd.13                up  1.00000
>>  1.00000
>>  14  2.00000             osd.14                up  1.00000
>>  1.00000
>>  15  2.00000             osd.15                up  1.00000
>>  1.00000
>>  16  2.00000             osd.16                up  1.00000
>>  1.00000
>>  17  2.00000             osd.17                up  1.00000
>>  1.00000
>>  18  2.00000             osd.18                up  1.00000
>>  1.00000
>>  19  2.00000             osd.19                up  1.00000
>>  1.00000
>>  -9        0         host or1010051251045
>>
>>  -6 20.00000     rack rack_C1
>>
>>  -7 20.00000         host or1010051251042
>>
>>  20  2.00000             osd.20                up  1.00000
>>  1.00000
>>  21  2.00000             osd.21                up  1.00000
>>  1.00000
>>  22  2.00000             osd.22                up  1.00000
>>  1.00000
>>  23  2.00000             osd.23                up  1.00000
>>  1.00000
>>  24  2.00000             osd.24                up  1.00000
>>  1.00000
>>  25  2.00000             osd.25                up  1.00000
>>  1.00000
>>  26  2.00000             osd.26                up  1.00000
>>  1.00000
>>  27  2.00000             osd.27                up  1.00000
>>  1.00000
>>  28  2.00000             osd.28                up  1.00000
>>  1.00000
>>  29  2.00000             osd.29                up  1.00000
>>  1.00000
>> -10        0         host or1010051251046
>>
>> -11        0         host or1010051251023
>>
>> root@or1010051251044:~#
>>
>>
>>
>>
>>
>> root@or1010051251044:~# ceph -s
>>     cluster 6eacac66-087a-464d-94cb-9ca2585b98d5
>>      health HEALTH_OK
>>      monmap e3: 3 mons at {or1010051251037=10.51.251.37:
>> 6789/0,or1010051251038=10.51.251.38:6789/0,or1010051251039=
>> 10.51.251.39:6789/0}
>>             election epoch 144, quorum 0,1,2 or1010051251037,
>> or1010051251038,or1010051251039
>>      osdmap e1814: 40 osds: 40 up, 40 in
>>       pgmap v446581: 1728 pgs, 4 pools, 7389 GB data, 1847 kobjects
>>             22221 GB used, 57472 GB / 79793 GB avail
>>                 1728 active+clean
>>   client io 61472 kB/s wr, 30 op/s
>> root@or1010051251044:~#
>>
>>
>> Thanks,
>> Pardhiv Karri
>>
>> On Tue, May 22, 2018 at 5:01 AM, David Turner <drakonst...@gmail.com>
>> wrote:
>>
>>> What are your `ceph osd tree` and `ceph status` as well?
>>>
>>> On Tue, May 22, 2018, 3:05 AM Pardhiv Karri <meher4in...@gmail.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> We are using Ceph Hammer 0.94.9. Some of our OSDs never get any data or
>>>> PGs even at their full crush weight, up and running. Rest of the OSDs are
>>>> at 50% full. Is there a bug in Hammer that is causing this issue? Does
>>>> upgrading to Jewel or Luminous fix this issue?
>>>>
>>>> I tried deleting and recreating this OSD N number of times and still
>>>> the same issue. I am seeing this in 3 of our 4 ceph clusters in different
>>>> datacenters. We are using HDD as OSD and SSD as Journal drive.
>>>>
>>>> The below is from our lab and OSD 38 is the one that never fills.
>>>>
>>>>
>>>> ID  WEIGHT   REWEIGHT SIZE   USE    AVAIL  %USE  VAR  TYPE NAME
>>>>
>>>>  -1 80.00000        -      0      0      0     0    0 root default
>>>>
>>>>  -2 40.00000        - 39812G  6190G 33521G 15.55 0.68     rack rack_A1
>>>>
>>>>  -3 20.00000        - 19852G  3718G 16134G 18.73 0.82         host
>>>> or1010051251040
>>>>   0  2.00000  1.00000  1861G   450G  1410G 24.21 1.07             osd.0
>>>>
>>>>   1  2.00000  1.00000  1999G   325G  1673G 16.29 0.72             osd.1
>>>>
>>>>   2  2.00000  1.00000  1999G   336G  1662G 16.85 0.74             osd.2
>>>>
>>>>   3  2.00000  1.00000  1999G   386G  1612G 19.35 0.85             osd.3
>>>>
>>>>   4  2.00000  1.00000  1999G   385G  1613G 19.30 0.85             osd.4
>>>>
>>>>   5  2.00000  1.00000  1999G   364G  1634G 18.21 0.80             osd.5
>>>>
>>>>   6  2.00000  1.00000  1999G   319G  1679G 15.99 0.70             osd.6
>>>>
>>>>   7  2.00000  1.00000  1999G   434G  1564G 21.73 0.96             osd.7
>>>>
>>>>   8  2.00000  1.00000  1999G   352G  1646G 17.63 0.78             osd.8
>>>>
>>>>   9  2.00000  1.00000  1999G   362G  1636G 18.12 0.80             osd.9
>>>>
>>>>  -8 20.00000        - 19959G  2472G 17387G 12.39 0.55         host
>>>> or1010051251044
>>>>  30  2.00000  1.00000  1999G   362G  1636G 18.14 0.80
>>>> osd.30
>>>>  31  2.00000  1.00000  1999G   293G  1705G 14.66 0.65
>>>> osd.31
>>>>  32  2.00000  1.00000  1999G   202G  1796G 10.12 0.45
>>>> osd.32
>>>>  33  2.00000  1.00000  1999G   215G  1783G 10.76 0.47
>>>> osd.33
>>>>  34  2.00000  1.00000  1999G   192G  1806G  9.61 0.42
>>>> osd.34
>>>>  35  2.00000  1.00000  1999G   337G  1661G 16.90 0.74
>>>> osd.35
>>>>  36  2.00000  1.00000  1999G   206G  1792G 10.35 0.46
>>>> osd.36
>>>>  37  2.00000  1.00000  1999G   266G  1732G 13.33 0.59
>>>> osd.37
>>>>  38  2.00000  1.00000  1999G 55836k  1998G  0.00    0
>>>> osd.38
>>>>  39  2.00000  1.00000  1968G   396G  1472G 20.12 0.89
>>>> osd.39
>>>>  -4 20.00000        -      0      0      0     0    0     rack rack_B1
>>>>
>>>>  -5 20.00000        - 19990G  5978G 14011G 29.91 1.32         host
>>>> or1010051251041
>>>>  10  2.00000  1.00000  1999G   605G  1393G 30.27 1.33
>>>> osd.10
>>>>  11  2.00000  1.00000  1999G   592G  1406G 29.62 1.30
>>>> osd.11
>>>>  12  2.00000  1.00000  1999G   539G  1460G 26.96 1.19
>>>> osd.12
>>>>  13  2.00000  1.00000  1999G   684G  1314G 34.22 1.51
>>>> osd.13
>>>>  14  2.00000  1.00000  1999G   510G  1488G 25.56 1.13
>>>> osd.14
>>>>  15  2.00000  1.00000  1999G   590G  1408G 29.52 1.30
>>>> osd.15
>>>>  16  2.00000  1.00000  1999G   595G  1403G 29.80 1.31
>>>> osd.16
>>>>  17  2.00000  1.00000  1999G   652G  1346G 32.64 1.44
>>>> osd.17
>>>>  18  2.00000  1.00000  1999G   544G  1454G 27.23 1.20
>>>> osd.18
>>>>  19  2.00000  1.00000  1999G   665G  1333G 33.27 1.46
>>>> osd.19
>>>>  -9        0        -      0      0      0     0    0         host
>>>> or1010051251045
>>>>  -6 20.00000        -      0      0      0     0    0     rack rack_C1
>>>>
>>>>  -7 20.00000        - 19990G  5956G 14033G 29.80 1.31         host
>>>> or1010051251042
>>>>  20  2.00000  1.00000  1999G   701G  1297G 35.11 1.55
>>>> osd.20
>>>>  21  2.00000  1.00000  1999G   573G  1425G 28.70 1.26
>>>> osd.21
>>>>  22  2.00000  1.00000  1999G   652G  1346G 32.64 1.44
>>>> osd.22
>>>>  23  2.00000  1.00000  1999G   612G  1386G 30.62 1.35
>>>> osd.23
>>>>  24  2.00000  1.00000  1999G   614G  1384G 30.74 1.35
>>>> osd.24
>>>>  25  2.00000  1.00000  1999G   561G  1437G 28.11 1.24
>>>> osd.25
>>>>  26  2.00000  1.00000  1999G   558G  1440G 27.93 1.23
>>>> osd.26
>>>>  27  2.00000  1.00000  1999G   610G  1388G 30.52 1.34
>>>> osd.27
>>>>  28  2.00000  1.00000  1999G   515G  1483G 25.81 1.14
>>>> osd.28
>>>>  29  2.00000  1.00000  1999G   555G  1443G 27.78 1.22
>>>> osd.29
>>>> -10        0        -      0      0      0     0    0         host
>>>> or1010051251046
>>>> -11        0        -      0      0      0     0    0         host
>>>> or1010051251023
>>>>                 TOTAL 79793G 18126G 61566G 22.72
>>>>
>>>> MIN/MAX VAR: 0/1.55  STDDEV: 8.26
>>>>
>>>>
>>>> Thanks
>>>> Pardhiv karri
>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>
>>
>>
>> --
>> *Pardhiv Karri*
>> "Rise and Rise again until LAMBS become LIONS"
>>
>>
>>


-- 
*Pardhiv Karri*
"Rise and Rise again until LAMBS become LIONS"
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to