Thanks Igor, Here is a link to the ceph perf data on several osds. 

https://paste.ee/p/IzDMy 

In terms of the object sizes. We use rgw to backup the data from various 
workstations and servers. So, the sizes would be from a few kb to a few gig per 
individual file. 

Cheers 

> From: "Igor Fedotov" <ifedo...@suse.de>
> To: "andrei" <and...@arhont.com>
> Cc: "ceph-users" <ceph-users@lists.ceph.com>
> Sent: Wednesday, 3 July, 2019 12:29:33
> Subject: Re: [ceph-users] troubleshooting space usage

> Hi Andrei,

> Additionally I'd like to see performance counters dump for a couple of HDD 
> OSDs
> (obtained through 'ceph daemon osd.N perf dump' command).

> W.r.t average object size - I was thinking that you might know what objects 
> had
> been uploaded... If not then you might want to estimate it by using "rados 
> get"
> command on the pool: retrieve some random object set and check their sizes. 
> But
> let's check performance counters first - most probably they will show loses
> caused by allocation.

> Also I've just found similar issue (still unresolved) in our internal tracker 
> -
> but its root cause is definitely different from allocation overhead. Looks 
> like
> some orphaned objects in the pool. Could you please compare and share the
> amounts of objects in the pool reported by "ceph (or rados) df detail" and
> radosgw tools?

> Thanks,

> Igor

> On 7/3/2019 12:56 PM, Andrei Mikhailovsky wrote:

>> Hi Igor,

>> Many thanks for your reply. Here are the details about the cluster:

>> 1. Ceph version - 13.2.5-1xenial (installed from Ceph repository for ubuntu
>> 16.04)

>> 2. main devices for radosgw pool - hdd. we do use a few ssds for the other 
>> pool,
>> but it is not used by radosgw

>> 3. we use BlueStore

>> 4. Average rgw object size - I have no idea how to check that. Couldn't find 
>> a
>> simple answer from google either. Could you please let me know how to check
>> that?

>> 5. Ceph osd df tree:

>> 6. Other useful info on the cluster:

>> # ceph osd df tree
>> ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME

>> -1 112.17979 - 113 TiB 90 TiB 23 TiB 79.25 1.00 - root uk
>> -5 112.17979 - 113 TiB 90 TiB 23 TiB 79.25 1.00 - datacenter ldex
>> -11 112.17979 - 113 TiB 90 TiB 23 TiB 79.25 1.00 - room ldex-dc3
>> -13 112.17979 - 113 TiB 90 TiB 23 TiB 79.25 1.00 - row row-a
>> -4 112.17979 - 113 TiB 90 TiB 23 TiB 79.25 1.00 - rack ldex-rack-a5
>> -2 28.04495 - 28 TiB 22 TiB 6.2 TiB 77.96 0.98 - host arh-ibstorage1-ib

>> 0 hdd 2.73000 0.79999 2.8 TiB 2.3 TiB 519 GiB 81.61 1.03 145 osd.0
>> 1 hdd 2.73000 1.00000 2.8 TiB 1.9 TiB 847 GiB 70.00 0.88 130 osd.1
>> 2 hdd 2.73000 1.00000 2.8 TiB 2.2 TiB 561 GiB 80.12 1.01 152 osd.2
>> 3 hdd 2.73000 1.00000 2.8 TiB 2.3 TiB 469 GiB 83.41 1.05 160 osd.3
>> 4 hdd 2.73000 1.00000 2.8 TiB 1.8 TiB 983 GiB 65.18 0.82 141 osd.4
>> 32 hdd 5.45999 1.00000 5.5 TiB 4.4 TiB 1.1 TiB 80.68 1.02 306 osd.32
>> 35 hdd 2.73000 1.00000 2.8 TiB 1.7 TiB 1.0 TiB 62.89 0.79 126 osd.35
>> 36 hdd 2.73000 1.00000 2.8 TiB 2.3 TiB 464 GiB 83.58 1.05 175 osd.36
>> 37 hdd 2.73000 0.89999 2.8 TiB 2.5 TiB 301 GiB 89.34 1.13 160 osd.37
>> 5 ssd 0.74500 1.00000 745 GiB 642 GiB 103 GiB 86.15 1.09 65 osd.5

>> -3 28.04495 - 28 TiB 24 TiB 4.5 TiB 84.03 1.06 - host arh-ibstorage2-ib
>> 9 hdd 2.73000 0.95000 2.8 TiB 2.4 TiB 405 GiB 85.65 1.08 158 osd.9
>> 10 hdd 2.73000 0.89999 2.8 TiB 2.4 TiB 352 GiB 87.52 1.10 169 osd.10
>> 11 hdd 2.73000 1.00000 2.8 TiB 2.0 TiB 783 GiB 72.28 0.91 160 osd.11
>> 12 hdd 2.73000 0.84999 2.8 TiB 2.4 TiB 359 GiB 87.27 1.10 153 osd.12
>> 13 hdd 2.73000 1.00000 2.8 TiB 2.4 TiB 348 GiB 87.69 1.11 169 osd.13
>> 14 hdd 2.73000 1.00000 2.8 TiB 2.5 TiB 283 GiB 89.97 1.14 170 osd.14
>> 15 hdd 2.73000 1.00000 2.8 TiB 2.2 TiB 560 GiB 80.18 1.01 155 osd.15
>> 16 hdd 2.73000 0.95000 2.8 TiB 2.4 TiB 332 GiB 88.26 1.11 178 osd.16
>> 26 hdd 5.45999 1.00000 5.5 TiB 4.4 TiB 1.0 TiB 81.04 1.02 324 osd.26
>> 7 ssd 0.74500 1.00000 745 GiB 607 GiB 138 GiB 81.48 1.03 62 osd.7

>> -15 28.04495 - 28 TiB 22 TiB 6.4 TiB 77.40 0.98 - host arh-ibstorage3-ib
>> 18 hdd 2.73000 0.95000 2.8 TiB 2.5 TiB 312 GiB 88.96 1.12 156 osd.18
>> 19 hdd 2.73000 1.00000 2.8 TiB 2.0 TiB 771 GiB 72.68 0.92 162 osd.19
>> 20 hdd 2.73000 1.00000 2.8 TiB 2.0 TiB 733 GiB 74.04 0.93 149 osd.20
>> 21 hdd 2.73000 1.00000 2.8 TiB 2.2 TiB 533 GiB 81.12 1.02 155 osd.21
>> 22 hdd 2.73000 1.00000 2.8 TiB 2.1 TiB 692 GiB 75.48 0.95 144 osd.22
>> 23 hdd 2.73000 1.00000 2.8 TiB 1.6 TiB 1.1 TiB 58.43 0.74 130 osd.23
>> 24 hdd 2.73000 1.00000 2.8 TiB 2.2 TiB 579 GiB 79.51 1.00 146 osd.24
>> 25 hdd 2.73000 1.00000 2.8 TiB 1.9 TiB 886 GiB 68.63 0.87 147 osd.25
>> 31 hdd 5.45999 1.00000 5.5 TiB 4.7 TiB 758 GiB 86.50 1.09 326 osd.31
>> 6 ssd 0.74500 0.89999 744 GiB 640 GiB 104 GiB 86.01 1.09 61 osd.6

>> -17 28.04494 - 28 TiB 22 TiB 6.3 TiB 77.61 0.98 - host arh-ibstorage4-ib
>> 8 hdd 2.73000 1.00000 2.8 TiB 1.9 TiB 909 GiB 67.80 0.86 141 osd.8
>> 17 hdd 2.73000 1.00000 2.8 TiB 1.9 TiB 904 GiB 67.99 0.86 144 osd.17
>> 27 hdd 2.73000 1.00000 2.8 TiB 2.1 TiB 654 GiB 76.84 0.97 152 osd.27
>> 28 hdd 2.73000 1.00000 2.8 TiB 2.3 TiB 481 GiB 82.98 1.05 153 osd.28
>> 29 hdd 2.73000 1.00000 2.8 TiB 1.9 TiB 829 GiB 70.65 0.89 137 osd.29
>> 30 hdd 2.73000 1.00000 2.8 TiB 2.0 TiB 762 GiB 73.03 0.92 142 osd.30
>> 33 hdd 2.73000 1.00000 2.8 TiB 2.3 TiB 501 GiB 82.25 1.04 166 osd.33
>> 34 hdd 5.45998 1.00000 5.5 TiB 4.5 TiB 968 GiB 82.77 1.04 325 osd.34
>> 39 hdd 2.73000 0.95000 2.8 TiB 2.4 TiB 402 GiB 85.77 1.08 162 osd.39
>> 38 ssd 0.74500 1.00000 745 GiB 671 GiB 74 GiB 90.02 1.14 68 osd.38
>> TOTAL 113 TiB 90 TiB 23 TiB 79.25
>> MIN/MAX VAR: 0.74/1.14 STDDEV: 8.14

>> # for i in $(radosgw-admin bucket list | jq -r '.[]'); do radosgw-admin 
>> bucket
>> stats --bucket=$i | jq '.usage | ."rgw.main" | .size_kb' ; done | awk '{ SUM 
>> +=
>> $1} END { print SUM/1024/1024/1024 }'
>> 6.59098

>> # ceph df

>> GLOBAL:
>> SIZE AVAIL RAW USED %RAW USED
>> 113 TiB 23 TiB 90 TiB 79.25

>> POOLS:
>> NAME ID USED %USED MAX AVAIL OBJECTS
>> Primary-ubuntu-1 5 27 TiB 87.56 3.9 TiB 7302534
>> .users.uid 15 6.8 KiB 0 3.9 TiB 39
>> .users 16 335 B 0 3.9 TiB 20
>> .users.swift 17 14 B 0 3.9 TiB 1
>> .rgw.buckets 19 15 TiB 79.88 3.9 TiB 8787763
>> .users.email 22 0 B 0 3.9 TiB 0
>> .log 24 109 MiB 0 3.9 TiB 102301
>> .rgw.buckets.extra 37 0 B 0 2.6 TiB 0
>> .rgw.root 44 2.9 KiB 0 2.6 TiB 16
>> .rgw.meta 45 1.7 MiB 0 2.6 TiB 6249
>> .rgw.control 46 0 B 0 2.6 TiB 8
>> .rgw.gc 47 0 B 0 2.6 TiB 32
>> .usage 52 0 B 0 2.6 TiB 0
>> .intent-log 53 0 B 0 2.6 TiB 0
>> default.rgw.buckets.non-ec 54 0 B 0 2.6 TiB 0
>> .rgw.buckets.index 55 0 B 0 2.6 TiB 11485
>> .rgw 56 491 KiB 0 2.6 TiB 1686
>> Primary-ubuntu-1-ssd 57 1.2 TiB 92.39 105 GiB 379516

>> I am not too sure if the issue relates to the BlueStore overhead as I would
>> probably have seen the discrepancy in my Primary-ubuntu-1 pool as well.
>> However, the data usage on Primary-ubuntu-1 pool seems to be consistent with 
>> my
>> expectations (precise numbers to be verified soon). The issues seems to be 
>> only
>> with the .rgw-buckets pool where the "ceph df " output shows 15TB of usage 
>> and
>> the sum of all buckets in that pool shows just over 6.5TB.

>> Cheers

>> Andrei

>>> From: "Igor Fedotov" [ mailto:ifedo...@suse.de | <ifedo...@suse.de> ]
>>> To: "andrei" [ mailto:and...@arhont.com | <and...@arhont.com> ] , 
>>> "ceph-users" [
>>> mailto:ceph-users@lists.ceph.com | <ceph-users@lists.ceph.com> ]
>>> Sent: Tuesday, 2 July, 2019 10:58:54
>>> Subject: Re: [ceph-users] troubleshooting space usage

>>> Hi Andrei,

>>> The most obvious reason is space usage overhead caused by BlueStore 
>>> allocation
>>> granularity, e.g. if bluestore_min_alloc_size is 64K and average object 
>>> size is
>>> 16K one will waste 48K per object in average. This is rather a speculation 
>>> so
>>> far as we lack key the information about your cluster:

>>> - Ceph version

>>> - What are the main devices for OSD: hdd or ssd.

>>> - BlueStore or FileStore.

>>> - average RGW object size.

>>> You might also want to collect and share performance counter dumps (ceph 
>>> daemon
>>> osd.N perf dump) and "
>>> " reports from a couple of your OSDs.

>>> Thanks,

>>> Igor

>>> On 7/2/2019 11:43 AM, Andrei Mikhailovsky wrote:

>>>> Bump!

>>>>> From: "Andrei Mikhailovsky" [ mailto:and...@arhont.com | 
>>>>> <and...@arhont.com> ]
>>>>> To: "ceph-users" [ mailto:ceph-users@lists.ceph.com |
>>>>> <ceph-users@lists.ceph.com> ]
>>>>> Sent: Friday, 28 June, 2019 14:54:53
>>>>> Subject: [ceph-users] troubleshooting space usage

>>>>> Hi

>>>>> Could someone please explain / show how to troubleshoot the space usage 
>>>>> in Ceph
>>>>> and how to reclaim the unused space?

>>>>> I have a small cluster with 40 osds, replica of 2, mainly used as a 
>>>>> backend for
>>>>> cloud stack as well as the S3 gateway. The used space doesn't make any 
>>>>> sense to
>>>>> me, especially the rgw pool, so I am seeking help.

>>>>> Here is what I found from the client:

>>>>> Ceph -s shows the

>>>>> usage: 89 TiB used, 24 TiB / 113 TiB avail

>>>>> Ceph df shows:

>>>>> Primary-ubuntu-1 5 27 TiB 90.11 3.0 TiB 7201098
>>>>> Primary-ubuntu-1-ssd 57 1.2 TiB 89.62 143 GiB 359260
>>>>> .rgw.buckets 19 15 TiB 83.73 3.0 TiB 8742222

>>>>> the usage of the Primary-ubuntu-1 and Primary-ubuntu-1-ssd is in line 
>>>>> with my
>>>>> expectations. However, the .rgw.buckets pool seems to be using way too 
>>>>> much.
>>>>> The usage of all rgw buckets shows 6.5TB usage (looking at the size_kb 
>>>>> values
>>>>> from the "radosgw-admin bucket stats"). I am trying to figure out why
>>>>> .rgw.buckets is using 15TB of space instead of the 6.5TB as shown from the
>>>>> bucket usage.

>>>>> Thanks

>>>>> Andrei

>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> [ mailto:ceph-users@lists.ceph.com | ceph-users@lists.ceph.com ]
>>>>> [ http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com |
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ]

>>>> _______________________________________________
>>>> ceph-users mailing list [ mailto:ceph-users@lists.ceph.com |
>>>> ceph-users@lists.ceph.com ] [
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com |
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ]
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to