In the same pool with compression enabled, I have a 1TB RBD filled with a
10GB /dev/urandom file repeating through the entire RBD.  Deleting both of
these RBDs didn't change the number of  bluestore_extent_compress.  I'm
also pretty certain that's the same number I saw there before starting
these tests, but I have no proof.  There are 6 nodes with 5 OSDs in each.
5 of the nodes have bluestore OSDs and 1 has filestore (a part of the test
was to see the behavior of a pool with compression enabled with filestore
and bluestore OSDs in it.

On Tue, Jun 26, 2018 at 10:07 AM Blair Bethwaite <blair.bethwa...@gmail.com>
wrote:

> Hi,
>
> Zeros are not a great choice of data for testing a storage system unless
> you are specifically testing what it does with zeros. Ceph knows that other
> higher layers in the storage stack use zero-fill for certain things and
> will probably optimise for it. E.g., it's common for thin-provisioning
> systems to not actually store runs of 0's (they'll still return them when
> read). I don't know what Ceph does in this specific case, but I could
> imagine it allocating relevant space within the RBD object-map, but never
> bothering to send anything to the actual objects.
>
> Also, note that the description of bluestore_extent_compress is "Sum for
> extents that have been removed due to compression", so sounds to me like
> something is working. Perhaps you can provide for detail of the overall
> stats, cluster config, and whether anything else is stored in it?
>
> On Tue, 26 Jun 2018 at 23:53, David Turner <drakonst...@gmail.com> wrote:
>
>> ceph daemon osd.1 perf dump | grep bluestore | grep compress
>>         "bluestore_compressed": 0,
>>         "bluestore_compressed_allocated": 0,
>>         "bluestore_compressed_original": 0,
>>         "bluestore_extent_compress": 35372,
>>
>> I filled up an RBD in a compressed pool (aggressive) in my test cluster
>> with 1TB of zeros.  The bluestore OSDs in the cluster all show similar to
>> this. Am I missing something here?  Is there any other method to
>> determining compression ratios?
>>
>> On Tue, Dec 5, 2017 at 1:25 AM Rafał Wądołowski <
>> rwadolow...@cloudferro.com> wrote:
>>
>>> Finally, I've founded the command:
>>>
>>>
>>> ceph daemon osd.1 perf dump | grep bluestore
>>>
>>>
>>> And there you have compressed data
>>> Regards,
>>>
>>> Rafał Wądołowski
>>>
>>>
>>> http://cloudferro.com/
>>> On 04.12.2017 14:17, Rafał Wądołowski wrote:
>>>
>>> Hi,
>>>
>>> Is there any command or tool to show effectiveness of bluestore
>>> compression?
>>>
>>> I see the difference (in ceph osd df tree), while uploading a object to
>>> ceph, but maybe there are more friendly method to do it.
>>>
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
> --
> Cheers,
> ~Blairo
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to