Hi,

You have 2 ways:
First is using s3vrowser app and in the menu select the multipart uploads and 
clean it up.
The other is like this:

Set lifecycle policy
        On the client:
                vim lifecyclepolicy
                <?xml version="1.0" ?>
                <LifecycleConfiguration 
xmlns="http://s3.amazonaws.com/doc/2006-03-01/";>
                                <Rule>
                                                <ID>Incomplete Multipart 
Uploads</ID>
                                                <Prefix/>
                                                <Status>Enabled</Status>
                                                <AbortIncompleteMultipartUpload>
                                                                
<DaysAfterInitiation>1</DaysAfterInitiation>
                                                
</AbortIncompleteMultipartUpload>
                                </Rule>
                </LifecycleConfiguration>

                /bin/s3cmd setlifecycle lifecyclepolicy  s3://bucketname
        On mon node process manually
                radosgw-admin lc list
                radosgw-admin lc process

Istvan Szabo
Senior Infrastructure Engineer
---------------------------------------------------
Agoda Services Co., Ltd.
e: istvan.sz...@agoda.com
---------------------------------------------------

-----Original Message-----
From: Boris Behrens <b...@kervyn.de>
Sent: Monday, April 19, 2021 4:10 PM
To: ceph-users@ceph.io
Subject: [Suspicious newsletter] [ceph-users] cleanup multipart in radosgw

Hi,
is there a way to remove multipart uploads that are older than X days?

It doesn't need to be build into ceph or is automated to the end. Just 
something I don't need to build on my own.

I currently try to debug a problem where ceph reports a lot more used space 
than it actually requires ( 
https://www.mail-archive.com/ceph-users@ceph.io/msg09810.html).

I came across a lot of old _multipart_ files in some buckets and now I want to 
clean them up.
I don't know if this will fix my problem but I would love to rule that out.

radosgw-admin bucket check --bucket=bucket --check-objects --fix does not work 
because it is a shareded bucket.

I have also some buckets that look like this, and contain 100% _multipart_ 
files which are >2 years old:
        "buckets": [
            {
                "bucket": "ncprod",
                "tenant": "",
                "num_objects": -482,
                "num_shards": 0,
                "objects_per_shard": -482,
                "fill_status": "OVER 180143985094819%"
            }
        ]

--
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im 
groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to 
ceph-users-le...@ceph.io

________________________________
This message is confidential and is for the sole use of the intended 
recipient(s). It may also be privileged or otherwise protected by copyright or 
other legal rules. If you have received it by mistake please let us know by 
reply email and delete it from your system. It is prohibited to copy this 
message or disclose its content to anyone. Any confidentiality or privilege is 
not waived or lost by any mistaken delivery or unauthorized disclosure of the 
message. All messages sent to and from Agoda may be monitored to ensure 
compliance with company policies, to protect the company's interests and to 
remove potential malware. Electronic messages may be intercepted, amended, lost 
or deleted, or contain viruses.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to