Is anyone aware of a way to either reconcile or remove possible orphaned
"shadow" files in a federated gateway configuration?  The issue we're seeing
is the number of chunks/shadow files on the slave has many more "shadow"
files than the master, the breakdown is as follows:

master zone:

.region-1.zone-1.rgw.buckets = 1737 "shadow" files of which there are 10
distinct sets of tags, an example of 1 distinct set is:

alph-1.80907.1__shadow_.VTZYW5ubV53wCHAKcnGwrD_yGkyGDuG_1 through
alph-1.80907.1__shadow_.VTZYW5ubV53wCHAKcnGwrD_yGkyGDuG_516

 

slave zone:

.region-1.zone-2.rgw.buckets = 331961 "shadow" files, of which there are 652
distinct sets of  tags, examples:

1 set having 516 "shadow" files:

alph-1.80907.1__shadow_.yPT037fjWhTi_UtHWSYPcRWBanaN9Oy_1 through
alph-1.80907.1__shadow_.yPT037fjWhTi_UtHWSYPcRWBanaN9Oy_516

 

236 sets having 515 "shadow" files apiece:

alph-1.80907.1__shadow_.RA9KCc_U5T9kBN_ggCUx8VLJk36RSiw_1 through
alph-1.80907.1__shadow_.RA9KCc_U5T9kBN_ggCUx8VLJk36RSiw_515

alph-1.80907.1__shadow_.aUWuanLbJD5vbBSD90NWwjkuCxQmvbQ_1 through
alph-1.80907.1__shadow_.aUWuanLbJD5vbBSD90NWwjkuCxQmvbQ_515

..

 

The number of shadow files in zone-2 is taking quite a bit of space from the
OSD's in the cluster.   Without being able to trace back to the original
file name from an s3 or rados tag, I have no way of knowing which files
these are.  Is it possible that the same file may have been replicated
multiple times, due to network or connectivity issues?

 

I can provide any logs or other information that may provide some help,
however at this point we're not seeing any real errors.

 

Thanks in advance for any help that can be provided,

MLM

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to