Hi Eric,

> 1. I recommend that you *not* issue another bucket reshard until you figure 
> out what’s going on.

Thanks, noted!

> 2. Which version of Ceph are you using?
17.2.5
I wanted to get the Cluster to Health OK before upgrading. I didn't
see anything that led me to believe that an upgrade could fix the
reshard issue.

> 3. Can you issue a `radosgw-admin metadata get bucket:<bucket-name>` so we 
> can verify what the current marker is?

# radosgw-admin metadata get bucket:sql20
{
    "key": "bucket:sql20",
    "ver": {
        "tag": "_hGhtgzjcWY9rO9JP7YlWzt8",
        "ver": 3
    },
    "mtime": "2023-07-12T15:56:55.226784Z",
    "data": {
        "bucket": {
            "name": "sql20",
            "marker": "3caabb9a-4e3b-4b8a-8222-34c33dd63210.10610190.9",
            "bucket_id": "3caabb9a-4e3b-4b8a-8222-34c33dd63210.10648356.1",
            "tenant": "",
            "explicit_placement": {
                "data_pool": "",
                "data_extra_pool": "",
                "index_pool": ""
            }
        },
        "owner": "S3user",
        "creation_time": "2023-04-26T09:22:01.681646Z",
        "linked": "true",
        "has_bucket_info": "false"
    }
}

> 4. After you resharded previously, did you get command-line output along the 
> lines of:
> 2023-07-24T13:33:50.867-0400 7f10359f2a80 1 execute INFO: reshard of bucket 
> “<bucket-name>" completed successfully

I think so, at least for the second reshard. But I wouldn't bet my
life on it. I fear I might have missed an error on the first one since
I have done a radosgw-admin bucket reshard so often and never seen it
fail.

Christian
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to