Hi Anthony,
here are the config values we've set or with their defaults. There is
no rgw_keystone_token_cache_ttl (neither in the documentation, nor can I
set it via ceph config set client.rgw rgw_keystone_token_cache_ttl 3600):
~# ceph config show-with-defaults rgw.rgw1 | grep rgw_keystone | column -t
rgw_keystone_accepted_admin_roles default
rgw_keystone_accepted_roles objectstore_operator
mon
rgw_keystone_admin_domain default
mon
rgw_keystone_admin_password yyyyyyyy
mon
rgw_keystone_admin_password_path default
rgw_keystone_admin_project services
mon
rgw_keystone_admin_tenant default
rgw_keystone_admin_token default
rgw_keystone_admin_token_path default
rgw_keystone_admin_user xxxxxxx
mon
rgw_keystone_api_version 3
mon
rgw_keystone_barbican_domain default
rgw_keystone_barbican_password default
rgw_keystone_barbican_project default
rgw_keystone_barbican_tenant default
rgw_keystone_barbican_user default
rgw_keystone_expired_token_cache_expiration 3600
default
rgw_keystone_implicit_tenants false
default
rgw_keystone_service_token_accepted_roles admin
default
rgw_keystone_service_token_enabled false
default
rgw_keystone_token_cache_size 100000
mon <-- i've set this to test if this solves the problem, but
this is the default value
rgw_keystone_url https://auth.tld
mon
rgw_keystone_verify_ssl true
default
Am Mo., 3. Nov. 2025 um 16:40 Uhr schrieb Anthony D'Atri <
[email protected]>:
> Check the values of rgw_keystone_token_cache_size and
> rgw_keystone_token_cache_ttl and other rgw_keystone options.
>
> I've seen at least one deployment tool that disabled Keystone caching for
> dev purposes, but leaked that into the release code, which deployed RGW
> with Rook with a configmap override.
>
>
> > On Nov 3, 2025, at 9:52 AM, Boris <[email protected]> wrote:
> >
> > Hi,
> > I am currently debugging a problem that the radosgw keystone token cache
> > seems not to work properly. Or at all. I tried to debug it and attached
> the
> > rgw_debug log set to 10. I've truncated to only show the part from "No
> > stored secret string, cache miss" until the request is done.
> >
> > The failed request hits a rate limit on the keystone which currently
> takes
> > around 2k answered requests per minute.
> > Any ideas what I did wrong?
> >
> > * All requests were done within 10 seconds and were only an ls to show
> > buckets.
> > * This particular RGW only took my requests during testing.
> > * We didn't set any timeouts or special cache configs in ceph
> > * system time is correct
> >
> >
> > First request worked instantly:
> >
> > req 8122732607072897744 0.106001295s s3:list_buckets No stored secret
> > string, cache miss
> > [4.0K blob data]
> > req 8122732607072897744 0.315003842s s3:list_buckets s3 keystone:
> validated
> > token: 8144848695793469:user-9XGYcbFNUVTQ expires: 1762266594
> > req 8122732607072897744 0.315003842s s3:list_buckets cache get:
> >
> name=eu-central-lz.rgw.meta+users.uid+a13f0472be744104ad1f64bb2855cdee$a13f0472be744104ad1f64bb2855cdee
> > : hit (negative entry)
> > req 8122732607072897744 0.315003842s s3:list_buckets cache get:
> > name=eu-central-lz.rgw.meta+users.uid+a13f0472be744104ad1f64bb2855cdee :
> > hit (requested=0x13, cached=0x13)
> > req 8122732607072897744 0.315003842s s3:list_buckets normalizing buckets
> > and tenants
> > req 8122732607072897744 0.315003842s s->object=<NULL> s->bucket=
> > req 8122732607072897744 0.315003842s s3:list_buckets init permissions
> > req 8122732607072897744 0.315003842s s3:list_buckets cache get:
> > name=eu-central-lz.rgw.meta+users.uid+a13f0472be744104ad1f64bb2855cdee :
> > hit (requested=0x13, cached=0x13)
> > req 8122732607072897744 0.315003842s s3:list_buckets recalculating target
> > req 8122732607072897744 0.315003842s s3:list_buckets reading permissions
> > req 8122732607072897744 0.315003842s s3:list_buckets init op
> > req 8122732607072897744 0.315003842s s3:list_buckets verifying op mask
> > req 8122732607072897744 0.315003842s s3:list_buckets verifying op
> > permissions
> > req 8122732607072897744 0.315003842s s3:list_buckets verifying op params
> > req 8122732607072897744 0.315003842s s3:list_buckets pre-executing
> > req 8122732607072897744 0.315003842s s3:list_buckets check rate limiting
> > req 8122732607072897744 0.315003842s s3:list_buckets executing
> > req 8122732607072897744 0.315003842s s3:list_buckets completing
> > req 8122732607072897744 0.315003842s cache get:
> > name=eu-central-lz.rgw.log++script.postrequest. : hit (negative entry)
> > req 8122732607072897744 0.315003842s s3:list_buckets op status=0
> > req 8122732607072897744 0.315003842s s3:list_buckets http status=200
> > ====== req done req=0x74659e51b6f0 op status=0 http_status=200
> > latency=0.315003842s ======
> >
> > 2nd request failed
> >
> > req 10422983006485317789 0.061000749s s3:list_buckets cache get:
> > name=eu-central-lz.rgw.meta+users.keys+05917cf2ee9d4fdea8baf6a3348ca33a :
> > hit (negative entry)
> > req 10422983006485317789 0.061000749s s3:list_buckets error reading user
> > info, uid=05917cf2ee9d4fdea8baf6a3348ca33a can't authenticate
> > req 10422983006485317789 0.061000749s s3:list_buckets Failed the auth
> > strategy, reason=-5
> > failed to authorize request
> > WARNING: set_req_state_err err_no=5 resorting to 500
> > req 10422983006485317789 0.061000749s cache get:
> > name=eu-central-lz.rgw.log++script.postrequest. : hit (negative entry)
> > req 10422983006485317789 0.061000749s s3:list_buckets op status=0
> > req 10422983006485317789 0.061000749s s3:list_buckets http status=500
> > ====== req done req=0x74659e51b6f0 op status=0 http_status=500
> > latency=0.061000749s ======
> >
> > 3rd requests went through again
> >
> > req 13123970335019889535 0.000000000s s3:list_buckets No stored secret
> > string, cache miss
> > [250B blob data]
> > req 13123970335019889535 0.204002500s s3:list_buckets s3 keystone:
> > validated token: 8144848695793469:user-9XGYcbFNUVTQ expires: 1762266602
> > req 13123970335019889535 0.204002500s s3:list_buckets cache get:
> >
> name=eu-central-lz.rgw.meta+users.uid+a13f0472be744104ad1f64bb2855cdee$a13f0472be744104ad1f64bb2855cdee
> > : hit (negative entry)
> > req 13123970335019889535 0.204002500s s3:list_buckets cache get:
> > name=eu-central-lz.rgw.meta+users.uid+a13f0472be744104ad1f64bb2855cdee :
> > hit (requested=0x13, cached=0x13)
> > req 13123970335019889535 0.204002500s s3:list_buckets normalizing buckets
> > and tenants
> > req 13123970335019889535 0.204002500s s->object=<NULL> s->bucket=
> > req 13123970335019889535 0.204002500s s3:list_buckets init permissions
> > req 13123970335019889535 0.204002500s s3:list_buckets cache get:
> > name=eu-central-lz.rgw.meta+users.uid+a13f0472be744104ad1f64bb2855cdee :
> > hit (requested=0x13, cached=0x13)
> > req 13123970335019889535 0.204002500s s3:list_buckets recalculating
> target
> > req 13123970335019889535 0.204002500s s3:list_buckets reading permissions
> > req 13123970335019889535 0.204002500s s3:list_buckets init op
> > req 13123970335019889535 0.204002500s s3:list_buckets verifying op mask
> > req 13123970335019889535 0.204002500s s3:list_buckets verifying op
> > permissions
> > req 13123970335019889535 0.204002500s s3:list_buckets verifying op params
> > req 13123970335019889535 0.204002500s s3:list_buckets pre-executing
> > req 13123970335019889535 0.204002500s s3:list_buckets check rate limiting
> > req 13123970335019889535 0.204002500s s3:list_buckets executing
> > req 13123970335019889535 0.204002500s s3:list_buckets completing
> > req 13123970335019889535 0.204002500s cache get:
> > name=eu-central-lz.rgw.log++script.postrequest. : hit (negative entry)
> > req 13123970335019889535 0.204002500s s3:list_buckets op status=0
> > req 13123970335019889535 0.204002500s s3:list_buckets http status=200
> > ====== req done req=0x74659e51b6f0 op status=0 http_status=200
> > latency=0.204002500s ======
> >
> >
> >
> >
> >
> > --
> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
> > groüen Saal.
> > _______________________________________________
> > ceph-users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
>
>
--
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]