[ceph-users] Re: "ceph orch restart mgr" creates manager daemon restart loop

2021-11-23 Thread Adam King
Hi Roman, what ceph version are you on? Also, when you ran the
restart command originally, did you get a message about scheduling the
restarts or no output?



On Tue, Nov 23, 2021 at 6:04 AM Roman Steinhart  wrote:

> Hi all,
>
> while digging down another issue I had with the managers I restarted them
> using "ceph orch restart mgr".
> After running that command the main manager is now in a restart loop.
> The only way for me to stop this is running "ceph orch pause", as soon as I
> do "ceph orch resume" the loop starts again.
>
> Has anyone a suggestion on how I can remove that stuck "restart" job?
> "ceph orch cancel" does not work, it returns "This Orchestrator does not
> support `orch cancel`"
>
> While googling for this issue it seems I'm not the first one having that
> issue:
>
> https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/IRC4ZPSSZYELWPU5D2FHKWJ2VU7IP3JG/
>
> https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/WZUWZGVYPFII5HTUTW6OAQXIZ6VNT2E2/
>
> See the logs below:
> user@ceph1:~# journalctl -fu
> ceph-6d588189-f434-4cb1-8c60-6e48cbf43a2a@mgr.ceph1.service --since "1 day
> ago" -g "Ceph mgr.ceph1 for 6d588189-f434-4cb1-8c60-6e48cbf43a2a"
> Nov 22 15:31:38 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 15:31:51 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:31:51 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:34:49 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 15:35:00 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:35:00 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:38:39 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 15:38:50 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:38:50 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:43:35 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 15:43:46 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:43:46 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:50:19 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 15:50:30 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:50:30 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:58:26 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 15:58:37 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 15:58:37 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:10:16 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 16:10:28 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:10:28 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:17:17 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 16:17:29 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:17:29 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:40:52 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 16:41:07 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:41:07 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:45:34 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 16:45:47 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:45:47 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:49:34 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 16:49:39 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:49:39 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:51:47 ceph1 systemd[1]: Stopping Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a...
> Nov 22 16:51:51 ceph1 systemd[1]: Stopped Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:51:51 ceph1 systemd[1]: Started Ceph mgr.ceph1 for
> 6d588189-f434-4cb1-8c60-6e48cbf43a2a.
> Nov 22 16:54:02 ceph1 systemd[1]: Stopping Ceph mgr.c

[ceph-users] Re: have buckets with low number of shards

2021-11-23 Thread mahnoosh shahidi
Hi Josh

Thanks for your response. Do you have any advice how to reshard these big
buckets so it doesn't cause any down time in our cluster? Resharding these
buckets makes a lots of slow ops in deleting old shard phase and the
cluster can't responde to any requests till resharding is completely done.

Regards,
Mahnoosh

On Tue, Nov 23, 2021, 5:28 PM Josh Baergen 
wrote:

> Hey Mahnoosh,
>
> > Running cluster in octopus 15.2.12 . We have a big bucket with about 800M
> > objects and resharding this bucket makes many slow ops in our bucket
> index
> > osds. I wanna know what happens if I don't reshard this bucket any more?
> > How does it affect the performance? The performance problem would be only
> > for that bucket or it affects the entire bucket index pool?
>
> Unfortunately, if you don't reshard the bucket, it's likely that
> you'll see widespread index pool performance and stability issues,
> generally manifesting as one or more OSDs becoming very busy to the
> point of holding up traffic for multiple buckets or even flapping (the
> OSD briefly gets marked down), leading to recovery. Recovering large
> index shards can itself cause issues like this to occur. Although the
> official recommendation, IIRC, is 100K objects per index shard, the
> exact objects per shard count at which one starts to experience these
> sorts of issues highly depends on the hardware involved and user
> workload.
>
> Josh
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: have buckets with low number of shards

2021-11-23 Thread DHilsbos
Manoosh;

You can't reshard a bucket without downtime.  During a reshard RGW creates new 
RADOS objects to match the new shard number.  Then all the RGW objects are 
moved from the old RADOS objects to the new RADOS objects, and the original 
RADOS objects are destroyed.  The reshard locks the bucket for the duration.

Thank you,

Dominic L. Hilsbos, MBA
Vice President - Information Technology
Perform Air International Inc.
dhils...@performair.com
www.PerformAir.com


-Original Message-
From: mahnoosh shahidi [mailto:mahnooosh@gmail.com] 
Sent: Tuesday, November 23, 2021 8:20 AM
To: Josh Baergen
Cc: Ceph Users
Subject: [ceph-users] Re: have buckets with low number of shards

Hi Josh

Thanks for your response. Do you have any advice how to reshard these big
buckets so it doesn't cause any down time in our cluster? Resharding these
buckets makes a lots of slow ops in deleting old shard phase and the
cluster can't responde to any requests till resharding is completely done.

Regards,
Mahnoosh

On Tue, Nov 23, 2021, 5:28 PM Josh Baergen 
wrote:

> Hey Mahnoosh,
>
> > Running cluster in octopus 15.2.12 . We have a big bucket with about 800M
> > objects and resharding this bucket makes many slow ops in our bucket
> index
> > osds. I wanna know what happens if I don't reshard this bucket any more?
> > How does it affect the performance? The performance problem would be only
> > for that bucket or it affects the entire bucket index pool?
>
> Unfortunately, if you don't reshard the bucket, it's likely that
> you'll see widespread index pool performance and stability issues,
> generally manifesting as one or more OSDs becoming very busy to the
> point of holding up traffic for multiple buckets or even flapping (the
> OSD briefly gets marked down), leading to recovery. Recovering large
> index shards can itself cause issues like this to occur. Although the
> official recommendation, IIRC, is 100K objects per index shard, the
> exact objects per shard count at which one starts to experience these
> sorts of issues highly depends on the hardware involved and user
> workload.
>
> Josh
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: RGW support IAM user authentication

2021-11-23 Thread Michael Breen
Hi Pritha - or anyone who knows,

I too have problems with IAM, in particular with AssumeRoleWithWebIdentity.

I am running the master branch version of Ceph because it looks like it
includes code related to the functionality described at
https://docs.ceph.com/en/latest/radosgw/STS/ - code which is not in any
released version, even 17.0.

Looking at the code on that page, there appear to be at least two errors:
(1) an instance of "client" should be "sts_client" (or vice versa)
(2) an access key and secret key are specified when creating sts_client,
which is unnecessary and therefore confusing: only the access token is used
or should be required for assume_role_with_web_identity

But I still cannot get the AssumeRoleWithWebIdentity code example to work.
The RGW debug logs show

debug 2021-11-23T15:51:22.247+ 7fad6e351700  0 evaluating policy for
op: 93 returned deny/pass

In my case, the policy_document and role_policy are

policy_document =
'''{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":{"Federated":["arn:aws:iam:::oidc-provider/proteus.ves.corp/auth/realms/cno"]},"Action":["sts:AssumeRoleWithWebIdentity"],"Condition":{"StringEquals":{"proteus.ves.corp/auth/realms/cno:app_id":"ceph_rgw"}}}]}'''
role_policy =
'''{"Version":"2012-10-17","Statement":{"Effect":"Allow","Action":"s3:*","Resource":"arn:aws:s3:::*"}}'''

(I assume it is only the former that may be relevant here, but maybe I'm
wrong.)

In /etc/ceph/ceph.conf I have
[client.radosgw.gateway]
rgw sts key = abcdefghijklmnop
rgw s3 auth use sts = true

In the debug I can see the token from Keycloak looks like (after formatting
it)

{
  "exp": 1637677729,
  "iat": 1637677429,
  "jti": "06e5422e-8395-4727-9366-a851c3f5930f",
  "iss": "https://proteus.ves.corp/auth/realms/cno";,
  "aud": "account",
  "sub": "f45bae70-1517-48f6-9d75-af7f421f4a0c",
  "typ": "Bearer",
  "azp": "ceph_rgw",
  "session_state": "1413beec-9785-4e63-947f-72eb26da9daf",
  "acr": "1",
  "allowed-origins": [
"*"
  ],
  "realm_access": {
"roles": [
  "offline_access",
  "uma_authorization"
]
  },
  "resource_access": {
"ceph_rgw": {
  "roles": [
"arn:aws:iam:::role/S3Access",
"S3Access"
  ]
},
"account": {
  "roles": [
"manage-account",
"manage-account-links",
"view-profile"
  ]
}
  },
  "scope": "openid profile email",
  "email_verified": true,
  "name": "testuser",
  "preferred_username": "testuser",
  "given_name": "testuser",
  "email": "test-u...@help-me-please.com"
}

Please, if you are familiar with this, can you tell me what step is
missing? There is no description on that page of what should be done at
Keycloak, so I'm guessing the problem may be there. (Keycloak screens are
shown elsewhere, but for a different example.) I have spent a good deal of
time trying to understand this, so if you could help I would greatly
appreciate it.
Kind regards,
Michael

On Tue, 23 Nov 2021 at 06:22, Pritha Srivastava  wrote:

> Hi Nio,
>
> Can you provide more details around what you are trying to do?
>
> RGW supports attaching IAM policies to users that aid in managing their
> permissions.
>
> Thanks,
> Pritha
>
> On Tue, Nov 23, 2021 at 11:43 AM nio  wrote:
>
> > hi,all:
> > In the process of using RGW, I still cannot authenticate users
> through
> > IAM. In the near future, will RGW support IAM to manage user permissions
> > and authentication functions?
> >
> >
> > Looking forward to your reply 😁
> > ___
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
> >
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>

-- 
CONFIDENTIALITY
This e-mail message and any attachments thereto, is 
intended only for use by the addressee(s) named herein and may contain 
legally privileged and/or confidential information. If you are not the 
intended recipient of this e-mail message, you are hereby notified that any 
dissemination, distribution or copying of this e-mail message, and any 
attachments thereto, is strictly prohibited.  If you have received this 
e-mail message in error, please immediately notify the sender and 
permanently delete the original and any copies of this email and any prints 
thereof.
ABSENT AN EXPRESS STATEMENT TO THE CONTRARY HEREINABOVE, THIS 
E-MAIL IS NOT INTENDED AS A SUBSTITUTE FOR A WRITING.  Notwithstanding the 
Uniform Electronic Transactions Act or the applicability of any other law 
of similar substance and effect, absent an express statement to the 
contrary hereinabove, this e-mail message its contents, and any attachments 
hereto are not intended to represent an offer or acceptance to enter into a 
contract and are not otherwise intended to bind the sender, Sanmina 
Corporation (or any of its subsidiaries), or any other person or entity.
___

[ceph-users] Re: have buckets with low number of shards

2021-11-23 Thread mahnoosh shahidi
Hi Dominic,

Thanks for explanation but I didn't mean the bucket lock which happens
during the reshard. My problem is when number of objects in a bucket is
about 500M and more than that, deleting those old RADOS objects in the
reshard process, causes slow ops which results in osd failures so we
experience down time in the whole cluster not only in the resharded bucket.

Thanks

On Tue, Nov 23, 2021, 7:43 PM  wrote:

> Manoosh;
>
> You can't reshard a bucket without downtime.  During a reshard RGW creates
> new RADOS objects to match the new shard number.  Then all the RGW objects
> are moved from the old RADOS objects to the new RADOS objects, and the
> original RADOS objects are destroyed.  The reshard locks the bucket for the
> duration.
>
> Thank you,
>
> Dominic L. Hilsbos, MBA
> Vice President - Information Technology
> Perform Air International Inc.
> dhils...@performair.com
> www.PerformAir.com
>
>
> -Original Message-
> From: mahnoosh shahidi [mailto:mahnooosh@gmail.com]
> Sent: Tuesday, November 23, 2021 8:20 AM
> To: Josh Baergen
> Cc: Ceph Users
> Subject: [ceph-users] Re: have buckets with low number of shards
>
> Hi Josh
>
> Thanks for your response. Do you have any advice how to reshard these big
> buckets so it doesn't cause any down time in our cluster? Resharding these
> buckets makes a lots of slow ops in deleting old shard phase and the
> cluster can't responde to any requests till resharding is completely done.
>
> Regards,
> Mahnoosh
>
> On Tue, Nov 23, 2021, 5:28 PM Josh Baergen 
> wrote:
>
> > Hey Mahnoosh,
> >
> > > Running cluster in octopus 15.2.12 . We have a big bucket with about
> 800M
> > > objects and resharding this bucket makes many slow ops in our bucket
> > index
> > > osds. I wanna know what happens if I don't reshard this bucket any
> more?
> > > How does it affect the performance? The performance problem would be
> only
> > > for that bucket or it affects the entire bucket index pool?
> >
> > Unfortunately, if you don't reshard the bucket, it's likely that
> > you'll see widespread index pool performance and stability issues,
> > generally manifesting as one or more OSDs becoming very busy to the
> > point of holding up traffic for multiple buckets or even flapping (the
> > OSD briefly gets marked down), leading to recovery. Recovering large
> > index shards can itself cause issues like this to occur. Although the
> > official recommendation, IIRC, is 100K objects per index shard, the
> > exact objects per shard count at which one starts to experience these
> > sorts of issues highly depends on the hardware involved and user
> > workload.
> >
> > Josh
> >
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] DACH Ceph Meetup

2021-11-23 Thread Mike Perez
Hi everyone,

There will be a virtual Ceph Meetup taking place on November 30th at
16:00 UTC. Take a look at the excellent lineup of speakers we have and
register.

https://ceph.io/en/community/events/2021/meetup-dach-2021-11-30/

P.S. This is an opportunity to claim a free Ceph Pacific release
shirt. More details to follow during the event.

-- 
Mike Perez

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: RGW support IAM user authentication

2021-11-23 Thread Pritha Srivastava
Hi Michael,

My responses are inline:

On Tue, Nov 23, 2021 at 10:07 PM Michael Breen <
michael.br...@vikingenterprise.com> wrote:

> Hi Pritha - or anyone who knows,
>
> I too have problems with IAM, in particular with AssumeRoleWithWebIdentity.
>
> I am running the master branch version of Ceph because it looks like it
> includes code related to the functionality described at
> https://docs.ceph.com/en/latest/radosgw/STS/ - code which is not in any
> released version, even 17.0.
>
> Looking at the code on that page, there appear to be at least two errors:
> (1) an instance of "client" should be "sts_client" (or vice versa)
> (2) an access key and secret key are specified when creating sts_client,
> which is unnecessary and therefore confusing: only the access token is used
> or should be required for assume_role_with_web_identity
>

I agree, these can be corrected. However I don't understand what you mean
by saying that this -  "code which is not in any released version, even
17.0."

>
> But I still cannot get the AssumeRoleWithWebIdentity code example to work.
> The RGW debug logs show
>
> debug 2021-11-23T15:51:22.247+ 7fad6e351700  0 evaluating policy for
> op: 93 returned deny/pass
>
> In my case, the policy_document and role_policy are
>
> policy_document =
> '''{"Version":"2012-10-17","Statement":[{"Effect":"Allow","Principal":{"Federated":["arn:aws:iam:::oidc-provider/proteus.ves.corp/auth/realms/cno"]},"Action":["sts:AssumeRoleWithWebIdentity"],"Condition":{"StringEquals":{"proteus.ves.corp/auth/realms/cno:app_id":"ceph_rgw"}}}]}'''
> role_policy =
> '''{"Version":"2012-10-17","Statement":{"Effect":"Allow","Action":"s3:*","Resource":"arn:aws:s3:::*"}}'''
>
>

> (I assume it is only the former that may be relevant here, but maybe I'm
> wrong.)
>

The former is only relevant here. And have you created an openid connect
provider in RGW? ( I am assuming you must have, since it is there in the
examples given). You have set the condition as app_id: ceph_rgw. Whereas
the documentation says that - "The app_id in the condition above must match
the ‘aud’ claim of the incoming token." in the example which uses "app_id"
as a condition element, and the value of "aud" is "account" in the web
token. So please modify the condition accordingly. Also note that other
claims can also now be used in the condition element of the trust policy.

>
> In /etc/ceph/ceph.conf I have
> [client.radosgw.gateway]
> rgw sts key = abcdefghijklmnop
> rgw s3 auth use sts = true
>
> In the debug I can see the token from Keycloak looks like (after
> formatting it)
>
> {
>   "exp": 1637677729,
>   "iat": 1637677429,
>   "jti": "06e5422e-8395-4727-9366-a851c3f5930f",
>   "iss": "https://proteus.ves.corp/auth/realms/cno";,
>   "aud": "account",
>   "sub": "f45bae70-1517-48f6-9d75-af7f421f4a0c",
>   "typ": "Bearer",
>   "azp": "ceph_rgw",
>   "session_state": "1413beec-9785-4e63-947f-72eb26da9daf",
>   "acr": "1",
>   "allowed-origins": [
> "*"
>   ],
>   "realm_access": {
> "roles": [
>   "offline_access",
>   "uma_authorization"
> ]
>   },
>   "resource_access": {
> "ceph_rgw": {
>   "roles": [
> "arn:aws:iam:::role/S3Access",
> "S3Access"
>   ]
> },
> "account": {
>   "roles": [
> "manage-account",
> "manage-account-links",
> "view-profile"
>   ]
> }
>   },
>   "scope": "openid profile email",
>   "email_verified": true,
>   "name": "testuser",
>   "preferred_username": "testuser",
>   "given_name": "testuser",
>   "email": "test-u...@help-me-please.com"
> }
>
> Please, if you are familiar with this, can you tell me what step is
> missing? There is no description on that page of what should be done at
> Keycloak, so I'm guessing the problem may be there. (Keycloak screens are
> shown elsewhere, but for a different example.) I have spent a good deal of
> time trying to understand this, so if you could help I would greatly
> appreciate it.
> Kind regards,
> Michael
>

If you can paste further logs here then I will be able to help you. Also
Ceph documentation explains how to get a token from Keycloak and proceed
with that. For any other Keycloak configurations, please refer to Keycloak
documentation.

Thanks,
Pritha

>
> On Tue, 23 Nov 2021 at 06:22, Pritha Srivastava 
> wrote:
>
>> Hi Nio,
>>
>> Can you provide more details around what you are trying to do?
>>
>> RGW supports attaching IAM policies to users that aid in managing their
>> permissions.
>>
>> Thanks,
>> Pritha
>>
>> On Tue, Nov 23, 2021 at 11:43 AM nio  wrote:
>>
>> > hi,all:
>> > In the process of using RGW, I still cannot authenticate users
>> through
>> > IAM. In the near future, will RGW support IAM to manage user permissions
>> > and authentication functions?
>> >
>> >
>> > Looking forward to your reply 😁
>> > ___
>> > ceph-users mailing list -- ceph-users@ceph.io
>> > To unsubscribe send an email to ceph-users-le...@ceph.io
>>

[ceph-users] Re: One pg stuck in active+undersized+degraded after OSD down

2021-11-23 Thread David Tinker
Fiddling with the crush weights sorted this out and I was able to remove
the OSD from the cluster. I set all the big weights down to 1

ceph osd crush reweight osd.7 1.0
etc.

Tx for all the help

On Tue, Nov 23, 2021 at 9:35 AM Stefan Kooman  wrote:

> On 11/23/21 08:21, David Tinker wrote:
> > Yes it recovered when I put the OSD back in. The issue is that it fails
> > to sort itself out when I remove that OSD even though I have loads of
> > space and 8 other OSDs in 4 different zones to choose from. The weights
> > are very different (some 3.2 others 0.36) and that post I found
> > suggested that this might cause trouble for the crush algorithm. So I
> > was thinking about making them more even before removing the OSD.
>
> Sure, that sounds like a good plan. If the balancer is not able to
> optimize at this point, the improved balancer bij Jonas Jelten might do
> the trick here [1], also see ML thread [2].
>
> Do you have many different pools?
>
> What Weiwen Hu said in reply to this threads sounds very plausible. You
> might dump your crush map, remove OSD.7, and check if the mappings can
> still be made.
>
> Gr. Stefan
>
> [1]: https://github.com/TheJJ/ceph-balancer
> [2]:
>
> https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/QXTFTYRD43XB3JRDSTMH657ISSMZ6QTU/
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io