[ceph-users] Re: Storage class usage stats

2021-10-28 Thread Engelmann Florian
Is there any PR ongoing to add such counters to bucket stats? rados-level is not an option if those counters are needd to do, eg. rating/billing. From: Casey Bodley Sent: Wednesday, September 9, 2020 7:50:12 PM To: Tobias Urdin Cc: ceph-users@ceph.io Sub

[ceph-users] Re: RBD quota per namespace

2021-10-28 Thread Stefan Kooman
On 9/24/20 16:01, Jason Dillaman wrote: On Thu, Sep 24, 2020 at 9:53 AM Stefan Kooman wrote: On 2020-09-24 14:34, Eugen Block wrote: Hi *, I'm curious if this idea [1] of quotas on namespace level for rbd will be implemented. I couldn't find any existing commands in my lab Octopus cluster so

[ceph-users] Re: 16.2.6 OSD down, out but container running....

2021-10-28 Thread Eugen Block
Hi, === START OF READ SMART DATA SECTION === SMART Health Status: FIRMWARE IMPENDING FAILURE TOO MANY BLOCK REASSIGNS [asc=5d, ascq=64] this indicates a slowly failing drive. You should contact the vendor and replace the drive. This can happen on new drives, too. Zitat von Marco Pizzolo :

[ceph-users] Minimal requirements for ceph csi users?

2021-10-28 Thread Burkhard Linke
Hi, I'm currently setting up ceph CSI for our kubernetes cluster. What are the minimum requirements / capabilities needed for the rbd and cephfs users? The current setup is working well with admin privileges, but I would like to reduce it to the necessary minimum. Regards, Burkhard _

[ceph-users] Re: octupus: stall i/o during recovery

2021-10-28 Thread Peter Lieven
Hi Istvan, I have not given Octopus another try yet. But as far as I remember Manuel figured out the root cause. Maybe he can give more insights. Best, Peter Am 28.10.21 um 13:07 schrieb Szabo, Istvan (Agoda): Hi Peter, Have you figured out what was the issue? Istvan Szabo Senior Infrastruc

[ceph-users] Re: MDS and OSD Problems with cephadm@rockylinux solved

2021-10-28 Thread Sebastian Wagner
In case you still have the error messages and additional info, do you want to create a tracker issue for this? https://tracker.ceph.com/projects/orchestrator/issues/new . To me this sounds like a network issue and not like a rockylinux issue. Am 26.10.21 um 13:17 schrieb Magnus Harlander: > Hi, >

[ceph-users] Re: cephadm does not find podman objects for osds

2021-10-28 Thread Sebastian Wagner
Some thoughts: * Do you have any error messages form the MDS daemons? https://docs.ceph.com/en/latest/cephadm/troubleshooting/#gathering-log-filesĀ  * Do you have any error messages form the OSDs? * What do you mean by "osd podman object"? * Try downgrading to 3.0.1 Am 25.10.21 um 23

[ceph-users] Re: OSD spend too much time on "waiting for readable" -> slow ops -> laggy pg -> rgw stop -> worst case osd restart

2021-10-28 Thread Manuel Lausch
Hello Istvan, the state "waiting for readable" seems to be related to this read_lease topic documented here: https://docs.ceph.com/en/latest/dev/osd_internals/stale_read/ The only parameter to tune around this I know about, is the "osd_pool_default_read_lease_ratio" which is defaulted 0.8 My clu

[ceph-users] Re: [Suspicious newsletter] Re: slow ops at restarting OSDs (octopus)

2021-10-28 Thread Manuel Lausch
Hello Istvan, as described on the ceph docu page this ratio will be multiplied with the value from osd_heartbeat_grace which is defaulted to 20 seconds. The product is 16 seconds which is the lease ticket time. So your requests can theoretically be blocked up to 16 seconds. For our usecase this i

[ceph-users] Re: Storage class usage stats

2021-10-28 Thread Casey Bodley
On Thu, Oct 28, 2021 at 3:46 AM Engelmann Florian wrote: > > Is there any PR ongoing to add such counters to bucket stats? rados-level is > not an option if those counters are needd to do, eg. rating/billing. i'm not aware of any work in progress here. a feature request at https://tracker.ceph.

[ceph-users] [IMPORTANT NOTICE] Potential data corruption in Pacific

2021-10-28 Thread Igor Fedotov
Dear Ceph users. On behalf of Ceph's developers community I have to inform about a recently discovered severe bug which might cause data corruption. The issue occurs during OMAP format conversion for clusters upgraded to Pacific, new clusters aren't affected. OMAP format conversion's trigger

[ceph-users] Re: Minimal requirements for ceph csi users?

2021-10-28 Thread Konstantin Shalygin
Hi, Try to use profile cap, like 'allow profile rbd' k Sent from my iPhone > On 28 Oct 2021, at 12:47, Burkhard Linke > wrote: > > I'm currently setting up ceph CSI for our kubernetes cluster. What are the > minimum requirements / capabilities needed for the rbd and cephfs users? The > cur

[ceph-users] Re: [IMPORTANT NOTICE] Potential data corruption in Pacific

2021-10-28 Thread Igor Fedotov
On 10/28/2021 7:13 PM, Eneko Lacunza wrote: Hi all, For those looking for the exact commands, I think they are: * Check value for bluestore_fsck_quick_fix_on_mount: ceph config get osd bluestore_fsck_quick_fix_on_mount * Set bluestore_fsck_quick_fix_on_mount to false: ceph config set osd b

[ceph-users] Re: Minimal requirements for ceph csi users?

2021-10-28 Thread Burkhard Linke
Hi, On 28.10.21 18:10, Konstantin Shalygin wrote: Hi, Try to use profile cap, like 'allow profile rbd' That's fine for csi rbd, thx. Works like a charm so far. But cephfs is a little different beast. As far as I understand the source code, it uses the mgr interface to create subvolumes an

[ceph-users] Ceph User + Dev Monthly Meetup

2021-10-28 Thread Neha Ojha
Hi everyone, We are kicking off a new monthly meeting for Ceph users to directly interact with Ceph Developers. The high-level aim of this meeting is to provide users with a forum to: - share their experience running Ceph clusters - provide feedback on Ceph versions they are using - ask questions

[ceph-users] Re: [Ceph] Recovery is very Slow

2021-10-28 Thread Christian Wuerdig
Yes, just expose each disk as an individual OSD and you'll already be better off. Depending what type of SSD they are - if they can sustain high random write IOPS you may even want to consider partitioning each disk and create 2 OSDs per SSD to make better use of the available IO capacity. For all-

[ceph-users] Hardware VAR for 500TB cluster with RedHat support.

2021-10-28 Thread accounts@stargate.services
Hello to the list. Who is everyone using for a hardware VAR that can build a Ceph cluster with RedHat support? Looking to buy a 500TB cluster. 45Drives is high on the list, but they don't (yet) offer 24/7 phone support and that is a requirement for the business. We have existing support cont

[ceph-users] Re: 2 OSDs Near Full, Others Under 50%

2021-10-28 Thread Janne Johansson
Den tors 28 okt. 2021 kl 22:25 skrev Dave Hall : > Hello, > I have a Nautilus 14.2.21 cluster with 48 x 12TB OSDs across 6 nodes, with > 3 new nodes and 24 more OSDs ready to come online. The bulk of my pools > are EC 8+2 with a failure domain of OSD. > Until yesterday one of the original 48 OSDs