Is there any PR ongoing to add such counters to bucket stats? rados-level is
not an option if those counters are needd to do, eg. rating/billing.
From: Casey Bodley
Sent: Wednesday, September 9, 2020 7:50:12 PM
To: Tobias Urdin
Cc: ceph-users@ceph.io
Sub
On 9/24/20 16:01, Jason Dillaman wrote:
On Thu, Sep 24, 2020 at 9:53 AM Stefan Kooman wrote:
On 2020-09-24 14:34, Eugen Block wrote:
Hi *,
I'm curious if this idea [1] of quotas on namespace level for rbd will
be implemented. I couldn't find any existing commands in my lab Octopus
cluster so
Hi,
=== START OF READ SMART DATA SECTION ===
SMART Health Status: FIRMWARE IMPENDING FAILURE TOO MANY BLOCK REASSIGNS
[asc=5d, ascq=64]
this indicates a slowly failing drive. You should contact the vendor
and replace the drive. This can happen on new drives, too.
Zitat von Marco Pizzolo :
Hi,
I'm currently setting up ceph CSI for our kubernetes cluster. What are
the minimum requirements / capabilities needed for the rbd and cephfs
users? The current setup is working well with admin privileges, but I
would like to reduce it to the necessary minimum.
Regards,
Burkhard
_
Hi Istvan,
I have not given Octopus another try yet. But as far as I remember Manuel
figured out the root cause.
Maybe he can give more insights.
Best,
Peter
Am 28.10.21 um 13:07 schrieb Szabo, Istvan (Agoda):
Hi Peter,
Have you figured out what was the issue?
Istvan Szabo
Senior Infrastruc
In case you still have the error messages and additional info, do you
want to create a tracker issue for this?
https://tracker.ceph.com/projects/orchestrator/issues/new . To me this
sounds like a network issue and not like a rockylinux issue.
Am 26.10.21 um 13:17 schrieb Magnus Harlander:
> Hi,
>
Some thoughts:
* Do you have any error messages form the MDS daemons?
https://docs.ceph.com/en/latest/cephadm/troubleshooting/#gathering-log-filesĀ
* Do you have any error messages form the OSDs?
* What do you mean by "osd podman object"?
* Try downgrading to 3.0.1
Am 25.10.21 um 23
Hello Istvan,
the state "waiting for readable" seems to be related to this read_lease
topic documented here:
https://docs.ceph.com/en/latest/dev/osd_internals/stale_read/
The only parameter to tune around this I know about, is the
"osd_pool_default_read_lease_ratio" which is defaulted 0.8
My clu
Hello Istvan,
as described on the ceph docu page this ratio will be multiplied with
the value from osd_heartbeat_grace which is defaulted to 20 seconds.
The product is 16 seconds which is the lease ticket time. So your
requests can theoretically be blocked up to 16 seconds. For our usecase
this i
On Thu, Oct 28, 2021 at 3:46 AM Engelmann Florian
wrote:
>
> Is there any PR ongoing to add such counters to bucket stats? rados-level is
> not an option if those counters are needd to do, eg. rating/billing.
i'm not aware of any work in progress here. a feature request at
https://tracker.ceph.
Dear Ceph users.
On behalf of Ceph's developers community I have to inform about a
recently discovered severe bug which might cause data corruption. The
issue occurs during OMAP format conversion for clusters upgraded to
Pacific, new clusters aren't affected. OMAP format conversion's trigger
Hi,
Try to use profile cap, like 'allow profile rbd'
k
Sent from my iPhone
> On 28 Oct 2021, at 12:47, Burkhard Linke
> wrote:
>
> I'm currently setting up ceph CSI for our kubernetes cluster. What are the
> minimum requirements / capabilities needed for the rbd and cephfs users? The
> cur
On 10/28/2021 7:13 PM, Eneko Lacunza wrote:
Hi all,
For those looking for the exact commands, I think they are:
* Check value for bluestore_fsck_quick_fix_on_mount:
ceph config get osd bluestore_fsck_quick_fix_on_mount
* Set bluestore_fsck_quick_fix_on_mount to false:
ceph config set osd b
Hi,
On 28.10.21 18:10, Konstantin Shalygin wrote:
Hi,
Try to use profile cap, like 'allow profile rbd'
That's fine for csi rbd, thx. Works like a charm so far.
But cephfs is a little different beast. As far as I understand the
source code, it uses the mgr interface to create subvolumes an
Hi everyone,
We are kicking off a new monthly meeting for Ceph users to directly
interact with Ceph Developers. The high-level aim of this meeting is
to provide users with a forum to:
- share their experience running Ceph clusters
- provide feedback on Ceph versions they are using
- ask questions
Yes, just expose each disk as an individual OSD and you'll already be
better off. Depending what type of SSD they are - if they can sustain
high random write IOPS you may even want to consider partitioning each
disk and create 2 OSDs per SSD to make better use of the available IO
capacity.
For all-
Hello to the list.
Who is everyone using for a hardware VAR that can build a Ceph cluster
with RedHat support? Looking to buy a 500TB cluster. 45Drives is high on
the list, but they don't (yet) offer 24/7 phone support and that is a
requirement for the business.
We have existing support cont
Den tors 28 okt. 2021 kl 22:25 skrev Dave Hall :
> Hello,
> I have a Nautilus 14.2.21 cluster with 48 x 12TB OSDs across 6 nodes, with
> 3 new nodes and 24 more OSDs ready to come online. The bulk of my pools
> are EC 8+2 with a failure domain of OSD.
> Until yesterday one of the original 48 OSDs
18 matches
Mail list logo