[ceph-users] force-create-pg not working

2022-09-19 Thread Jesper Lykkegaard Karlsen
Dear all, System: latest Octopus, 8+3 erasure Cephfs I have a PG that has been driving me crazy. It had gotten to a bad state after heavy backfilling, combined with OSD going down in turn. State is: active+recovery_unfound+undersized+degraded+remapped I have tried repairing it with ceph-objec

[ceph-users] Re: tcmu-runner lock failure

2022-09-19 Thread Xiubo Li
On 19/09/2022 23:32, j.rasakunasin...@autowork.com wrote: Hi, we have 3x controller and 6xstorage Ceph Cluster running. We use iscsi/tcmu runner (16.2.9) to connect VMware to Ceph. We face an issue, that we lost the connection to the iscsi gateways, that ESXi is connected not works properly.

[ceph-users] Re: rbd mirroring - journal growing and snapshot high io load

2022-09-19 Thread Ilya Dryomov
On Thu, Sep 15, 2022 at 3:33 PM Arthur Outhenin-Chalandre wrote: > > Hi Ronny, > > > On 15/09/2022 14:32 ronny.lippold wrote: > > hi arthur, some time went ... > > > > i would like to know, if there are some news of your setup. > > do you have replication active running? > > No, there was no chan

[ceph-users] Re: tcmu-runner lock failure

2022-09-19 Thread j . rasakunasingam
Hi, we have 3x controller and 6xstorage Ceph Cluster running. We use iscsi/tcmu runner (16.2.9) to connect VMware to Ceph. We face an issue, that we lost the connection to the iscsi gateways, that ESXi is connected not works properly. After restarting the servers it works again, but later the

[ceph-users] Re: quincy v17.2.4 QE Validation status

2022-09-19 Thread Guillaume Abrioux
ceph-volume approved. (errors in CI can be safely ignored). On Sat, 17 Sept 2022 at 12:52, Venky Shankar wrote: > On Wed, Sep 14, 2022 at 1:33 AM Yuri Weinstein > wrote: > > > > Details of this release are summarized here: > > > > https://tracker.ceph.com/issues/57472#note-1 > > Release Notes -

[ceph-users] Bluestore config issue with ceph orch

2022-09-19 Thread Aristide Bekroundjo
Hi dear, I’m running a cluster under 16.2.10. I’m trying to to configure a bluestore device with ceph orch command but I failed. [ceph: root@monnode1 /]# ceph orch daemon add osd osdnode1:data_devices=/dev/sdd,wal_devices=/dev/sde,db_devices=/dev/sdf Error EINVAL: Traceback (most recent c

[ceph-users] tcmu-runner lock failure

2022-09-19 Thread j . rasakunasingam
Hi,   we have 3x controller and 6xstorage Ceph Cluster running. We use iscsi/tcmu runner (16.2.9) to connect VMware to Ceph.    We face an issue, that we lost the connection to the iscsi gateways, that ESXi is connected not works properly. After restarting the servers it works again, but late

[ceph-users] Re: CephFS Mirroring failed

2022-09-19 Thread Aristide Bekroundjo
Hi Milind, Thanks, I’ll test and revert back to you. Best regards, De : Milind Changire Envoyé le :lundi 19 septembre 2022 08:59 À : Aristide Bekroundjo Cc : ceph-users@ceph.io Objet :Re: [ceph-users] Re: CephFS Mi

[ceph-users] Re: [ceph-users] OSD Crash in recovery: SST file contains data beyond the point of corruption.

2022-09-19 Thread Benjamin Naber
Hi Igor, in case of antoher corrupted record in rocksdb "kSkipAnyCorruptedRecords" did the job at the end :) Regards and many thanks again! Ben Am Montag, September 19, 2022 10:51 CEST, schrieb Igor Fedotov :    Hi Benjamin, good to know, thanks for the feedback. Curious which exact mode (kS

[ceph-users] Re: Public RGW access without any LB in front?

2022-09-19 Thread Casey Bodley
hi Boris, it looks like your other questions have been covered but i'll snipe this one: On Fri, Sep 16, 2022 at 7:55 AM Boris Behrens wrote: > > How good is it handling bad HTTP request, sent by an attacker?) rgw relies on the boost.beast library to parse these http requests. that library has ha

[ceph-users] tcmu-runner

2022-09-19 Thread j . rasakunasingam
Hi,   we have 3x controller and 6xstorage Ceph Cluster running. We use iscsi/tcmu runner (16.2.9) to connect VMware to Ceph.    We face an issue, that we lost the connection to the iscsi gateways, that ESXi is connected not works properly. After restarting the servers it works again, but later

[ceph-users] Re: Any disadvantage to go above the 100pg/osd or 4osd/disk?

2022-09-19 Thread Szabo, Istvan (Agoda)
Sorry, 96vcore is a typo, 2vcore/osd but can be 4 also. > > On 2022. Sep 19., at 19:50, Szabo, Istvan (Agoda) > wrote: > > Hi, > > My question is, is there any technical limit to have 8osd/ssd and on each of > them 100pg if the memory and cpu resource available (8gb memory/osd and > 96vcor

[ceph-users] Any disadvantage to go above the 100pg/osd or 4osd/disk?

2022-09-19 Thread Szabo, Istvan (Agoda)
Hi, My question is, is there any technical limit to have 8osd/ssd and on each of them 100pg if the memory and cpu resource available (8gb memory/osd and 96vcore)? The iops and bandwidth on the disks are very low so I don’t see any issue to go with this. In my cluster I’m using 15.3TB ssds. We

[ceph-users] Re: Public RGW access without any LB in front?

2022-09-19 Thread Tobias Urdin
I was assuming it had to do with scaling, ofcourse there is multiple ways to do it. Personally I don’t find scaling that way is reasonable but that’s design decision, that way you would still have some control to do traffic engineering On 19 Sept 2022, at 10:23, Konstantin Shalygin mailto:k0...

[ceph-users] Re: OSD Crash in recovery: SST file contains data beyond the point of corruption.

2022-09-19 Thread Igor Fedotov
Hi Benjamin, good to know, thanks for the feedback. Curious which exact mode (kSkipAnyCorruptedRecords or kTolerateCorruptedTailRecords) did the trick? Regards, Igor On 9/17/2022 4:57 PM, Benjamin Naber wrote: Hey Igor, i just wanted to thank you for the help! With the Flag you told me,

[ceph-users] Re: Public RGW access without any LB in front?

2022-09-19 Thread Konstantin Shalygin
Hi, > On 19 Sep 2022, at 10:38, Tobias Urdin wrote: > > Why not scaleout HAproxy by adding multiple ones and use a TCP load balancer > in front of multiple HAproxy instances or use BGP ECMP routing directly to > split > load between multiple HAproxy? Because you can do this without "TCP load b

[ceph-users] Re: CephFS Mirroring failed

2022-09-19 Thread Milind Changire
All paths mentioned while configuring cephfs snapshot mirroring start at the respective cephfs file-system root eg. if you typically mount the cephfs file-system at /mnt/folderfs, then the path "/mnt/folderfs" is meaningless to cephfs snapshot mirroring unless you indeed have a folder hierarchy /mn

[ceph-users] Re: Public RGW access without any LB in front?

2022-09-19 Thread Tobias Urdin
Hello, Why not scaleout HAproxy by adding multiple ones and use a TCP load balancer in front of multiple HAproxy instances or use BGP ECMP routing directly to split load between multiple HAproxy? Best regards > On 16 Sept 2022, at 13:54, Boris Behrens wrote: > > Hi, > does someone got experien

[ceph-users] Re: Public RGW access without any LB in front?

2022-09-19 Thread Konstantin Shalygin
Hi, Actually rgw can handle SSL traffic, and updates of certs is just a restarting of service. For client it will be reset of connection, client will make a new one We use keeaplived DR method for RGW's for a years The only bottleneck in this setup is input traffic limited by LB. This also can