[ceph-users] MDS crash on FAILED ceph_assert(cur->is_auth())

2021-06-28 Thread Peter van Heusden
I am running Ceph 15.2.13 on CentOS 7.9.2009 and recently my MDS servers have started failing with the error message In function 'void Server::handle_client_open(MDRequestRef&)' thread 7f0ca9908700 time 2021-06-28T09:21:11.484768+0200 /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAI

[ceph-users] Re: [Suspicious newsletter] Nic bonding (lacp) settings for ceph

2021-06-28 Thread Szabo, Istvan (Agoda)
We are using it with 3+4. Istvan Szabo Senior Infrastructure Engineer --- Agoda Services Co., Ltd. e: istvan.sz...@agoda.com --- -Original Message- From: mhnx Sent: Saturday, June 26, 2021 8:4

[ceph-users] Re: Can we deprecate FileStore in Quincy?

2021-06-28 Thread Kai Börnert
If you want to go cheap and somewhat questionable, there are some asrock mainboards with a soldered in atom cpu, that support up to 32gb memory (officially only 8, but the controller does more) and have 2 sata directly + a free 16x pcie port, Those boards are usually less than 90€, not as chea

[ceph-users] Re: MDS crash on FAILED ceph_assert(cur->is_auth())

2021-06-28 Thread Peter van Heusden
Yes it keeps crashing in a loop. I ran again with debug set to 20 and the last 100,000 lines of that log are here: https://gist.github.com/pvanheus/33eb22b179a9cbd68a460984de8ef24a On Mon, 28 Jun 2021 at 15:29, Stefan Kooman wrote: > On 6/28/21 10:52 AM, Peter van Heusden wrote: > > I am runnin

[ceph-users] Re: radosgw user "check_on_raw" setting

2021-06-28 Thread Konstantin Shalygin
Hi, I think is not possible to make this with CLI, like for placements configs Actually this is settings of period: "period_config": { "bucket_quota": { "enabled": false, "check_on_raw": false, "max_size": -1, "max_size_kb": 0,

[ceph-users] Re: Nic bonding (lacp) settings for ceph

2021-06-28 Thread Marc 'risson' Schmitt
Hi, On Sat, 26 Jun 2021 16:47:19 +0300 mhnx wrote: > I've changed ad_select to bandwitdh and both nic is in use now but > layer2 hash prevents dual nic usage for between two nodes (because > layer2 using only Mac ). As I understand it, setting ad_select to bandwidth is only going to be useful if

[ceph-users] Re: MDS crash on FAILED ceph_assert(cur->is_auth())

2021-06-28 Thread Stefan Kooman
On 6/28/21 10:52 AM, Peter van Heusden wrote: I am running Ceph 15.2.13 on CentOS 7.9.2009 and recently my MDS servers have started failing with the error message Do they keep crashing (in a loop)? Can you set ms / mds debug to ... say 20/20? debug_ms = 20/20 debug_mds = 20/20 Gr. Stefan __

[ceph-users] Re: Nic bonding (lacp) settings for ceph

2021-06-28 Thread mhnx
Thanks for the answer. I'm into ad_select bandwitdh because we use osd nodes as rgw gateways, VMs and different applications. I have seperate cluster (10+10Gbe) and public (10+10Gbe) network. I tested stable, bandwitdh and count. Results are clearly good with bandwitdh. Count is the worst option.

[ceph-users] Re: Nic bonding (lacp) settings for ceph

2021-06-28 Thread Andrew Walker-Brown
HI, I think ad_select is only relevant in the scenario below I.e where you have more than one port-channel being presented to the Linux bond. So below, you have 2 port channels, one from each switch, but at the Linux side all the ports involved are slaves in the same bond. In your scenario it

[ceph-users] Re: MDS crash on FAILED ceph_assert(cur->is_auth())

2021-06-28 Thread Stefan Kooman
On 6/28/21 3:52 PM, Peter van Heusden wrote: Yes it keeps crashing in a loop.  I ran again with debug set to 20 and the last 100,000 lines of that log are here: https://gist.github.com/pvanheus/33eb22b179a9cbd68a460984de8ef24a

[ceph-users] Re: Nic bonding (lacp) settings for ceph

2021-06-28 Thread mhnx
To be clear. I have stacked switch and this is my configuration. Bonding cluster: (hash 3+4) Cluster nic1(10Gbe) -> Switch A Cluster nic2(10Gbe) -> Switch B Bonding public: (hash 3+4) Public nic1(10Gbe) -> Switch A Public nic2(10Gbe) -> Switch B Data distribution wasn't good at the begining du

[ceph-users] Re: Nic bonding (lacp) settings for ceph

2021-06-28 Thread Marc 'risson' Schmitt
On Mon, 28 Jun 2021 22:35:36 +0300 mhnx wrote: > To be clear. > I have stacked switch and this is my configuration. > > Bonding cluster: (hash 3+4) > Cluster nic1(10Gbe) -> Switch A > Cluster nic2(10Gbe) -> Switch B > > Bonding public: (hash 3+4) > Public nic1(10Gbe) -> Switch A > Public nic2(

[ceph-users] Where did links to official MLs are moved?

2021-06-28 Thread Satoru Takeuchi
Today I visited Ceph's official site and found that links to `resources` page seemed to be missing. https://ceph.io/en/ In addition, this page is no longer exists. https://ceph.io/resources/ Could you tell me where did they are moved? Thanks, Satoru

[ceph-users] Re: Why you might want packages not containers for Ceph deployments

2021-06-28 Thread Sage Weil
On Fri, Jun 25, 2021 at 10:27 AM Nico Schottelius wrote: > Hey Sage, > > Sage Weil writes: > > Thank you for bringing this up. This is in fact a key reason why the > > orchestration abstraction works the way it does--to allow other > > runtime environments to be supported (FreeBSD! > > sysvinit/

[ceph-users] upgrading from Nautilus on CentOS7 to Octopus on Ubuntu 20.04.2

2021-06-28 Thread Jean-Philippe Méthot
Hi, I’ve been running a staging Ceph environment on CentOS 7/Nautilus for quite a while now. Because of many good reasons that you can probably guess, I am currently trying to move this staging environment to Octopus on Ubuntu 20.04.2. Since I’m trying to keep the data, but don’t mind downtime

[ceph-users] Re: upgrading from Nautilus on CentOS7 to Octopus on Ubuntu 20.04.2

2021-06-28 Thread Marc
> I’ve been running a staging Ceph environment on CentOS 7/Nautilus for quite a > while now. Because of many good reasons that you can probably guess, I am > currently trying to move this staging environment to Octopus on Ubuntu > 20.04.2. > > What made you decide to chose Ubuntu and not this ro

[ceph-users] Re: Can we deprecate FileStore in Quincy?

2021-06-28 Thread Eric Petit
>> At a Ceph Day in Hillsboro someone, forgive me for not remembering >> who, spoke of running production on servers with 2GB RAM per OSD. He >> said that it was painful, required a lot of work, and would not >> recommend it. ymmv. > > Yeah, I wouldn't want to go below 4GB RAM. FWIW, I have bee