[ceph-users] CephFS 16.2.10 problem

2024-11-24 Thread Alexey.Tsivinsky
Good evening!

The following problem occurred.
There is a cluster ceph 16.2.10
The cluster was operating normally on Friday. Shut down cluster:
-Excluded all clients
Executed commands:
ceph osd set noout
ceph osd set nobackfill
ceph osd set norecover
ceph osd set norebalance
ceph osd set nodown
ceph osd set pause
Turned off the cluster, checked server maintenance.
Enabled cluster. He gathered himself, found all the nodes, and here the problem 
began. After all OSD went up and all pg became available, cephfs refused to 
start.
Now mds are in the replay status, and do not go to the ready status.
Previously, one of them was in the replay (laggy) status, but we executed 
command:  ceph config set mds mds_wipe_sessions true
After that, mds switched to the status of replays, the third in standby status 
started, and mds crashes with an error stopped.
But cephfs is still unavailable.
What else can we do?
The cluster is very large, almost 200 million files.


Best regards


A.Tsivinsky

e-mail: alexey.tsivin...@baikalelectronics.com
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: CephFS 16.2.10 problem

2024-11-24 Thread Marc
> 
> The following problem occurred.
> There is a cluster ceph 16.2.10
> The cluster was operating normally on Friday. Shut down cluster:
> -Excluded all clients
> Executed commands:
> ceph osd set noout
> ceph osd set nobackfill
> ceph osd set norecover
> ceph osd set norebalance
> ceph osd set nodown
> ceph osd set pause
> Turned off the cluster, checked server maintenance.
> Enabled cluster. He gathered himself, found all the nodes, and here the
> problem began. After all OSD went up and all pg became available, cephfs
> refused to start.
> Now mds are in the replay status, and do not go to the ready status.
> Previously, one of them was in the replay (laggy) status, but we
> executed command:  ceph config set mds mds_wipe_sessions true
> After that, mds switched to the status of replays, the third in standby
> status started, and mds crashes with an error stopped.
> But cephfs is still unavailable.
> What else can we do?
> The cluster is very large, almost 200 million files.
> 

I assume you tried to start just one mds and wait until it would come up as 
active (before starting the others)?




___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Migrated to cephadm, rgw logs to file even when rgw_ops_log_rados is true

2024-11-24 Thread Paul JURCO
Hi!
Currently I have limitted the optput of rgw log to syslog from rsyslog (as
suggested by
Anthony), limitted docker logs from daemon.json.
I still get ops logs written to both logs pool and ops log file
(ops-log-ceph-client.rgw.hostname.log).

How to stop logging ops log on rgw disk and keep logs on logs pool?
Current config:
globaladvanced  debug_rgw
0/0
globaladvanced
 rgw_enable_ops_log true
globaladvanced
 rgw_ops_log_rados   true

Thank you!

-- 
Paul Jurco


On Fri, Nov 22, 2024 at 6:11 PM Paul JURCO  wrote:

> Hi,
> we recently migrated to cephadm from ceph-deploy a 18.2.2 ceph cluster
> (Ubuntu with docker).
> RGWs are separate vms.
> We noticed syslog increased a lot due to rgw's access logs sent to it.
> And because we use to log ops, a huge ops log file on
> /var/log/ceph/cluster-id/ops-log-ceph-client.rgw.hostname-here.log.
>
> While having "rgw_ops_log_rados": "true", oplogs goes to both file and
> rados pool for logs.
> If false it doesn't log anything, as expected.
> How to stop dockered rgws to log to syslog and to a file on disk, but to
> keep opslog in logs pool?
>
> Config is:
> globalbasic
> log_to_journald  false
> globaladvanced
>  rgw_enable_ops_log  false
> globaladvanced
>  rgw_ops_log_radostrue
>
> A few hours later after after enabling it back, after massive cleanup, it
> does logging ops but only to files.
> How to get ops logs in rados pool and access log to a file on disk but not
> on syslog?
> I have add this to daemon.json to limit access logs to accumulate on
> /var/log/docker/containers/rand/rand/json.log file:
>
> {
>   "log-driver": "local",
>   "log-opts": {
> "max-size": "512m",
> "max-file": "3"
>   }
> }
>
>
> Thank you!
> Paul
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Separate gateway for bucket lifecycle

2024-11-24 Thread Paul JURCO
Hi, just remove them from balancer.
Also, there are two configs you want to be true on the rgws for LC and GC
processing and false on the rgws that are exposed to clients:

rgw_enable_lc_threads = true

rgw_enable_gc_threads = true


-- 
Paul


On Mon, Nov 25, 2024 at 8:40 AM Szabo, Istvan (Agoda) <
istvan.sz...@agoda.com> wrote:

> Hi,
>
> I remember some discussion that someone is using separated gateways for
> bucket lifecycles but I couldn't find it.
>
> How is that can be possible?
>
> Thank you
>
> 
> This message is confidential and is for the sole use of the intended
> recipient(s). It may also be privileged or otherwise protected by copyright
> or other legal rules. If you have received it by mistake please let us know
> by reply email and delete it from your system. It is prohibited to copy
> this message or disclose its content to anyone. Any confidentiality or
> privilege is not waived or lost by any mistaken delivery or unauthorized
> disclosure of the message. All messages sent to and from Agoda may be
> monitored to ensure compliance with company policies, to protect the
> company's interests and to remove potential malware. Electronic messages
> may be intercepted, amended, lost or deleted, or contain viruses.
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: CephFS 16.2.10 problem

2024-11-24 Thread Alexey.Tsivinsky
Good afternoon

We tried to leave only one mds, stopped others, even deleted one, and turned 
off the requirement for stand-by mds. Nothing helped, mds remained in the 
status of replays.
Current situation: we now have two active mds in the status of replays, and one 
in stand-by.
At the same time, in the logs we see a message
mds.0.660178  waiting for osdmap 123138 (which blocklists prior instance)
At the same time, there is no activity on both mds.
The launch of the cephfs-journal-tool journal inspect utility does not produce 
any results - the utility worked for 12 hours and did not produce anything, we 
stopped it.

Maybe the problem is this blocking? How to remove it?

Best regards!

Alexey Tsivinsky
e-mail: a.tsivin...@baikalelectronics.com

От: Marc 
Отправлено: 25 ноября 2024 г. 1:47
Кому: Цивинский Алексей Александрович; ceph-users@ceph.io
Тема: RE: CephFS 16.2.10 problem

>
> The following problem occurred.
> There is a cluster ceph 16.2.10
> The cluster was operating normally on Friday. Shut down cluster:
> -Excluded all clients
> Executed commands:
> ceph osd set noout
> ceph osd set nobackfill
> ceph osd set norecover
> ceph osd set norebalance
> ceph osd set nodown
> ceph osd set pause
> Turned off the cluster, checked server maintenance.
> Enabled cluster. He gathered himself, found all the nodes, and here the
> problem began. After all OSD went up and all pg became available, cephfs
> refused to start.
> Now mds are in the replay status, and do not go to the ready status.
> Previously, one of them was in the replay (laggy) status, but we
> executed command:  ceph config set mds mds_wipe_sessions true
> After that, mds switched to the status of replays, the third in standby
> status started, and mds crashes with an error stopped.
> But cephfs is still unavailable.
> What else can we do?
> The cluster is very large, almost 200 million files.
>

I assume you tried to start just one mds and wait until it would come up as 
active (before starting the others)?




___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Separate gateway for bucket lifecycle

2024-11-24 Thread Szabo, Istvan (Agoda)
Hi,

I remember some discussion that someone is using separated gateways for bucket 
lifecycles but I couldn't find it.

How is that can be possible?

Thank you


This message is confidential and is for the sole use of the intended 
recipient(s). It may also be privileged or otherwise protected by copyright or 
other legal rules. If you have received it by mistake please let us know by 
reply email and delete it from your system. It is prohibited to copy this 
message or disclose its content to anyone. Any confidentiality or privilege is 
not waived or lost by any mistaken delivery or unauthorized disclosure of the 
message. All messages sent to and from Agoda may be monitored to ensure 
compliance with company policies, to protect the company's interests and to 
remove potential malware. Electronic messages may be intercepted, amended, lost 
or deleted, or contain viruses.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io