[ceph-users] list cephfs dirfrags

2023-11-06 Thread Ben
Hi, I used this but all returns "directory inode not in cache" ceph tell mds.* dirfrag ls path I would like to pin some subdirs to a rank after dynamic subtree partitioning. Before that, I need to know where are they exactly Thank you, Ben __

[ceph-users] Re: list cephfs dirfrags

2023-11-08 Thread Ben
Hi, this directory is very busy: ceph tell mds.* dirfrag ls /volumes/csi/csi-vol-3a69d51a-f3cd-11ed-b738-964ec15fdba7/ while running it, all mds output: [ { "value": 0, "bits": 0, "str": "0/0" } ] Thank you, Ben Patrick

[ceph-users] one cephfs volume becomes very slow

2023-11-08 Thread Ben
Dear cephers, we have a cephfs volume, that will be mounted by many clients with concurrent read/write capability. From time to time, maybe when concurrency goes as high as 100 clients' access, accessing it will become very slow to be useful at all. the cluster has multiple active mds. All disks a

[ceph-users] Re: one cephfs volume becomes very slow

2023-11-10 Thread Ben
e was a recent thread [1] where Zakhar explained nicely how many > IOPS you can expect from a hdd-only cluster. Maybe that helps. > > [1] > > https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/FPMCNPYIHBHIJLWVVG2ECI2DSTR6DZIO/ > > Zitat von Ben : > > > Dear ce

[ceph-users] Re: mds slow request with “failed to authpin, subtree is being exported"

2023-11-26 Thread Ben
with the same mds configuration, we see exactly the same(problem, log and solution) with 17.2.5, constantly happening again and again in couples days intervals. MDS servers are stuck somewhere, ceph status reports no issue however. We need to restart some of the mds (if not all of them) to restore

[ceph-users] Quincy patch upgrade

2023-11-26 Thread Ben
Dear Cephers, With improvement on dmclock in 17.2.7, we are considering upgrade from 17.2.5 to 17.2.7. seeing this which is worrysome: https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/SG7CKALU3AIWEIVN7QENIY3KRETUQKM7/ any suggestions for an easy upgrade? Regards, Ben

[ceph-users] rook module not working with Quincy 17.2.3

2022-10-23 Thread Ben
ssl.c:897)'),)) debug 2022-10-21T09:07:22.079+ 7f019f728700 0 [dashboard INFO dashboard.module] Engine started. It looks like the rook module complains the k8s api server has a self sighed certification. Any configuration can be done to resolve the issue? Thanks, Ben ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] nfs RGW export makes nfs-gnaesha server in crash loop

2023-01-16 Thread Ben
ceph nfs export delete nfs4rgw /bucketexport Ganesha servers go back normal: rook-ceph-nfs-nfs1-a-679fdb795-82tcx 2/2 Running 0 4h30m rook-ceph-nfs-nfs4rgw-a-5c594d67dc-nlr42 2/2 Running 10 4h33m Any ideas to make it work? Thanks

[ceph-users] restoring ceph cluster from osds

2023-03-04 Thread Ben
d that in the course. Thanks, Ben osd0 log: debug 2023-03-04T04:41:50.620+ 7f824d7043c0 0 set uid:gid to 167:167 (ceph:ceph) debug 2023-03-04T04:41:50.620+ 7f824d7043c0 0 ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable), process ceph-osd, pid 1 debug 202

[ceph-users] restoring ceph cluster from osds

2023-03-07 Thread Ben
restore that in the course. Thanks, Ben ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: restoring ceph cluster from osds

2023-03-09 Thread Ben
initialize The fact is original monitor clusters ids are k,m,o, however the new ones are a,b,d. It was deployed by rook. Any ideas to make this work? Ben Eugen Block 于2023年3月9日周四 16:00写道: > Hi, > > there's no attachment to your email, please use something like > pastebin to

[ceph-users] Re: restoring ceph cluster from osds

2023-03-09 Thread Ben
I could start a mon daemon cluster with exact ID as original one( something like k,m,o). Any thoughts on this? Ben Eugen Block 于2023年3月9日周四 20:56写道: > Hi, > > I'm not familiar with rook so the steps required may vary. If you try > to reuse the old mon stores you'll have

[ceph-users] ERROR: Distro uos version 20 not supported

2023-04-26 Thread Ben
d.service', 'ntpd.service', 'ntp.service', 'ntpsec.service', 'openntpd.service'] ERROR: Distro uos version 20 not supported uname -a Linux 4.19.0-91.82.42.uelc20.x86_64 #1 SMP Sat May 15 13:50:04 CST 2021 x86_64 x86_64 x86_64 GNU/Linux Thank you in

[ceph-users] non root deploy ceph 17.2.5 failed

2023-05-08 Thread Ben
) File "/lib/python3.6/site-packages/asyncssh/scp.py", line 343, in _make_cd_request self._fs.basename(path)) File "/lib/python3.6/site-packages/asyncssh/scp.py", line 224, in make_request raise exc Any ideas on this? Thanks, Ben ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: non root deploy ceph 17.2.5 failed

2023-05-08 Thread Ben
. Ben Eugen Block 于2023年5月8日周一 21:01写道: > Hi, > > could you provide some more details about your host OS? Which cephadm > version is it? I was able to bootstrap a one-node cluster with both > 17.2.5 and 17.2.6 with a non-root user with no such error on openSUSE > Leap 15.4: >

[ceph-users] Re: ERROR: Distro uos version 20 not supported

2023-05-08 Thread Ben
Get one of the time services up and running, you then get through this. The error message is quite of misleading? Ben 于2023年4月26日周三 15:07写道: > Hi, > This seems not very relevant since all ceph components are running in > containers. Any ideas to get over this issue? Any other

[ceph-users] Re: non root deploy ceph 17.2.5 failed

2023-05-09 Thread Ben
appen when > they had changed the umask to 027. We had patched in > https://github.com/ceph/ceph/pull/50736 to address it, which I don't > think was merged too late for the 17.2.6 release. > > On Mon, May 8, 2023 at 5:24 AM Ben wrote: > >> Hi, >> >> with foll

[ceph-users] Re: non root deploy ceph 17.2.5 failed

2023-05-09 Thread Ben
rtunately, I > can't use the curl-based cephadm but the contents are identical, I > compared. Not sure what it could be at the moment. > > Zitat von Ben : > > > Hi, It is uos v20(with kernel 4.19), one linux distribution among others. > > no matter since cephadm depl

[ceph-users] docker restarting lost all managers accidentally

2023-05-10 Thread Ben
mgr up running again through cephadm? Thanks, Ben ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: docker restarting lost all managers accidentally

2023-05-11 Thread Ben
hem for obvious reasons (these are just on a test cluster > I'll tear back down so it's fine for me) but those are the first couple > things I'd check. You could also try to make adjustments directly to the > unit.run file if you have other things you'd like to

[ceph-users] Re: Grafana service fails to start due to bad directory name after Quincy upgrade

2023-05-17 Thread Ben
you could check owner of /var/lib/ceph on host with grafana container running. If its owner is root, change to 167:167 recursively. Then systemctl daemon-reload and restart the service. Good luck. Ben Adiga, Anantha 于2023年5月17日周三 03:57写道: > Hi > > Upgraded from Pacific 16.2.5 to 17.2

[ceph-users] Re: Grafana service fails to start due to bad directory name after Quincy upgrade

2023-05-17 Thread Ben
use this to get relevant long lines in log: journalctl -u ceph-d0a3b6e0-d2c3-11ed-be05-a7a3a1d7a87e@grafana-fl31ca104ja0201 | less -S it is '--user 472' by content of unit.run, not the default ceph user 167. Maybe set the directory owner to 472 could help. Hope it helps Ben Adig

[ceph-users] slow mds requests with random read test

2023-05-30 Thread Ben
0 PM [INF] Health check cleared: MDS_SLOW_METADATA_IO (was: 1 MDSs report slow metadata IOs) 5/30/23 8:12:40 PM [INF] MDS health message cleared (mds.?): 1 slow metadata IOs are blocked > 30 secs, oldest blocked for 38 secs However, random write test is performing very good. Any sugg

[ceph-users] Re: slow mds requests with random read test

2023-05-31 Thread Ben
something with cephfs kernel client? Any other thoughts? Patrick Donnelly 于2023年5月31日周三 00:58写道: > On Tue, May 30, 2023 at 8:42 AM Ben wrote: > > > > Hi, > > > > We are performing couple performance tests on CephFS using fio. fio is > run > > in k8s pod an

[ceph-users] alerts in dashboard

2023-06-20 Thread Ben
Hi, I got many critical alerts in ceph dashboard. Meanwhile the cluster shows health ok status. See attached screenshot for detail. My questions are, are they real alerts? How to get rid of them? Thanks Ben ___ ceph-users mailing list -- ceph-users

[ceph-users] Re: alerts in dashboard

2023-06-25 Thread Ben
CephFS filesystem is degraded CephFilesystemFailureNoStandby critical mds 1 minute MDS daemon failed, no further standby available Meanwhile the cluster status is green ok. What should we do for this? Thanks, Ben ___ ceph-users mailing list -- ceph-users

[ceph-users] MDSs report slow metadata IOs

2023-07-07 Thread Ben
Hi, see many of this in cluster log channel. many are blocked with long period of seconds. It should hurt client access performance. Any ideas to get rid of them? Thanks, Ben - 7/7/23 4:48:50 PM [WRN] Health check update: 8 MDSs report slow metadata IOs

[ceph-users] mon log file grows huge

2023-07-10 Thread Ben
Hi, In our cluster monitors' log grows to couple GBs in days. There are quite many debug message from rocksdb, osd, mgr and mds. These should not be necessary with a well-run cluster. How could I close these logging? Thanks, Ben ___ ceph-users ma

[ceph-users] Re: mon log file grows huge

2023-07-10 Thread Ben
just rechecked debug_mon is by default 1/5. mgr/cephadm log_to_cluster level has been set to critical from debug. Wonder how to set others' level. Haven't got a clue to do that. Thanks, Ben Wesley Dillingham 于2023年7月10日周一 23:21写道: > At what level do you have logging set to for yo

[ceph-users] question about OSD onode hits ratio

2023-08-02 Thread Ben
Hi, We have a cluster running for a while. From grafana ceph dashboard, I saw OSD onode hits ratio 92% when cluster was just up and running. After couple month, it says now 70%. This is not a good trend I think. Just wondering what should be done to stop this trend. Many thank, Ben

[ceph-users] outdated mds slow requests

2023-09-24 Thread Ben
Hi, It is running 17.2.5. there are slow requests warnings in cluster log. ceph tell mds.5 dump_ops_in_flight, get the following. These look like outdated and clients were k8s pods. There are warning of the kind in other mds as well. How could they be cleaned from warnings safely? Many thanks.

[ceph-users] cephfs health warn

2023-09-26 Thread Ben
. It looks like bugs with client library. And rebooting nodes with problem will fix it for short period of time? Any suggestions from community for fixing? Thanks, Ben [root@8cd2c0657c77 /]# ceph health detail HEALTH_WARN 6 hosts fail cephadm check; 2 clients failing to respond to capability

[ceph-users] Re: cephfs health warn

2023-09-27 Thread Ben
195341004/893374309813 to expire debug 2023-09-27T15:27:33.024+ 7f8f047cf700 10 mds.5.log _trim_expired_segments waiting for 189942575/642272326541 to expire Regards, Ben Ben 于2023年9月26日周二 20:31写道: > Hi, > see below for details of warnings. > the cluster is running 17.2.5. the warn

[ceph-users] Re: cephfs health warn

2023-09-28 Thread Ben
definitely interrupt business. Any thoughts to stop the warnings? Best wishes, Ben Venky Shankar 于2023年9月28日周四 11:56写道: > Hi Ben, > > On Tue, Sep 26, 2023 at 6:02 PM Ben wrote: > > > > Hi, > > see below for details of warnings. > > the cluster is running 17.2.5. th

[ceph-users] Re: cephfs health warn

2023-10-02 Thread Ben
Dear cephers: more log captures(see below) show the full segments list(more than 3 to be trimmed stuck, growing over time). any ideas to get out of this? Thanks, Ben debug 2023-09-30T14:34:14.557+ 7f9c29bb1700 5 mds.4.log trim already expiring segment 195341004/893374309813, 180 events

[ceph-users] Re: cephfs health warn

2023-10-03 Thread Ben
Yes, I am. 8 active + 2 standby, no subtree pinning. What if I restart the mds with trimming issues? Trying to figure out what happens with restarting. Venky Shankar 于2023年10月3日周二 12:39写道: > Hi Ben, > > Are you using multimds without subtree pinning? > > On Tue, Oct 3, 2023

[ceph-users] Re: cephfs health warn

2023-10-03 Thread Ben
egment list continually grows overtime. Any ideas for having the segment list to be processed well normal again? The issue has been around for weeks and haven't seen complaints from storage client side so far. Best wishes, Ben Venky Shankar 于2023年10月4日周三 13:31写道: > Hi Ben, > > On T

[ceph-users] Re: outdated mds slow requests

2023-10-04 Thread Ben
draw concerns overtime. Any other ideas? Thanks, Ben Eugen Block 于2023年10月4日周三 16:44写道: > Hi, > > is this still an issue? If so, I would try to either evict the client > via admin socket: > > ceph tell mds.5 client evict [...] --- Evict client > session(s) based on a filter &g

[ceph-users] Re: outdated mds slow requests

2023-10-10 Thread Ben
Hi, It get cleared by restarting ceph client with issues. It works. to do that, you would umount problematic cephfs volume and remount. All ceph warning is gone in couple minutes, trimming well now. Indeed I wouldn't restart mds unless I had to. Many thanks for help, Ben Eugen Block 于2023年

[ceph-users] slow recovery with Quincy

2023-10-10 Thread Ben
, deployed by cephadm. The slowness can do harm in peak hours of usage. Best wishes, Ben - volumes: 1/1 healthy pools: 8 pools, 209 pgs objects: 93.04M objects, 4.8 TiB usage: 15 TiB used, 467 TiB / 482 TiB avail pgs

[ceph-users] Re: slow recovery with Quincy

2023-10-11 Thread Ben
not aware of? 胡 玮文 于2023年10月11日周三 00:18写道: > Hi Ben, > > Please see this thread > https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/PWHG6QJ6N2TJEYD2U4AXJAJ23CRPJG4E/#7ZMBM23GXYFIGY52ZWJDY5NUSYSDSYL6 > for > possible workaround. > > 发自我的 iPad > > 在 2023年1

[ceph-users] Fwd: MDS Daemon Damaged

2020-05-29 Thread Ben
I have a 3 node ceph cluster for my house that I have been using for a few years now without issue. Each node is a MON, MGR, and MDS, and has 2-3 OSDs on them. It has, however been slow. I decided to finally move the bluestore DBs to SSDs. I did one OSD as a test case to make sure everything was go

[ceph-users] Cluster inaccessible

2021-10-10 Thread Ben Timby
A colleague asked me to take a look at a Ceph cluster that has stopped working. The "ceph -s" command (any ceph command) just times out. Of three monitors two are crashing with: (gdb) bt #0 0x7fffee17b7bb in raise () from /lib/x86_64-linux-gnu/libc.so.6 #1 0x7fffee166535 in abort () f

[ceph-users] Re: Cluster inaccessible

2021-10-10 Thread Ben Timby
Thanks for your reply. I tried something similar (but wrong) based on your messages in the referenced threads, but I was missing the "gdb commands...end" sequence so I just kept hitting my breakpoint over and over. However, I was able to get the monitor running with your guidance. Indeed, ceph rm

[ceph-users] Re: Cluster inaccessible

2021-10-10 Thread Ben Timby
Yes, I have all MDS running now and I am in the process of building ceph sources with your patch from the original thread. I also tried the command: ceph fs compat prod add_incompat 7 "mds uses inline data" Since I noticed that option was defined for two of the three file systems (the working one

[ceph-users] Re: Cluster inaccessible

2021-10-10 Thread Ben Timby
I am having little luck building ceph on Debian bullseye. First of all if I install the package ninja-build and run ninja in the build directory, I get an error that ninja.build file is not found after running ./do_cmake.sh. I am thus just executing make, which is failing to compile some tests. I a

[ceph-users] Re: Cluster inaccessible

2021-10-10 Thread Ben Timby
Building without tests worked and I am now operational once again. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] OSDs failed to start after host reboot | Cephadm

2023-01-12 Thread Ben Meinhart
h (~6 months) and I haven’t emailed this list before. Sorry in advance if I’ve mistakenly broken any roles and thanks for the help! - Ben M ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] nfs RGW export makes nfs-gnaesha server in crash loop

2023-01-16 Thread Ben Gao
skip) 4, delete the export: ceph nfs export delete nfs4rgw /bucketexport Ganesha servers go back normal: rook-ceph-nfs-nfs1-a-679fdb795-82tcx 2/2 Running 0 4h30m rook-ceph-nfs-nfs4rgw-a-5c594d67dc-nlr42 2/2 Running 10 4h

[ceph-users] Massive OMAP remediation

2023-04-25 Thread Ben . Zieglmeier
Hi All, We have a RGW cluster running Luminous (12.2.11) that has one object with an extremely large OMAP database in the index pool. Listomapkeys on the object returned 390 Million keys to start. Through bilog trim commands, we’ve whittled that down to about 360 Million. This is a bucket index

[ceph-users] Re: [EXTERNAL] Re: Massive OMAP remediation

2023-04-27 Thread Ben . Zieglmeier
. Thanks again! -Ben From: Dan van der Ster Date: Wednesday, April 26, 2023 at 11:11 AM To: Ben.Zieglmeier Cc: ceph-users@ceph.io Subject: [EXTERNAL] Re: [ceph-users] Massive OMAP remediation Hi Ben, Are you compacting the relevant osds periodically? ceph tell osd.x compact (for the three osds

[ceph-users] Upgrading standard Debian packages

2023-06-14 Thread Ben Thompson
/nautilus/#upgrading-from-mimic-or-luminous the procedure is to upgrade each daemon one by one (MON -> MGR -> OSD). However, it seems that the Debian packages are all interdependent on ceph-base which means that upgrading one at a time is not possible. I would be grateful for any advice.

[ceph-users] Unable to start MDS and access CephFS after upgrade to 17.2.6

2023-06-14 Thread Ben Stöver
seem to be a solution mentioned there. https://bugzilla.redhat.com/show_bug.cgi?id=2056935 ) Below are some parts of the MDS logs that seem relevant to us for this issue. We are thankful for any ideas. :-) Best Ben Log excerpt of Active MDS (replay):   -140> 2023-06-14T07:51:59.585+ 7feb58

[ceph-users] RGWs offline after upgrade to Nautilus

2023-07-20 Thread Ben . Zieglmeier
ialization timeout to 600, and tried removing all non-essential config from ceph.conf. All produce the same results. I would greatly appreciate any other ideas or insight. Thanks, Ben ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe

[ceph-users] Re: Ceph Octopus packages missing at download.ceph.com

2024-11-14 Thread Ben Zieglmeier
I was able to get what I needed from http://mirrors.gigenet.com/ceph/ (one of the mirrors listed in the Ceph doco). On Thu, Nov 14, 2024, 6:05 AM Frank Schilder wrote: > Hi all, > > +1 from me > > this is a really bad issue. We need access to these packages very soon. > Please restore this folde