[ceph-users] Re: How to improve write latency on R3 pool Images

2025-08-01 Thread Devender Singh
; On Aug 1, 2025, at 11:03 PM, Devender Singh wrote: > > No all’s are ssd and none > > Regards > Dev > > > On Fri, 1 Aug 2025 at 7:58 PM, Anthony D'Atri > wrote: > >> What is this the output of? Is this reading from HDDs? >> >> > On Aug 1,

[ceph-users] How to improve write latency on R3 pool Images

2025-08-01 Thread Devender Singh
Hello All Using R3 pool with images running ubuntu vms. Having three nodes… I have default settings for ms_asyc threads to 3 and rbd_op_thread to 1 How to improve below latency…mainly write # sed 's/^.\{60\}//' a WRRDWR_BYTESRD_BYTES WR_LAT RD_LAT 50/s 0/s

[ceph-users] Re: bdev_ioring -- true - Drives failing

2025-07-31 Thread Devender Singh
e pr. Although it seems to be present in 19.2.3. I'm not quite > sure if there is a workaround for this in the meantime. > > > On Thu, 24 Jul 2025 at 11:03, Devender Singh <mailto:deven...@netskrt.io>> wrote: >> Thanks for your reply… >> >> Currently

[ceph-users] Re: MDS Mount Issue - 19.2.2

2025-07-25 Thread Devender Singh
> mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/ /mnt/cephfs -o > name=proxmoxvolfs,fs=van2-dev,secret=AQBw8X9o8b3GHRAAw58xzwqCP6Zt+WZZbjdb/A== > > Did you check you key file content (I've done this mistake with an empty > key file generated by a wrong command 🙁) > > Patrick &g

[ceph-users] Re: CEPH performance all Flash lower than local

2025-07-24 Thread Devender Singh
Hello Thanks for your reply… Just to clear.. currently no load, nothing running on ceph… the fit test was done on a fresh installed vm to have a fit test output for reference before deploying any further vms.. Below is the bench result, yes its replicated pool… # rados bench -p vms-os-r3

[ceph-users] Re: bdev_ioring -- true - Drives failing

2025-07-24 Thread Devender Singh
Thanks for your reply… Currently I disabled that option,, but below are the logs…. # zcat ceph-osd.*.log.1.gz |egrep -i "faile|error|down" 2025-07-11T17:57:27.303+ 7f2aeacb4680 4 rocksdb: Options.error_if_exists: 0 2025-07-11T17:57:27.303+ 7f2aeacb4680 4 rocks

[ceph-users] bdev_ioring -- true - Drives failing

2025-07-24 Thread Devender Singh
Hello all Using 19.2.2, when enabling bdev_ioring and rebooting host then osd never comes online as below…. Why? Also Dasboard not matching with cli? Host output in dashboard… showing running. Tried failing manager too but it same.  Dashboard showing down  # ceph health detail HEALTH

[ceph-users] CEPH performance all Flash lower than local

2025-07-23 Thread Devender Singh
Hello all I tried doing some fio test on local disk(NVME) and ceph rbd. Why ceph is having low IO whereas it’s also on all NVME. What to tune to reach equal amount of IO? root@node01:~/fio-cdm# python3 fio-cdm ./ tests: 5, size: 1.0GiB, target: /root/fio-cdm 6.3GiB/64.4GiB |Name| Read

[ceph-users] Re: MDS Mount Issue - 19.2.2

2025-07-22 Thread Devender Singh
Dev > On Jul 22, 2025, at 1:19 PM, Devender Singh wrote: > > Thanks Patrick > > But still the same issue > > # mount -t ceph > 10.101.1.246,10.101.1.247,10.101.1.248:/volumes/_nogroup/test/e393e142-977f-4d4a-817d-41bb721cb09c >/mnt/cephfs -o name=proxmoxv

[ceph-users] 19.2.2. OSDs are up but still showing error on daemon

2025-07-22 Thread Devender Singh
Hello all Need some help to clean up below… # ceph orch ps |grep -v running NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID osd.0

[ceph-users] Re: MDS Mount Issue - 19.2.2

2025-07-22 Thread Devender Singh
my keefile is /etc/ceph/ceph.client.sebfolder.keyring > * my cephfs volume is seb_vol > > >> >> It works >> >> Patrick >> >> Le 21/07/2025 à 23:32, Devender Singh a écrit : >>> Hello Dario >>> >>> Thanks for your reply. >

[ceph-users] Re: MDS Mount Issue - 19.2.2

2025-07-21 Thread Devender Singh
-t ceph admin@.van2-dev=/ /cephfs/ -o > noatime,ms_mode=prefer-crc,_netdev > > Regards. > > [1] > https://www.google.com/url?q=https://docs.ceph.com/en/squid/man/8/mount.ceph/&source=gmail-imap&ust=175345708700&usg=AOvVaw3RPvx2e7p7nZypeiYOOf9c > > On Wed,

[ceph-users] Re: MDS Mount Issue - 19.2.2 - seems its Bug

2025-07-15 Thread Devender Singh
On Jul 15, 2025, at 4:02 PM, Devender Singh wrote: > > Hello > > > Tried creating EC pool, still facing same issue… > > https://bugzilla.redhat.com/show_bug.cgi?id=2255030 > > ceph osd erasure-code-profile set ec-21-profile-isa k=2 m=1 > crush-failure-domain=ho

[ceph-users] Re: MDS Mount Issue - 19.2.2 - seems its Bug

2025-07-15 Thread Devender Singh
e=admin,fs=van2devcephfs # mount -t ceph :/ /new_cephfs -o name=admin,fs=van2devcephfs mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you may not be authorized Regards Dev > On Jul 15, 2025, at 3:08 PM, Devender Singh wrote: > > Tried below too but

[ceph-users] Re: MDS Mount Issue - 19.2.2

2025-07-15 Thread Devender Singh
-t ceph :/ /cephfs/ -o noatime,ms_mode=prefer-crc,_netdev,fs=van2-dev mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you may not be authorized Regards Dev > On Jul 15, 2025, at 2:17 PM, Devender Singh wrote: > > Hello all > > I have cephfs wo

[ceph-users] MDS Mount Issue - 19.2.2

2025-07-15 Thread Devender Singh
Hello all I have cephfs working fine with running ads but still not able to mount, getting error mount error: no mds (Metadata Server) is up Below are details # ceph osd dump | grep min_compat_client require_min_compat_client reef min_compat_client reef # ceph balancer status { "active"

[ceph-users] MDS Client Request Load

2025-07-11 Thread Devender Singh
Hello How to lower this load? Cephfs is on EC pool sitting over replicated data and meta.. What settings can improve it?  Regards Dev ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Recovery Not happening..

2025-07-10 Thread Devender Singh
Sorry ! Yes its 3+2, Now I remember still one is not yet replaced which went down… Regards Dev > On Jul 10, 2025, at 3:51 PM, Peter Eisch wrote: > > ill likely be juggling % to get you ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscri

[ceph-users] Recovery Not happening..

2025-07-10 Thread Devender Singh
Hello all Need some help why nothing is happening… version is 19.2.2. Tried repairing the pgs, repair and force-backfill and force-recovery but nothing initiating backfills. # ceph -v ceph version 19.2.2 # ceph -s cluster: id: 79d4fade-ffa9-11eb-a57d-f398f000cfc1 health: HEAL

[ceph-users] Re: Disk Utilisation Graph

2025-07-02 Thread Devender Singh
Hello Peter I hope you are not keeping images on EC Pool? In my case my all data and images majorly on ec pools Regards Dev On Wed, 2 Jul 2025 at 1:29 PM, Anthony D'Atri wrote: > We need to see `ceph osd crush rule dump` and `ceph osd pool ls detail` to > see which pools are using which CRUSH r

[ceph-users] Incomplete PG's

2025-06-23 Thread Devender Singh
Hello Seeking some help to recover these incomplete pgs… I made osd, down then tried but nothing seems working …. Ceph version 18.2.7, we have multiple disk failed in ceph cluster on different nodes… for i in 19.5a 19.5d 19.5e 19.6b 19.6f 19.74 19.7b 34.48 34.69; do ceph pg map $i | grep

[ceph-users] Re: CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%

2025-06-22 Thread Devender Singh
ght one to query an OSD for the > cluster network, it just shows the public address of an OSD (like a client > would need to). Just use 'ceph osd dump' and look at the OSD output. > > > Zitat von Devender Singh : > >> Hello >> >> I checked on my al

[ceph-users] Re: CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%

2025-06-22 Thread Devender Singh
s not the right one to query an OSD for the > cluster network, it just shows the public address of an OSD (like a client > would need to). Just use 'ceph osd dump' and look at the OSD output. > > > Zitat von Devender Singh : > >> Hello >> >> I chec

[ceph-users] Re: CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%

2025-06-22 Thread Devender Singh
Hello I checked on my all clusters everywhere OSD’s not using cluster network. Here is another example in my lab where I have three hosts in vlan1 and vlan2 running on one proxmox server and here also same thing… No MTU change, these are default to 1500. I don’t understand what I am missing?

[ceph-users] Re: CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%

2025-06-21 Thread Devender Singh
ph orch device ls --hostname=phl-prod-host04n.example.comis --refresh > > and see if the upgrade resumes. It if doesn't, check > > $ ceph log last 1000 debug cephadm > > and run > > $ ceph orch upgrade pause > $ ceph orch upgrade resume > > again, see if the upg

[ceph-users] OSD network Issue -- Not using cluster Network

2025-06-21 Thread Devender Singh
Hello All Found some network issue on OSD’s, why they are not using cluster network? Only using public network? My mons are on public network… How can I resolve it? Changing this network will lead any issue like need to stop all osd’s or restart cluster or rolling restart osd’s & how, please

[ceph-users] Re: CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%

2025-06-20 Thread Devender Singh
ou need to compact OSD. >> >> I am not sure what you adjusted but as long as the cluster works, I would >> not have changed parameters and try to fix the mentioned problems. >> >> Good luck, >> >> Michel >> Sent from my mobile >&

[ceph-users] Re: Cluster Not recovering after multiple disk failure

2025-06-19 Thread Devender Singh
Thanks for your reply.. There were multiple OSD’s failure on different hosts and had to zap and read. Which lead us to this issue… and upgrade was running too.. Now I see three options for upgrade “Back, Pause & Stop” — is back for rollback? And which one I should use the best option…. Tried r

[ceph-users] Re: CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%

2025-06-19 Thread Devender Singh
s", "services_complete": [ "crash", "mgr", "mon" ], "progress": "74/113 daemons upgraded", "message": "", "is_paused": false } Regards Dev > On Jun 19, 2025, at 8:06 PM

[ceph-users] CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%

2025-06-19 Thread Devender Singh
Hello all I have a cluster where my cluster is in hung state, Some back fills are there but I reduced it to 1 but still upgrade not progressing… Please help… ```# ceph health detail HEALTH_WARN 8 OSD(s) experiencing slow operations in BlueStore; Failed to apply 2 service(s): osd.all-availab

[ceph-users] Re: 4k IOPS: miserable performance in All-SSD cluster

2025-06-18 Thread Devender Singh
Hello All Any resolution found for this? I am also facing same issue, with slow Ops, tried restating odd’s but no luck? Regards Dev > On Nov 26, 2024, at 3:46 AM, Sergio Rabellino wrote: > > hen I have strange (and bad) behaviours like these, I double check name > resolution/DNS configurat

[ceph-users] Cluster Not recovering after multiple disk failure

2025-06-18 Thread Devender Singh
Hello all Need urgent help on below… I tried reducing min_size but still showing same… cluster: id: 15688cb4-044a-11ec-942e-516035adea04 health: HEALTH_ERR 3 failed cephadm daemon(s) 1 filesystem is degraded 1 MDSs report slow metadata IOs

[ceph-users] 18.2.7 - IOPS issue

2025-06-16 Thread Devender Singh
Dear All I upgrade my clusters from 18.2.4 & 18.2.6 to 18.2.7. 1. I started facing multiple disks went down.. 2. Fio test showing 5-35 IOPS on cephfs? $ fio --name=latency-test --ioengine=libaio --rw=randread --bs=4k --size=512M --numjobs=1 --iodepth=1 --direct=1 --runtime=60

[ceph-users] Re: Scrubbing + Deep crubbing

2025-05-15 Thread Devender Singh
May 2025 at 9:14 AM, Devender Singh wrote: > Hello All > > > Need some help to tune scrubbing. > > 1. *How to control not run scrubbing and deep scrubbing together and > which one should be started first?* > > * 124 active+clean+remapped* > > *

[ceph-users] Slow osd + 30% deviates (toward higher pgs or lower pgs numbers on osd)

2025-05-12 Thread Devender Singh
Hello all Seeking some help to solve below… 1. # ceph health detail HEALTH_WARN 2 OSD(s) experiencing slow operations in BlueStore [WRN] BLUESTORE_SLOW_OP_ALERT: 2 OSD(s) experiencing slow operations in BlueStore osd.10 observed slow operation indications in BlueStore osd.21 observe

[ceph-users] Scrubbing + Deep crubbing

2025-05-12 Thread Devender Singh
Hello All Need some help to tune scrubbing. 1. How to control not run scrubbing and deep scrubbing together and which one should be started first? 124 active+clean+remapped 35 active+clean+scrubbing+deep 24 active+clean+scrubbing 4acti

[ceph-users] Re: 18.2.6 : OSD(s) experiencing slow operations in BlueStore

2025-05-02 Thread Devender Singh
Sorry some typo. It is mclock. Exact parameter osd_recovery_max_active_ssd/hdd. Is 10, to reduce you have to override mclock to true. Restarting osd daemon alone will solve your issue. Regards Dev On Fri, 2 May 2025 at 9:07 AM, Devender Singh wrote: > Hello > > Try restarting osd

[ceph-users] Re: 18.2.6 : OSD(s) experiencing slow operations in BlueStore

2025-05-02 Thread Devender Singh
Hello Try restarting osds showing slow ops. Also if any recovery going on then max recovery drives for Malcom is 10 try reducing it. Will resolve this issue. If it persist for a drive then check for smart TK for errors and replace that drive. Regards Dev On Fri, 2 May 2025 at 8:45 AM, Maged Mokh

[ceph-users] Re: Heads up -- 18.2.6 regression

2025-05-01 Thread Devender Singh
Hello Alex As per Dan bug is in Reef which is v18.2.6 in open ceph. I also upgraded my cluster to 18.2.6 before I saw the first message of this mail chain and I am having 120osds but yet I have not seen any issue where’s as one of my host remain down for 24hrs with 24 osds on it and now I joined

[ceph-users] S3 Bucket Upload - Boto3 - Disable or Enable Checksum on bucket/rgw

2025-02-25 Thread Devender Singh
Hello Using boto3 trying to upload files to s3 bucket, when I am using boto version 1.35 it works fine, when upgrading it to 1.36 doesn’t work, facing below error, python3 s3_upload.py Error testing the new bucket: An error occurred (XAmzContentSHA256Mismatch) when calling the PutObject opera

[ceph-users] Ceph calculator

2025-02-13 Thread Devender Singh
Hello all Do we have a good cluster design calculator which can suggest failure domain and pool size and min size according the number of nodes and drive and their size for both replicated and EC pools. Regards Dev ___ ceph-users mailing list -- ceph-u

[ceph-users] Grafana certificate issue

2025-02-11 Thread Devender Singh
Hello all Recently upgraded to 19.2.1 and started facing .. Invalid grafana certificate on hostnode1: Invalid certificate key: [('PEM routines', '', 'no start line’)] Tried suggest steps, even re-deployed grafana but no solution? 1. [WRN] CEPHADM_CERT_ERROR: Invalid grafana certific

[ceph-users] Re: v19.2.1 Squid released

2025-02-06 Thread Devender Singh
👌🥳🕺🕺 On Thu, 6 Feb 2025 at 2:56 PM, Yuri Weinstein wrote: > We're happy to announce the 1st backport release in the Squid series. > > https://ceph.io/en/news/blog/2025/v19-2-1-squid-released/ > > Notable Changes > --- > * CephFS: The command `fs subvolume create` now allows tagging >

[ceph-users] Re: EC pool - Profile change

2025-02-05 Thread Devender Singh
Hello all Thanks for your reply. I already tried few things.. 1. Tried deleting the old ec profile and thought to create it with same name but with osd failure domain, it did not allow deleting from pool. 2. Changed crush rule with osd too. It reverted automatically to host. 3. Then Create

[ceph-users] Backfills Not Increasing.

2025-02-05 Thread Devender Singh
Hello all I tried disabling osd_mclock_override_recovery_settings to true and increased backfill from 1 to 15 after disabling scrub and deep-scrub. There was no change I started from 1 then 3 then 5 then 7 then 10 then 13 and 15. But no change. So it a default behaviour or bug? # ceph config

[ceph-users] EC pool - Profile change

2025-02-04 Thread Devender Singh
Hello all Is there a way to change ec pool profile from host to osd? Regrds Dev ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Error ENOENT: Module not found

2025-01-31 Thread Devender Singh
uot; "config-history/5150/+client.ceph-exporter.node03/container_image" "config-history/5156/-client.ceph-exporter.node03/container_image" "config-history/5179/+osd/host:node03/osd_memory_target" "config-history/5179/-osd/host:node03/osd_memory_target&qu

[ceph-users] Re: Error ENOENT: Module not found

2025-01-31 Thread Devender Singh
Please ignore, worked after setting blank file with []. Regards Dev > On Jan 31, 2025, at 10:55 PM, Devender Singh wrote: > > Hello > > Need some help.. > > Tried draining host but it got stuck and now orchestrator is not running. > But cluster health is OK

[ceph-users] Error ENOENT: Module not found

2025-01-31 Thread Devender Singh
Hello Need some help.. Tried draining host but it got stuck and now orchestrator is not running. But cluster health is OK , also I adde host back. # ceph health detail HEALTH_OK Tried adding blank queue file,, ceph config-key set mgr/cephadm/osd_remove_queue -I osd_remove_queue_blank.json

[ceph-users] Suggestions

2025-01-31 Thread Devender Singh
Hello all Few more suggestions.. if can be added to further releases. 1. We faced some issue, can we add more command to control clients using watcher, rbd status pool/image Watchers: watcher=10.160.0.245:0/2076588905 client.12541259 cookie=140446370329088 Some commands to control w

[ceph-users] Re: OSD latency

2025-01-29 Thread Devender Singh
Hello Please find attached again..  Regards Dev > On Jan 29, 2025, at 10:58 AM, Afreen Misbah wrote: > > Hi Devender, > > Seems like the graphs (screenhots) are not visible. Can you attach those > again ? > > Thanks > > Afreen > > On Thu, Jan

[ceph-users] OSD latency

2025-01-29 Thread Devender Singh
Hello Friends Seeking some help on OSD write latency in "osd performance graph" whereas "osd graph on dashboard" is not showing any high latency. So, is dashboard graph showing only read latency as I can see low latency in osd perf graph which matches the dashboard osd perf graph? OSD per

[ceph-users] Re: Error ENOENT: Module not found

2025-01-25 Thread Devender Singh
to add it for > the sake of completeness. I strongly suggest to re-add the fifth host (and > think about adding a sixth). > > Regards, > Eugen > > Zitat von Devender Singh mailto:deven...@netskrt.io>>: > >> +Eugen >> Lets follow “No recovery after removing

[ceph-users] Re: Error ENOENT: Module not found

2025-01-25 Thread Devender Singh
+Eugen Lets follow “No recovery after removing node - active+undersized+degraded-- removed osd using purge…”. Here. Sorry I missed ceph version which is 18.2.4. (with 5 nodes, 22osd each, where I removed one node and all mess.) Regards Dev > On Jan 25, 2025, at 11:34 AM, Devender Si

[ceph-users] Re: Watcher Issue

2025-01-25 Thread Devender Singh
Hello Eugen Thanks for your reply. ceph osd set nodeep-scrub is not stopping if repairs are running. Reapir started another set for deepscrub+repair which is not controlled using this command. When I started my cluster utilization as 74% and when it finished now my cluster is showin

[ceph-users] Re: Error ENOENT: Module not found

2025-01-25 Thread Devender Singh
t;, "config-history/9710/-osd/host:node3/osd_memory_target", "config/osd/host:node3/osd_memory_target”, Regards Dev > On Jan 25, 2025, at 4:39 AM, Frédéric Nass > wrote: > > Hi, > > I've seen this happening on a test cluster after draining a host

[ceph-users] No recovery after removing node - active+undersized+degraded-- removed osd using purge...

2025-01-24 Thread Devender Singh
Hello all Urgent help needed. No recovery happening. Tried repairing pg and redeploy or create. Rebooted cluster but no luck.. data: volumes: 2/2 healthy pools: 18 pools, 817 pgs objects: 6.06M objects, 20 TiB usage: 30 TiB used, 302 TiB / 332 TiB avail pgs: 284

[ceph-users] Re: Watcher Issue

2025-01-24 Thread Devender Singh
Hello all Sorry for late reply.. I tried two things, 1. My cluster was using swap, I made swap off. 2. I started repair command on pool/image and seems it worked . But after that there was no command found to pause repair as it started deep scrubs too. How to unpause the repairs… Regard

[ceph-users] Re: Error ENOENT: Module not found

2025-01-24 Thread Devender Singh
big error is Error ENOENT: Regards Dev > On Jan 24, 2025, at 4:59 PM, Fnu Virender Kumar > wrote: > > Did you try > > Ceph mgr module enable orchestrator > Ceph orch set backend > Ceph orch ls > > Check the mgr service daemon as well > Ceph -s > &

[ceph-users] Error ENOENT: Module not found

2025-01-24 Thread Devender Singh
Hello all Any quick fix for … root@sea-devnode1:~# ceph orch ls Error ENOENT: Module not found Regards Dev ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Watcher Issue

2025-01-22 Thread Devender Singh
9 - /dev/rbd0 > sh-4.4# umount /dev/rbd/k8s-rgnl-disks/csi-vol-945c6a66-9129 > sh-4.4# rbd unmap /dev/rbd/k8s-rgnl-disks/csi-vol-945c6a66-9129 > sh-4.4# rbd device list > sh-4.4# > > Hope there's no typo. > > Regards, > Frédéric. > > - Le 21 Jan 25,

[ceph-users] Re: Seeking Participation! Take the new Ceph User Stores Survey!

2025-01-22 Thread Devender Singh
Hello Laura Few more suggestions.. 1. As we are facing some issue, can we add more command to control clients using watcher, rbd status pool/image Watchers: watcher=10.160.0.245:0/2076588905 client.12541259 cookie=140446370329088 Some commands to control watcher and kill client.id

[ceph-users] Re: Watcher Issue

2025-01-22 Thread Devender Singh
Dear all Sending again. Our mail server was having some issue, not sure if you have received it or not. Exactly Similar issue, https://github.com/ceph/ceph-csi/discussions/4410 Regards Dev > On Jan 21, 2025, at 2:33 PM, Devender Singh wrote: > > Hello Eugen > > Thanks

[ceph-users] Re: Watcher Issue

2025-01-21 Thread Devender Singh
Similar issue https://github.com/ceph/ceph-csi/discussions/4410 Regards Dev On Tue, 21 Jan 2025 at 2:33 PM, Devender Singh wrote: > Hello Eugen > > Thanks for your reply. > I have the image available and it’s not under trash. > > When scaling a pod to different node usin

[ceph-users] Re: Watcher Issue

2025-01-21 Thread Devender Singh
list the client > to release the watcher, then delete the image again. > > I have to do that from time to time on a customer’s openstack cluster. > > Zitat von Devender Singh : > >> Hello >> >> Seeking some help if I can clean the client mounting my volume? >>

[ceph-users] Watcher Issue

2025-01-21 Thread Devender Singh
Hello Seeking some help if I can clean the client mounting my volume? rbd status pool/image Watchers: watcher=10.160.0.245:0/2076588905 client.12541259 cookie=140446370329088 Issue: pod is failing in init- state. Events: Type Reason Age From Message

[ceph-users] Re: Changing crush map result in > 100% objects degraded

2025-01-21 Thread Devender Singh
You moved some oSD’s I believe it’s looking for peer osds data too. But as long as you keep nobackfill,norebalance,norecover it will take longer and keep showing more data to balance, and kept filling or writing data to volumes it will accumulate. So unset and wait for some time to finish it. Rega

[ceph-users] Cephfs mirroring --

2024-09-20 Thread Devender Singh
Hello all I tried configuring cephfs mirroring between cluster as per doc https://docs.ceph.com/en/reef/dev/cephfs-mirroring/ using bootstrap. But My replication is not working. Using two mirror daemons on each cluster. [Enabled module/mirroring/snap mirroring/aded directory/allow_new_snaps]

[ceph-users] Re: squid release codename

2024-08-15 Thread Devender Singh
Few More If you like. Metaverse Jumbo Regards Dev > On Aug 15, 2024, at 10:12 AM, Tarrago, Eli (RIS-BCT) > wrote: > > Additional names that could be considered: > > Sea Monster > Sea Bear > Star (for Patrick star) > > And for the best underwater squirrel friend, Sandy. > > > From: Boris

[ceph-users] Re: 18.2.4 regression: 'diskprediction_local' has failed: No module named 'sklearn'

2024-08-01 Thread Devender Singh
What is the fix for above error. My OS is : ubuntu 20.04 and ceph version is 18.2.4 Currently I disabled the prediction. : ceph mgr module disable diskprediction_local Regards Dev ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send