[ceph-users] Repair/Rebalance slows down

2022-01-06 Thread Ray Cunningham
Hi Everyone! I have a 16 node, 640 OSD (5 to 1 SSD) bluestore cluster which is mainly used for RGW services. It has its own backend cluster network for IO separate from the customer network. Whenever we add or remove an OSD the rebalance or repair IO starts off very fast 4GB/s+ but it will con

[ceph-users] Re: Repair/Rebalance slows down

2022-01-06 Thread Ray Cunningham
I can do that for adding new drives, but what about when removing failed disks? There’s not one OSD to run an iostat on? The performance profile is the same for adds and removes. Thank you, Ray From: David Orman Sent: Thursday, January 6, 2022 9:18 AM To: Ray Cunningham Cc: ceph-users

[ceph-users] Scrubbing

2022-03-09 Thread Ray Cunningham
stions for increasing scrub throughput? Thank you, Ray Cunningham Systems Engineering and Services Manager keepertechnology<http://www.keepertech.com/> (571) 223-7242 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email

[ceph-users] Re: Scrubbing

2022-03-10 Thread Ray Cunningham
n get answers for specific questions. Off the top of my head, we have set: osd_max_scrubs 20 osd_scrub_auto_repair true osd _scrub_load_threashold 0.6 We do not limit srub hours. Thank you, Ray     -Original Message- From: norman.kern Sent: Wednesday, March 9, 2022 7:28 PM To: Ray Cunni

[ceph-users] Re: Scrubbing

2022-03-10 Thread Ray Cunningham
We have that set to 20 at the moment. Thank you, Ray Cunningham Systems Engineering and Services Manager keepertechnology<http://www.keepertech.com/> (571) 223-7242 From: Szabo, Istvan (Agoda) Sent: Wednesday, March 9, 2022 7:35 PM To: Ray Cunningham Cc: ceph-users@ceph.io Subject: Re:

[ceph-users] Re: Scrubbing

2022-03-10 Thread Ray Cunningham
load is? Is that a ceph command, or an OS command? Thank you, Ray   -Original Message- From: Ray Cunningham Sent: Thursday, March 10, 2022 7:59 AM To: norman.kern Cc: ceph-users@ceph.io Subject: RE: [ceph-users] Scrubbing We have 16 Storage Servers each with 16TB HDDs and 2TB SSDs for

[ceph-users] Re: Scrubbing

2022-03-10 Thread Ray Cunningham
Well that was incorrect. Someone changed it back to 1. I have now set our max scrubs to 2. We’ll see if that makes a difference. Thank you, Ray From: Ray Cunningham Sent: Thursday, March 10, 2022 8:00 AM To: Szabo, Istvan (Agoda) Cc: ceph-users@ceph.io Subject: RE: [ceph-users] Scrubbing We

[ceph-users] Re: Scrubbing

2022-03-11 Thread Ray Cunningham
;ve been bottlenecked by those? Thank you, Ray   -Original Message- From: norman.kern Sent: Thursday, March 10, 2022 9:27 To: Ray Cunningham Cc: ceph-users@ceph.io Subject: Re: [ceph-users] Re: Scrubbing Ray, You can use node-exporter+prom+grafana  to collect the load of CPUs statistics.

[ceph-users] Re: Scrubbing

2022-03-14 Thread Ray Cunningham
mance impact. dp On 3/11/22 17:32, Ray Cunningham wrote: > That's what I thought. We looked at the cluster storage nodes and found them > all to be less than .2 normalized maximum load. > > Our 'normal' BW for client IO according to ceph -s is around 60MB/s-100MB/s. &g

[ceph-users] Stop Rebalancing

2022-04-12 Thread Ray Cunningham
Hi Everyone, We just upgraded our 640 OSD cluster to Ceph 16.2.7 and the resulting rebalancing of misplaced objects is overwhelming the cluster and impacting MON DB compaction, deep scrub repairs and us upgrading legacy bluestore OSDs. We have to pause the rebalancing if misplaced objects or we

[ceph-users] Re: Stop Rebalancing

2022-04-12 Thread Ray Cunningham
including the term premerge so you are probably right about this being caused by a PG merge. Thank you, Ray   -Original Message- From: Matt Vandermeulen Sent: Tuesday, April 12, 2022 2:39 PM To: Ray Cunningham Cc: ceph-users@ceph.io Subject: Re: [ceph-users] Stop Rebalancing It soun

[ceph-users] Re: Stop Rebalancing

2022-04-12 Thread Ray Cunningham
il 12, 2022 2:46 PM To: Ray Cunningham Cc: ceph-users@ceph.io Subject: Re: [ceph-users] Stop Rebalancing Hi Ray, Disabling the autoscaler on all pools is probably a good idea. At least until https://tracker.ceph.com/issues/53729 is fixed. (You are likely not susceptible to that -- but better safe

[ceph-users] Re: Stop Rebalancing

2022-04-13 Thread Ray Cunningham
: Dan van der Ster Sent: Wednesday, April 13, 2022 10:00 AM To: Ray Cunningham Cc: ceph-users@ceph.io Subject: Re: [ceph-users] Stop Rebalancing One more thing, could you please also share the `ceph osd pool autoscale-status` ? On Tue, Apr 12, 2022 at 9:50 PM Ray Cunningham wrote: > > Th

[ceph-users] Re: Stop Rebalancing

2022-04-13 Thread Ray Cunningham
All pools have gone backfillfull. Thank you, Ray Cunningham Systems Engineering and Services Manager keepertechnology<http://www.keepertech.com/> (571) 223-7242 From: Ray Cunningham Sent: Wednesday, April 13, 2022 10:15:56 AM To: Dan van der Ster Cc

[ceph-users] Re: Stop Rebalancing

2022-04-13 Thread Ray Cunningham
No repair IO and misplaced objects increasing with norebalance and nobackfill set. Thank you, Ray From: Ray Cunningham Sent: Wednesday, April 13, 2022 10:38:29 AM To: Dan van der Ster Cc: ceph-users@ceph.io Subject: Re: [ceph-users] Stop Rebalancing All

[ceph-users] Re: Stop Rebalancing

2022-04-13 Thread Ray Cunningham
#x27;ll have to mute a warning in the meantime -- check the docs after the warning appears). Cheers, dan On Wed, Apr 13, 2022 at 5:16 PM Ray Cunningham wrote: > > Perfect timing, I was just about to reply. We have disabled autoscaler on all > pools now. > > Unfortunately, I can&#

[ceph-users] Re: Stop Rebalancing

2022-04-13 Thread Ray Cunningham
ng for the moment.. Thank you, Ray -Original Message- From: Gregory Farnum Sent: Wednesday, April 13, 2022 12:07 PM To: Dan van der Ster Cc: Ray Cunningham ; Ceph Users Subject: Re: [ceph-users] Re: Stop Rebalancing On Wed, Apr 13, 2022 at 10:01 AM Dan van der Ster wrote: > >

[ceph-users] Re: Stop Rebalancing

2022-04-13 Thread Ray Cunningham
We've done that, I'll update with what happens overnight. Thanks everyone! Thank you, Ray From: Anthony D'Atri Sent: Wednesday, April 13, 2022 4:49 PM To: Ceph Users Subject: [ceph-users] Re: Stop Rebalancing > In any case, isn't this still the best approac