[ceph-users] Re: [Urgent suggestion needed] New Prod Cluster Hardware recommendation

2025-07-09 Thread Alex Gorbachev
oxmox just needs a network and keyring to talk to this cluster. You can run deployment and automation functions from a VM in Proxmox that runs on local storage. -- Alex Gorbachev https://alextelescope.blogspot.com On Wed, Jul 9, 2025 at 10:28 AM Anthony D'Atri wrote: > > > >

[ceph-users] Re: OBJECT_UNFOUND and PG_DAMAGED (recovery_unfound) errors

2025-05-15 Thread Alex
I'll check. Thanks again. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: OBJECT_UNFOUND and PG_DAMAGED (recovery_unfound) errors

2025-05-15 Thread Alex
Hi Maged I don't think our min size = 1 No power outage fingers crossed. #3 Not sure I understand, what's EC? ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Repo name bug?

2025-05-15 Thread Alex
Hi guys. Any feedback on my CR? Thanks, Alex. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: OBJECT_UNFOUND and PG_DAMAGED (recovery_unfound) errors

2025-05-14 Thread Alex
Sorry I didn't make it clear. We marked the osd as out, the error cleared, then we replaced the drive. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: OBJECT_UNFOUND and PG_DAMAGED (recovery_unfound) errors

2025-05-14 Thread Alex
Thanks Eugen. The cluster is currently HEALTHY, we have 3 replicas of all the data and currently only the one failed HDD is out, the rest are in/up. If I'm understanding you correctly it's likely that there are other bad drives which have not yet been detected causing the failure? That's very like

[ceph-users] OBJECT_UNFOUND and PG_DAMAGED (recovery_unfound) errors

2025-05-12 Thread Alex
Hi everyone. Help me settle a debate. My coworker is seeing OBJECT_UNFOUND and PG_DAMAGED (recovery_unfound) errors. We both agree they are caused by bad drives. The fix is to mark the drive as out, replace it and add it back in. Whenever we see this error on Ceph we see corresponding read errors

[ceph-users] Re: Heads up -- 18.2.6 regression

2025-05-01 Thread Alex
Thanks Michael. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Heads up -- 18.2.6 regression

2025-05-01 Thread Alex
Thanks. According to Red Hat Ceph 6 is Quincy Ceph 7 is Reef Ceph 8 is Squid Is the bug in Reef or Squid? ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Heads up -- 18.2.6 regression

2025-05-01 Thread Alex
Thanks for the notice Dan. We just finished upgrading to Red Hat Ceph Storage 6 and the next phase is for us to upgrade to 7. Is 18.2.6 Red Hat Ceph Storage 7.x? In that case I'll be sure to not upgrade until we get the "green light" from you. I assume the fix is on the scale of weeks out, not mo

[ceph-users] Re: v18.2.6 Reef released

2025-04-29 Thread Alex Gorbachev
Ditto, I had a "fun" 18.2.5 experience upgrading our lab with dm-crypt OSDs, but even with 18.2.5 bricking the OSD startup, the orchestrator prevented any downtime by blocking further updates, and as soon as 18.2.6 became available, the upgrade finished fine. Thanks for the quick fi

[ceph-users] Re: Repo name bug?

2025-04-18 Thread Alex
Thanks John. Doing some quick Googling it seems that I can clean up an existing PR without making a new one? - Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Preflight Playbook

2025-04-18 Thread Alex
Hi Frédéric. I have more Ceph servers to update (unsure when) and I'll try to do more troubleshooting. Thanks. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Repo name bug?

2025-04-17 Thread Alex
Hi. I cleaned up and squashed my commit. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Preflight Playbook

2025-04-17 Thread Alex
If you like I can make a Pull Request. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Preflight Playbook

2025-04-17 Thread Alex
For some reason `dnf update podman` did not update crun. I'm proposing adding crun to main.yml: infra_pkgs: - chrony - podman - lvm2 - sos - rpcbind - firewalld ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to

[ceph-users] Re: Preflight Playbook

2025-04-17 Thread Alex
Thanks. Nothing is broken so there's no support case needed. Dnf update is one of the steps which probably updates crun as well. We didn't run that since we didn't want to update all of the software on the server, we just wanted to update the Ceph related packages. I manually updated podman but n

[ceph-users] Re: Preflight Playbook

2025-04-16 Thread Alex
Yeah. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Preflight Playbook

2025-04-16 Thread Alex
I was upgrading from Ceph 5 to 6 on RHEL 9. Looking at the dnf logs it seems that i was running crun 1.8.7 and now on 1.16.1 Thanks. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Repo name bug?

2025-04-16 Thread Alex
Hi. Any feedback on my pull request? Thanks, Alex. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Preflight Playbook

2025-04-16 Thread Alex
Good morning. Should the Preflight playbook maybe update `crun` as well? I was updating the Ceph version yesterday and it wouldn't work until I updated `crun`. Thanks. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to cep

[ceph-users] Re: Upcoming Ceph Days - CFP and Join Us

2025-04-15 Thread Alex
Anything coming up in NYC? ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Upcoming Ceph Days - CFP and Join Us

2025-04-14 Thread Alex
Thanks! ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Repo name bug?

2025-04-11 Thread Alex
x it and if I can actually make this VERY small contribution to the Ceph codebase that would be amazing. Thanks! Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-11 Thread Alex
Sounds good to me. I responded to your comment in the PR. Thanks. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: nodes with high density of OSDs

2025-04-11 Thread Alex from North
Hello Tim! First of all, thanks for the detailed answer! Yes, probably in set up of 4 nodes by 116 OSD it looks a bit overloaded, but what if I have 10 nodes? Yes, nodes itself are still heavy but in a row it seems to be not that dramatic, no? However, in a docu I see that it is quite common for

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-10 Thread Alex
I made a Pull Request for cephadm.log set DEBUG. Not sure if I should merge it. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Repo name bug?

2025-04-10 Thread Alex
I created a pull request, not sure what the etiquette is if I can merge it. First timer here. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-10 Thread Alex
I think it's the same block of code Eugen found. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-10 Thread Alex
Is this bit of code responsible for hardcoding DEBUG to cephadm.log? 'loggers': { '': { 'level': 'DEBUG', 'handlers': ['console', 'log_file'], } } in /var/lib/ceph//cephadm.* ? ___ ceph-users mailing list

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-10 Thread Alex
I did have to add "su root root" to the log rotate script to fix the permissions issue. There's a RH KB article and Ceph github pull requests to fix it. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@cep

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-10 Thread Alex
Thanks! ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: nodes with high density of OSDs

2025-04-10 Thread Alex from North
Hello Dominique! Os is quite new - Ubuntu 22.04 with all the latest upgrades. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] nodes with high density of OSDs

2025-04-10 Thread Alex from North
Hello everybody! I have a 4 nodes with 112 OSDs each and 18.2.4. OSD consist of db on SSD and data on HDD For some reason, when I reboot node, not all OSDs get up because some VG or LV are not active. To make it alive again I manually do vgchange -ay $VG_NAME or lvchange -ay $LV_NAME. I suspect

[ceph-users] Repo name bug?

2025-04-09 Thread Alex
um repo file then gets named ceph_stable_$basearch. Shouldn't it be basearch without the $ ? - Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-09 Thread Alex
How does that work? ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-09 Thread Alex
Thanks. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-09 Thread Alex
Haha yeah think that's what we're doing. I'm just going to add it to logrotate. Do you use copytruncate option or postrotate to restart ceph? ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-09 Thread Alex
Thanks Eugen! I think you're right since support had me grep for the same code. Seems crazy though that it's hardcoded doesn't it? I guess we can mod the Python file, but you'd think that wouldn't be necessary. Should we make a feature request or modify the code ourselves and make a pull request?

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-09 Thread Alex
Official IBM and RH "fix" is to replace DEBUG with INFO in /var/lib/ceph//cephadm.hash ¯\_ (ツ) _/¯ ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
I mean this bit 'log_file': { 'level': 'DEBUG', 'class': 'logging.handlers.WatchedFileHandler', 'formatter': 'cephadm', 'filename': '%s/cephadm.log' % LOG_DIR, } ___ ceph-users mailing list

[ceph-users] Re: NIH Datasets

2025-04-08 Thread Alex Buie
If you have a line on the data, I have connections that can store it or I can consult pro bono on building a system to store it. However wan-ceph is not the answer here. On Sun, Apr 6, 2025, 11:08 PM Linas Vepstas wrote: > OK what you will read below might sound insane but I am obliged to ask.

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
What about Pacific and Quincy? ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
Can someone paste in here their copy of logrotate? The trick always with rotating logs is that the service writing to it needs to be restarted or told to stop writing so the file handle gets closed. Otherwise it stays open and the free disc space isn't recovered. ___

[ceph-users] Re: NIH Datasets

2025-04-08 Thread Alex Gorbachev
to be copied. If not, what are the restrictions on obtaining and copying the data? - Organization: how will the storage and maintenance of data be organized (and funded)? A foundation, a SETI-at-home like network, a blockchain (to preserve data veracity)? - Legal support? -- Alex Gorbachev

[ceph-users] Re: NIH Datasets

2025-04-08 Thread Alex Gorbachev
, and there is a TAPAS service from SpectraLogic. I would imagine questions would arise about custody of the data, legal implications etc. The easiest is for the organization already hosting the data to just preserve it by archiving, and thereby claim a significant cost reduction. -- Alex

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
Awesome. Wish I knew that before I spent half a day trying to overwrite it. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
Interesting. So it's like that for everybody? Meaning cephadm.log logs debug messages. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Cephadm flooding /var/log/ceph/cephadm.log

2025-04-07 Thread Alex
Hi everyone. My company has paid Ceph support. The tech is saying: "...cephadm package of ceph 5 is having bug. So It generate debug logs even its set for "info" logs..." I have two clusters running, one Ceph 5 and the other Ceph 6 (Quincy). Both of them are sending "DEBUG -" messages to c

[ceph-users] Re: NIH Datasets

2025-04-07 Thread Alex Buie
preservation project technically in a volunteer capacity. I can build a cutting edge resilient distributed storage system for cheaper than anything currently on the market. Please reach out or pass along my email. Alex On Sun, Apr 6, 2025 at 11:08 PM Linas Vepstas wrote: > OK what you will read be

[ceph-users] Major version upgrades with CephADM

2025-04-05 Thread Alex Petty
grade ? Thanks, Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Division by zero while upgrading

2025-03-19 Thread Alex
Think I figured it out. There's still a bug, but when updating the OSD nodes I just need to specify the --hosts, no need for --services. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Division by zero while upgrading

2025-03-19 Thread Alex
Upgrading from Pacific to Quincy, anyone ever see this? This one looks FUN! I upgraded mgrs, mons, and 1/2 of the osd nodes. Hit this when upgrading the last OSD node. It's a test cluster. Running the command below: ceph orch upgrade start ... --hosts --services osd.all-available-devices $ ceph h

[ceph-users] Re: Experience with 100G Ceph in Proxmox

2025-03-15 Thread Alex Gorbachev
How about testing the actual network throughput with iperf? Even today there are speed/duplex mismatches on switch ports. And what everyone else said about saturation etc. We get, at absolute worst, 600 MB/s on a 10G connection. -- Alex Gorbachev https://alextelescope.blogspot.com On Tue

[ceph-users] Re: Module 'devicehealth' has failed

2025-03-06 Thread Alex from North
Thanks for the help, buddy! I really appreciate it! Will try to wait. Maybe someone else jumps in. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Module 'devicehealth' has failed

2025-03-06 Thread Alex from North
I did. It says more or less the same Mar 06 10:44:05 node1.ec.mts conmon[10588]: 2025-03-06T10:44:05.769+ 7faca5624640 -1 log_channel(cephadm) log [ERR] : Failed to apply osd.node1.ec.mts_all_disks spec DriveGroupSpec.from_json(yaml.safe_load('''service_type: osd Mar 06 10:44:05 node1.ec.m

[ceph-users] Re: Module 'devicehealth' has failed

2025-03-06 Thread Alex from North
a bit more details. Now I've notices that ceph health detail signals to me that [WRN] CEPHADM_APPLY_SPEC_FAIL: Failed to apply 1 service(s): osd.node1.ec.all_disks osd.node1.ec.all_disks: Expecting value: line 1 column 2311 (char 2310) Okay, I checked my spec but do not see anything suspicious

[ceph-users] Re: Module 'devicehealth' has failed

2025-03-06 Thread Alex from North
I will provide you any info you need, just gimme a sign. My starter post was related to 19.2.0. Now I downgraded (full reinstall as this is completely new cluster I wanna run) to 18.2.4 and the same story Mar 06 09:37:41 node1.ec.mts conmon[10588]: failed to collect metrics: Mar 06 09:37:41 nod

[ceph-users] Re: Module 'devicehealth' has failed

2025-03-04 Thread Alex from North
yes, I do .mgr 10 1 769 KiB2 2.3 MiB 04.7 PiB ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Module 'devicehealth' has failed

2025-03-04 Thread Alex from North
Hello everybody! Running 19.2.0 faced an issued still cannot struggle. And this is Module 'devicehealth' has failed: Expecting value: line 1 column 2378 (char 2377) in MGR log I see Mar 04 12:48:07 node2.ec.mts ceph-mgr[3821449]: Traceback (most recent call last):

[ceph-users] Re: mgr module 'orchestrator' is not enabled/loaded

2025-03-04 Thread Alex from North
found the answer here https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/QJHES7GKTI6O7BT6UBGCHK6WFTJRNJHE/ ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] mgr module 'orchestrator' is not enabled/loaded

2025-03-04 Thread Alex from North
Hi all! Faced sad situation I don't know where to dig to. That is why I am here again, hoping for hints. Situation: all the "orch" commands aint available and give error in mgr log mgr.server reply reply (95) Operation not supported Module 'orchestrator' is not enabled/loaded (required by comm

[ceph-users] Re: User new to ceph

2025-02-22 Thread Alex Gorbachev
those volume groups, for example: vgs|grep ceph|awk '{print $1}'|xargs -I {} vgremove -y {} We run these procedures often and they always work. -- Alex Gorbachev https://alextelescope.blogspot.com On Sat, Feb 22, 2025 at 4:12 PM Christian Hansen wrote: > So i have battled the

[ceph-users] Re: NFS recommendations

2025-02-05 Thread Alex Buie
that install as it was a WORM media server. Cephfs just wasn’t a great fit for reliability. Been working wonders across the WAN now tho keeping all our sites in sync. *Alex Buie*Senior Cloud Operations Engineer 450 Century Pkwy # 100 Allen, TX 75013 <https://maps.google.com/?q=450+Century+Pkwy+

[ceph-users] ceph orch ls --refresh

2025-01-20 Thread Alex Hussein-Kershaw (HE/HIM)
t of date in normal operation. Is that an accurate understanding? I'm looking to use this command in some scripts so would like to be sure I understand what I am doing. Thanks, Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send

[ceph-users] Re: [EXTERNAL] Re: Cephadm: Specifying RGW Certs & Keys By Filepath

2025-01-17 Thread Alex Hussein-Kershaw (HE/HIM)
omplicate using this also. I'm probably going to shy away from this for now. I still think a sensible fix here is to not add the defaulted ssl_certificate option if it's passed as an extra_frontend_arg, which allows my use case without a mainline change to

[ceph-users] Re: [EXTERNAL] Re: Cephadm: Specifying RGW Certs & Keys By Filepath

2025-01-17 Thread Alex Hussein-Kershaw (HE/HIM)
pping the "ssl_port" just brings it up as a HTTP endpoint.     beast port=7480 ssl_certificate=/etc/ssl/certs/server.crt ssl_private_key=/etc/ssl/private/server.key So I don't have a combination that allows me to do just HTTPS with my cert/key provided as a file path. T

[ceph-users] Re: [EXTERNAL] Re: Cephadm: Specifying RGW Certs & Keys By Filepath

2025-01-16 Thread Alex Hussein-Kershaw (HE/HIM)
Oh actually I have spoke to soon. That does work, but it also exposes port HTTP over port 80. 🙁   beast port=80 ssl_port=7480 ssl_certificate=/etc/ssl/certs/server.crt ssl_private_key=/etc/ssl/private/server.key From: Alex Hussein-Kershaw (HE/HIM) Sent

[ceph-users] Re: [EXTERNAL] Re: Cephadm: Specifying RGW Certs & Keys By Filepath

2025-01-16 Thread Alex Hussein-Kershaw (HE/HIM)
Amazing. How did I miss that. Dropping "ssl: true" and adding "ssl_port=1234" to the rgw_frontend_extra_args values has me sorted. Many thanks! From: Redouane Kachach Sent: Thursday, January 16, 2025 4:39 PM To: Alex Hussein-Kershaw (HE

[ceph-users] Re: Cephadm: Specifying RGW Certs & Keys By Filepath

2025-01-16 Thread Alex Hussein-Kershaw (HE/HIM)
.com/issues/69567>, and will submit a PR. ____ From: Alex Hussein-Kershaw (HE/HIM) Sent: Thursday, January 16, 2025 3:36 PM To: ceph-users Subject: Cephadm: Specifying RGW Certs & Keys By Filepath Hi Folks, Looking for some advice on RGW service specs and Cephadm.

[ceph-users] Cephadm: Specifying RGW Certs & Keys By Filepath

2025-01-16 Thread Alex Hussein-Kershaw (HE/HIM)
hat I can specify the key and certificate by file path and not have to manually change the config to make it work? Thanks, Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: 4k IOPS: miserable performance in All-SSD cluster

2024-11-26 Thread Alex Gorbachev
Martin, are MONs set up on the same hosts, or is there latency to them by any chance? -- Alex Gorbachev https://alextelescope.blogspot.com On Tue, Nov 26, 2024 at 5:20 AM Martin Gerhard Loschwitz < martin.loschw...@true-west.com> wrote: > Hi Alex, > > thank you for the reply. H

[ceph-users] Re: 4k IOPS: miserable performance in All-SSD cluster

2024-11-25 Thread Alex Gorbachev
iagnostics. I know you guys have played with Ceph before, and can do the math of what the IOPS values should be - three clusters all seeing the same problem would most likely indicate a non-default configuration value that is not correct. -- Alex Gorbachev ISS On Mon, Nov 25, 2024 at 9:34 PM Mart

[ceph-users] NFS and Service Dependencies

2024-11-09 Thread Alex Buie
to run on the same node so that the virtual IP is present for nfs to bind to? Or, am I doing something wrong here? Thanks a bunch! *Alex* ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: [EXTERNAL] Re: Ceph Multisite Version Compatibility

2024-11-06 Thread Alex Hussein-Kershaw (HE/HIM)
hat: * full sync is working * incremental sync is not ____ From: Alex Hussein-Kershaw (HE/HIM) Sent: Wednesday, November 6, 2024 3:27 PM To: Eugen Block ; ceph-users@ceph.io Subject: Re: [EXTERNAL] [ceph-users] Re: Ceph Multisite Version Compatibility Hi Eugen, Than

[ceph-users] Re: [EXTERNAL] Re: Ceph Multisite Version Compatibility

2024-11-06 Thread Alex Hussein-Kershaw (HE/HIM)
site sync between Squid and Quincy does not work in one direction - rgw - Ceph<https://tracker.ceph.com/issues/68819>. Best wishes, Alex From: Eugen Block Sent: Wednesday, November 6, 2024 11:35 AM To: ceph-users@ceph.io Subject: [EXTERNAL] [ceph-users] Re:

[ceph-users] Re: Ceph Multisite Version Compatibility

2024-11-01 Thread Alex Hussein-Kershaw (HE/HIM)
objects to SiteB, they appear on siteA shortly after. * Write objects to SiteA, they never appear on siteB. It seems to be 100% reproducible. Suspect I need to raise a tracker. I welcome any suggestions that I'm doing this wrong meanwhile. ____ From: Alex Hussein-Ke

[ceph-users] Ceph Multisite Version Compatibility

2024-11-01 Thread Alex Hussein-Kershaw (HE/HIM)
eference that claims something like "must be within +/- 1 version of the other zones"). Thanks, Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: no recovery running

2024-10-30 Thread Alex Walender
Hey Joffrey, try to switch back to the wpq scheduler in ceph.conf: osd_op_queue = wpq ...and restart all OSDs. I also had issues where recovery was very very slow (10kb/s). Best Regards, Alex Walender Am 17.10.24 um 11:44 schrieb Joffrey: HI, This is my cluster: cluster: id

[ceph-users] Re: Install on Ubuntu Noble on Arm64?

2024-10-28 Thread Alex Closs
Ubuntu noble *is* an LTS release, 24.04 > On Oct 28, 2024, at 06:40, Robert Sander wrote: > > Hi > >> On 10/25/24 19:57, Daniel Brown wrote: >> Think I’ve asked this before but — has anyone attempted to use a cephadm >> type install with Debian Nobel running on Arm64? Have tried both Reef an

[ceph-users] Re: failed to load OSD map for epoch 2898146, got 0 bytes

2024-10-24 Thread Alex Walender
st Regards, Alex Walender Am 21.10.24 um 22:31 schrieb Vladimir Sigunov: Hi Dan and Frank, From my experience, if an osd was down for a long period of time, it could take more than one manual restart for this osd to catch up an actual epoch. Under manual restart I mean systemctl re

[ceph-users] Re: [EXTERNAL] Re: How to Speed Up Draining OSDs?

2024-10-21 Thread Alex Hussein-Kershaw (HE/HIM)
My pool size is indeed 3. Operator error 🙂 Thanks again, Alex From: Eugen Block Sent: Monday, October 21, 2024 3:08 PM To: Alex Hussein-Kershaw (HE/HIM) Cc: ceph-users@ceph.io Subject: Re: [EXTERNAL] [ceph-users] Re: How to Speed Up Draining OSDs? If your

[ceph-users] Re: [EXTERNAL] Re: How to Speed Up Draining OSDs?

2024-10-21 Thread Alex Hussein-Kershaw (HE/HIM)
that it's stuck rather than slow. I wonder if it's actually possible to drop below 3 OSDs in this manner? Thanks, Alex From: Eugen Block Sent: Monday, October 21, 2024 2:20 PM To: ceph-users@ceph.io Subject: [EXTERNAL] [ceph-users] Re: How to Speed Up D

[ceph-users] How to Speed Up Draining OSDs?

2024-10-21 Thread Alex Hussein-Kershaw (HE/HIM)
ot; to check I won't cause any PGs to go offline before hand. Are there any tricks here that I'm missing? Thanks, Alex ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Forced upgrade OSD from Luminous to Pacific

2024-10-09 Thread Alex Rydzewski
nning Pacific. What does a 'ceph versions' says? Did you restart all OSDs after the upgrade? Regards, Frédéric. - Le 9 Oct 24, à 14:39, Alex Rydzewski rydzewski...@gmail.com a écrit : I thought so too, Frédéric But when I try change it, I get this error: root@helper:~# ceph osd requ

[ceph-users] Re: Forced upgrade OSD from Luminous to Pacific

2024-10-09 Thread Alex Rydzewski
equire-osd-release pacific' should get you out of this. But maybe someone can confirm? Regards, Frédéric. - Le 9 Oct 24, à 14:07, Alex Rydzewski rydzewski...@gmail.com a écrit : Of course, Frédéric, root@helper:~# ceph osd dump | head -13 epoch 45887 fsid 96b6ff1d-25bf-403f-be3d-78c2fb0ff7

[ceph-users] Re: Forced upgrade OSD from Luminous to Pacific

2024-10-09 Thread Alex Rydzewski
Of course, Frédéric, root@helper:~# ceph osd dump | head -13 epoch 45887 fsid 96b6ff1d-25bf-403f-be3d-78c2fb0ff747 created 2018-06-02T13:12:54.207727+0300 modified 2024-10-09T11:08:53.638661+0300 flags sortbitwise,recovery_deletes,purged_snapdirs crush_version 82 full_ratio 0.95 backfillfull_rati

[ceph-users] Re: Forced upgrade OSD from Luminous to Pacific

2024-10-09 Thread Alex Rydzewski
4, à 15:24, Alex rydzewskirydzewski...@gmail.com a écrit : Hello, dear community! I kindly ask for your help in resolving my issue. I have a server with a single-node CEPH setup with 5 OSDs. This server has been powered off for about two years, and when I needed the data from it, I found that the

[ceph-users] Forced upgrade OSD from Luminous to Pacific

2024-10-08 Thread Alex Rydzewski
Hello, dear community! I kindly ask for your help in resolving my issue. I have a server with a single-node CEPH setup with 5 OSDs. This server has been powered off for about two years, and when I needed the data from it, I found that the SSD where the system was installed had died. I tried

[ceph-users] Re: All monitors fall down simultaneously when I try to map rbd on client

2024-09-27 Thread Alex from North
yes, this is a bug, indeed. https://www.spinics.net/lists/ceph-users/msg82468.html > Remove mappings by: > $ `ceph osd dump` > For each pg_upmap_primary entry in the above output: > $ `ceph osd rm-pg-upmap-primary ` ___ ceph-users mailing list -- ceph-u

[ceph-users] Re: All monitors fall down simultaneously when I try to map rbd on client

2024-09-27 Thread Alex from North
fixed by https://www.spinics.net/lists/ceph-users/msg82468.html CLOSED. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: All monitors fall down simultaneously when I try to map rbd on client

2024-09-27 Thread Alex from North
By increasing debulg level I found out the following but have no idea how to fix this issue. ``` src/osd/OSDMap.cc: 3242: FAILED ceph_assert(pg_upmap_primaries.empty()) ``` There is only one topic in google and with no answer ___ ceph-users mailing lis

[ceph-users] All monitors fall down simultaneously when I try to map rbd on client

2024-09-27 Thread Alex from North
Hello everybody, found intresting thing: for some reason ALL the monitors crash when I try to rbd map on client host. here is my pool: root@ceph1:~# ceph osd pool ls iotest Here is my rbd in this pool: root@ceph1:~# rbd ls -p iotest test1 this is a client creds to connect to this pool: [cli

[ceph-users] Re: [EXTERNAL] Re: Backup strategies for rgw s3

2024-09-26 Thread Alex Hussein-Kershaw (HE/HIM)
We have been using Amazon S3 (rclone.org) to copy all the data to a filesystem nightly to provide an S3 backup mechanism. It has Ceph support out the box (added by one of my colleagues a few years ago). From: Adam Prycki Sent: Wednes

[ceph-users] Re: [EXTERNAL] Re: Bucket Notifications v2 & Multisite Redundancy

2024-09-25 Thread Alex Hussein-Kershaw (HE/HIM)
multisite even with notifications_v2 disabled - rgw - Ceph<https://tracker.ceph.com/issues/68227?next_issue_id=68211>. From: Alex Hussein-Kershaw (HE/HIM) Sent: Wednesday, September 25, 2024 10:42 AM To: Yuval Lifshitz ; Ceph Users Subject: Re: [EXTERNAL] Re:

[ceph-users] Re: [EXTERNAL] Re: Bucket Notifications v2 & Multisite Redundancy

2024-09-25 Thread Alex Hussein-Kershaw (HE/HIM)
he inability of notification config to remain single site on the latest Squid RC. Given the conversation we had below I think this is a bug. Happy to raise a tracker. Welcome any thoughts. I'll try to repro this on Reef shortly. Thanks, Alex From: Alex Hussein-

[ceph-users] Re: [EXTERNAL] Multisite sync: is metadata transferred in plain text?

2024-09-23 Thread Alex Hussein-Kershaw (HE/HIM)
Feels like you answered your own question here - why not just use HTTPS for your multisite sync? I'm not aware of any other encryption mechanisms for metadata (I guess using SSE encryption for your data may save that going over in plain text) and would assume that just using HTTP would indeed b

[ceph-users] Re: [EXTERNAL] Deploy rgw different version using cephadm

2024-09-19 Thread Alex Hussein-Kershaw (HE/HIM)
I think the advice is not to use floating tags (i.e. "latest") and use specific tags if possible. I believe you can achieve what you want with either: "ceph orch upgrade --image " not sure if this allows you to downgrade, but certainly lets you upgrade and change image, see Upgrading Ceph — Cep

[ceph-users] Re: [EXTERNAL] Re: Bucket Notifications v2 & Multisite Redundancy

2024-09-17 Thread Alex Hussein-Kershaw (HE/HIM)
tion configuration - rgw - Ceph<https://tracker.ceph.com/issues/68104> Also re-including the mailing list as it was dropped. From: Yuval Lifshitz Sent: Tuesday, September 17, 2024 10:36 AM To: Alex Hussein-Kershaw (HE/HIM) Subject: Re: [EXTERNAL] Re:

[ceph-users] Re: [EXTERNAL] Re: Bucket Notifications v2 & Multisite Redundancy

2024-09-16 Thread Alex Hussein-Kershaw (HE/HIM)
ot; }, "arn": "arn:aws:sns:geored_zg::ahk2", "opaqueData": "", "policy": "" } ] } ] } So maybe a bug in Squid? __

  1   2   3   >