[ceph-users] Re: Ceph Octopus packages missing at download.ceph.com / eu.ceph.com permission problem

2024-11-16 Thread Oliver Freyermuth
___ ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io> To unsubscribe send an email to ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io> ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an emai

[ceph-users] Re: Snaptrim issue after nautilus to octopus upgrade

2024-08-26 Thread Oliver Freyermuth
as I know and I'm not sure how smooth upgrade can be. The cluster size is 4 petabyte and taking backup is not an option :) The worst part is I have to change OS and Kernel from arch-linux to custom made ubuntu to be able to apply this change. Oliver Freyermuth mailto:freyerm...@ph

[ceph-users] Re: Snaptrim issue after nautilus to octopus upgrade

2024-08-26 Thread Oliver Freyermuth
t; > Thank you for the advice, I will share the results when it's done. > > Regards. > > Oliver Freyermuth mailto:freyerm...@physik.uni-bonn.de>>, 23 Ağu 2024 Cum, 18:48 > tarihinde şunu yazdı: > >> Hi Özkan, >> >

[ceph-users] Re: Snaptrim issue after nautilus to octopus upgrade

2024-08-23 Thread Oliver Freyermuth
I will try OSD compaction. May I learn did you do online compaction or offline? Because I have 2 side and I can shutdown 1 entire rack and do the offline compaction and do the same thing other side when its done. What do you think? Regards. Oliver Freyermuth mailto:freyerm...@physik.uni-bon

[ceph-users] Re: Snaptrim issue after nautilus to octopus upgrade

2024-08-23 Thread Oliver Freyermuth
or an HDD with lots of omap data). You can disable this automatic conversion with: What should I do to solve this problem? Thanks. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io -- Oliver Freyermuth

[ceph-users] Re: RBD Journaling seemingly getting stuck for some VMs after upgrade to Octopus

2024-08-13 Thread Oliver Freyermuth
Am 13.08.24 um 15:02 schrieb Ilya Dryomov: On Mon, Aug 12, 2024 at 1:17 PM Oliver Freyermuth wrote: Am 12.08.24 um 12:16 schrieb Ilya Dryomov: On Mon, Aug 12, 2024 at 11:28 AM Oliver Freyermuth wrote: Am 12.08.24 um 11:09 schrieb Ilya Dryomov: On Mon, Aug 12, 2024 at 10:20 AM Oliver

[ceph-users] Re: RBD Journaling seemingly getting stuck for some VMs after upgrade to Octopus

2024-08-12 Thread Oliver Freyermuth
Am 12.08.24 um 12:16 schrieb Ilya Dryomov: On Mon, Aug 12, 2024 at 11:28 AM Oliver Freyermuth wrote: Am 12.08.24 um 11:09 schrieb Ilya Dryomov: On Mon, Aug 12, 2024 at 10:20 AM Oliver Freyermuth wrote: Dear Cephalopodians, we've successfully operated a "good old" Mim

[ceph-users] Re: RBD Journaling seemingly getting stuck for some VMs after upgrade to Octopus

2024-08-12 Thread Oliver Freyermuth
Am 12.08.24 um 11:09 schrieb Ilya Dryomov: On Mon, Aug 12, 2024 at 10:20 AM Oliver Freyermuth wrote: Dear Cephalopodians, we've successfully operated a "good old" Mimic cluster with primary RBD images, replicated via journaling to a "backup cluster" with Octopus

[ceph-users] RBD Journaling seemingly getting stuck for some VMs after upgrade to Octopus

2024-08-12 Thread Oliver Freyermuth
cluster running Octopus before, and also did not happen when both were running Mimic. We plan to switch to snapshot-based mirroring in the future anyways, but it would be good to understand this strange issue in any case. Cheers, Oliver [0] https://docs.ceph.com/en/octopu

[ceph-users] Re: Mirror de.ceph.com broken?

2022-10-27 Thread Oliver Freyermuth
istian ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io -- Oliver Freyermuth Universität Bonn Physikalisches Institut, Raum 1.047 Nußallee 12 53115 Bonn -- Tel.: +49 228 73 2367 Fax: +49 228 7

[ceph-users] Re: are you using nfs-ganesha builds from download.ceph.com

2022-01-12 Thread Oliver Freyermuth
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io -- Oliver Freyermuth Universität Bonn Physikalisches Institut, Raum 1.047 Nußallee 12 53115 Bonn -- Tel.: +49 228 73 2367 Fax: +49 228 73 7869 -- ___

[ceph-users] Re: Why you might want packages not containers for Ceph deployments

2021-06-25 Thread Oliver Freyermuth
ttps://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/PPLJIHT6WKYPDJ45HVJ3Z37375WIGKDW/ [2] Of course, you can (and should) use a staged rollout. [3] Well, that's a presumption, but the fact that you mentioned user concerns about this in the survey seems to strengthen that point. --

[ceph-users] Re: Why you might want packages not containers for Ceph deployments

2021-06-02 Thread Oliver Freyermuth
a 'boring' Ceph cluster for the medium term, I'd still advise on using packages. I don't think this makes me a luddite :) Regards, and apologies for the wall of text, Matthew [0] I think that's a fair summary! [1] This hasn't always been true... [2] Simple (s

[ceph-users] Re: Cephfs: Migrating Data to a new Data Pool

2021-04-05 Thread Oliver Freyermuth
FS and copy things over. CephFS mirroring (of snapshots) in Pacific may make this easier by allowing to reduce any actual downtime for users, but any other solution would be much appreciated. Cheers, Oliver On Mon, Apr 5, 2021 at 2:55 PM Oliver Freyermuth mailto:freyerm...@physik.uni-bon

[ceph-users] Re: Cephfs: Migrating Data to a new Data Pool

2021-04-05 Thread Oliver Freyermuth
Hi, that really looks like a useful tool, thanks for mentioning this on the list :-). However, I'd also love to learn about a different way ­— as documentation states: "You may notice that object counts in your primary data pool (the one passed to fs new) continue to increase, even if files a

[ceph-users] Re: ceph-ansible in Pacific and beyond?

2021-03-17 Thread Oliver Freyermuth
Am 17.03.21 um 20:09 schrieb Stefan Kooman: On 3/17/21 7:51 PM, Martin Verges wrote:   I am still not convinced that containerizing everything brings any benefits except the collocation of services. Is there even a benefit? Decoupling from underlying host OS. On a test cluster I'm running U

[ceph-users] Re: ceph-volume lvm cannot zap???

2020-09-19 Thread Oliver Freyermuth
Hi, we have also seen such cases, it seems that sometimes (when the controller / device is broken in special ways), device mapper keeps the volume locked. You can check as follows: 1) Check if lvs / pvs / vgs show some undefined device. In that case, you may have to flush the lvmeta cache:

[ceph-users] Re: OSDs and tmpfs

2020-09-11 Thread Oliver Freyermuth
Hi together, I believe the deciding factor is whether the OSD was deployed using ceph-disk (in "ceph-volume" speak, a "simple" OSD), which means the metadata will be on a separate partition, or whether it was deployed with "ceph-volume lvm". The latter stores the metadata in LVM tags, so the e

[ceph-users] Re: RadosGW and DNS Round-Robin

2020-09-04 Thread Oliver Freyermuth
Hi, Am 04.09.20 um 18:20 schrieb dhils...@performair.com: All; We've been running RadosGW on our nautilus cluster for a while, and we're going to be adding iSCSI capabilities to our cluster, via 2 additional servers. I intend to also run RadosGW on these servers. That begs the question of ho

[ceph-users] Re: Ceph-deploy on rhel.

2020-07-25 Thread Oliver Freyermuth
ar, so the ceph-deploy issue (whatever it was) seems solved :-). It seems like you might have missed this step: https://docs.ceph.com/docs/nautilus/mgr/dashboard/#enabling-the-object-gateway-management-frontend which is necessary to let the dashboard manage the Object Gateways. Cheers,

[ceph-users] Re: Ceph-deploy on rhel.

2020-07-25 Thread Oliver Freyermuth
s (or doing that manually). Cheers, Oliver > > > On Sat, 25 Jul 2020, 21:36 Oliver Freyermuth, <mailto:freyerm...@physik.uni-bonn.de>> wrote: > > Hi, > > Am 22.07.20 um 12:13 schrieb sathvik vutukuri: > > Hi, > > > &

[ceph-users] Re: Ceph-deploy on rhel.

2020-07-25 Thread Oliver Freyermuth
Hi, Am 22.07.20 um 12:13 schrieb sathvik vutukuri: > Hi, > > Did any one installed ceph-deploy on rhel7 with rados gate way. > > I see there are no rpms available for rhel7 in ceph-deploy in > download.ceph.com for nautilis , luminous, octopus versions. where exactly did you look? I find the

[ceph-users] Re: Ceph SSH orchestrator?

2020-07-03 Thread Oliver Freyermuth
Am 03.07.20 um 20:29 schrieb Dimitri Savineau: > You can try to use ceph-ansible which supports baremetal and containerized > deployment. > > https://github.com/ceph/ceph-ansible Thanks for the pointer! I know about ceph-ansible. The problem is that our full infrastructure is Puppet-based, so

[ceph-users] Re: Ceph SSH orchestrator?

2020-07-03 Thread Oliver Freyermuth
Am 03.07.20 um 10:00 schrieb Sebastian Wagner: Am 02.07.20 um 19:57 schrieb Oliver Freyermuth: Dear Cephalopodians, as we all know, ceph-deploy is on its demise since a while and essentially in "maintenance mode". We've been eyeing the "ssh orchestrator" whi

[ceph-users] Ceph SSH orchestrator?

2020-07-02 Thread Oliver Freyermuth
Dear Cephalopodians, as we all know, ceph-deploy is on its demise since a while and essentially in "maintenance mode". We've been eyeing the "ssh orchestrator" which was in Nautilus as the "successor in spirit" of ceph-deploy. While we have not tried it out just yet, I find this module seems

[ceph-users] Re: State of SMR support in Ceph?

2020-05-06 Thread Oliver Freyermuth
Am 06.05.20 um 16:12 schrieb brad.swan...@adtran.com: Take a look at the available SMR drives: https://www.ixsystems.com/community/resources/list-of-known-smr-drives.141/ Thanks for this nice overview link! Indeed, my question was more thinking about "the next years". While currently, SMR is c

[ceph-users] Re: State of SMR support in Ceph?

2020-05-06 Thread Oliver Freyermuth
Dear Janne, Am 06.05.20 um 09:18 schrieb Janne Johansson: Den ons 6 maj 2020 kl 00:58 skrev Oliver Freyermuth mailto:freyerm...@physik.uni-bonn.de>>: Dear Cephalopodians, seeing the recent moves of major HDD vendors to sell SMR disks targeted for use in consumer NAS devices (inc

[ceph-users] State of SMR support in Ceph?

2020-05-05 Thread Oliver Freyermuth
Dear Cephalopodians, seeing the recent moves of major HDD vendors to sell SMR disks targeted for use in consumer NAS devices (including RAID systems), I got curious and wonder what the current status of SMR support in Bluestore is. Of course, I'd expect disk vendors to give us host-managed SMR

[ceph-users] Re: RBD-mirror instabilities

2020-02-15 Thread Oliver Freyermuth
Dear Jason, Am 13.02.20 um 01:28 schrieb Jason Dillaman: > On Wed, Feb 12, 2020 at 2:53 PM Oliver Freyermuth > wrote: >> >> Dear Jason, >> >> Am 12.02.20 um 19:29 schrieb Jason Dillaman: >>> On Wed, Feb 12, 2020 at 11:55 AM Oliver Freyermuth &g

[ceph-users] Re: RBD-mirror instabilities

2020-02-12 Thread Oliver Freyermuth
Dear Jason, Am 12.02.20 um 19:29 schrieb Jason Dillaman: > On Wed, Feb 12, 2020 at 11:55 AM Oliver Freyermuth > wrote: >> >> Dear Cephalopodians, >> >> for those on the list also fighting rbd mirror process instabilities: With >> 14.2.7 (but maybe it was als

[ceph-users] Re: RBD-mirror instabilities

2020-02-12 Thread Oliver Freyermuth
o restarts or short "hangs" of many OSD processes). Cheers and hope this helps somebody with similar issues, Oliver Am 27.12.19 um 02:43 schrieb Oliver Freyermuth: Dear Cephalopodians, for those following along through the holiday season, here's my "quick hack" for

[ceph-users] Re: RBD-mirror instabilities

2019-12-26 Thread Oliver Freyermuth
trailing brace. I created a tracker issue here: https://tracker.ceph.com/issues/43429 Cheers, Oliver Am 24.12.19 um 04:39 schrieb Oliver Freyermuth: > Dear Cephalopodians, > > running 13.2.6 on the source cluster and 14.2.5 on the rbd mirror nodes and > the target

[ceph-users] RBD-mirror instabilities

2019-12-23 Thread Oliver Freyermuth
Dear Cephalopodians, running 13.2.6 on the source cluster and 14.2.5 on the rbd mirror nodes and the target cluster, I observe regular failures of rbd-mirror processes. With failures, I mean that traffic stops, but the daemons are still listed as active rbd-mirror daemons in "ceph -s", and the

[ceph-users] Re: atime with cephfs

2019-12-16 Thread Oliver Freyermuth
ers and thanks, Oliver > > On Sat, Dec 14, 2019 at 5:42 AM Oliver Freyermuth > wrote: >> >> Hi together, >> >> I had a look at ceph-fuse code and if I read it correctly, it does indeed >> not seem to have the relatime behaviour since kernels 2.6.30 i

[ceph-users] Re: Ceph on CentOS 8?

2019-12-14 Thread Oliver Freyermuth
Dear Sage, Am 13.12.19 um 17:43 schrieb Sage Weil: > Ceph Octopus will be based on CentOS 8. It's due out in March. > > The centos8 transition is awkward because our python 2 dependencies don't > exist on in centos8, and it is a huge amount of effort to produce them. > Octopus switches to pyt

[ceph-users] Re: atime with cephfs

2019-12-13 Thread Oliver Freyermuth
Hi together, I had a look at ceph-fuse code and if I read it correctly, it does indeed not seem to have the relatime behaviour since kernels 2.6.30 implemented. Should I open a ticket on this? Cheers, Oliver Am 02.12.19 um 14:31 schrieb Oliver Freyermuth: > I was thinking about

[ceph-users] Re: atime with cephfs

2019-12-02 Thread Oliver Freyermuth
; /dev/null # stat /var/log/gpu-manager.log ... Access: 2019-12-02 14:28:20.643034594 +0100 Modify: 2019-11-20 01:11:11.594634547 +0100 Change: 2019-11-20 01:11:11.594634547 +0100 I would also be fine with lazytime, though, but does ceph-fuse support it? Cheers, Oliv

[ceph-users] atime with cephfs

2019-12-02 Thread Oliver Freyermuth
Dear Cephers, we are currently mounting CephFS with relatime, using the FUSE client (version 13.2.6): ceph-fuse on /cephfs type fuse.ceph-fuse (rw,relatime,user_id=0,group_id=0,allow_other) For the first time, I wanted to use atime to identify old unused data. My expectation with "relatime"

[ceph-users] Re: Device Health Metrics on EL 7

2019-11-07 Thread Oliver Freyermuth
Dear Benjeman, dear all, indeed, after waiting a bit longer and an mgr restart, it now works (for the single case where I temporarily had SELinux off)! So at least we now know the remaining issues with health metrics :-). Cheers, Oliver Am 07.11.19 um 18:51 schrieb Oliver Freyermuth

[ceph-users] Re: Device Health Metrics on EL 7

2019-11-07 Thread Oliver Freyermuth
s there as well. I think we did something kind of similar to what is attached to this tracker issue: https://tracker.ceph.com/issues/40683 That seemed to get us as far as hosts being able to report disk health to the module. thanks, Ben On Sat, Nov 2, 2019 at 11:38 PM Oliver Freyermuth wro

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-07 Thread Oliver Freyermuth
us/dists/bionic/InRelease 500  Internal Server Error [IP: 147.204.6.136 8080] W: Some index files failed to download. They have been ignored, or old ones used instead. Regards Thomas Am 07.11.2019 um 10:04 schrieb Oliver Freyermuth: Dear Thomas, the most correct thing to do is probably to add th

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-07 Thread Oliver Freyermuth
ddities in the logs while looking through them, like https://tracker.ceph.com/issues/42666 which will hopefully have a fix ready for 14.2.5. I'm not sure about that auth error message, though! sage On Sat, 2 Nov 2019, Oliver Freyermuth wrote: Dear Sage, good news - it happened a

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-06 Thread Oliver Freyermuth
mething similar after running for a week or more with Nautilus on old and slow hardware? Cheers, Oliver Am 02.11.19 um 18:20 schrieb Oliver Freyermuth: Dear Sage, good news - it happened again, with debug logs! There's nothing obvious to my eye, it's uploaded as: 0b2d0c09-46f3-41

[ceph-users] Device Health Metrics on EL 7

2019-11-02 Thread Oliver Freyermuth
Dear Cephers, I went through some of the OSD logs of our 14.2.4 nodes and found this: -- Nov 01 01:22:25 sudo[1087697]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -a --json /dev/sds Nov 01 01:22:51 sudo[1087729]: pam_unix(sudo:auth): conv

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-02 Thread Oliver Freyermuth
e-149689> > > Related: https://tracker.ceph.com/issues/39264 > <https://tracker.ceph.com/issues/39264#change-149689> > > On 02/11/2019 17:34, Oliver Freyermuth wrote: >> Dear Reed, >> >> yes, also the balancer is on for me - but the instabilities vanished

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-02 Thread Oliver Freyermuth
of the log. I'll now go back to "stable operation", in case you need anything else, just let me know. Cheers and all the best, Oliver Am 02.11.19 um 17:38 schrieb Oliver Freyermuth: > Dear Sage, > > at least for the simple case: > ceph device get-health-me

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-02 Thread Oliver Freyermuth
and am waiting for them to become silent again. Let's hope the issue reappears before the disks run full of logs ;-). Cheers, Oliver Am 02.11.19 um 02:56 schrieb Sage Weil: > On Sat, 2 Nov 2019, Oliver Freyermuth wrote: >> Dear Cephers, >> >> interestingly, afte

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-02 Thread Oliver Freyermuth
ne pool), > but the balancer is what appeared to be my big driver of instability. > > Reed > >> On Nov 2, 2019, at 11:24 AM, Oliver Freyermuth >> wrote: >> >> Hi Thomas, >> >> indeed, I also had the dashboard open at these times - but right now

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-02 Thread Oliver Freyermuth
but > not working. > In the logs of the standby MGR nodes I found an error (after restarting > service) that pointed to Ceph Dashboard. > > Since disabling the dashboard my MGRs are stable again. > > Regards > Thomas > > Am 02.11.2019 um 02:48 schrieb Oliver F

[ceph-users] Re: mgr daemons becoming unresponsive

2019-11-01 Thread Oliver Freyermuth
09 schrieb Oliver Freyermuth: > Dear Cephers, > > this is a 14.2.4 cluster with device health metrics enabled - since about a > day, all mgr daemons go "silent" on me after a few hours, i.e. "ceph -s" > shows: > > cluster: > id: 269cf2b2

[ceph-users] mgr daemons becoming unresponsive

2019-11-01 Thread Oliver Freyermuth
Dear Cephers, this is a 14.2.4 cluster with device health metrics enabled - since about a day, all mgr daemons go "silent" on me after a few hours, i.e. "ceph -s" shows: cluster: id: 269cf2b2-7e7c-4ceb-bd1b-a33d915ceee9 health: HEALTH_WARN no active mgr 1/3

[ceph-users] Re: Static website hosting with RGW

2019-10-28 Thread Oliver Freyermuth
Am 28.10.19 um 15:48 schrieb Casey Bodley: > > On 10/24/19 8:38 PM, Oliver Freyermuth wrote: >> Dear Cephers, >> >> I have a question concerning static websites with RGW. >> To my understanding, it is best to run >=1 RGW client for "classic" S3 and

[ceph-users] Re: [EXTERNAL] Static website hosting with RGW

2019-10-26 Thread Oliver Freyermuth
of a tenant bucket: /: Hope this helps. -Ben On 10/24/19, 7:38 PM, "Oliver Freyermuth" wrote: Dear Cephers, I have a question concerning static websites with RGW. To my understanding, it is best to run >=1 RGW client for "classic" S3 and in addition operate

[ceph-users] Static website hosting with RGW

2019-10-24 Thread Oliver Freyermuth
Dear Cephers, I have a question concerning static websites with RGW. To my understanding, it is best to run >=1 RGW client for "classic" S3 and in addition operate >=1 RGW client for website serving (potentially with HAProxy or its friends in front) to prevent messup of requests via the differe

[ceph-users] Re: upmap supported in SLES 12SPx

2019-09-16 Thread Oliver Freyermuth
Am 16.09.19 um 11:06 schrieb Konstantin Shalygin: On 9/16/19 3:59 PM, Thomas wrote: I tried to run this command with failure: root@ld3955:/mnt/rbd# ceph osd set-require-min-compat-client luminous Error EPERM: cannot set require_min_compat_client to luminous: 6 connected client(s) look like jewel

[ceph-users] Re: CEPH Cluster Backup - Options on my solution

2019-08-17 Thread Oliver Freyermuth
Hi together, Am 17.08.19 um 13:31 schrieb Mike O'Connor: > >> [SNIP script] >> >> Hi mike >> >> When looking for backup solutions, did you come across benji [1][2] >> and the orginal backy2 [3][4] solutions ? >> I have been running benji for a while now, and it seems solid. I use a >> second clus