___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an emai
as I know and I'm not sure how smooth upgrade can
be.
The cluster size is 4 petabyte and taking backup is not an option :)
The worst part is I have to change OS and Kernel from arch-linux to custom made
ubuntu to be able to apply this change.
Oliver Freyermuth mailto:freyerm...@ph
t;
> Thank you for the advice, I will share the results when it's done.
>
> Regards.
>
> Oliver Freyermuth mailto:freyerm...@physik.uni-bonn.de>>, 23 Ağu 2024 Cum, 18:48
> tarihinde şunu yazdı:
>
>> Hi Özkan,
>>
>
I will
try OSD compaction.
May I learn did you do online compaction or offline?
Because I have 2 side and I can shutdown 1 entire rack and do the offline
compaction and do the same thing other side when its done.
What do you think?
Regards.
Oliver Freyermuth mailto:freyerm...@physik.uni-bon
or an HDD with lots of omap data). You can disable
this automatic conversion with:
What should I do to solve this problem?
Thanks.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
--
Oliver Freyermuth
Am 13.08.24 um 15:02 schrieb Ilya Dryomov:
On Mon, Aug 12, 2024 at 1:17 PM Oliver Freyermuth
wrote:
Am 12.08.24 um 12:16 schrieb Ilya Dryomov:
On Mon, Aug 12, 2024 at 11:28 AM Oliver Freyermuth
wrote:
Am 12.08.24 um 11:09 schrieb Ilya Dryomov:
On Mon, Aug 12, 2024 at 10:20 AM Oliver
Am 12.08.24 um 12:16 schrieb Ilya Dryomov:
On Mon, Aug 12, 2024 at 11:28 AM Oliver Freyermuth
wrote:
Am 12.08.24 um 11:09 schrieb Ilya Dryomov:
On Mon, Aug 12, 2024 at 10:20 AM Oliver Freyermuth
wrote:
Dear Cephalopodians,
we've successfully operated a "good old" Mim
Am 12.08.24 um 11:09 schrieb Ilya Dryomov:
On Mon, Aug 12, 2024 at 10:20 AM Oliver Freyermuth
wrote:
Dear Cephalopodians,
we've successfully operated a "good old" Mimic cluster with primary RBD images,
replicated via journaling to a "backup cluster" with Octopus
cluster running Octopus before, and also did not happen when both were running
Mimic.
We plan to switch to snapshot-based mirroring in the future anyways, but it
would be good to understand this strange issue in any case.
Cheers,
Oliver
[0] https://docs.ceph.com/en/octopu
istian
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
--
Oliver Freyermuth
Universität Bonn
Physikalisches Institut, Raum 1.047
Nußallee 12
53115 Bonn
--
Tel.: +49 228 73 2367
Fax: +49 228 7
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
--
Oliver Freyermuth
Universität Bonn
Physikalisches Institut, Raum 1.047
Nußallee 12
53115 Bonn
--
Tel.: +49 228 73 2367
Fax: +49 228 73 7869
--
___
ttps://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/PPLJIHT6WKYPDJ45HVJ3Z37375WIGKDW/
[2] Of course, you can (and should) use a staged rollout.
[3] Well, that's a presumption, but the fact that you mentioned user concerns
about this in the survey seems to strengthen that point.
--
a 'boring' Ceph
cluster for the medium term, I'd still advise on using packages. I don't think
this makes me a luddite :)
Regards, and apologies for the wall of text,
Matthew
[0] I think that's a fair summary!
[1] This hasn't always been true...
[2] Simple (s
FS and copy things over.
CephFS mirroring (of snapshots) in Pacific may make this easier by allowing to
reduce any actual downtime for users, but any other solution would be much
appreciated.
Cheers,
Oliver
On Mon, Apr 5, 2021 at 2:55 PM Oliver Freyermuth mailto:freyerm...@physik.uni-bon
Hi,
that really looks like a useful tool, thanks for mentioning this on the list
:-).
However, I'd also love to learn about a different way — as documentation
states:
"You may notice that object counts in your primary data pool (the one passed to fs new) continue to increase, even if files a
Am 17.03.21 um 20:09 schrieb Stefan Kooman:
On 3/17/21 7:51 PM, Martin Verges wrote:
I am still not convinced that containerizing everything brings any
benefits except the collocation of services.
Is there even a benefit?
Decoupling from underlying host OS. On a test cluster I'm running U
Hi,
we have also seen such cases, it seems that sometimes (when the controller /
device is broken in special ways),
device mapper keeps the volume locked.
You can check as follows:
1) Check if lvs / pvs / vgs show some undefined device. In that case, you may
have to flush the lvmeta cache:
Hi together,
I believe the deciding factor is whether the OSD was deployed using ceph-disk
(in "ceph-volume" speak, a "simple" OSD),
which means the metadata will be on a separate partition, or whether it was
deployed with "ceph-volume lvm".
The latter stores the metadata in LVM tags, so the e
Hi,
Am 04.09.20 um 18:20 schrieb dhils...@performair.com:
All;
We've been running RadosGW on our nautilus cluster for a while, and we're going
to be adding iSCSI capabilities to our cluster, via 2 additional servers.
I intend to also run RadosGW on these servers. That begs the question of ho
ar, so the ceph-deploy issue (whatever it was) seems solved :-).
It seems like you might have missed this step:
https://docs.ceph.com/docs/nautilus/mgr/dashboard/#enabling-the-object-gateway-management-frontend
which is necessary to let the dashboard manage the Object Gateways.
Cheers,
s (or doing
that manually).
Cheers,
Oliver
>
>
> On Sat, 25 Jul 2020, 21:36 Oliver Freyermuth, <mailto:freyerm...@physik.uni-bonn.de>> wrote:
>
> Hi,
>
> Am 22.07.20 um 12:13 schrieb sathvik vutukuri:
> > Hi,
> >
> &
Hi,
Am 22.07.20 um 12:13 schrieb sathvik vutukuri:
> Hi,
>
> Did any one installed ceph-deploy on rhel7 with rados gate way.
>
> I see there are no rpms available for rhel7 in ceph-deploy in
> download.ceph.com for nautilis , luminous, octopus versions.
where exactly did you look?
I find the
Am 03.07.20 um 20:29 schrieb Dimitri Savineau:
> You can try to use ceph-ansible which supports baremetal and containerized
> deployment.
>
> https://github.com/ceph/ceph-ansible
Thanks for the pointer!
I know about ceph-ansible. The problem is that our full infrastructure is
Puppet-based, so
Am 03.07.20 um 10:00 schrieb Sebastian Wagner:
Am 02.07.20 um 19:57 schrieb Oliver Freyermuth:
Dear Cephalopodians,
as we all know, ceph-deploy is on its demise since a while and essentially in
"maintenance mode".
We've been eyeing the "ssh orchestrator" whi
Dear Cephalopodians,
as we all know, ceph-deploy is on its demise since a while and essentially in
"maintenance mode".
We've been eyeing the "ssh orchestrator" which was in Nautilus as the
"successor in spirit" of ceph-deploy.
While we have not tried it out just yet, I find this module seems
Am 06.05.20 um 16:12 schrieb brad.swan...@adtran.com:
Take a look at the available SMR drives:
https://www.ixsystems.com/community/resources/list-of-known-smr-drives.141/
Thanks for this nice overview link!
Indeed, my question was more thinking about "the next years". While currently,
SMR is c
Dear Janne,
Am 06.05.20 um 09:18 schrieb Janne Johansson:
Den ons 6 maj 2020 kl 00:58 skrev Oliver Freyermuth mailto:freyerm...@physik.uni-bonn.de>>:
Dear Cephalopodians,
seeing the recent moves of major HDD vendors to sell SMR disks targeted for
use in consumer NAS devices (inc
Dear Cephalopodians,
seeing the recent moves of major HDD vendors to sell SMR disks targeted for use
in consumer NAS devices (including RAID systems),
I got curious and wonder what the current status of SMR support in Bluestore
is.
Of course, I'd expect disk vendors to give us host-managed SMR
Dear Jason,
Am 13.02.20 um 01:28 schrieb Jason Dillaman:
> On Wed, Feb 12, 2020 at 2:53 PM Oliver Freyermuth
> wrote:
>>
>> Dear Jason,
>>
>> Am 12.02.20 um 19:29 schrieb Jason Dillaman:
>>> On Wed, Feb 12, 2020 at 11:55 AM Oliver Freyermuth
&g
Dear Jason,
Am 12.02.20 um 19:29 schrieb Jason Dillaman:
> On Wed, Feb 12, 2020 at 11:55 AM Oliver Freyermuth
> wrote:
>>
>> Dear Cephalopodians,
>>
>> for those on the list also fighting rbd mirror process instabilities: With
>> 14.2.7 (but maybe it was als
o restarts or short
"hangs" of many OSD processes).
Cheers and hope this helps somebody with similar issues,
Oliver
Am 27.12.19 um 02:43 schrieb Oliver Freyermuth:
Dear Cephalopodians,
for those following along through the holiday season, here's my "quick hack" for
trailing brace.
I created a tracker issue here:
https://tracker.ceph.com/issues/43429
Cheers,
Oliver
Am 24.12.19 um 04:39 schrieb Oliver Freyermuth:
> Dear Cephalopodians,
>
> running 13.2.6 on the source cluster and 14.2.5 on the rbd mirror nodes and
> the target
Dear Cephalopodians,
running 13.2.6 on the source cluster and 14.2.5 on the rbd mirror nodes and the
target cluster,
I observe regular failures of rbd-mirror processes.
With failures, I mean that traffic stops, but the daemons are still listed as
active rbd-mirror daemons in
"ceph -s", and the
ers and thanks,
Oliver
>
> On Sat, Dec 14, 2019 at 5:42 AM Oliver Freyermuth
> wrote:
>>
>> Hi together,
>>
>> I had a look at ceph-fuse code and if I read it correctly, it does indeed
>> not seem to have the relatime behaviour since kernels 2.6.30 i
Dear Sage,
Am 13.12.19 um 17:43 schrieb Sage Weil:
> Ceph Octopus will be based on CentOS 8. It's due out in March.
>
> The centos8 transition is awkward because our python 2 dependencies don't
> exist on in centos8, and it is a huge amount of effort to produce them.
> Octopus switches to pyt
Hi together,
I had a look at ceph-fuse code and if I read it correctly, it does indeed not
seem to have the relatime behaviour since kernels 2.6.30 implemented.
Should I open a ticket on this?
Cheers,
Oliver
Am 02.12.19 um 14:31 schrieb Oliver Freyermuth:
> I was thinking about
; /dev/null
# stat /var/log/gpu-manager.log
...
Access: 2019-12-02 14:28:20.643034594 +0100
Modify: 2019-11-20 01:11:11.594634547 +0100
Change: 2019-11-20 01:11:11.594634547 +0100
I would also be fine with lazytime, though, but does ceph-fuse support it?
Cheers,
Oliv
Dear Cephers,
we are currently mounting CephFS with relatime, using the FUSE client (version
13.2.6):
ceph-fuse on /cephfs type fuse.ceph-fuse
(rw,relatime,user_id=0,group_id=0,allow_other)
For the first time, I wanted to use atime to identify old unused data. My expectation
with "relatime"
Dear Benjeman, dear all,
indeed, after waiting a bit longer and an mgr restart, it now works
(for the single case where I temporarily had SELinux off)!
So at least we now know the remaining issues with health metrics :-).
Cheers,
Oliver
Am 07.11.19 um 18:51 schrieb Oliver Freyermuth
s there as
well. I think we did something kind of similar to what is attached to
this tracker issue:
https://tracker.ceph.com/issues/40683
That seemed to get us as far as hosts being able to report disk health
to the module.
thanks,
Ben
On Sat, Nov 2, 2019 at 11:38 PM Oliver Freyermuth
wro
us/dists/bionic/InRelease
500 Internal Server Error [IP: 147.204.6.136 8080]
W: Some index files failed to download. They have been ignored, or old
ones used instead.
Regards
Thomas
Am 07.11.2019 um 10:04 schrieb Oliver Freyermuth:
Dear Thomas,
the most correct thing to do is probably to add th
ddities in the logs while looking through them,
like
https://tracker.ceph.com/issues/42666
which will hopefully have a fix ready for 14.2.5. I'm not sure about that
auth error message, though!
sage
On Sat, 2 Nov 2019, Oliver Freyermuth wrote:
Dear Sage,
good news - it happened a
mething similar after running for a week or more with
Nautilus on old and slow hardware?
Cheers,
Oliver
Am 02.11.19 um 18:20 schrieb Oliver Freyermuth:
Dear Sage,
good news - it happened again, with debug logs!
There's nothing obvious to my eye, it's uploaded as:
0b2d0c09-46f3-41
Dear Cephers,
I went through some of the OSD logs of our 14.2.4 nodes and found this:
--
Nov 01 01:22:25 sudo[1087697]: ceph : TTY=unknown ; PWD=/ ; USER=root ;
COMMAND=/sbin/smartctl -a --json /dev/sds
Nov 01 01:22:51 sudo[1087729]: pam_unix(sudo:auth): conv
e-149689>
>
> Related: https://tracker.ceph.com/issues/39264
> <https://tracker.ceph.com/issues/39264#change-149689>
>
> On 02/11/2019 17:34, Oliver Freyermuth wrote:
>> Dear Reed,
>>
>> yes, also the balancer is on for me - but the instabilities vanished
of the log.
I'll now go back to "stable operation", in case you need anything else, just
let me know.
Cheers and all the best,
Oliver
Am 02.11.19 um 17:38 schrieb Oliver Freyermuth:
> Dear Sage,
>
> at least for the simple case:
> ceph device get-health-me
and am waiting for them to become silent again. Let's hope the issue reappears
before the disks run full of logs ;-).
Cheers,
Oliver
Am 02.11.19 um 02:56 schrieb Sage Weil:
> On Sat, 2 Nov 2019, Oliver Freyermuth wrote:
>> Dear Cephers,
>>
>> interestingly, afte
ne pool),
> but the balancer is what appeared to be my big driver of instability.
>
> Reed
>
>> On Nov 2, 2019, at 11:24 AM, Oliver Freyermuth
>> wrote:
>>
>> Hi Thomas,
>>
>> indeed, I also had the dashboard open at these times - but right now
but
> not working.
> In the logs of the standby MGR nodes I found an error (after restarting
> service) that pointed to Ceph Dashboard.
>
> Since disabling the dashboard my MGRs are stable again.
>
> Regards
> Thomas
>
> Am 02.11.2019 um 02:48 schrieb Oliver F
09 schrieb Oliver Freyermuth:
> Dear Cephers,
>
> this is a 14.2.4 cluster with device health metrics enabled - since about a
> day, all mgr daemons go "silent" on me after a few hours, i.e. "ceph -s"
> shows:
>
> cluster:
> id: 269cf2b2
Dear Cephers,
this is a 14.2.4 cluster with device health metrics enabled - since about a
day, all mgr daemons go "silent" on me after a few hours, i.e. "ceph -s" shows:
cluster:
id: 269cf2b2-7e7c-4ceb-bd1b-a33d915ceee9
health: HEALTH_WARN
no active mgr
1/3
Am 28.10.19 um 15:48 schrieb Casey Bodley:
>
> On 10/24/19 8:38 PM, Oliver Freyermuth wrote:
>> Dear Cephers,
>>
>> I have a question concerning static websites with RGW.
>> To my understanding, it is best to run >=1 RGW client for "classic" S3 and
of a tenant bucket:
/:
Hope this helps.
-Ben
On 10/24/19, 7:38 PM, "Oliver Freyermuth" wrote:
Dear Cephers,
I have a question concerning static websites with RGW.
To my understanding, it is best to run >=1 RGW client for "classic" S3 and in addition operate
Dear Cephers,
I have a question concerning static websites with RGW.
To my understanding, it is best to run >=1 RGW client for "classic" S3 and in
addition operate >=1 RGW client for website serving
(potentially with HAProxy or its friends in front) to prevent messup of
requests via the differe
Am 16.09.19 um 11:06 schrieb Konstantin Shalygin:
On 9/16/19 3:59 PM, Thomas wrote:
I tried to run this command with failure:
root@ld3955:/mnt/rbd# ceph osd set-require-min-compat-client luminous
Error EPERM: cannot set require_min_compat_client to luminous: 6
connected client(s) look like jewel
Hi together,
Am 17.08.19 um 13:31 schrieb Mike O'Connor:
>
>> [SNIP script]
>>
>> Hi mike
>>
>> When looking for backup solutions, did you come across benji [1][2]
>> and the orginal backy2 [3][4] solutions ?
>> I have been running benji for a while now, and it seems solid. I use a
>> second clus
56 matches
Mail list logo