Hi David,
What about memory usage?
1] 23 OSD nodes: 15x 10TB Seagate Ironwolf filestore with journals on Intel
DC P3700, 70% full cluster, Dual Socket E5-2620 v4 @ 2.10GHz, 128GB RAM.
If you upgrade to bluestore, memory usage will likely increase. 15x10TB ~~
150GB RAM needed especially in recove
Hi Tim,
I wanted to share our experience here as we've been in a situation in
the past (on a friday afternoon of course...) that injecting a snaptrim
priority of 40 to all OSDs in the cluster (to speed up snaptimming)
resulted in alls OSD nodes crashing at the same time, in all 3
datacenters.
Hi,
I am happy to announce our next meetup on March 26, we will have a talk
about openATTIC presented by Jan from SuSE.
Please RSVP at https://www.meetup.com/Ceph-Berlin/events/qbpxrhyxfbjc/
Regards
--
Robert Sander
Heinlein Support GmbH
Schwedter Str. 8/9b, 10119 Berlin
http://www.heinlein-su
Hi,
the error looks like there might be something wrong with the device classes
(which are managed via separate trees with magic names behind the scenes).
Can you post your crush map and the command that you are trying to run?
Paul
2018-03-15 16:27 GMT+01:00 :
> Hi All,
>
>
>
> Having some int
Hallo,
I am on Jewel 10.2.10 and willing to upgrade to Luminous. I thought
I'd proceed same as for the upgrade to Jewel, by running ceph-ansible on
OSD nodes one by one, then on MON nodes one by one.
---> Is this a sensible way to upgrade to Luminous?
Problem: on first OSD node I
Hi,
2018-03-16 15:18 GMT+01:00 Fulvio Galeazzi :
> Hallo,
> I am on Jewel 10.2.10 and willing to upgrade to Luminous. I thought
> I'd proceed same as for the upgrade to Jewel, by running ceph-ansible on
> OSD nodes one by one, then on MON nodes one by one.
> ---> Is this a sensible wa
Hi Paul
Many thanks for the reply.
The command is: crush move rack04 room=R80-Upper
Crush map is here: https://pastebin.com/CX7GKtBy
I’ve done some more testing, and the following all work:
· Moving machines between the racks under the default root.
· Renaming racks/hosts und
Hallo Paul,
You're correct of course, thanks!
Ok tried to upgrade one MON (out of 3) to Luminous by:
- removing the MON from the cluster
- wiping ceph-common
- running ceph-ansible with "ceph_stable_release: luminous"
but I am now stuck at "[ceph-mon : collect admin and bootstrap keys]"
Hi all,
I have a very small cluster consisting of 1 overloaded OSD node and a
couple MON/MGR/MDS nodes. I will be adding new OSD nodes to the cluster and
need to move 36 drives from the existing node to a new one. I'm running
Luminous 12.2.2 on Ubuntu 16.04 and everything was created with ceph-dep
Hi All,
Can someone confirm please that, for a perfect performance/safety
compromise, the following would be the best settings ( id 0 is SSD, id 1
is HDD )
Alternatively, any suggestions / sharing configuration / advice would be
greatly appreciated
Note
server is a DELL R620 with PERC 710 , 1GB
Hi,
looks like it fails to adjust the number of weight set entries when moving
the entries. The good news is that this is 100% reproducible with your
crush map:
you should open a bug at http://tracker.ceph.com/ to get this fixed.
Deleting the weight set fixes the problem. Moving the item manually
Hallo Paul, thanks for your tip which guided me to success.
I just needed to manually update via yum and restart services: MONs
first, then OSDs. I am happily running Luminous, now, and verified
ceph-ansible can add new disks.
Thanks
Fulvio
Original Messa
Hi Paul,
Many thanks for the super quick replys and analysis on this.
Is it a case of removing the weights from the new hosts and there osds then
moving them? After reweighing them correctly?
I already have a bug open, I will get this email chain added to this.
Warren
_
Hi jon,
Am 16. März 2018 17:00:09 MEZ schrieb Jon Light :
>Hi all,
>
>I have a very small cluster consisting of 1 overloaded OSD node and a
>couple MON/MGR/MDS nodes. I will be adding new OSD nodes to the cluster
>and
>need to move 36 drives from the existing node to a new one. I'm running
>Lumino
I have no logging options configured in ceph.conf, yet I get syslog
entries like these
Mar 16 23:55:42 c01 ceph-osd: 2018-03-16 23:55:42.535796 7f2f5c53a700 -1
osd.0 pg_epoch: 18949 pg[17.21( v 18949'4044827
(18949'4043279,18949'4044827] local-lis/les=18910/18911 n=3125
ec=3636/3636 lis/c 1
Hope this helps someone if your recovery is impacting client traffic.
We have been migrating OSD hosts and experiencing massive client
timeouts due to overwhelming recovery traffic in Jewel (3-4 GB/s), to
the point where Areca HBAs would seize up and crash the hosts.
Setting osd_recovery_sleep =
16 matches
Mail list logo