Re: [ceph-users] Backfilling on Luminous

2018-03-16 Thread Caspar Smit
Hi David, What about memory usage? 1] 23 OSD nodes: 15x 10TB Seagate Ironwolf filestore with journals on Intel DC P3700, 70% full cluster, Dual Socket E5-2620 v4 @ 2.10GHz, 128GB RAM. If you upgrade to bluestore, memory usage will likely increase. 15x10TB ~~ 150GB RAM needed especially in recove

Re: [ceph-users] Disk write cache - safe?

2018-03-16 Thread Frédéric Nass
Hi Tim, I wanted to share our experience here as we've been in a situation in the past (on a friday afternoon of course...) that injecting a snaptrim priority of 40 to all OSDs in the cluster (to speed up snaptimming) resulted in alls OSD nodes crashing at the same time, in all 3 datacenters.

[ceph-users] Berlin Ceph MeetUp March 26 - openATTIC

2018-03-16 Thread Robert Sander
Hi, I am happy to announce our next meetup on March 26, we will have a talk about openATTIC presented by Jan from SuSE. Please RSVP at https://www.meetup.com/Ceph-Berlin/events/qbpxrhyxfbjc/ Regards -- Robert Sander Heinlein Support GmbH Schwedter Str. 8/9b, 10119 Berlin http://www.heinlein-su

Re: [ceph-users] Crush Bucket move crashes mons

2018-03-16 Thread Paul Emmerich
Hi, the error looks like there might be something wrong with the device classes (which are managed via separate trees with magic names behind the scenes). Can you post your crush map and the command that you are trying to run? Paul 2018-03-15 16:27 GMT+01:00 : > Hi All, > > > > Having some int

[ceph-users] Luminous "ceph-disk activate" issue

2018-03-16 Thread Fulvio Galeazzi
Hallo, I am on Jewel 10.2.10 and willing to upgrade to Luminous. I thought I'd proceed same as for the upgrade to Jewel, by running ceph-ansible on OSD nodes one by one, then on MON nodes one by one. ---> Is this a sensible way to upgrade to Luminous? Problem: on first OSD node I

Re: [ceph-users] Luminous "ceph-disk activate" issue

2018-03-16 Thread Paul Emmerich
Hi, 2018-03-16 15:18 GMT+01:00 Fulvio Galeazzi : > Hallo, > I am on Jewel 10.2.10 and willing to upgrade to Luminous. I thought > I'd proceed same as for the upgrade to Jewel, by running ceph-ansible on > OSD nodes one by one, then on MON nodes one by one. > ---> Is this a sensible wa

Re: [ceph-users] Crush Bucket move crashes mons

2018-03-16 Thread warren.jeffs
Hi Paul Many thanks for the reply. The command is: crush move rack04 room=R80-Upper Crush map is here: https://pastebin.com/CX7GKtBy I’ve done some more testing, and the following all work: · Moving machines between the racks under the default root. · Renaming racks/hosts und

Re: [ceph-users] Luminous "ceph-disk activate" issue

2018-03-16 Thread Fulvio Galeazzi
Hallo Paul, You're correct of course, thanks! Ok tried to upgrade one MON (out of 3) to Luminous by: - removing the MON from the cluster - wiping ceph-common - running ceph-ansible with "ceph_stable_release: luminous" but I am now stuck at "[ceph-mon : collect admin and bootstrap keys]"

[ceph-users] Moving OSDs between hosts

2018-03-16 Thread Jon Light
Hi all, I have a very small cluster consisting of 1 overloaded OSD node and a couple MON/MGR/MDS nodes. I will be adding new OSD nodes to the cluster and need to move 36 drives from the existing node to a new one. I'm running Luminous 12.2.2 on Ubuntu 16.04 and everything was created with ceph-dep

Re: [ceph-users] Disk write cache - safe?

2018-03-16 Thread Steven Vacaroaia
Hi All, Can someone confirm please that, for a perfect performance/safety compromise, the following would be the best settings ( id 0 is SSD, id 1 is HDD ) Alternatively, any suggestions / sharing configuration / advice would be greatly appreciated Note server is a DELL R620 with PERC 710 , 1GB

Re: [ceph-users] Crush Bucket move crashes mons

2018-03-16 Thread Paul Emmerich
Hi, looks like it fails to adjust the number of weight set entries when moving the entries. The good news is that this is 100% reproducible with your crush map: you should open a bug at http://tracker.ceph.com/ to get this fixed. Deleting the weight set fixes the problem. Moving the item manually

[ceph-users] SOLVED Re: Luminous "ceph-disk activate" issue

2018-03-16 Thread Fulvio Galeazzi
Hallo Paul, thanks for your tip which guided me to success. I just needed to manually update via yum and restart services: MONs first, then OSDs. I am happily running Luminous, now, and verified ceph-ansible can add new disks. Thanks Fulvio Original Messa

Re: [ceph-users] Crush Bucket move crashes mons

2018-03-16 Thread warren.jeffs
Hi Paul, Many thanks for the super quick replys and analysis on this. Is it a case of removing the weights from the new hosts and there osds then moving them? After reweighing them correctly? I already have a bug open, I will get this email chain added to this. Warren _

Re: [ceph-users] Moving OSDs between hosts

2018-03-16 Thread ceph
Hi jon, Am 16. März 2018 17:00:09 MEZ schrieb Jon Light : >Hi all, > >I have a very small cluster consisting of 1 overloaded OSD node and a >couple MON/MGR/MDS nodes. I will be adding new OSD nodes to the cluster >and >need to move 36 drives from the existing node to a new one. I'm running >Lumino

[ceph-users] Syslog logging date/timestamp

2018-03-16 Thread Marc Roos
I have no logging options configured in ceph.conf, yet I get syslog entries like these Mar 16 23:55:42 c01 ceph-osd: 2018-03-16 23:55:42.535796 7f2f5c53a700 -1 osd.0 pg_epoch: 18949 pg[17.21( v 18949'4044827 (18949'4043279,18949'4044827] local-lis/les=18910/18911 n=3125 ec=3636/3636 lis/c 1

[ceph-users] osd recovery sleep helped us with limiting recovery impact

2018-03-16 Thread Alex Gorbachev
Hope this helps someone if your recovery is impacting client traffic. We have been migrating OSD hosts and experiencing massive client timeouts due to overwhelming recovery traffic in Jewel (3-4 GB/s), to the point where Areca HBAs would seize up and crash the hosts. Setting osd_recovery_sleep =