> Op 16 juli 2016 om 9:06 schreef Zoltan Arnold Nagy > <zol...@linux.vnet.ibm.com>: > > > I’ve also upgraded last weekend our Hammer cluster to Jewel except for RGW. > Reading the initial user stories on the list where you had to run some > special script to create the default built-in zone discouraged me from doing > so. > > So on .2, the rgw upgrade went without any hitch just by upgrading radosgw, > or did you run any helper scripts as well? >
I did not personally upgrade the RGW. Due to time constraints somebody else did that. In this case we had two RGWs in production we were upgraded. When I asked the other day he said all went fine and no problems came up. I checked afterwards and couldn't see a problem. Wido > Thanks! > > > On 14 Jul 2016, at 08:33, Wido den Hollander <w...@42on.com> wrote: > > > > To add, the RGWs upgraded just fine as well. > > > > No regions in use here (yet!), so that upgraded as it should. > > > > Wido > > > >> Op 13 juli 2016 om 16:56 schreef Wido den Hollander <w...@42on.com>: > >> > >> > >> Hello, > >> > >> The last 3 days I worked at a customer with a 1800 OSD cluster which had > >> to be upgraded from Hammer 0.94.5 to Jewel 10.2.2 > >> > >> The cluster in this case is 99% RGW, but also some RBD. > >> > >> I wanted to share some of the things we encountered during this upgrade. > >> > >> All 180 nodes are running CentOS 7.1 on a IPv6-only network. > >> > >> ** Hammer Upgrade ** > >> At first we upgraded from 0.94.5 to 0.94.7, this went well except for the > >> fact that the monitors got spammed with these kind of messages: > >> > >> "Failed to encode map eXXX with expected crc" > >> > >> Some searching on the list brought me to: > >> > >> ceph tell osd.* injectargs -- --clog_to_monitors=false > >> > >> This reduced the load on the 5 monitors and made recovery succeed smoothly. > >> > >> ** Monitors to Jewel ** > >> The next step was to upgrade the monitors from Hammer to Jewel. > >> > >> Using Salt we upgraded the packages and afterwards it was simple: > >> > >> killall ceph-mon > >> chown -R ceph:ceph /var/lib/ceph > >> chown -R ceph:ceph /var/log/ceph > >> > >> Now, a systemd quirck. 'systemctl start ceph.target' does not work, I had > >> to manually enabled the monitor and start it: > >> > >> systemctl enable ceph-mon@srv-zmb04-05.service > >> systemctl start ceph-mon@srv-zmb04-05.service > >> > >> Afterwards the monitors were running just fine. > >> > >> ** OSDs to Jewel ** > >> To upgrade the OSDs to Jewel we initially used Salt to update the packages > >> on all systems to 10.2.2, we then used a Shell script which we ran on one > >> node at a time. > >> > >> The failure domain here is 'rack', so we executed this in one rack, then > >> the next one, etc, etc. > >> > >> Script can be found on Github: > >> https://gist.github.com/wido/06eac901bd42f01ca2f4f1a1d76c49a6 > >> > >> Be aware that the chown can take a long, long, very long time! > >> > >> We ran into the issue that some OSDs crashed after start. But after trying > >> again they would start. > >> > >> "void FileStore::init_temp_collections()" > >> > >> I reported this in the tracker as I'm not sure what is happening here: > >> http://tracker.ceph.com/issues/16672 > >> > >> ** New OSDs with Jewel ** > >> We also had some new nodes which we wanted to add to the Jewel cluster. > >> > >> Using Salt and ceph-disk we ran into a partprobe issue in combination with > >> ceph-disk. There was already a Pull Request for the fix, but that was not > >> included in Jewel 10.2.2. > >> > >> We manually applied the PR and it fixed our issues: > >> https://github.com/ceph/ceph/pull/9330 > >> > >> Hope this helps other people with their upgrades to Jewel! > >> > >> Wido > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com