Good to know. We must have misconfigured our router when we were testing
this.

On Tue, Sep 5, 2017, 3:00 AM Morrice Ben <ben.morr...@epfl.ch> wrote:

> Hi all,
>
> Thanks for your responses. I managed to re-ip the OSDs
>
> I did not need to set cluster network or public network in [global], just
> changing the address in the [osd.#] section was sufficient.
>
>
> In my environment, the catalyst was a misconfiguration on the network
> side. After I provided iperf results between servers in the OLD and NEW,
> our network team resolved the issue and then ceph 'just worked'.
>
>
> Cheers,
>
>
> Ben​
>
> ------------------------------
>
> *From:* ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Jake
> Young <jak3...@gmail.com>
> *Sent:* Wednesday, August 30, 2017 11:37 PM
> *To:* Jeremy Hanmer; ceph-users
> *Subject:* Re: [ceph-users] Ceph re-ip of OSD node
>
> Hey Ben,
>
> Take a look at the osd log for another OSD who's ip you did not change.
>
> What errors does it show related the re-ip'd OSD?
>
> Is the other OSD trying to communicate with the re-ip'd OSD's old ip
> address?
>
> Jake
>
>
> On Wed, Aug 30, 2017 at 3:55 PM Jeremy Hanmer <jeremy.han...@dreamhost.com>
> wrote:
>
>> This is simply not true. We run quite a few ceph clusters with
>> rack-level layer2 domains (thus routing between racks) and everything
>> works great.
>>
>> On Wed, Aug 30, 2017 at 10:52 AM, David Turner <drakonst...@gmail.com>
>> wrote:
>> > ALL OSDs need to be running the same private network at the same time.
>> ALL
>> > clients, RGW, OSD, MON, MGR, MDS, etc, etc need to be running on the
>> same
>> > public network at the same time.  You cannot do this as a one at a time
>> > migration to the new IP space.  Even if all of the servers can still
>> > communicate via routing, it just won't work.  Changing the
>> public/private
>> > network addresses for a cluster requires full cluster down time.
>> >
>> > On Wed, Aug 30, 2017 at 11:09 AM Ben Morrice <ben.morr...@epfl.ch>
>> wrote:
>> >>
>> >> Hello
>> >>
>> >> We have a small cluster that we need to move to a different network in
>> >> the same datacentre.
>> >>
>> >> My workflow was the following (for a single OSD host), but I failed
>> >> (further details below)
>> >>
>> >> 1) ceph osd set noout
>> >> 2) stop ceph-osd processes
>> >> 3) change IP, gateway, domain (short hostname is the same), VLAN
>> >> 4) change references of OLD IP (cluster and public network) in
>> >> /etc/ceph/ceph.conf with NEW IP (see [1])
>> >> 5) start a single OSD process
>> >>
>> >> This seems to work as the NEW IP can communicate with mon hosts and osd
>> >> hosts on the OLD network, the OSD is booted and is visible via 'ceph
>> -w'
>> >> however after a few seconds the OSD drops with messages such as the
>> >> below in it's log file
>> >>
>> >> heartbeat_check: no reply from 10.1.1.100:6818 osd.14 ever on either
>> >> front or back, first ping sent 2017-08-30 16:42:14.692210 (cutoff
>> >> 2017-08-30 16:42:24.962245)
>> >>
>> >> There are logs like the above for every OSD server/process
>> >>
>> >> and then eventually a
>> >>
>> >> 2017-08-30 16:42:14.486275 7f6d2c966700  0 log_channel(cluster) log
>> >> [WRN] : map e85351 wrongly marked me down
>> >>
>> >>
>> >> Am I missing something obvious to reconfigure the network on a OSD
>> host?
>> >>
>> >>
>> >>
>> >> [1]
>> >>
>> >> OLD
>> >> [osd.0]
>> >>     host = sn01
>> >>     devs = /dev/sdi
>> >>     cluster addr = 10.1.1.101
>> >>     public addr = 10.1.1.101
>> >> NEW
>> >> [osd.0]
>> >>     host = sn01
>> >>     devs = /dev/sdi
>> >>     cluster addr = 10.1.2.101
>> >>     public addr = 10.1.2.101
>> >>
>> >> --
>> >> Kind regards,
>> >>
>> >> Ben Morrice
>> >>
>> >> ______________________________________________________________________
>> >> Ben Morrice | e: ben.morr...@epfl.ch | t: +41-21-693-9670
>> >> EPFL / BBP
>> >> Biotech Campus
>> >> Chemin des Mines 9
>> >> 1202 Geneva
>> >> Switzerland
>> >>
>> >> _______________________________________________
>> >> ceph-users mailing list
>> >> ceph-users@lists.ceph.com
>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to