We're running Ceph entirely in Docker containers, so we couldn't use ceph-deploy due to the requirement of having a process management daemon (upstart, in Ubuntu's case). So, I wrote things out and templated them myself following the documentation.
Thanks for linking the bug, Christian! You saved us a lot of time and troubleshooting. I'll post a comment on the bug. Chris On Wed, Dec 10, 2014 at 8:18 PM, Christian Balzer <ch...@gol.com> wrote: > On Wed, 10 Dec 2014 20:09:01 -0800 Christopher Armstrong wrote: > > > Christian, > > > > That indeed looks like the bug! We tried with moving the monitor > > host/address into global and everything works as expected - see > > https://github.com/deis/deis/issues/2711#issuecomment-66566318 > > > > This seems like a potentially bad bug - how has it not come up before? > > Ah, but as you can see from the issue report is has come up before. > But that discussion as well as that report clearly fell through the cracks. > > It's another reason I dislike ceph-deploy, as people using just it > (probably the vast majority) will be unaffected as it stuffs everything > into [global]. > > People reading the documentation examples or coming from older versions > (and making changes to their config) will get bitten. > > Christian > > > Anything we can do to help with a patch? > > > > Chris > > > > On Wed, Dec 10, 2014 at 5:14 PM, Christian Balzer <ch...@gol.com> wrote: > > > > > > > > Hello, > > > > > > I think this might very well be my poor, unacknowledged bug report: > > > http://tracker.ceph.com/issues/10012 > > > > > > People with a mon_hosts entry in [global] (as created by ceph-deploy) > > > will be fine, people with mons specified outside of [global] will not. > > > > > > Regards, > > > > > > Christian > > > > > > On Thu, 11 Dec 2014 00:49:03 +0000 Joao Eduardo Luis wrote: > > > > > > > On 12/10/2014 09:05 PM, Gregory Farnum wrote: > > > > > What version is he running? > > > > > > > > > > Joao, does this make any sense to you? > > > > > > > > From the MonMap code I'm pretty sure that the client should have > > > > built the monmap from the [mon.X] sections, and solely based on 'mon > > > > addr'. > > > > > > > > 'mon_initial_members' is only useful to the monitors anyway, so it > > > > can be disregarded. > > > > > > > > Thus, there are two ways for a client to build a monmap: > > > > 1) based on 'mon_hosts' on the config (or -m on cli); or > > > > 2) based on 'mon addr = ip1,ip2...' from the [mon.X] sections > > > > > > > > I don't see a 'mon hosts = ip1,ip2,...' on the config file, and I'm > > > > assuming a '-m ip1,ip2...' has been supplied on the cli, so we would > > > > have been left with the 'mon addr' options on each individual [mon.X] > > > > section. > > > > > > > > We are left with two options here: assume there was unexpected > > > > behavior on this code path -- logs or steps to reproduce would be > > > > appreciated in this case! -- or assume something else failed: > > > > > > > > - are the ips on the remaining mon sections correct (nodo-1 && > > > > nodo-2)? > > > > - were all the remaining monitors up and running when the failure > > > > occurred? > > > > - were the remaining monitors reachable by the client? > > > > > > > > In case you are able to reproduce this behavior, would be nice if you > > > > could provide logs with 'debug monc = 10' and 'debug ms = 1'. > > > > > > > > Cheers! > > > > > > > > -Joao > > > > > > > > > > > > > -Greg > > > > > > > > > > On Wed, Dec 10, 2014 at 11:54 AM, Christopher Armstrong > > > > > <ch...@opdemand.com> wrote: > > > > >> Thanks Greg - I thought the same thing, but confirmed with the > > > > >> user that it appears the radosgw client is indeed using initial > > > > >> members - when he added all of his hosts to initial members, > > > > >> things worked just fine. In either event, all of the monitors > > > > >> were always fully enumerated later in the config file. Is this > > > > >> potentially a bug specific to radosgw? Here's his config file: > > > > >> > > > > >> [global] > > > > >> fsid = fc0e2e09-ade3-4ff6-b23e-f789775b2515 > > > > >> mon initial members = nodo-3 > > > > >> auth cluster required = cephx > > > > >> auth service required = cephx > > > > >> auth client required = cephx > > > > >> osd pool default size = 3 > > > > >> osd pool default min_size = 1 > > > > >> osd pool default pg_num = 128 > > > > >> osd pool default pgp_num = 128 > > > > >> osd recovery delay start = 15 > > > > >> log file = /dev/stdout > > > > >> mon_clock_drift_allowed = 1 > > > > >> > > > > >> > > > > >> [mon.nodo-1] > > > > >> host = nodo-1 > > > > >> mon addr = 192.168.2.200:6789 > > > > >> > > > > >> [mon.nodo-2] > > > > >> host = nodo-2 > > > > >> mon addr = 192.168.2.201:6789 > > > > >> > > > > >> [mon.nodo-3] > > > > >> host = nodo-3 > > > > >> mon addr = 192.168.2.202:6789 > > > > >> > > > > >> > > > > >> > > > > >> [client.radosgw.gateway] > > > > >> host = deis-store-gateway > > > > >> keyring = /etc/ceph/ceph.client.radosgw.keyring > > > > >> rgw socket path = /var/run/ceph/ceph.radosgw.gateway.fastcgi.sock > > > > >> log file = /dev/stdout > > > > >> > > > > >> > > > > >> On Wed, Dec 10, 2014 at 11:40 AM, Gregory Farnum > > > > >> <g...@gregs42.com> wrote: > > > > >>> > > > > >>> On Tue, Dec 9, 2014 at 3:11 PM, Christopher Armstrong > > > > >>> <ch...@opdemand.com> wrote: > > > > >>>> Hi folks, > > > > >>>> > > > > >>>> I think we have a bit of confusion around how initial members is > > > > >>>> used. I understand that we can specify a single monitor (or a > > > > >>>> subset of monitors) so > > > > >>>> that the cluster can form a quorum when it first comes up. This > > > > >>>> is how we're > > > > >>>> using the setting now - so the cluster can come up with just one > > > > >>>> monitor, > > > > >>>> with the other monitors to follow later. > > > > >>>> > > > > >>>> However, a Deis user reported that when the monitor in his > > > > >>>> initial members > > > > >>>> list went down, radosgw stopped functioning, even though there > > > > >>>> are three mons in his config file. I would think that the > > > > >>>> radosgw client would connect > > > > >>>> to any of the nodes in the config file to get the state of the > > > > >>>> cluster, and > > > > >>>> that the initial members list is only used when the monitors > > > > >>>> first come up > > > > >>>> and are trying to achieve quorum. > > > > >>>> > > > > >>>> The issue he filed is here: > > > https://github.com/deis/deis/issues/2711 > > > > >>>> > > > > >>>> He also found this Ceph issue filed: > > > > >>>> https://github.com/ceph/ceph/pull/1233 > > > > >>> > > > > >>> Nope, this has nothing to do with it. > > > > >>> > > > > >>>> > > > > >>>> Is that what we're seeing here? Can anyone point us in the right > > > > >>>> direction? > > > > >>> > > > > >>> I didn't see the actual conf file posted anywhere to look at, > > > > >>> but my guess is simply that (since it looks like you're using > > > > >>> generated conf files which can differ across hosts) that the one > > > > >>> on the server(s) in question don't have the monitors listed in > > > > >>> them. I'm only skimming the code, but from it and my > > > > >>> recollection, when a Ceph client starts up it will try to > > > > >>> assemble a list of monitors to contact from: 1) the contents of > > > > >>> the "mon host" config entry 2) the "mon addr" value in any of > > > > >>> the "global", "mon" or "mon.X" sections > > > > >>> > > > > >>> The clients don't even look at mon_initial_members that I can > > > > >>> see, actually — so perhaps your client config only lists the > > > > >>> initial monitor, without adding the others? > > > > >>> -Greg > > > > >> > > > > >> > > > > > _______________________________________________ > > > > > ceph-users mailing list > > > > > ceph-users@lists.ceph.com > > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > > > > > > > > > > > > > > -- > > > Christian Balzer Network/Systems Engineer > > > ch...@gol.com Global OnLine Japan/Fusion Communications > > > http://www.gol.com/ > > > > > > -- > Christian Balzer Network/Systems Engineer > ch...@gol.com Global OnLine Japan/Fusion Communications > http://www.gol.com/ >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com