This is a similar issue that we ran into, the root cause was that
ceph-deploy doesn't set the partition type guid (that is used to auto
activate the volume) on an existing partition. Setting this beforehand
while pre-creating the partition is a must or you have you put entries in
fstab.


On Mon, Dec 9, 2013 at 8:11 AM, Alfredo Deza <alfredo.d...@inktank.com>wrote:

> On Mon, Dec 9, 2013 at 6:49 AM, Matthew Walster <matt...@walster.org>
> wrote:
> > I'm having a play with ceph-deploy after some time away from it (mainly
> > relying on the puppet modules).
> >
> > With a test setup of only two debian testing servers, I do the following:
> >
> > ceph-deploy new host1 host2
> > ceph-deploy install host1 host2 (installs emperor)
> > ceph-deploy mon create host1 host2
> > ceph-deploy osd prepare host1:/dev/sda4 host2:/dev/sda4
> > ceph-deploy osd activate host1:/dev/sda4 host2:/dev/sda4
> > ceph-deploy mds create host1 host2
> >
> > Everything is running fine -- copy some files into CephFS, everything it
> > looking great.
> >
> > host1: /etc/init.d/ceph stop osd
> >
> > Still fine.
> >
> > host1: /etc/init.d/ceph stop mds
> >
> > Fails over to the standby mds after a few seconds. Little outage, but to
> be
> > expected. Everything fine.
> >
> > host1: /etc/init.d/ceph start osd
> > host1: /etc/init.d/ceph start mds
> >
> > Everything recovers, everything is fine.
> >
> > Now, let's do something drastic:
> >
> > host1: reboot
> > host2: reboot
> >
> > Both hosts come back up, but the mds never recovers -- it always says it
> is
> > replaying.
>
> That is something I would not expect having deployed with ceph-deploy.
>
> >
> > On closer inspection, host2's osd never came back into action. Doing:
> >
> > ceph-deploy osd activate host2:/dev/sda4 fixed the issue, and the mds
> > recovered, as well as the osd now reporting both "up" and "in".
> >
> > Is there something obvious I'm missing? The ceph.conf seemed remarkably
> > empty, do I have to re-deploy the configuration file to the monitors or
> > similar?
>
> ceph-deploy doesn't create specific entries for mon/mds/osd/'s I think
> it barely adds something in the global section for the mon initial
> members
>
> So that is actually normal ceph-deploy behavior.
>
>  I've never noticed a problem with puppet deployed hosts, but that
> > manually writes out the ceph.conf as part of the puppet run.
>
> Are you able to reproduce this in a different host from scratch? I
> just tried on a CentOS 6.4 box and everything came back after a
> reboot.
>
> It would also be very helpful to have all the output from ceph-deploy
> as you try to reproduce this behavior.
>
> >
> > Many thanks in advance,
> >
> > Matthew Walster
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>



-- 
If google has done it, Google did it right!
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to