I believe your problem is that you haven't created bootstrap-osd key and distributed it to your OSD node in /var/lib/ceph/bootstrap-osd/.
On Wed, Apr 22, 2015 at 5:41 AM, Daniel Piddock <dgp-c...@corefiling.co.uk> wrote: > Hi, > > I'm a ceph newbie setting up some trial installs for evaluation. > > Using Debian stable (Wheezy) with Ceph Firefly from backports > (0.80.7-1~bpo70+1). > > I've been following the instructions at > http://docs.ceph.com/docs/firefly/install/manual-deployment/ and first > time through went well, using a partition on the same drive as the OS. I > then migrated to having data on separate harddrives and that worked too. > > I'm currently trying to get an OSD set up with the journal on an SSD > partition that's separate from the data drive. ceph-disk is not playing > ball and I've been getting various forms of failure. My greatest success > was getting the OSD created but it would never go "up". I'm struggling > to find anything useful in the logs or really what to look for. > > I purged the ceph package and wiped the storage drives to give me a > blank slate and tried again. > > Steps performed: > > camel (MON server): > $ apt-get install ceph > $ uuidgen #= 8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 > # created /etc/ceph/ceph.conf, attached > $ ceph-authtool --create-keyring /tmp/ceph.mon.keyring --gen-key -n mon. \ > --cap mon 'allow *' > $ ceph-authtool --create-keyring /etc/ceph/ceph.client.admin.keyring > --gen-key \ > -n client.admin --set-uid=0 --cap mon 'allow *' --cap osd 'allow *' \ > --cap mds 'allow' > $ ceph-authtool /tmp/ceph.mon.keyring --import-keyring \ > /etc/ceph/ceph.client.admin.keyring > $ monmaptool --create --add a 10.1.0.3 --fsid \ > 8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 /tmp/monmap > $ ceph-mon --mkfs -i a --monmap /tmp/monmap --keyring /tmp/ceph.mon.keyring > $ /etc/init.d/ceph start mon > $ ceph osd lspools #= 0 data,1 metadata,2 rbd, > > storage node 1: > $ apt-get install ceph > $ rsync -a camel:/etc/ceph/ceph.conf /etc/ceph/ > $ rsync -a camel:/var/lib/ceph/bootstrap-osd/ceph.keyring \ > /var/lib/ceph/bootstrap-osd/ > $ ceph-disk prepare --cluster ceph --cluster-uuid \ > 8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 /dev/sdb /dev/sdc > > Output: > cannot read partition index; assume it isn't present > (Error: Command '/sbin/parted' returned non-zero exit status 1) > WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the > same device as the osd data > Creating new GPT entries. > Information: Moved requested sector from 34 to 2048 in > order to align on 2048-sector boundaries. > The operation has completed successfully. > Creating new GPT entries. > Information: Moved requested sector from 34 to 2048 in > order to align on 2048-sector boundaries. > The operation has completed successfully. > meta-data=/dev/sdb1 isize=2048 agcount=4, agsize=15262347 > blks > = sectsz=512 attr=2, projid32bit=0 > data = bsize=4096 blocks=61049385, imaxpct=25 > = sunit=0 swidth=0 blks > naming =version 2 bsize=4096 ascii-ci=0 > log =internal log bsize=4096 blocks=29809, version=2 > = sectsz=512 sunit=0 blks, lazy-count=1 > realtime =none extsz=4096 blocks=0, rtextents=0 > The operation has completed successfully. > > $ ceph-disk activate /dev/sdb1 > Hangs > > Looking at ps -efH I can see that ceph-disk launched: > /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 0 --monmap > /var/lib/ceph/tmp/mnt.ST6Kz_/activate.monmap --osd-data > /var/lib/ceph/tmp/mnt.ST6Kz_ --osd-journal > /var/lib/ceph/tmp/mnt.ST6Kz_/journal --osd-uuid > 636f694a-3677-44f0-baaf-4d74195b1806 --keyring > /var/lib/ceph/tmp/mnt.ST6Kz_/keyring > > /var/lib/ceph/tmp/mnt.ST6Kz_ contains: > activate.monmap current/ journal magic superblock > ceph_fsid fsid journal_uuid store_version whoami > > journal is a symlink to > /dev/disk/by-partuuid/798fa1c5-9751-403c-9d5a-5f7665a60d4b (sdc1) > > ceph osd tree: > # id weight type name up/down reweight > -1 0 root default > 0 0 osd.0 down 0 > > If I Ctrl-C ceph-disk, kill the ceph-osd process and try again it still > hangs. > > Please can somebody help? I've also attached the ceph-osd.0.log > > Dan > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com