Sorry, reading too fast. That key isn't from a previous attempt, correct? But I doubt that is the problem as you would receive an access denied message in the logs.
Try running Ceph-disk zap and recreate the OSD. Also remove the Auth key and the osd (ceph osd rm <id>) then do a ceph-disk prepare. I don't think the first stay up should be trying to create file systems, that should have been done with prepare. Robert LeBlanc Sent from a mobile device please excuse any typos. On Apr 23, 2015 3:43 AM, "Daniel Piddock" <dgp-c...@corefiling.co.uk> wrote: > On 22/04/15 20:32, Robert LeBlanc wrote: > > I believe your problem is that you haven't created bootstrap-osd key and > distributed it to your OSD node in /var/lib/ceph/bootstrap-osd/. > > > Hi Robert, > > Thank you for your reply. > > In my original post, steps performed, I did include copying over the > bootstrap-osd key. Also "ceph-disk activate" fails with an obvious error > when that file is missing: > > 2015-04-23 10:16:47.245951 7fccc5a9c700 -1 monclient(hunting): ERROR: > missing keyring, cannot use cephx for authentication > 2015-04-23 10:16:47.245955 7fccc5a9c700 0 librados: client.bootstrap-osd > initialization error (2) No such file or directory > Error connecting to cluster: ObjectNotFound > ERROR:ceph-disk:Failed to activate > ceph-disk: Error: ceph osd create failed: Command '/usr/bin/ceph' returned > non-zero exit status 1: > > This is not the source of my issue. > > Dan > > > > On Wed, Apr 22, 2015 at 5:41 AM, Daniel Piddock <dgp-c...@corefiling.co.uk > > wrote: > >> Hi, >> >> I'm a ceph newbie setting up some trial installs for evaluation. >> >> Using Debian stable (Wheezy) with Ceph Firefly from backports >> (0.80.7-1~bpo70+1). >> >> I've been following the instructions at >> http://docs.ceph.com/docs/firefly/install/manual-deployment/ and first >> time through went well, using a partition on the same drive as the OS. I >> then migrated to having data on separate harddrives and that worked too. >> >> I'm currently trying to get an OSD set up with the journal on an SSD >> partition that's separate from the data drive. ceph-disk is not playing >> ball and I've been getting various forms of failure. My greatest success >> was getting the OSD created but it would never go "up". I'm struggling >> to find anything useful in the logs or really what to look for. >> >> I purged the ceph package and wiped the storage drives to give me a >> blank slate and tried again. >> >> Steps performed: >> >> camel (MON server): >> $ apt-get install ceph >> $ uuidgen #= 8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 >> # created /etc/ceph/ceph.conf, attached >> $ ceph-authtool --create-keyring /tmp/ceph.mon.keyring --gen-key -n mon. \ >> --cap mon 'allow *' >> $ ceph-authtool --create-keyring /etc/ceph/ceph.client.admin.keyring >> --gen-key \ >> -n client.admin --set-uid=0 --cap mon 'allow *' --cap osd 'allow *' \ >> --cap mds 'allow' >> $ ceph-authtool /tmp/ceph.mon.keyring --import-keyring \ >> /etc/ceph/ceph.client.admin.keyring >> $ monmaptool --create --add a 10.1.0.3 --fsid \ >> 8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 /tmp/monmap >> $ ceph-mon --mkfs -i a --monmap /tmp/monmap --keyring >> /tmp/ceph.mon.keyring >> $ /etc/init.d/ceph start mon >> $ ceph osd lspools #= 0 data,1 metadata,2 rbd, >> >> storage node 1: >> $ apt-get install ceph >> $ rsync -a camel:/etc/ceph/ceph.conf /etc/ceph/ >> $ rsync -a camel:/var/lib/ceph/bootstrap-osd/ceph.keyring \ >> /var/lib/ceph/bootstrap-osd/ >> $ ceph-disk prepare --cluster ceph --cluster-uuid \ >> 8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 /dev/sdb /dev/sdc >> >> Output: >> cannot read partition index; assume it isn't present >> (Error: Command '/sbin/parted' returned non-zero exit status 1) >> WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the >> same device as the osd data >> Creating new GPT entries. >> Information: Moved requested sector from 34 to 2048 in >> order to align on 2048-sector boundaries. >> The operation has completed successfully. >> Creating new GPT entries. >> Information: Moved requested sector from 34 to 2048 in >> order to align on 2048-sector boundaries. >> The operation has completed successfully. >> meta-data=/dev/sdb1 isize=2048 agcount=4, agsize=15262347 >> blks >> = sectsz=512 attr=2, projid32bit=0 >> data = bsize=4096 blocks=61049385, imaxpct=25 >> = sunit=0 swidth=0 blks >> naming =version 2 bsize=4096 ascii-ci=0 >> log =internal log bsize=4096 blocks=29809, version=2 >> = sectsz=512 sunit=0 blks, lazy-count=1 >> realtime =none extsz=4096 blocks=0, rtextents=0 >> The operation has completed successfully. >> >> $ ceph-disk activate /dev/sdb1 >> Hangs >> >> Looking at ps -efH I can see that ceph-disk launched: >> /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 0 --monmap >> /var/lib/ceph/tmp/mnt.ST6Kz_/activate.monmap --osd-data >> /var/lib/ceph/tmp/mnt.ST6Kz_ --osd-journal >> /var/lib/ceph/tmp/mnt.ST6Kz_/journal --osd-uuid >> 636f694a-3677-44f0-baaf-4d74195b1806 --keyring >> /var/lib/ceph/tmp/mnt.ST6Kz_/keyring >> >> /var/lib/ceph/tmp/mnt.ST6Kz_ contains: >> activate.monmap current/ journal magic superblock >> ceph_fsid fsid journal_uuid store_version whoami >> >> journal is a symlink to >> /dev/disk/by-partuuid/798fa1c5-9751-403c-9d5a-5f7665a60d4b (sdc1) >> >> ceph osd tree: >> # id weight type name up/down reweight >> -1 0 root default >> 0 0 osd.0 down 0 >> >> If I Ctrl-C ceph-disk, kill the ceph-osd process and try again it still >> hangs. >> >> Please can somebody help? I've also attached the ceph-osd.0.log >> >> Dan >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com