Adam,

I believe you need the command 'ceph osd create' prior to 'ceph-osd -i X --mkfs --mkkey' for each OSD you add.

http://ceph.com/docs/master/rados/operations/add-or-rm-osds/#adding-an-osd-manual

Cheers,
Mike

On 4/5/2014 7:37 PM, Adam Clark wrote:
HI all,
   I am trying to setup a Ceph cluster for the first time.

I am following the manual deployment at I want to orchestrate it with
puppet.

http://ceph.com/docs/master/install/manual-deployment/

All is going well until I want to add the OSD to the crush map.

I get the following error:
ceph osd crush add osd.0 1.0 host=ceph-osd133
Error ENOENT: osd.0 does not exist.  create it before updating the crush map

Here is the process that I went through:
ceph -v
ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)

cat /etc/ceph/ceph.conf
[global]
osd_pool_default_pgp_num = 100
osd_pool_default_min_size = 1
auth_service_required = cephx
mon_initial_members = ceph-mon01,ceph-mon02,ceph-mon03
fsid = 983a74a9-1e99-42ef-8a1d-097553c3e6ce
cluster_network = 172.16.34.0/24 <http://172.16.34.0/24>
auth_supported = cephx
auth_cluster_required = cephx
mon_host = 172.16.33.20,172.16.33.21,172.16.33.22
auth_client_required = cephx
osd_pool_default_size = 2
osd_pool_default_pg_num = 100
public_network = 172.16.33.0/24 <http://172.16.33.0/24>

ceph -s
     cluster 983a74a9-1e99-42ef-8a1d-097553c3e6ce
      health HEALTH_ERR 192 pgs stuck inactive; 192 pgs stuck unclean;
no osds
      monmap e3: 3 mons at
{ceph-mon01=172.16.33.20:6789/0,ceph-mon02=172.16.33.21:6789/0,ceph-mon03=172.16.33.22:6789/0
<http://172.16.33.20:6789/0,ceph-mon02=172.16.33.21:6789/0,ceph-mon03=172.16.33.22:6789/0>},
election epoch 6, quorum 0,1,2 ceph-mon01,ceph-mon02,ceph-mon03
      osdmap e3: 0 osds: 0 up, 0 in
       pgmap v4: 192 pgs, 3 pools, 0 bytes data, 0 objects
             0 kB used, 0 kB / 0 kB avail
                  192 creating

ceph-disk list
/dev/fd0 other, unknown
/dev/sda :
  /dev/sda1 other, ext2, mounted on /boot
  /dev/sda2 other
  /dev/sda5 other, LVM2_member
/dev/sdb :
  /dev/sdb1 ceph data, active, cluster ceph, osd.0, journal /dev/sdb2
  /dev/sdb2 ceph journal, for /dev/sdb1
/dev/sr0 other, unknown

mount /dev/sdb1 /var/lib/ceph/osd/ceph-0
ceph-osd -i 0 --mkfs --mkkey
ceph auth add osd.0 osd 'allow *' mon 'allow rwx' -i
/var/lib/ceph/osd/ceph-0/keyring
ceph osd crush add-bucket ceph-osd133 host
ceph osd crush move ceph-osd133 root=default
ceph osd crush add osd.0 1.0 host=ceph-osd133
Error ENOENT: osd.0 does not exist.  create it before updating the crush map

I have seen that in earlier versions, it can show this message but
happily proceeds.

Is the doco out of date, or am I missing something?

Cheers

Adam


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to