Please disregard the earlier message. I found the culprit: `osd_crush_update_on_start` was set to false.
*Mami Hayashida* *Research Computing Associate* Univ. of Kentucky ITS Research Computing Infrastructure On Wed, Jun 26, 2019 at 11:37 AM Hayashida, Mami <mami.hayash...@uky.edu> wrote: > I am trying to build a Ceph cluster using ceph-deploy. To add OSDs, I > used the following command (which I had successfully used before to build > another cluster): > > ceph-deploy osd create --block-db=ssd0/db0 --data=/dev/sdh osd0 > ceph-deploy osd create --block-db=ssd0/db1 --data=/dev/sdi osd0 > etc. > > Prior to running those commands, I did manually create LVs on /dev/sda for > DB/WAL with: > > *** on osd0 node*** > sudo pvcreate /dev/sda > sudo vgcreate ssd0 /dev/sda; > for i in {0..9}; do > sudo lvcreate -L 40G -n db${i} ssd0; > done > ****** > But I just realized (after creating over 240 OSDs!) neither the host nor > each osd weight was added to the CRUSH map as far as I can tell (expected > weight for each osd is 3.67799): > > cephuser@admin_node:~$ ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -1 0 root default > 0 hdd 0 osd.0 up 1.00000 1.00000 > 1 hdd 0 osd.1 up 1.00000 1.00000 > (... and so on) > > And checking the cruch map with `ceph osd crush dump` also confirms that > there are no host entries or weight (capacity) of each osd. At the same > time, > `ceph -s` and the dashboard correctly shows ` usage: 9.7 TiB used, 877 TiB > / 886 TiB avail` (correct number for all the OSDs added so far). In fact, > the dashboard even correctly groups OSDs into correct hosts. > > One additional info: I have been able to create a test pool `ceph osd pool > create mytest 8` but cannot create an object in the pool. > > I am running Ceph version mimic 13.2.6 which I installed using ceph-deploy > version 2.0.1, all servers running Ubuntu 18.0.4.2. > > Any help/advice is appreciated. > > *Mami Hayashida* > *Research Computing Associate* > Univ. of Kentucky ITS Research Computing Infrastructure > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com