Hello,

Osds are not starting on any of the nodes after I upgraded ceph-0.67.4 to
emperor 0.72.2. Tried to start osd see the following verbose output. The
same error comes up on all nodes when starting osds.

[root@ceph2 ~]# service ceph -v start osd.20
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "user"
=== osd.20 ===
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "run dir"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "pid file"
--- ceph2# mkdir -p /var/run/ceph
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "log dir"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "auto start"
--- ceph2# [ -e /var/run/ceph/osd.20.pid ] || exit 1   # no pid, presumably
not running
        pid=`cat /var/run/ceph/osd.20.pid`
        [ -e /proc/$pid ] && grep -q ceph-osd /proc/$pid/cmdline && grep
-qwe -i.20 /proc/$pid/cmdline && exit 0 # running
        exit 1  # pid is something else
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "copy executable to"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd data"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "fs path"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "devs"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "btrfs devs"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "lock file"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "admin socket"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "max open files"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "restart on core dump"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "valgrind"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd crush update on
start"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd crush location"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd crush initial
weight"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "keyring"
--- ceph2# timeout 10 /usr/bin/ceph                     --name=osd.20
            --keyring=/var/lib/ceph/osd/ceph-20/keyring
osd crush create-or-move                        --        20
        --- ceph2# df /var/lib/ceph/osd/ceph-20/. | tail -1 | awk '{ d=
$2/1073741824 ; r = sprintf("%.2f", d); print r }'
0.45                    root=default                    host=ceph2
Invalid command:  --- doesn't represent a float
osd crush create-or-move <osdname (id|osd.id)> <float[0.0-]> <args>
[<args>...] :  create entry or move existing entry for <name> <weight>
at/to location <args>
Error EINVAL: invalid command
bash: line 1: 0.45: command not found
failed: 'timeout 10 /usr/bin/ceph                       --name=osd.20
            --keyring=/var/lib/ceph/osd/ceph-20/keyring
osd crush create-or-move                        --        20
        --- ceph2# df /var/lib/ceph/osd/ceph-20/. | tail -1 | awk '{ d=
$2/1073741824 ; r = sprintf("%.2f", d); print r }'
0.45                    root=default                    host=ceph2



However, Osds start when I use ceph-osd -c /etc/ceph/ceph.conf -i <osdnum>
 but not through service ceph or /etc/init.d/ceph. After I stared all the
osds, ceph warning comes up with a message that a  "pool has too few pgs".
I deleted the pool as there wasn't any important data in it. The same
warning now comes up on a different pool.

[root@ceph1 ~]# ceph -s
    cluster c0459c67-e2cd-45f7-b580-dec1afc9dea5
     health HEALTH_WARN pool vmware-backups has too few pgs
     monmap e3: 3 mons at {a=
192.168.6.101:6789/0,b=192.168.6.102:6789/0,c=192.168.6.103:6789/0},
election epoch 17684, quorum 0,1,2 a,b,c
     mdsmap e28128: 1/1/1 up {0=a=up:active}, 1 up:standby
     osdmap e7053: 30 osds: 30 up, 30 in
      pgmap v16242514: 6348 pgs, 12 pools, 9867 GB data, 2543 kobjects
            19775 GB used, 58826 GB / 78602 GB avail
                6343 active+clean
                   5 active+clean+scrubbing+deep
  client io 0 B/s rd, 617 kB/s wr, 81 op/s

[root@ceph1 ~]# ceph osd pool delete vmware-backups vmware-backups
--yes-i-really-really-mean-it
pool 'vmware-backups' deleted
[root@ceph1 ~]# ceph -s
    cluster c0459c67-e2cd-45f7-b580-dec1afc9dea5
     health HEALTH_WARN pool centaur-backups has too few pgs
     monmap e3: 3 mons at {a=
192.168.6.101:6789/0,b=192.168.6.102:6789/0,c=192.168.6.103:6789/0},
election epoch 17684, quorum 0,1,2 a,b,c
     mdsmap e28128: 1/1/1 up {0=a=up:active}, 1 up:standby
     osdmap e7054: 30 osds: 30 up, 30 in
      pgmap v16243076: 6048 pgs, 12 pools, 4437 GB data, 1181 kobjects
            19775 GB used, 58826 GB / 78602 GB avail
                6047 active+clean
                   1 active+clean+scrubbing+deep
  client io 54836 kB/s rd, 699 op/s


Regards,
Gagan
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to