HI Guys

Finally writing ..after loosing my patience to configure my cluster
multiple times but still not able to achieve active+clean .. looks like its
almost impossible to configure this on centos 6.5.

As I have to prepare a POC ceph+cinder but with this config difficult to
convince someone. Also wher ethe is no udev rules for centos 6.5 which I
copied from git also why ceph-deploy dosent create required directories on
ceph-nodes ..like /var/lib/ceph/osd , /var/lib/ceph/bootstrap-osd for
person who is configuring it for first time almost get mad what went wrong

Q1 - why it start creating pages after creating a cluster even there is no
OSD added to cluster ,, without osd where it tries to write below out bfr
adding osd to cluster

[root@ceph-node1 my-cluster]# ceph-deploy mon create-initial
[root@ceph-node1 my-cluster]# ceph -s
    cluster fbf07780-f7bf-4d92-a144-a931ef5cd4a9
     health HEALTH_ERR 192 pgs stuck inactive; 192 pgs stuck unclean; no
osds
     monmap e1: 1 mons at {ceph-node1=192.168.10.41:6789/0}, election epoch
2, quorum 0 ceph-node1
     osdmap e1: 0 osds: 0 up, 0 in
      pgmap v2: 192 pgs, 3 pools, 0 bytes data, 0 objects
            0 kB used, 0 kB / 0 kB avail
                 *192 creating*


*After adding 1st OSD*

[root@ceph-node1 my-cluster]# ceph-deploy osd --zap-disk create
ceph-node2:sdb
[root@ceph-node1 my-cluster]# ceph -w
    cluster fbf07780-f7bf-4d92-a144-a931ef5cd4a9
     health HEALTH_WARN 192 pgs degraded; 192 pgs stuck unclean
     monmap e1: 1 mons at {ceph-node1=192.168.10.41:6789/0}, election epoch
2, quorum 0 ceph-node1
     osdmap e5: 1 osds: 1 up, 1 in
      pgmap v7: 192 pgs, 3 pools, 0 bytes data, 0 objects
            35116 kB used, 5074 MB / 5108 MB avail
                 192 active+degraded

*After 2nd OSD*

[root@ceph-node1 my-cluster]# ceph-deploy osd --zap-disk create
ceph-node3:sdb
[root@ceph-node1 my-cluster]# ceph -w
    cluster fbf07780-f7bf-4d92-a144-a931ef5cd4a9
     health HEALTH_WARN 192 pgs degraded; 192 pgs stuck unclean
     monmap e1: 1 mons at {ceph-node1=192.168.10.41:6789/0}, election epoch
2, quorum 0 ceph-node1
     osdmap e8: 2 osds: 2 up, 2 in
      pgmap v13: 192 pgs, 3 pools, 0 bytes data, 0 objects
            68828 kB used, 10150 MB / 10217 MB avail
                 192 active+degraded
2014-06-06 23:29:47.358646 mon.0 [INF] pgmap v13: 192 pgs: 192
active+degraded; 0 bytes data, 68828 kB used, 10150 MB / 10217 MB avail
2014-06-06 23:31:46.711047 mon.0 [INF] pgmap v14: 192 pgs: 192
active+degraded; 0 bytes data, 68796 kB used, 10150 MB / 10217 MB avail

[root@ceph-node1 my-cluster]# cat /etc/ceph/ceph.conf
[global]
osd_pool_default_pgp_num = 100
auth_service_required = cephx
osd_pool_default_size = 2
filestore_xattr_use_omap = true
auth_client_required = cephx
osd_pool_default_pg_num = 100
auth_cluster_required = cephx
mon_host = 192.168.10.41
public_network = 192.168.10.0/24
mon_clock_drift_allowed = .3
mon_initial_members = ceph-node1
cluster_network = 192.168.10.0/24
fsid = fbf07780-f7bf-4d92-a144-a931ef5cd4a9

[root@ceph-node1 my-cluster]# ceph osd dump
epoch 8
fsid fbf07780-f7bf-4d92-a144-a931ef5cd4a9
created 2014-06-06 23:21:47.665510
modified 2014-06-06 23:29:41.411379
flags
pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0 flags hashpspool
crash_replay_interval 45 stripe_width 0
pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0 flags hashpspool
stripe_width 0
pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0 flags hashpspool
stripe_width 0
max_osd 2
osd.0 up   in  weight 1 up_from 4 up_thru 4 down_at 0 last_clean_interval
[0,0) 192.168.10.42:6800/5848 192.168.10.42:6801/5848
192.168.10.42:6802/5848 192.168.10.42:6803/5848 exists,up
0f55a826-fa5b-44b2-b2f8-7b83d15526bf
osd.1 up   in  weight 1 up_from 8 up_thru 0 down_at 0 last_clean_interval
[0,0) 192.168.10.43:6800/7758 192.168.10.43:6801/7758
192.168.10.43:6802/7758 192.168.10.43:6803/7758 exists,up
5c701240-51a2-407a-b32a-9830935c1567
[root@ceph-node1 my-cluster]# ceph osd tree
# id    weight  type name       up/down reweight
-1      0       root default
-2      0               host ceph-node2
0       0                       osd.0   up      1
-3      0               host ceph-node3
1       0                       osd.1   up      1

Thanks
Anil
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to