3 HOST:
1 CPU + 4Disks(3T SATA Disk)
Ceph version: 0.80.6
OS:  Redhat 6.5
Cluster:  3 host,  and have 3 MONs + 9 OSDs( One OSD hold one Disk)

1. When cluster status is Health_OK,  I write a little  data,   then  I can 
find some block file in PG directory.

[root@rhls-test2 release]# ll data/osd/ceph-0/current/2.106_head/
total 4100
-rw-r--r--. 1 root root 4194304 Dec 17 16:25 
rb.0.1021.6b8b4567.000000000024__head_753F3906__2

2.  Before replace the osd disk , we set the cluster NOOUT flag. 

3. We stop one OSD.2 which response the PG(2.106) as replica node,  and replace 
the disk with empty disk.

4. and we format the disk with xfs filesystem, and  use the ceph-osd --mkfs 
format.
    ceph-osd -i 2 --mkfs --set-osd-fsid 86828adf-7579-4127-8789-cb5e8266f15c
    note:
    For simply to replace disk,    
    we modify the ceph-osd code , add  -set-osd-fsid option for ceph-osd to set 
the osd use the old fsid.   

5. the osd start is OK , and we can find all PG's statues is  active+clean.
cluster 7c731223-9637-4e21-a6f5-c576a9cf92a4
     health HEALTH_OK
    monmap e1: 3 mons at 
{a=192.169.1.84:6789/0,b=192.169.1.85:6789/0,c=192.169.1.86:6789/0}, election 
epoch 78, quorum 0,1,2 a,b,c
     osdmap e808: 9 osds: 9 up, 9 in
      pgmap v36218: 3072 pgs, 3 pools, 7069 MB data, 8254 objects
            48063 MB used, 22298 GB / 22345 GB avail
                3072 active+clean

6. but I find the  osd.2 disk have not any data block except the meta 
data(omap, superblock ,etc).  and I can find the  all PG's directory, but is 
empty.
[root@rhls-test2 release]# ll data/osd/ceph-2/current/2.106_head/
total 0








_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to