Your probably running into issues with sysvinit / upstart / whatever. 

Try partitioning the DM and then mapping it directly in your ceph.conf under 
the osd section. 

It should work, ceph is just a process using the filesystem. 





        

Tyler Bishop 
Chief Technical Officer 
513-299-7108 x10 



tyler.bis...@beyondhosting.net 


If you are not the intended recipient of this transmission you are notified 
that disclosing, copying, distributing or taking any action in reliance on the 
contents of this information is strictly prohibited. 




From: "Andrus, Brian Contractor" <bdand...@nps.edu> 
To: ceph-users@lists.ceph.com 
Sent: Thursday, February 11, 2016 5:35:30 PM 
Subject: [ceph-users] Multipath devices with infernalis 



All, 



I have a set of hardware with a few systems connected via IB along with a DDN 
SFA12K. 

There are 4 IB/SRP paths to each block device. Those show up as 
/dev/mapper/mpath[b-d] 



I am trying to do an initial install/setup of ceph on 3 nodes. Each will be a 
monitor as well as host a single OSD. 



I am using the ceph-deploy to do most of the heavy lifting (using CentOS 
7.2.1511). 



Everything is quite successful installing monitors and even the first OSD. 



ceph status shows: 

cluster 0d9e68e4-176d-4229-866b-d408f8055e5b 

health HEALTH_OK 

monmap e1: 3 mons at 
{ceph-1-35a=10.100.1.35:6789/0,ceph-1-35b=10.100.1.85:6789/0,ceph-1-36a=10.100.1.36:6789/0}
 

election epoch 8, quorum 0,1,2 ceph-1-35a,ceph-1-36a,ceph-1-35b 

osdmap e5: 1 osds: 1 up, 1 in 

flags sortbitwise 

pgmap v8: 64 pgs, 1 pools, 0 bytes data, 0 objects 

40112 kB used, 43888 GB / 43889 GB avail 

64 active+clean 



But as soon as I try to add the next OSD on the next system using 

ceph-deploy osd create ceph-1-35b:/dev/mapper/mpathc 

things start acting up. 

The last bit from the output seems ok: 
[ceph-1-35b][INFO ] checking OSD status... 

[ceph-1-35b][INFO ] Running command: ceph --cluster=ceph osd stat --format=json 

[ceph-1-35b][WARNIN] there is 1 OSD down 

[ceph-1-35b][WARNIN] there is 1 OSD out 

[ceph_deploy.osd][DEBUG ] Host ceph-1-35b is now ready for osd use. 



But, ceph status is now: 

cluster 0d9e68e4-176d-4229-866b-d408f8055e5b 

health HEALTH_OK 

monmap e1: 3 mons at 
{ceph-1-35a=10.100.1.35:6789/0,ceph-1-35b=10.100.1.85:6789/0,ceph-1-36a=10.100.1.36:6789/0}
 

election epoch 8, quorum 0,1,2 ceph-1-35a,ceph-1-36a,ceph-1-35b 

osdmap e6: 2 osds: 1 up, 1 in 

flags sortbitwise 

pgmap v10: 64 pgs, 1 pools, 0 bytes data, 0 objects 

40120 kB used, 43888 GB / 43889 GB avail 

64 active+clean 



And ceph osd tree: 

ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY 

-1 42.86040 root default 

-2 42.86040 host ceph-1-35a 

0 42.86040 osd.0 up 1.00000 1.00000 

1 0 osd.1 down 0 1.00000 



I don’t understand why ceph-deploy didn’t activate this one when it did for the 
first. The OSD is not mounted on the other box. 

I can try to activate the down OSD (ceph-deploy disk activate 
ceph-1-35b:/dev/mapper/mpathc1:/dev/mapper/mpathc2) 

Things look good for a bit: 

cluster 0d9e68e4-176d-4229-866b-d408f8055e5b 

health HEALTH_OK 

monmap e1: 3 mons at 
{ceph-1-35a=10.100.1.35:6789/0,ceph-1-35b=10.100.1.85:6789/0,ceph-1-36a=10.100.1.36:6789/0}
 

election epoch 8, quorum 0,1,2 ceph-1-35a,ceph-1-36a,ceph-1-35b 

osdmap e8: 2 osds: 2 up, 2 in 

flags sortbitwise 

pgmap v14: 64 pgs, 1 pools, 0 bytes data, 0 objects 

74804 kB used, 87777 GB / 87778 GB avail 

64 active+clean 



But after about 1 minute, it goes down: 

ceph osd tree 

ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY 

-1 85.72079 root default 

-2 42.86040 host ceph-1-35a 

0 42.86040 osd.0 up 1.00000 1.00000 

-3 42.86040 host ceph-1-35b 

1 42.86040 osd.1 down 1.00000 1.00000 



ceph status 

cluster 0d9e68e4-176d-4229-866b-d408f8055e5b 

health HEALTH_WARN 

1/2 in osds are down 

monmap e1: 3 mons at 
{ceph-1-35a=10.100.1.35:6789/0,ceph-1-35b=10.100.1.85:6789/0,ceph-1-36a=10.100.1.36:6789/0}
 

election epoch 8, quorum 0,1,2 ceph-1-35a,ceph-1-36a,ceph-1-35b 

osdmap e9: 2 osds: 1 up, 2 in 

flags sortbitwise 

pgmap v15: 64 pgs, 1 pools, 0 bytes data, 0 objects 

74804 kB used, 87777 GB / 87778 GB avail 

64 active+clean 



Has anyone played with getting multipath devices to work? 
Of course it could be something completely different and I need to step back 
and see what step is failing. Any insight into where to dig would be 
appreciated. 



Thanks in advance, 

Brian Andrus 

ITACS/Research Computing 

Naval Postgraduate School 

Monterey, California 

voice: 831-656-6238 



_______________________________________________ 
ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to