Hi all,

I did a bit of an experiment with multi-mds on firefly, and it worked fine 
until one of the MDS crashed when rebalancing. It's not the end of the world, 
and I could just start fresh with the cluster, but I'm keen to see if this can 
be fixed as running multi-mds is something I would like to do in production, as 
when it was working, it did reduce load and improve response time significantly.

The output of ceph mds dump is:

dumped mdsmap epoch 1232
epoch   1232
flags   0
created 2014-03-24 23:24:35.584469
modified        2014-06-06 00:17:54.336201
tableserver     0
root    0
session_timeout 60
session_autoclose       300
max_file_size   1099511627776
last_failure    1227
last_failure_osd_epoch  24869
compat  compat={},rocompat={},incompat={1=base v0.20,2=client writeable 
ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses 
versioned encoding,6=dirfrag is stored in omap}
max_mds 2
in      0,1
up      {1=578616}
failed
stopped
data_pools      
0,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,101,105
metadata_pool   1
inline_data     disabled
578616: 10.60.8.18:6808/252227 'c' mds.1.36 up:resolve seq 2
577576: 10.60.8.19:6800/58928 'd' mds.-1.0 up:standby seq 1
577603: 10.60.8.2:6801/245281 'a' mds.-1.0 up:standby seq 1
578623: 10.60.8.3:6800/75325 'b' mds.-1.0 up:standby seq 1

Modifying max_mds has no effect, and restarting/rebooting the cluster has no 
effect. No matter what combination of commands I try with the ceph-mds binary, 
or via the ceph tool, can I make a second MDS startup, causing mds.1 to leave 
resolve and move to the next step. Running with -debug_mds 10 provides no 
really enlightening information, nor does watching the mon logs. At a guess, 
it's looking for mds.0 to communicate with.

Anyone have some pointers?

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to