Hello, I have suddenly let 2 OSDs in our small 2 node cluster to be filled. Reading from the docs, i move 2 pgs dirs to another disk, so that free some disk space. Unfortunately after this the osd cannot start. Please advice! This happened before the 2:2 replication end, so it is absolutely needed to get the data back.
Thank you very much! Here's the log from the osd from which i moved directories: ceph version 0.67.4 (ad85b8bfafea6232d64cb7ba76a8b6e8252fa0c7) 1: /usr/bin/ceph-osd() [0x8fe702] 2: (()+0xf030) [0x7f90be6d3030] 3: (gsignal()+0x35) [0x7f90bcb7d475] 4: (abort()+0x180) [0x7f90bcb806f0] 5: (__gnu_cxx::__verbose_terminate_handler()+0x11d) [0x7f90bd3d289d] 6: (()+0x63996) [0x7f90bd3d0996] 7: (()+0x639c3) [0x7f90bd3d09c3] 8: (()+0x63bee) [0x7f90bd3d0bee] 9: (ceph::buffer::list::iterator::copy(unsigned int, char*)+0x127) [0x9c16a7] 10: (PG::peek_map_epoch(ObjectStore*, coll_t, hobject_t&, ceph::buffer::list*)+0x11a) [0x7f7fda] 11: (OSD::load_pgs()+0x57d) [0x79b73d] 12: (OSD::init()+0xd96) [0x79f396] 13: (main()+0x2251) [0x6bc1c1] 14: (__libc_start_main()+0xfd) [0x7f90bcb69ead] 15: /usr/bin/ceph-osd() [0x6bf0e9] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. --- logging levels --- 0/ 5 none 0/ 1 lockdep 0/ 1 context 1/ 1 crush 1/ 5 mds 1/ 5 mds_balancer 1/ 5 mds_locker 1/ 5 mds_log 1/ 5 mds_log_expire 1/ 5 mds_migrator 0/ 1 buffer 0/ 1 timer 0/ 1 filer 0/ 1 striper 0/ 1 objecter 0/ 5 rados 0/ 5 rbd 0/ 5 journaler 0/ 5 objectcacher 0/ 5 client 0/ 5 osd 0/ 5 optracker 0/ 5 objclass 1/ 3 filestore 1/ 3 journal 0/ 5 ms 1/ 5 mon 0/10 monc 1/ 5 paxos 0/ 5 tp 1/ 5 auth 1/ 5 crypto 1/ 1 finisher 1/ 5 heartbeatmap 1/ 5 perfcounter 1/ 5 rgw 1/ 5 hadoop 1/ 5 javaclient 1/ 5 asok 1/ 1 throttle -2/-2 (syslog threshold) -1/-1 (stderr threshold) max_recent 10000 max_new 1000 log_file /var/log/ceph/ceph-osd.3.log --- end dump of recent events --- Best regards, Kalin.
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com