----- Message from Igor Krstic <igor.z.krs...@gmail.com> ---------
   Date: Fri, 06 Jun 2014 13:23:19 +0200
   From: Igor Krstic <igor.z.krs...@gmail.com>
Subject: Re: [ceph-users] question about feature set mismatch
     To: Ilya Dryomov <ilya.dryo...@inktank.com>
     Cc: ceph-users@lists.ceph.com


On Fri, 2014-06-06 at 11:51 +0400, Ilya Dryomov wrote:
On Thu, Jun 5, 2014 at 10:38 PM, Igor Krstic <igor.z.krs...@gmail.com> wrote:
> Hello,
>
> dmesg:
> [  690.181780] libceph: mon1 192.168.214.102:6789 feature set mismatch, my
> 4a042a42 < server's 504a042a42, missing 5000000000
> [  690.181907] libceph: mon1 192.168.214.102:6789 socket error on read
> [  700.190342] libceph: mon0 192.168.214.101:6789 feature set mismatch, my
> 4a042a42 < server's 504a042a42, missing 5000000000
> [  700.190481] libceph: mon0 192.168.214.101:6789 socket error on read
> [  710.194499] libceph: mon1 192.168.214.102:6789 feature set mismatch, my
> 4a042a42 < server's 504a042a42, missing 5000000000
> [  710.194633] libceph: mon1 192.168.214.102:6789 socket error on read
> [  720.201226] libceph: mon1 192.168.214.102:6789 feature set mismatch, my
> 4a042a42 < server's 504a042a42, missing 5000000000
> [  720.201482] libceph: mon1 192.168.214.102:6789 socket error on read
>
> 5000000000 should be:
> CEPH_FEATURE_CRUSH_V2 36 1000000000
> and
> CEPH_FEATURE_OSD_ERASURE_CODES 38 4000000000
> CEPH_FEATURE_OSD_TMAP2OMAP 38* 4000000000
>
> That is happening on  two separate boxes that are just my nfs and block
> gateways (they are not osd/mon/mds). So I just need on them something like:
> sudo rbd map share2
> sudo mount -t xfs /dev/rbd1 /mnt/share2
>
> On ceph cluster and on those two separate boxes:
> ~$ ceph -v
> ceph version 0.80.1
>
> What could be the problem?

Which kernel version are you running?  Do you have any erasure coded
pools?

Thanks,

                Ilya
~$ uname -a
Linux ceph-gw1 3.13.0-24-generic #47~precise2-Ubuntu SMP Fri May 2
23:30:46 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux

Yes, one of the pools is erasure coded pool but the only thing I use on
that box is rbd and rbd pool is not ec pool. It is replicated pool. I to
not touch ec pool from there. Or at least, I believe so :)

Well, I saw something similar with CephFS: I didn't touch the pools in use by cephfs, but I created another pool with Erasure Code, and the ceph client (kernel 3.13, but not enough for EC) also stopped working with 'feature set mismatch'. Thus I guess the clients can't read the crushmap anymore when there is a 'erasure' mentioning in it:)


Thanks,
Igor

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


----- End message from Igor Krstic <igor.z.krs...@gmail.com> -----

--

Met vriendelijke groeten,
Kenneth Waegeman


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to