Hi, This is probably the min_size on your cephfs data and/or metadata pool. I believe the default is 2, if you have less than 2 replicas available I/O will stop. See: http://docs.ceph.com/docs/master/rados/operations/pools/#set-the-number-of-object-replicas
On Tue, Jun 28, 2016 at 10:23 AM, willi.feh...@t-online.de < willi.feh...@t-online.de> wrote: > Hello, > > I'm still very new to Ceph. I've created a small test Cluster. > > > > ceph-node1 > > osd0 > > osd1 > > osd2 > > ceph-node2 > > osd3 > > osd4 > > osd5 > > ceph-node3 > > osd6 > > osd7 > > osd8 > > > > My pool for CephFS has a replication count of 3. I've powered of 2 nodes(6 > OSDs went down) and my cluster status became critical and my ceph > clients(cephfs) run into a timeout. My data(I had only one file on my pool) > was still on one of the active OSDs. Is this the expected behaviour that > the Cluster status became critical and my Clients run into a timeout? > > > > Many thanks for your feedback. > > > > Regards - Willi > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com