Hi, You will need 2 mons to be online.
Thanks On 3 Jul 2016 8:58 a.m., "Willi Fehler" <willi.feh...@t-online.de> wrote: > Hello Tu, > > yes that's correct. The mon nodes run as well on the OSD nodes. So I have > > 3 nodes in total. OSD, MDS and Mon on each Node. > > Regards - Willi > > Am 03.07.16 um 09:56 schrieb Tu Holmes: > > Where are your mon nodes? > > Were you mixing mon and OSD together? > > Are 2 of the mon nodes down as well? > On Jul 3, 2016 12:53 AM, "Willi Fehler" <willi.feh...@t-online.de> wrote: > >> Hello Sean, >> >> I've powered down 2 nodes. So 6 of 9 OSD are down. But my client can't >> write and read anymore from my Ceph mount. Also 'ceph -s' hangs. >> >> pool 1 'cephfs_data' replicated size 3 min_size 1 crush_ruleset 0 >> object_hash rjenkins pg_num 300 pgp_num 300 last_change 447 flags >> hashpspool crash_replay_interval 45 stripe_width 0 >> pool 2 'cephfs_metadata' replicated size 3 min_size 1 crush_ruleset 0 >> object_hash rjenkins pg_num 300 pgp_num 300 last_change 445 flags >> hashpspool stripe_width 0 >> >> 2016-07-03 09:49:40.695953 7f3da56f9700 0 -- 192.168.0.5:0/2773396901 >> >> 192.168.0.7:6789/0 pipe(0x7f3da0001f50 sd=3 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7f3da0000f20).fault >> 2016-07-03 09:49:44.195029 7f3da57fa700 0 -- 192.168.0.5:0/2773396901 >> >> 192.168.0.6:6789/0 pipe(0x7f3da0005500 sd=4 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7f3da00067c0).fault >> 2016-07-03 09:49:50.205788 7f3da55f8700 0 -- 192.168.0.5:0/2773396901 >> >> 192.168.0.6:6789/0 pipe(0x7f3da0005500 sd=3 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7f3da0004c40).fault >> 2016-07-03 09:49:52.720116 7f3da57fa700 0 -- 192.168.0.5:0/2773396901 >> >> 192.168.0.7:6789/0 pipe(0x7f3da00023f0 sd=4 :0 s=1 pgs=0 cs=0 l=1 >> c=0x7f3da00036b0).fault >> >> Regards - Willi >> >> Am 03.07.16 um 09:36 schrieb Sean Redmond: >> >> It would need to be set to 1 >> On 3 Jul 2016 8:17 a.m., "Willi Fehler" <willi.feh...@t-online.de> wrote: >> >>> Hello David, >>> >>> so in a 3 node Cluster how should I set min_size if I want that 2 nodes >>> could fail? >>> >>> Regards - Willi >>> >>> Am 28.06.16 um 13:07 schrieb David: >>> >>> Hi, >>> >>> This is probably the min_size on your cephfs data and/or metadata pool. >>> I believe the default is 2, if you have less than 2 replicas available I/O >>> will stop. See: >>> http://docs.ceph.com/docs/master/rados/operations/pools/#set-the-number-of-object-replicas >>> >>> On Tue, Jun 28, 2016 at 10:23 AM, willi.feh...@t-online.de < >>> willi.feh...@t-online.de> wrote: >>> >>>> Hello, >>>> >>>> I'm still very new to Ceph. I've created a small test Cluster. >>>> >>>> >>>> >>>> ceph-node1 >>>> >>>> osd0 >>>> >>>> osd1 >>>> >>>> osd2 >>>> >>>> ceph-node2 >>>> >>>> osd3 >>>> >>>> osd4 >>>> >>>> osd5 >>>> >>>> ceph-node3 >>>> >>>> osd6 >>>> >>>> osd7 >>>> >>>> osd8 >>>> >>>> >>>> >>>> My pool for CephFS has a replication count of 3. I've powered of 2 >>>> nodes(6 OSDs went down) and my cluster status became critical and my ceph >>>> clients(cephfs) run into a timeout. My data(I had only one file on my pool) >>>> was still on one of the active OSDs. Is this the expected behaviour that >>>> the Cluster status became critical and my Clients run into a timeout? >>>> >>>> >>>> >>>> Many thanks for your feedback. >>>> >>>> >>>> >>>> Regards - Willi >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> >>> >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com