Not just the min_size of your cephfs data pool, but also your cephfs_metadata pool.
On Wed, Jun 14, 2017 at 4:07 PM David Turner <drakonst...@gmail.com> wrote: > Ceph recommends 1GB of RAM for ever 1TB of OSD space. Your 2GB nodes are > definitely on the low end. 50GB OSDs... I don't know what that will > require, but where you're running the mon and mds on the same node, I'd > still say that 2GB is low. The Ceph OSD daemon using 1GB of RAM is not > surprising, even at that size. > > When you say you increased the size of the pools to 3, what did you do to > the min_size? Is that still set to 2? > > On Wed, Jun 14, 2017 at 3:17 PM Daniel Carrasco <d.carra...@i2tic.com> > wrote: > >> Finally I've created three nodes, I've increased the size of pools to 3 >> and I've created 3 MDS (active, standby, standby). >> >> Today the server has decided to fail and I've noticed that failover is >> not working... The ceph -s command shows like everything was OK but the >> clients weren't able to connect and I had to restart the failing node and >> reconect the clients manually to make it work again (even I think that the >> active MDS was in another node). >> >> I don't know if maybe is because the server was not fully down, and only >> some connections were failing. I'll do some tests too see. >> >> Another question: How many memory needs a node to work?, because I've >> nodes with 2GB of RAM (one MDS, one MON and one OSD), and they have an high >> memory usage (more than 1GB on the OSD). >> The OSD size is 50GB and the data that contains is less than 3GB. >> >> Thanks, and Greetings!! >> >> 2017-06-12 23:33 GMT+02:00 Mazzystr <mazzy...@gmail.com>: >> >>> Since your app is an Apache / php app is it possible for you to >>> reconfigure the app to use S3 module rather than a posix open file()? Then >>> with Ceph drop CephFS and configure Civetweb S3 gateway? You can have >>> "active-active" endpoints with round robin dns or F5 or something. You >>> would also have to repopulate objects into the rados pools. >>> >>> Also increase that size parameter to 3. ;-) >>> >>> Lots of work for active-active but the whole stack will be much more >>> resilient coming from some with a ClearCase / NFS / stale file handles up >>> the wazoo background >>> >>> >>> >>> On Mon, Jun 12, 2017 at 10:41 AM, Daniel Carrasco <d.carra...@i2tic.com> >>> wrote: >>> >>>> 2017-06-12 16:10 GMT+02:00 David Turner <drakonst...@gmail.com>: >>>> >>>>> I have an incredibly light-weight cephfs configuration. I set up an >>>>> MDS on each mon (3 total), and have 9TB of data in cephfs. This data only >>>>> has 1 client that reads a few files at a time. I haven't noticed any >>>>> downtime when it fails over to a standby MDS. So it definitely depends on >>>>> your workload as to how a failover will affect your environment. >>>>> >>>>> On Mon, Jun 12, 2017 at 9:59 AM John Petrini <jpetr...@coredial.com> >>>>> wrote: >>>>> >>>>>> We use the following in our ceph.conf for MDS failover. We're running >>>>>> one active and one standby. Last time it failed over there was about 2 >>>>>> minutes of downtime before the mounts started responding again but it did >>>>>> recover gracefully. >>>>>> >>>>>> [mds] >>>>>> max_mds = 1 >>>>>> mds_standby_for_rank = 0 >>>>>> mds_standby_replay = true >>>>>> >>>>>> ___ >>>>>> >>>>>> John Petrini >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>> >>>> >>>> Thanks to both. >>>> Just now i'm working on that because I needs a very fast failover. For >>>> now the tests give me a very fast response when an OSD fails (about 5 >>>> seconds), but a very slow response when the main MDS fails (I've not tested >>>> the real time, but was not working for a long time). Maybe was because I >>>> created the other MDS after mount, because I've done some test just before >>>> send this email and now looks very fast (i've not noticed the downtime). >>>> >>>> Greetings!! >>>> >>>> >>>> -- >>>> _________________________________________ >>>> >>>> Daniel Carrasco Marín >>>> Ingeniería para la Innovación i2TIC, S.L. >>>> Tlf: +34 911 12 32 84 Ext: 223 >>>> www.i2tic.com >>>> _________________________________________ >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> >>> >> >> >> -- >> _________________________________________ >> >> Daniel Carrasco Marín >> Ingeniería para la Innovación i2TIC, S.L. >> Tlf: +34 911 12 32 84 Ext: 223 >> www.i2tic.com >> _________________________________________ >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com