I also would be interested in how bcache or flashcache would integrate.
On Mon, Oct 7, 2013 at 11:34 AM, Martin Catudal <mcatu...@metanor.ca> wrote: > Thank's Mike, > Kyle Bader suggest me also to use my large SSD (900 GB) as cache > drive using "bcache" or "flashcache". > Since I have already plan to use SSD for my journal, I would certainly > use also SSD as cache drive in addition. > > I will have to read documentation about "bcache" and his integration > with Ceph. > > Martin > > Martin Catudal > Responsable TIC > Ressources Metanor Inc > Ligne directe: (819) 218-2708 > > Le 2013-10-07 11:25, Mike Lowe a écrit : > > Based on my experience I think you are grossly underestimating the > expense and frequency of flushes issued from your vm's. This will be > especially bad if you aren't using the async flush from qemu >= 1.4.2 as > the vm is suspended while qemu waits for the flush to finish. I think your > best course of action until the caching pool work is completed (I think I > remember correctly that this is currently in development) is to either use > the ssd's as large caches with bcache or to use them for journal devices. > I'm sure there are some other more informed opinions out there on the best > use of ssd's in a ceph cluster and hopefully they will chime in. > > > > On Oct 6, 2013, at 9:23 PM, Martin Catudal <mcatu...@metanor.ca> wrote: > > > >> Hi Guys, > >> I read all Ceph documentation more than twice. I'm now very > >> comfortable with all the aspect of Ceph except for the strategy of using > >> my SSD and HDD. > >> > >> Here is my reflexion > >> > >> I've two approach in my understanding about use fast SSD (900 GB) for my > >> primary storage and huge but slower HDD (4 TB) for replicas. > >> > >> FIRST APPROACH > >> 1. I can use PG with cache write enable as my primary storage that's > >> goes on my SSD and let replicas goes on my 7200 RPM. > >> With the cache write enable, I will gain performance for my VM > >> user machine in VDI environment since Ceph client will not have to wait > >> for the replicas write confirmation on the slower HDD. > >> > >> SECOND APPROACH > >> 2. Use pools hierarchies and let's have one pool for the SSD as primary > >> and lets the replicas goes to a second pool name platter for HDD > >> replication. > >> As explain in the Ceph documentation > >> rule ssd-primary { > >> ruleset 4 > >> type replicated > >> min_size 5 > >> max_size 10 > >> step take ssd > >> step chooseleaf firstn 1 type host > >> step emit > >> step take platter > >> step chooseleaf firstn -1 type host > >> step emit > >> } > >> > >> At this point, I could not figure out what approach could have the most > >> advantage. > >> > >> Your point of view would definitely help me. > >> > >> Sincerely, > >> Martin > >> > >> -- > >> Martin Catudal > >> Responsable TIC > >> Ressources Metanor Inc > >> Ligne directe: (819) 218-2708 > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- -- *Jason Villalta* Co-founder [image: Inline image 1] 800.799.4407x1230 | www.RubixTechnology.com<http://www.rubixtechnology.com/>
<<EmailLogo.png>>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com