I found this without much effort.
http://www.sebastien-han.fr/blog/2012/11/15/make-your-rbd-fly-with-flashcache/


On Mon, Oct 7, 2013 at 11:39 AM, Jason Villalta <ja...@rubixnet.com> wrote:

> I also would be interested in how bcache or flashcache would integrate.
>
>
> On Mon, Oct 7, 2013 at 11:34 AM, Martin Catudal <mcatu...@metanor.ca>wrote:
>
>> Thank's Mike,
>>      Kyle Bader suggest me also to use my large SSD (900 GB) as cache
>> drive using "bcache" or "flashcache".
>> Since I have already plan to use SSD for my journal, I would certainly
>> use also SSD as cache drive in addition.
>>
>> I will have to read documentation about "bcache" and his integration
>> with Ceph.
>>
>> Martin
>>
>> Martin Catudal
>> Responsable TIC
>> Ressources Metanor Inc
>> Ligne directe: (819) 218-2708
>>
>> Le 2013-10-07 11:25, Mike Lowe a écrit :
>> > Based on my experience I think you are grossly underestimating the
>> expense and frequency of flushes issued from your vm's.  This will be
>> especially bad if you aren't using the async flush from qemu >= 1.4.2 as
>> the vm is suspended while qemu waits for the flush to finish.  I think your
>> best course of action until the caching pool work is completed (I think I
>> remember correctly that this is currently in development) is to either use
>> the ssd's as large caches with bcache or to use them for journal devices.
>>  I'm sure there are some other more informed opinions out there on the best
>> use of ssd's in a ceph cluster and hopefully they will chime in.
>> >
>> > On Oct 6, 2013, at 9:23 PM, Martin Catudal <mcatu...@metanor.ca> wrote:
>> >
>> >> Hi Guys,
>> >>      I read all Ceph documentation more than twice. I'm now very
>> >> comfortable with all the aspect of Ceph except for the strategy of
>> using
>> >> my SSD and HDD.
>> >>
>> >> Here is my reflexion
>> >>
>> >> I've two approach in my understanding about use fast SSD (900 GB) for
>> my
>> >> primary storage and huge but slower HDD (4 TB) for replicas.
>> >>
>> >> FIRST APPROACH
>> >> 1. I can use PG with cache write enable as my primary storage that's
>> >> goes on my SSD and let replicas goes on my 7200 RPM.
>> >>       With the cache write enable, I will gain performance for my VM
>> >> user machine in VDI environment since Ceph client will not have to wait
>> >> for the replicas write confirmation on the slower HDD.
>> >>
>> >> SECOND APPROACH
>> >> 2. Use pools hierarchies and let's have one pool for the SSD as primary
>> >> and lets the replicas goes to a second pool name platter for HDD
>> >> replication.
>> >>      As explain in the Ceph documentation
>> >>      rule ssd-primary {
>> >>                ruleset 4
>> >>                type replicated
>> >>                min_size 5
>> >>                max_size 10
>> >>                step take ssd
>> >>                step chooseleaf firstn 1 type host
>> >>                step emit
>> >>                step take platter
>> >>                step chooseleaf firstn -1 type host
>> >>                step emit
>> >>        }
>> >>
>> >> At this point, I could not figure out what approach could have the most
>> >> advantage.
>> >>
>> >> Your point of view would definitely help me.
>> >>
>> >> Sincerely,
>> >> Martin
>> >>
>> >> --
>> >> Martin Catudal
>> >> Responsable TIC
>> >> Ressources Metanor Inc
>> >> Ligne directe: (819) 218-2708
>> >> _______________________________________________
>> >> ceph-users mailing list
>> >> ceph-users@lists.ceph.com
>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>
> --
> --
> *Jason Villalta*
> Co-founder
> [image: Inline image 1]
> 800.799.4407x1230 | www.RubixTechnology.com<http://www.rubixtechnology.com/>
>



-- 
-- 
*Jason Villalta*
Co-founder
[image: Inline image 1]
800.799.4407x1230 | www.RubixTechnology.com<http://www.rubixtechnology.com/>

<<EmailLogo.png>>

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to