Thank's Mike,
     Kyle Bader suggest me also to use my large SSD (900 GB) as cache 
drive using "bcache" or "flashcache".
Since I have already plan to use SSD for my journal, I would certainly 
use also SSD as cache drive in addition.

I will have to read documentation about "bcache" and his integration 
with Ceph.

Martin

Martin Catudal
Responsable TIC
Ressources Metanor Inc
Ligne directe: (819) 218-2708

Le 2013-10-07 11:25, Mike Lowe a écrit :
> Based on my experience I think you are grossly underestimating the expense 
> and frequency of flushes issued from your vm's.  This will be especially bad 
> if you aren't using the async flush from qemu >= 1.4.2 as the vm is suspended 
> while qemu waits for the flush to finish.  I think your best course of action 
> until the caching pool work is completed (I think I remember correctly that 
> this is currently in development) is to either use the ssd's as large caches 
> with bcache or to use them for journal devices.  I'm sure there are some 
> other more informed opinions out there on the best use of ssd's in a ceph 
> cluster and hopefully they will chime in.
>
> On Oct 6, 2013, at 9:23 PM, Martin Catudal <mcatu...@metanor.ca> wrote:
>
>> Hi Guys,
>>      I read all Ceph documentation more than twice. I'm now very
>> comfortable with all the aspect of Ceph except for the strategy of using
>> my SSD and HDD.
>>
>> Here is my reflexion
>>
>> I've two approach in my understanding about use fast SSD (900 GB) for my
>> primary storage and huge but slower HDD (4 TB) for replicas.
>>
>> FIRST APPROACH
>> 1. I can use PG with cache write enable as my primary storage that's
>> goes on my SSD and let replicas goes on my 7200 RPM.
>>       With the cache write enable, I will gain performance for my VM
>> user machine in VDI environment since Ceph client will not have to wait
>> for the replicas write confirmation on the slower HDD.
>>
>> SECOND APPROACH
>> 2. Use pools hierarchies and let's have one pool for the SSD as primary
>> and lets the replicas goes to a second pool name platter for HDD
>> replication.
>>      As explain in the Ceph documentation
>>      rule ssd-primary {
>>                ruleset 4
>>                type replicated
>>                min_size 5
>>                max_size 10
>>                step take ssd
>>                step chooseleaf firstn 1 type host
>>                step emit
>>                step take platter
>>                step chooseleaf firstn -1 type host
>>                step emit
>>        }
>>
>> At this point, I could not figure out what approach could have the most
>> advantage.
>>
>> Your point of view would definitely help me.
>>
>> Sincerely,
>> Martin
>>
>> -- 
>> Martin Catudal
>> Responsable TIC
>> Ressources Metanor Inc
>> Ligne directe: (819) 218-2708
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to