Hi Sascha,

What is your file system type, XFS or Btrfs ?

Thanks
Wade
On Wed, Feb 3, 2016 at 7:01 AM Sascha Vogt <sascha.v...@gmail.com> wrote:

> Hi all,
>
> we recently tried adding a cache tier to our ceph cluster. We had 5
> spinning disks per hosts with a single journal NVMe disk, hosting the 5
> journals (1 OSD per spinning disk). We have 4 hosts up to now, so
> overall 4 NVMes hosting 20 journals for 20 spinning disks.
>
> As we had some space left on the NVMes so we made two additional
> partitions on each NVMe and created a 4 OSD cache tier.
>
> To our surprise the 4 OSD cache pool was able to deliver the same
> performance then the previous 20 OSD pool while reducing the OPs on the
> spinning disk to zero as long as the cache pool was sufficient to hold
> all / most data (ceph is used for very short living KVM virtual machines
> which do pretty heavy disk IO).
>
> As we don't need that much more storage right now we decided to extend
> our cluster by adding 8 additional NVMe disks solely as a cache pool and
> freeing the journal NVMes again. Now the question is: How to organize
> the OSDs on the NVMe disks (2 per host)?
>
> As the NVMes peak around 5-7 concurrent sequential writes (tested with
> fio) I thought about using 5 OSDs per NVMe. That would mean 10
> partitions (5 journals, 5 data). On the other hand the NVMes are only
> 400GB large, so that would result in OSD disk sizes for <80 GB
> (depending on the journal size).
>
> Would it make sense to skip the separate Journal partition and leave the
> journal on the data disk itself and limitting it to a rather small
> amount (lets say 1 GB or even less?) as SSDs typically don't like
> sequential writes anyway?
>
> Or, if I leave journal and data on separate partitions should I reduce
> the number of OSDs per disk to 3 as Ceph will most likly write to
> journal and data in parallel and I therefore already get 6 parallel
> "threads" of IO?
>
> Any feedback is highly appreciated :)
>
> Greetings
> -Sascha-
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to