On Wed, Aug 23, 2017 at 2:28 PM, Christian Balzer <ch...@gol.com> wrote:
> On Wed, 23 Aug 2017 13:38:25 +0800 Nick Tan wrote: > > > Thanks for the advice Christian. I think I'm leaning more towards the > > 'traditional' storage server with 12 disks - as you say they give a lot > > more flexibility with the performance tuning/network options etc. > > > > The cache pool is an interesting idea but as you say it can get quite > > expensive for the capacities we're looking at. I'm interested in how > > bluestore performs without a flash/SSD WAL/DB. In my small scale testing > > it seems much better than filestore so I was planning on building > something > > without any flash/SSD. There's always the option of adding it later if > > required. > > > Given the lack (for large writes) of double writes with Bluestore that's > to be expected. > > Since you're looking mostly at largish, sequential writes and reads, a > pure HDDs cluster may be feasible. > > > I have a final question. Assume I'll have 96 OSD's of 10TB each and I use erasure coding with k=5 and m=3 with the failure domain of host. It's my understanding that with erasure coding and host failure domains, the minimum number of hosts required is equal to k+m. Is this correct? So, if I use 8 hosts with 12 OSDs each this means I can lose any 3 OSD's but can I lose a host, assuming there's enough free space in the pool to accomodate the missing 12 OSD's? Or will the cluster block at this point because there's only 7 hosts which is less than k+m? If I use 48 hosts with 2 OSDs each then I can lose up to any 3 OSD's and the cluster can recover. And I could lose 1 host at a time assuming the recovery completes before the next host is lost (until there's 7 hosts left or there's not enough free space in the pool). Is this a correct assessment? Thanks, Nick
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com