10.11.2015 19:40, Paul Evans пишет:
> Mike - unless things have changed in the latest versions(s) of Ceph, I *not* 
> believe CRUSH will be successful in creating a valid PG map if the ’n' value 
> is 10 (k+m), your host count is 6, and your failure domain is set to host.  
> You’ll need to increase your host count to match or exceed ’n', change the 
> failure domain to OSD, or alter the k+m config to something that is more 
> compatible to your host count…otherwise you’ll end up with incomplete PG’s.
> Also note that having more failure domains (i.e. - hosts) than your ’n’ value 
> is recommended.
> 
> Beyond that, you’re likely to run operational challenges putting that many 
> drives behind a single CPU-complex when the host count is quite low. My $.02.
> --
> Paul

Thanks, Paul!
I didn't mentioned about it! It's a gold $.02 from you :)

> 
> On Nov 10, 2015, at 2:29 AM, Mike Almateia 
> <mike.almat...@gmail.com<mailto:mike.almat...@gmail.com>> wrote:
> 
> Hello.
> 
> For our CCTV storing streams project we decided to use Ceph cluster with EC 
> pool.
> Input requirements is not scary: max. 15Gbit/s input traffic from CCTV, 30 
> day storing,
> 99% write operations, a cluster must has grow up with out downtime.
> 
> By now our vision of architecture it like:
> * 6 JBOD with 90 HDD 8Tb capacity each (540 HDD total)
> * 6 Ceph servers connected to it own JBOD (we will have 6 pairs: 1 Server + 1 
> JBOD).
> 
> Ceph servers hardware details:
> * 2 x E5-2690v3 : 24 core (w/o HT), 2.6 Ghz each
> * 256 Gb RAM DDR4
> * 4 x 10Gbit/s NIC port (2 for Client network and 2 for Cluster Network)
> * servers also have 4 (8) x 2.5" HDD SATA on board for Cache Tiering Feature 
> (because ceph clients can't directly talk with EC pool)
> * Two HBA SAS controllers work with multipathing feature, for HA scenario.
> * For Ceph monitor functionality 3 servers have 2 SSD in Software RAID1
> 
> Some Ceph configuration rules:
> * EC pools with K=7 and M=3
> * EC plugin - ISA
> * technique = reed_sol_van
> * ruleset-failure-domain = host
> * near full ratio = 0.75
> * OSD journal partition on the same disk
> 
> We think that first and second problems it will be CPU and RAM on Ceph 
> servers.
> 
> Any ideas? it is can fly?
> 
> 

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to