10.11.2015 19:40, Paul Evans пишет: > Mike - unless things have changed in the latest versions(s) of Ceph, I *not* > believe CRUSH will be successful in creating a valid PG map if the ’n' value > is 10 (k+m), your host count is 6, and your failure domain is set to host. > You’ll need to increase your host count to match or exceed ’n', change the > failure domain to OSD, or alter the k+m config to something that is more > compatible to your host count…otherwise you’ll end up with incomplete PG’s. > Also note that having more failure domains (i.e. - hosts) than your ’n’ value > is recommended. > > Beyond that, you’re likely to run operational challenges putting that many > drives behind a single CPU-complex when the host count is quite low. My $.02. > -- > Paul
What if we make a buckets in each jbod, example; each jbod will be have 3 buckets witch 30HDD each? And failure domain make as bucket, not host? Yes, if jbod totaly fail - we will lost 3 buckets, but a chance of failure two disk in bucket it lower, than on whole jbod. It's reasonable? > > On Nov 10, 2015, at 2:29 AM, Mike Almateia > <mike.almat...@gmail.com<mailto:mike.almat...@gmail.com>> wrote: > > Hello. > > For our CCTV storing streams project we decided to use Ceph cluster with EC > pool. > Input requirements is not scary: max. 15Gbit/s input traffic from CCTV, 30 > day storing, > 99% write operations, a cluster must has grow up with out downtime. > > By now our vision of architecture it like: > * 6 JBOD with 90 HDD 8Tb capacity each (540 HDD total) > * 6 Ceph servers connected to it own JBOD (we will have 6 pairs: 1 Server + 1 > JBOD). > > Ceph servers hardware details: > * 2 x E5-2690v3 : 24 core (w/o HT), 2.6 Ghz each > * 256 Gb RAM DDR4 > * 4 x 10Gbit/s NIC port (2 for Client network and 2 for Cluster Network) > * servers also have 4 (8) x 2.5" HDD SATA on board for Cache Tiering Feature > (because ceph clients can't directly talk with EC pool) > * Two HBA SAS controllers work with multipathing feature, for HA scenario. > * For Ceph monitor functionality 3 servers have 2 SSD in Software RAID1 > > Some Ceph configuration rules: > * EC pools with K=7 and M=3 > * EC plugin - ISA > * technique = reed_sol_van > * ruleset-failure-domain = host > * near full ratio = 0.75 > * OSD journal partition on the same disk > > We think that first and second problems it will be CPU and RAM on Ceph > servers. > > Any ideas? it is can fly? > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > -- > Paul Evans > Principal Architect > Daystrom Technology Group > m: 707-479-1034 o: 800-656-3224 x511 > f: 650-472-4005 e: > paul.ev...@daystrom.com<mailto:paul.ev...@daystrom.com> > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com