Hi, Thanks for sharing :-) Have you published the tools that you used to gather these results ? It would be great to have a way to reproduce the same measures in different contexts.
Cheers On 11/05/2015 23:13, Somnath Roy wrote: > > > Hi Loic and community, > > > > I have gathered the following data on EC backend (all flash). I have decided > to use Jerasure since space saving is the utmost priority. > > > > Setup: > > -------- > > > > 41 OSDs (each on 8 TB flash), 5 node Ceph cluster. 48 core HT enabled cpu/64 > GB RAM. Tested with Rados Bench clients. > > > > > > EC plug-in > > > > EC ratio > > > > EC fault domain > > > > Workload > > > > Total clients > > > > Num client Host > > > > Runtime (Sec) > > > > QD > > (single client) > > > > Latency/client > > (avg/Max) > > > > BW (aggregated) > > > > Object_size > > > > Node Cpu usage % > > > > BW/HT core > > Jerasure > > > > 9,3 > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/1.2 > > > > 1786 MB/S > > > > 4M > > > > 28% > > > > 132 MB/s > > Jerasure > > > > 9,3 > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.9/2.1 > > > > 2174 MB/s > > > > 4M > > > > 35% > > > > 129 MB/s > > Jerasure > > > > 4,1 > > > > Host > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/2.3 > > > > 1737 MB/s > > > > 4M > > > > 14% > > > > 258 MB/s > > Jerasure > > > > 4,1 > > > > Host > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 1.0/25 (!) > > > > 1783 MB/s > > > > 4M > > > > 14% > > > > 265 MB/s > > Jerasure > > > > 15,3 > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.6/1.4 > > > > 1530 MB/s > > > > 4M > > > > 40% > > > > 79 MB/s > > Jerasure > > > > 15,3 > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 1.0/4.7 > > > > 1886 MB/s > > > > 4M > > > > 45% > > > > 87 MB/s > > Jerasure > > > > 6,2 > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/1.2 > > > > 1917 MB/s > > > > 4M > > > > 24% > > > > 166 MB/s > > Jerasure > > > > 6,2 > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.8/2.2 > > > > 2281 MB/s > > > > 4M > > > > 28% > > > > 170 MB/s > > Jerasure > > > > 6,2 (RS_r6_op) > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/1.2 > > > > 1876 MB/s > > > > 4M > > > > 25% > > > > 156 MB/s > > Jerasure > > > > 6,2 (RS_r6_op) > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.8/1.9 > > > > 2292 MB/s > > > > 4M > > > > 31% > > > > 154 MB/s > > *Jerasure* > > > > *6,2 (cauchy_orig)* > > > > *OSD* > > > > *PUT* > > > > *4* > > > > *1* > > > > *100* > > > > *64* > > > > *0.5/1.1* > > > > *2025 MB/s* > > > > *4M* > > > > *18%* > > > > *234 MB/s* > > *Jerasure* > > > > *6,2 (cauchy_orig)* > > > > *OSD* > > > > *PUT* > > > > *8* > > > > *2* > > > > *100* > > > > *64* > > > > *0.8/1.9* > > > > *2497 MB/s* > > > > *4M* > > > > *21%* > > > > *247 MB/s* > > Jerasure > > > > 6,2 (cauchy_good) > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/1.3 > > > > 1947MB/s > > > > 4M > > > > 18% > > > > 225 MB/s > > Jerasure > > > > 6,2 (cauchy_good) > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.9/8.5 > > > > 2336 MB/s > > > > 4M > > > > 21% > > > > 231 MB/s > > Jerasure > > > > 6,2 (liberation) > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.6/1.6 > > > > 1806 MB/s > > > > 4M > > > > 16% > > > > 235 MB/s > > Jerasure > > > > 6,2 (liberation) > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 1.1/12 > > > > 1969 MB/s > > > > 4M > > > > 17% > > > > 241 MB/s > > Jerasure > > > > 6,2 (blaum_roth) > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/1.5 > > > > 1859 MB/s > > > > 4M > > > > 17% > > > > 227 MB/s > > Jerasure > > > > 6,2 (blaum_roth) > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 1.0/5.8 > > > > 2042 MB/s > > > > 4M > > > > 19% > > > > 224 MB/s > > Jerasure > > > > 6,2 (liber8tion) > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/1.3 > > > > 1809 MB/s > > > > 4M > > > > 17% > > > > 221 MB/s > > Jerasure > > > > 6,2 (liber8tion) > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 1.1/15.7 > > > > 1749 MB/s > > > > 4M > > > > 16% > > > > 227 MB/s > > Jerasure > > > > 10,2 (cauchy_orig) > > > > OSD > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.5/1.3 > > > > 2066 MB/s > > > > 4M > > > > 20% > > > > 215 MB/s > > Jerasure > > > > 10,2 (cauchy_orig) > > > > OSD > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.9/6.2 > > > > 2019 MB/s > > > > 4M > > > > 24% > > > > 175 MB/s > > *Jerasure* > > > > *14,2 (cauchy_orig)* > > > > *OSD* > > > > *PUT* > > > > *4* > > > > *1* > > > > *100* > > > > *64* > > > > *0.5/1.5* > > > > *1872 MB/s* > > > > *4M* > > > > *18%* > > > > *216 MB/s* > > *Jerasure* > > > > *14,2 (cauchy_orig)* > > > > *OSD* > > > > *PUT* > > > > *8* > > > > *2* > > > > *100* > > > > *64* > > > > *1.0/7.4* > > > > *2043 Mb/s* > > > > *4M* > > > > *18%* > > > > *236 MB/s* > > Replication > > > > 2 replica > > > > Host > > > > PUT > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.7/8.8 > > > > 1198 MB/s > > > > 4M > > > > 8% > > > > 311 MB/s > > Replication > > > > 2 replica > > > > Host > > > > PUT > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 1.7/33 > > > > 1256 MB/s > > > > 4M > > > > 8% > > > > 327 MB/s > > Jerasure > > > > 9,3 > > > > OSD > > > > GET > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.2/0.6 > > > > 4338 MB/s > > > > 4M > > > > 24% > > > > 376 MB/s > > Jerasure > > > > 9,3 > > > > OSD > > > > GET > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.2/0.9 > > > > 8002 MB/s > > > > 4M > > > > 54% > > > > 308 MB/s > > Jerasure > > > > 4,1 > > > > Host > > > > GET > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.2/0.7 > > > > 4630 MB/s > > > > 4M > > > > 18% > > > > 535 MB/s > > Jerasure > > > > 4,1 > > > > Host > > > > GET > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.2/0.7 > > > > 8600 MB/s > > > > 4M > > > > 42% > > > > 426 MB/s > > *Jerasure* > > > > *14,2 (cauchy_orig)* > > > > *OSD* > > > > *GET* > > > > *4* > > > > *1* > > > > *100* > > > > *64* > > > > *0.2/0.7* > > > > *4329 MB/s* > > > > *4M* > > > > *24%* > > > > *375 MB/s* > > Jerasure > > > > 10,2 (cauchy_orig) > > > > OSD > > > > GET > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.2/0.6 > > > > 4366 MB/s > > > > 4M > > > > 19% > > > > 478 MB/s > > Jerasure > > > > 6,2 (cauchy_orig) > > > > OSD > > > > GET > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.2/0.7 > > > > 4370 MB/s > > > > 4M > > > > 16% > > > > 569 MB/s > > Jerasure > > > > 6,2 > > > > OSD > > > > GET > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.2/0.5 > > > > 4324 MB/s > > > > 4M > > > > 20% > > > > 450 MB/s > > Replication > > > > 2 replica > > > > Host > > > > GET > > > > 4 > > > > 1 > > > > 100 > > > > 64 > > > > 0.2/0.5 > > > > 4418 MB/s > > > > 4M > > > > 8% > > > > 1150 MB/s > > Replication > > > > 2 replica > > > > Host > > > > GET > > > > 8 > > > > 2 > > > > 100 > > > > 64 > > > > 0.2/0.9 > > > > 8935 MB/s > > > > 4M > > > > 18% > > > > 1034 MB/s > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Summary : > > ------------- > > > > 1. It is doing pretty good in Reads and 4 Rados Bench clients are saturating > 40 GB network. With more physical server, it is scaling almost linearly and > saturating 40 GbE on both the host. > > > > 2. As suspected with Ceph, problem is again with writes. Throughput wise it > is beating replicated pools in significant numbers. But, it is not scaling > with multiple clients and not saturating anything. > > > > > > So, my question is the following. > > > > 1. Probably, nothing to do with EC backend, we are suffering because of > filestore inefficiencies. Do you think any tunable like EC stipe size (or > anything else) will help here ? > > > > 2. I couldn’t make fault domain as ‘host’, because of HW limitation. Do you > think will that play a role in performance for bigger k values ? > > > > 3. Even though it is not saturating 40 GbE for writes, do you think > separating out public/private network will help in terms of performance ? > > > > > > Any feedback on this is much appreciated. > > > > > > Thanks & Regards > > Somnath > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ > > PLEASE NOTE: The information contained in this electronic mail message is > intended only for the use of the designated recipient(s) named above. If the > reader of this message is not the intended recipient, you are hereby notified > that you have received this message in error and that any review, > dissemination, distribution, or copying of this message is strictly > prohibited. If you have received this communication in error, please notify > the sender by telephone or e-mail (as shown above) immediately and destroy > any and all copies of this message in your possession (whether hard copies or > electronically stored copies). > -- Loïc Dachary, Artisan Logiciel Libre
signature.asc
Description: OpenPGP digital signature
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com