Hi,

Thanks for sharing :-) Have you published the tools that you used to gather 
these results ? It would be great to have a way to reproduce the same measures 
in different contexts.

Cheers

On 11/05/2015 23:13, Somnath Roy wrote:
>  
> 
> Hi Loic and community,
> 
>  
> 
> I have gathered the following data on EC backend (all flash). I have decided 
> to use Jerasure since space saving is the utmost priority.
> 
>  
> 
> Setup:
> 
> --------
> 
>  
> 
> 41 OSDs (each on 8 TB flash), 5 node Ceph cluster. 48 core HT enabled cpu/64 
> GB RAM. Tested with Rados Bench clients.
> 
>  
> 
>  
> 
> EC plug-in
> 
>       
> 
> EC ratio
> 
>       
> 
> EC fault domain
> 
>       
> 
> Workload
> 
>       
> 
> Total clients
> 
>       
> 
> Num client Host
> 
>       
> 
> Runtime (Sec)
> 
>       
> 
> QD
> 
> (single client)
> 
>       
> 
> Latency/client
> 
> (avg/Max)
> 
>       
> 
> BW (aggregated)
> 
>       
> 
> Object_size
> 
>       
> 
> Node Cpu usage %
> 
>       
> 
> BW/HT core
> 
> Jerasure
> 
>       
> 
> 9,3
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/1.2
> 
>       
> 
> 1786 MB/S
> 
>       
> 
> 4M
> 
>       
> 
> 28%
> 
>       
> 
> 132 MB/s
> 
> Jerasure
> 
>       
> 
> 9,3
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.9/2.1
> 
>       
> 
> 2174 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 35%
> 
>       
> 
> 129 MB/s
> 
> Jerasure
> 
>       
> 
> 4,1
> 
>       
> 
> Host
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/2.3
> 
>       
> 
> 1737 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 14%
> 
>       
> 
> 258 MB/s
> 
> Jerasure
> 
>       
> 
> 4,1
> 
>       
> 
> Host
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 1.0/25 (!)
> 
>       
> 
> 1783 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 14%
> 
>       
> 
> 265 MB/s
> 
> Jerasure
> 
>       
> 
> 15,3
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.6/1.4
> 
>       
> 
> 1530 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 40%
> 
>       
> 
> 79 MB/s
> 
> Jerasure
> 
>       
> 
> 15,3
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 1.0/4.7
> 
>       
> 
> 1886 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 45%
> 
>       
> 
> 87 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/1.2
> 
>       
> 
> 1917 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 24%
> 
>       
> 
> 166 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.8/2.2
> 
>       
> 
> 2281 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 28%
> 
>       
> 
> 170 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (RS_r6_op)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/1.2
> 
>       
> 
> 1876 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 25%
> 
>       
> 
> 156 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (RS_r6_op)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.8/1.9
> 
>       
> 
> 2292 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 31%
> 
>       
> 
> 154 MB/s
> 
> *Jerasure*
> 
>       
> 
> *6,2 (cauchy_orig)*
> 
>       
> 
> *OSD*
> 
>       
> 
> *PUT*
> 
>       
> 
> *4*
> 
>       
> 
> *1*
> 
>       
> 
> *100*
> 
>       
> 
> *64*
> 
>       
> 
> *0.5/1.1*
> 
>       
> 
> *2025 MB/s*
> 
>       
> 
> *4M*
> 
>       
> 
> *18%*
> 
>       
> 
> *234 MB/s*
> 
> *Jerasure*
> 
>       
> 
> *6,2 (cauchy_orig)*
> 
>       
> 
> *OSD*
> 
>       
> 
> *PUT*
> 
>       
> 
> *8*
> 
>       
> 
> *2*
> 
>       
> 
> *100*
> 
>       
> 
> *64*
> 
>       
> 
> *0.8/1.9*
> 
>       
> 
> *2497 MB/s*
> 
>       
> 
> *4M*
> 
>       
> 
> *21%*
> 
>       
> 
> *247 MB/s*
> 
> Jerasure
> 
>       
> 
> 6,2 (cauchy_good)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/1.3
> 
>       
> 
> 1947MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 18%
> 
>       
> 
> 225 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (cauchy_good)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.9/8.5
> 
>       
> 
> 2336 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 21%
> 
>       
> 
> 231 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (liberation)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.6/1.6
> 
>       
> 
> 1806 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 16%
> 
>       
> 
> 235 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (liberation)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 1.1/12
> 
>       
> 
> 1969 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 17%
> 
>       
> 
> 241 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (blaum_roth)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/1.5
> 
>       
> 
> 1859 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 17%
> 
>       
> 
> 227 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (blaum_roth)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 1.0/5.8
> 
>       
> 
> 2042 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 19%
> 
>       
> 
> 224 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (liber8tion)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/1.3
> 
>       
> 
> 1809 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 17%
> 
>       
> 
> 221 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (liber8tion)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 1.1/15.7
> 
>       
> 
> 1749 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 16%
> 
>       
> 
> 227 MB/s
> 
> Jerasure
> 
>       
> 
> 10,2 (cauchy_orig)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.5/1.3
> 
>       
> 
> 2066 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 20%
> 
>       
> 
> 215 MB/s
> 
> Jerasure
> 
>       
> 
> 10,2 (cauchy_orig)
> 
>       
> 
> OSD
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.9/6.2
> 
>       
> 
> 2019 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 24%
> 
>       
> 
> 175 MB/s
> 
> *Jerasure*
> 
>       
> 
> *14,2 (cauchy_orig)*
> 
>       
> 
> *OSD*
> 
>       
> 
> *PUT*
> 
>       
> 
> *4*
> 
>       
> 
> *1*
> 
>       
> 
> *100*
> 
>       
> 
> *64*
> 
>       
> 
> *0.5/1.5*
> 
>       
> 
> *1872 MB/s*
> 
>       
> 
> *4M*
> 
>       
> 
> *18%*
> 
>       
> 
> *216 MB/s*
> 
> *Jerasure*
> 
>       
> 
> *14,2 (cauchy_orig)*
> 
>       
> 
> *OSD*
> 
>       
> 
> *PUT*
> 
>       
> 
> *8*
> 
>       
> 
> *2*
> 
>       
> 
> *100*
> 
>       
> 
> *64*
> 
>       
> 
> *1.0/7.4*
> 
>       
> 
> *2043 Mb/s*
> 
>       
> 
> *4M*
> 
>       
> 
> *18%*
> 
>       
> 
> *236 MB/s*
> 
> Replication
> 
>       
> 
> 2 replica
> 
>       
> 
> Host
> 
>       
> 
> PUT
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.7/8.8
> 
>       
> 
> 1198 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 8%
> 
>       
> 
> 311 MB/s
> 
> Replication
> 
>       
> 
> 2 replica
> 
>       
> 
> Host
> 
>       
> 
> PUT
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 1.7/33
> 
>       
> 
> 1256 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 8%
> 
>       
> 
> 327 MB/s
> 
> Jerasure
> 
>       
> 
> 9,3
> 
>       
> 
> OSD
> 
>       
> 
> GET
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.6
> 
>       
> 
> 4338 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 24%
> 
>       
> 
> 376 MB/s
> 
> Jerasure
> 
>       
> 
> 9,3
> 
>       
> 
> OSD
> 
>       
> 
> GET
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.9
> 
>       
> 
> 8002 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 54%
> 
>       
> 
> 308 MB/s
> 
> Jerasure
> 
>       
> 
> 4,1
> 
>       
> 
> Host
> 
>       
> 
> GET
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.7
> 
>       
> 
> 4630 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 18%
> 
>       
> 
> 535 MB/s
> 
> Jerasure
> 
>       
> 
> 4,1
> 
>       
> 
> Host
> 
>       
> 
> GET
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.7
> 
>       
> 
> 8600 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 42%
> 
>       
> 
> 426 MB/s
> 
> *Jerasure*
> 
>       
> 
> *14,2 (cauchy_orig)*
> 
>       
> 
> *OSD*
> 
>       
> 
> *GET*
> 
>       
> 
> *4*
> 
>       
> 
> *1*
> 
>       
> 
> *100*
> 
>       
> 
> *64*
> 
>       
> 
> *0.2/0.7*
> 
>       
> 
> *4329 MB/s*
> 
>       
> 
> *4M*
> 
>       
> 
> *24%*
> 
>       
> 
> *375 MB/s*
> 
> Jerasure
> 
>       
> 
> 10,2 (cauchy_orig)
> 
>       
> 
> OSD
> 
>       
> 
> GET
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.6
> 
>       
> 
> 4366 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 19%
> 
>       
> 
> 478 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2 (cauchy_orig)
> 
>       
> 
> OSD
> 
>       
> 
> GET
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.7
> 
>       
> 
> 4370 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 16%
> 
>       
> 
> 569 MB/s
> 
> Jerasure
> 
>       
> 
> 6,2
> 
>       
> 
> OSD
> 
>       
> 
> GET
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.5
> 
>       
> 
> 4324 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 20%
> 
>       
> 
> 450 MB/s
> 
> Replication
> 
>       
> 
> 2 replica
> 
>       
> 
> Host
> 
>       
> 
> GET
> 
>       
> 
> 4
> 
>       
> 
> 1
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.5
> 
>       
> 
> 4418 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 8%
> 
>       
> 
> 1150 MB/s
> 
> Replication
> 
>       
> 
> 2 replica
> 
>       
> 
> Host
> 
>       
> 
> GET
> 
>       
> 
> 8
> 
>       
> 
> 2
> 
>       
> 
> 100
> 
>       
> 
> 64
> 
>       
> 
> 0.2/0.9
> 
>       
> 
> 8935 MB/s
> 
>       
> 
> 4M
> 
>       
> 
> 18%
> 
>       
> 
> 1034 MB/s
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
> Summary :
> 
> -------------
> 
>  
> 
> 1. It is doing pretty good in Reads and 4 Rados Bench clients are saturating 
> 40 GB network. With more physical server, it is scaling almost linearly and 
> saturating 40 GbE on both the host.
> 
>  
> 
> 2. As suspected with Ceph, problem is again with writes. Throughput wise it 
> is beating replicated pools in significant numbers. But, it is not scaling 
> with multiple clients and not saturating anything.
> 
>  
> 
>  
> 
> So, my question is the following.
> 
>  
> 
> 1. Probably, nothing to do with EC backend, we are suffering because of 
> filestore inefficiencies. Do you think any tunable like EC stipe size (or 
> anything else) will help here ?
> 
>  
> 
> 2. I couldn’t make fault domain as ‘host’, because of HW limitation. Do you 
> think will that play a role in performance for bigger k values ?
> 
>  
> 
> 3. Even though it is not saturating 40 GbE for writes, do you think 
> separating out public/private network will help in terms of performance ?
> 
>  
> 
>  
> 
> Any feedback on this is much appreciated.
> 
>  
> 
>  
> 
> Thanks & Regards
> 
> Somnath
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
>  
> 
> 
> ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
> 
> PLEASE NOTE: The information contained in this electronic mail message is 
> intended only for the use of the designated recipient(s) named above. If the 
> reader of this message is not the intended recipient, you are hereby notified 
> that you have received this message in error and that any review, 
> dissemination, distribution, or copying of this message is strictly 
> prohibited. If you have received this communication in error, please notify 
> the sender by telephone or e-mail (as shown above) immediately and destroy 
> any and all copies of this message in your possession (whether hard copies or 
> electronically stored copies).
> 

-- 
Loïc Dachary, Artisan Logiciel Libre

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to