Our compactions/repairs have already become nightmares and we have not approached the levels of data you describe here (~200 GB). Have any pointers/case studies for optimizing this?
On Nov 5, 2012, at 12:00 PM, Michael Kjellman wrote: > Rule of thumb is to try to keep nodes under 400GB. > Compactions/Repairs/Move operations etc become a nightmare otherwise. How > much data do you expect to have on each node? Also depends on caches, > bloom filters etc > > On 11/5/12 8:57 AM, "Oleg Dulin" <oleg.du...@gmail.com> wrote: > >> I have 4 nodes at my disposal. >> >> I can configure them like this: >> >> 1) RF=1, each node has 25% of the data. On random-reads, how big is the >> performance penalty if a node needs to look for data on another replica >> ? >> >> 2) RF=2, each node has 50% of the data. Same question ? >> >> >> >> -- >> Regards, >> Oleg Dulin >> NYC Java Big Data Engineer >> http://www.olegdulin.com/ >> >> > > > 'Like' us on Facebook for exclusive content and other resources on all > Barracuda Networks solutions. > Visit http://barracudanetworks.com/facebook > >