> On Nov 3, 2014, at 5:27 AM, Ergin Ozekes <ergin.oze...@gmail.com> wrote: > > Hi Ben; > > It will daily store approximately 2.8 Tb data. In our network 20.000 users > are exist. The ratio for cached vs new data is approximately %30. Also that > system generates 30 Gb log. > In according to these information what is your recommendations?
You expect to “write” 2.8TB of new data every day? How much of that is active typically? It’s difficult to say without knowing usage patterns, but if cost is a factor (and it usually is :), I’d probably stay with rotational drives over SSD for a very large data set, and use some of that money for more RAM. I’d probably go at least 64GB of RAM on such a system. As for ATS configs goes, the defaults would probably be just fine with that setup, at least initially, except for manually setting up the RAM cache. The one thing you can look at which can reduce memory usage on very large disk caches is proxy.config.cache.min_average_object_size. Increasing that will reduce memory consumption, but also reduce the total number of objects that the cache can hold. So, if you have a lot of very large objects, increasing this makes sense. For 10Gig NICs, there might be interesting things to do around IRQ balancing, and ring buffers and that sort of stuff. I should get our rocket scientist devops guys here to write something up. Ben and John, are you reading this? :). Cheers, — Leif