disk_acces_mode = mmap_index_only to use fewer maps (or disable it entirely
as appropriate).



On Thu, Apr 15, 2021 at 4:42 PM Kane Wilson <k...@raft.so> wrote:

> Cassandra mmaps SSTables into memory, of which there can be many files
> (including all their indexes and what not). Typically it'll do so greedily
> until you run out of RAM. 65k map areas tends to be quite low and can
> easily be exceeded - you'd likely need very low density nodes to avoid
> going over 65k, and thus you'd require lots of nodes (making management
> harder). I'd recommend figuring out a way to up your limits as the first
> course of action.
>
> raft.so - Cassandra consulting, support, and managed services
>
>
> On Fri, Apr 16, 2021 at 4:29 AM Jai Bheemsen Rao Dhanwada <
> jaibheem...@gmail.com> wrote:
>
>> Hello All,
>>
>> The recommended settings for Cassandra suggests to have a higher value
>> for vm.max_map_count than the default 65530
>>
>> WARN  [main] 2021-04-14 19:10:52,528 StartupChecks.java:311 - Maximum
>>> number of memory map areas per process (vm.max_map_count) 65530 is too
>>> low, recommended value: 1048575, you can change it with sysctl.
>>
>>
>> However, I am running Cassandra process as a container, where I don't
>> have access to change the value on Kubernetes worker node and the cassandra
>> pod runs with less privileges.  I would like to understand why Cassandra
>> needs a higher value of memory map? and is there a way to restrict
>> Cassandra to not use beyond the default value of 65530. If there is a way
>> please let me know how to restrict and also any side effects in making that
>> change?
>>
>> Thanks in advance
>>
>

Reply via email to