Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-04 Thread Reid Pinchback
. From: Shishir Kumar Reply-To: "user@cassandra.apache.org" Date: Wednesday, December 4, 2019 at 8:04 AM To: "user@cassandra.apache.org" Subject: Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB" Message from External Sender Corr

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-04 Thread Shishir Kumar
>>> >>>> >>>> >>>> One thing you’ll find out pretty quickly. There are a lot of knobs you >>>> can turn with C*, too many to allow for easy answers on what you should >>>> do. Figure out what your throughput and latency

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-04 Thread Hossein Ghiyasi Mehr
ould >>> do. Figure out what your throughput and latency SLAs are, and you’ll know >>> when to stop tuning. Otherwise you’ll discover that it’s a rabbit hole you >>> can dive into and not come out of for weeks. >>> >>> >>> >>> >>&

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-03 Thread Reid Pinchback
makes it hard to control the variables enough for my satisfaction. It can feel like a game of empirical whack-a-mole. From: Shishir Kumar Reply-To: "user@cassandra.apache.org" Date: Tuesday, December 3, 2019 at 9:23 AM To: "user@cassandra.apache.org" Subject: Re: &

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-03 Thread Shishir Kumar
and latency SLAs are, and you’ll know >> when to stop tuning. Otherwise you’ll discover that it’s a rabbit hole you >> can dive into and not come out of for weeks. >> >> >> >> >> >> *From: *Hossein Ghiyasi Mehr >> *Reply-To: *&

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-03 Thread John Belliveau
SLAs are, and you’ll know > when to stop tuning. Otherwise you’ll discover that it’s a rabbit hole you > can dive into and not come out of for weeks. > > > > > > *From: *Hossein Ghiyasi Mehr > *Reply-To: *"user@cassandra.apache.org" > *Date: *Monday, Decem

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-02 Thread Jeff Jirsa
ot;user@cassandra.apache.org" > *Date: *Monday, December 2, 2019 at 10:35 AM > *To: *"user@cassandra.apache.org" > *Subject: *Re: "Maximum memory usage reached (512.000MiB), cannot > allocate chunk of 1.000MiB" > > > > *Message from External Sender*

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-02 Thread Reid Pinchback
dive into and not come out of for weeks. From: Hossein Ghiyasi Mehr Reply-To: "user@cassandra.apache.org" Date: Monday, December 2, 2019 at 10:35 AM To: "user@cassandra.apache.org" Subject: Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000Mi

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-02 Thread Hossein Ghiyasi Mehr
It may be helpful: https://thelastpickle.com/blog/2018/08/08/compression_performance.html It's complex. Simple explanation, cassandra keeps sstables in memory based on chunk size and sstable parts. It manage loading new sstables to memory based on requests on different sstables correctly . You shou

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-02 Thread Rajsekhar Mallick
Hello Rahul, I would request Hossein to correct me if I am wrong. Below is how it works How will a application/database read something from the disk A request comes in for read> the application code internally would be invoking upon system calls-> these kernel level system calls will sche

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-02 Thread Rahul Reddy
Thanks Hossein, How does the chunks are moved out of memory (LRU?) if it want to make room for new requests to get chunks?if it has mechanism to clear chunks from cache what causes to cannot allocate chunk? Can you point me to any documention? On Sun, Dec 1, 2019, 12:03 PM Hossein Ghiyasi Mehr w

Re: "Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-01 Thread Hossein Ghiyasi Mehr
Chunks are part of sstables. When there is enough space in memory to cache them, read performance will increase if application requests it again. Your real answer is application dependent. For example write heavy applications are different than read heavy or read-write heavy. Real time application

"Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB"

2019-12-01 Thread Rahul Reddy
Hello, We are seeing memory usage reached 512 mb and cannot allocate 1MB. I see this because file_cache_size_mb by default set to 512MB. Datastax document recommends to increase the file_cache_size. We have 32G over all memory allocated 16G to Cassandra. What is the recommended value in my case

Commit sync issues and maximum memory usage reached messages seen in system log

2019-05-09 Thread Rajsekhar Mallick
Hello team, I am observing below warn and info message in system.log 1. Info log: maximum memory usage reached (1.000GiB), cannot allocate chunk of 1 MiB. I tried by increasing the file_cache_size_in_mb in Cassandra.yaml from 512 to 1024. But still this message shows up in logs. 2. Warn log

Re: Maximum memory usage reached

2019-03-07 Thread Kyrylo Lebediev
te: Wednesday, March 6, 2019 at 22:19 To: "user@cassandra.apache.org" Subject: Re: Maximum memory usage reached Also, that particular logger is for the internal chunk / page cache. If it can’t allocate from within that pool, it’ll just use a normal bytebuffer. It’s not really a

Re: Maximum memory usage reached

2019-03-06 Thread Jeff Jirsa
ghput = 40.492KiB/s, Row Throughput = >> ~106/s. 194 total partitions merged to 44. Partition merge counts were >> {1:18, 4:44, } >> >> INFO [IndexSummaryManager:1] 2019-03-06 11:00:22,007 >> IndexSummaryRedistribution.java:75 - Redistributing index summaries &g

Re: Maximum memory usage reached

2019-03-06 Thread Jonathan Haddad
artitions merged to 44. Partition merge counts were > {1:18, 4:44, } > > INFO [IndexSummaryManager:1] 2019-03-06 11:00:22,007 > IndexSummaryRedistribution.java:75 - Redistributing index summaries > > INFO [pool-1-thread-1] 2019-03-06 11:11:24,903 NoSpamLogger.java:91 - > Maximum

Maximum memory usage reached

2019-03-06 Thread Kyrylo Lebediev
-thread-1] 2019-03-06 11:11:24,903 NoSpamLogger.java:91 - Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB INFO [pool-1-thread-1] 2019-03-06 11:26:24,926 NoSpamLogger.java:91 - Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB INFO [pool-1

RE: Maximum memory usage

2019-02-10 Thread Kenneth Brotman
Can we the see “nodetool tablestats” for the biggest table as well. From: Kenneth Brotman [mailto:kenbrot...@yahoo.com.INVALID] Sent: Sunday, February 10, 2019 7:21 AM To: user@cassandra.apache.org Subject: RE: Maximum memory usage Okay, that’s at the moment it was calculated. Still need

RE: Maximum memory usage

2019-02-10 Thread Kenneth Brotman
Okay, that’s at the moment it was calculated. Still need to see histograms. From: Rahul Reddy [mailto:rahulreddy1...@gmail.com] Sent: Sunday, February 10, 2019 7:09 AM To: user@cassandra.apache.org Subject: Re: Maximum memory usage Thanks Kenneth, 110mb is the biggest partition in

Re: Maximum memory usage

2019-02-10 Thread Rahul Reddy
enneth Brotman > > > > *From:* Rahul Reddy [mailto:rahulreddy1...@gmail.com] > *Sent:* Sunday, February 10, 2019 6:43 AM > *To:* user@cassandra.apache.org > *Subject:* Re: Maximum memory usage > > > > ```Percentile SSTables Write Latency Read LatencyP

Re: Maximum memory usage

2019-02-10 Thread Rahul Reddy
> On Wednesday, February 6, 2019, 6:19:42 AM PST, Rahul Reddy < > rahulreddy1...@gmail.com> wrote: > > > Hello, > > I see maximum memory usage alerts in my system.log couple of times in a > day as INFO. So far I haven't seen any issue with db. Why those messag

RE: Maximum memory usage

2019-02-10 Thread Kenneth Brotman
Rahul, Those partitions are tiny. Could you give us the table histograms for the biggest tables. Thanks, Kenneth Brotman From: Rahul Reddy [mailto:rahulreddy1...@gmail.com] Sent: Sunday, February 10, 2019 6:43 AM To: user@cassandra.apache.org Subject: Re: Maximum memory usage

Re: Maximum memory usage

2019-02-10 Thread Rahul Reddy
ning > very often or logging a OOM. > > Dinesh > > > On Wednesday, February 6, 2019, 6:19:42 AM PST, Rahul Reddy < > rahulreddy1...@gmail.com> wrote: > > > Hello, > > I see maximum memory usage alerts in my system.log couple of times in a > day as INFO. So

Re: Maximum memory usage

2019-02-10 Thread Rahul Reddy
the “nodetool tablehistograms” > > > > Kenneth Brotman > > > > *From:* Rahul Reddy [mailto:rahulreddy1...@gmail.com] > *Sent:* Wednesday, February 06, 2019 6:19 AM > *To:* user@cassandra.apache.org > *Subject:* Maximum memory usage > > > > Hello, > >

Re: Maximum memory usage

2019-02-06 Thread dinesh.jo...@yahoo.com.INVALID
2-05 23:15:47,408 NoSpamLogger.java:91 - Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB Thanks in advance

RE: Maximum memory usage

2019-02-06 Thread Kenneth Brotman
Can you give us the “nodetool tablehistograms” Kenneth Brotman From: Rahul Reddy [mailto:rahulreddy1...@gmail.com] Sent: Wednesday, February 06, 2019 6:19 AM To: user@cassandra.apache.org Subject: Maximum memory usage Hello, I see maximum memory usage alerts in my system.log

Maximum memory usage

2019-02-06 Thread Rahul Reddy
Hello, I see maximum memory usage alerts in my system.log couple of times in a day as INFO. So far I haven't seen any issue with db. Why those messages are logged in system.log do we have any impact for reads/writes with those warnings? And what nerd to be looked INFO [RMI TCP Connection(1

Re: How to identify which table causing Maximum Memory usage limit

2018-06-11 Thread Nitan Kainth
e of all tables are even size. Max is always higher in all > tables. > > The question is, How do I identify, which table is throwing this "Maximum > memory usage reached (512.000MiB)" usage message? > > On Mon, Jun 11, 2018 at 5:59 AM, Martin Mačura wrote: > >> H

Re: How to identify which table causing Maximum Memory usage limit

2018-06-11 Thread Nitan Kainth
thanks Martin. 99 percentile of all tables are even size. Max is always higher in all tables. The question is, How do I identify, which table is throwing this "Maximum memory usage reached (512.000MiB)" usage message? On Mon, Jun 11, 2018 at 5:59 AM, Martin Mačura wrote: > Hi

Re: How to identify which table causing Maximum Memory usage limit

2018-06-11 Thread Martin Mačura
d be something else? > Do we need to worry about this message? > > INFO [CounterMutationStage-1] 2018-06-05 13:36:35,983 NoSpamLogger.java:91 > - Maximum memory usage reached (512.000MiB), cannot allocate chunk of > 1.000MiB > > -

How to identify which table causing Maximum Memory usage limit

2018-06-05 Thread learner dba
Hi, We see this message often, cluster has multiple keyspaces and column families;  How do I know which CF is causing this? Or it could be something else?Do we need to worry about this message? INFO  [CounterMutationStage-1] 2018-06-05 13:36:35,983 NoSpamLogger.java:91 - Maximum memory usage

Re: Maximum memory usage reached in cassandra!

2017-04-03 Thread Mark Rose
(Nokia - IN/Chennai) wrote: > Hi Cassandra users, > > > > I am getting “Maximum memory usage reached (536870912 bytes), cannot > allocate chunk of 1048576 bytes” . As a remedy I have changed the off heap > memory usage limit cap i.e file_cache_size_in_mb parameter in cassandra.ya

Maximum memory usage reached in cassandra!

2017-03-28 Thread Mokkapati, Bhargav (Nokia - IN/Chennai)
Hi Cassandra users, I am getting "Maximum memory usage reached (536870912 bytes), cannot allocate chunk of 1048576 bytes" . As a remedy I have changed the off heap memory usage limit cap i.e file_cache_size_in_mb parameter in cassandra.yaml from 512 to 1024. But now again the incre