Hello Stephen,
I've started tests with few angles, with larger and smaller memory setups, looked at Ignite unit tests and began looking for page eviction triggers. Once I found interesting code areas I realized that my test setup was not having proper page eviction set up. This lead me to further debugging which lead me piece of old code which was overriding persistence configuration, *effectively* reverting page eviction to default (no eviction). After clean up I successfully tested eviction with small caches (512MB max + direct buffer at 756MB). These worked stable under constant load, which proved the point that it is possible to have stable operation on fairly limited resources.

Guilt goes to spring configuration pre/post-processing which went under radar till today. It was really specific to test application itself, not Ignite. I thank you for feedback, pointers and assistance!

Kind regards,
Łukasz

On 30.01.2023 12:51, Stephen Darlington wrote:
I’m not saying you should test with 50Gb, but I am saying that 512Mb is too small. Maybe try a few gigs.

When you define eviction at a cache level, that is for the on-heap cache. Most people don’t need an on-heap cache.

You you define eviction at a data region level, that works on the off-heap data. As you observe, it works at a /page/ rather than a record level. (This is why a tiny region can be problematic. There are far few pages that can be chosen to evict.)

On 30 Jan 2023, at 11:37, Łukasz Dywicki <l...@code-house.org> wrote:

Dear Jeremy and Stephen,
Thank you for answers. I this issues I face is not bound exclusively to assigned memory but behavior of eviction policies. After all, without proper eviction or expiry policy, each amount of memory will eventually be exceeded. Small size of memory makes it simply easier to test. To be honest I do not see a point in testing eviction with 50GB of memory, if it cant be configured for 512MB. Expiry policy proved to work, but it will not preventing node from crash, as there always might be high write ratio causing memory to fill.

Both LRU and FIFO eviction policies in default configuration do not take into account overall memory allocation, they work only with on heap caches. Unless memory size is explicitly specified they will rely on over-provisioned memory, or assume that memory assignment is right. This is possible only if cache entry have a predictable. Another point is that above eviction policies are dedicated to on-heap caches and are told to have no affect to off-heap region, why is so? By looking at sources I see CacheOffheapEvictionManager and GridCacheEvictionManager, but I can't grasp what enables eviction of entries from off heap memory managed by Ignite. Given that my use case is continuous operation with variety of entries having various sizes I can't assume on-heap cache with LRU nor FIFO, especially that they do not impact off heap memory. I also can't risk filling in memory managed by Ignite due to lack of clear eviction policy during peak load. As far I can understand from DataRegionConfiguration, its eviction mechanism works for whole pages, but I haven't seen this happening. Could it be due to fact that single cache entry in my test exceeds maximum page size (16MB)?

Best regards,
Łukasz

On 30.01.2023 10:32, Stephen Darlington wrote:
Ignite is designed to work as a cluster with a lot of memory. A single node with 512Mb of memory just isn’t what most people are testing with. I’ve seen similar issues when people use tiny nodes. I don’t see anything immediately wrong with your configuration, but if it works with more memory, that’s your problem.
On 28 Jan 2023, at 00:35, Łukasz Dywicki <l...@code-house.org> wrote:

Hello again,
I've spent another day on this issue looking at configuration and with cluster larger than configured backup count ignite did the work. After that I reverted to testing of single node with minimal configuration and come to the point that the only way to keep Ignite survive load was setting ExpiryPolicy to very low value (10s).

To me it seems that Ignite behavior is to preserve all entries in memory at any cost, even if there is a risk of running into OOM. Is that true? I would like to change this behavior and make sure that entries do not expire because of time as long as there is memory available to store them. They should be evicted by appropriate EvictionPolicy only if memory fills up. To me it looks like DataStoreSettings do not make any impact in this regard. At least setting page eviction to LRU do not change it.

Please let me know if I am doing something wrong as I can not prove Ignite to be working stable. Even with such basic objectives I outlined in earlier mail.

Kind regards,
Łukasz

On 27.01.2023 00:58, Łukasz Dywicki wrote:
Dear all,
I come across use of Apache Ignite to cache results of expensive computation operation.
Objectives are basic:
- Keep most of "hot" data in memory
- Offload cold part to cache store
- Keep memory utilization under control (evict entries as needed)
While it sounds basic, it doesn't seem to fit Ignite defaults.
What I am testing now is behavior with large objects which can grow up to 10 mb (serialized) or 25 mb (json representation). Usually objects will stay far below that threshold, but we can't make assumption on that. I began testing various configurations of Ignite in order to facilitate offloading of memory contents to database. So far I am stuck for two days at Ignite/application itself running out of memory after processing several of such large objects. While I know that storing 10 mb blob in database is not the best idea, I have to test that behavior too. By observing database contents I see that number of entries there grows, but cache do not seem to be evicted. When I try to switch eviction, it does require onheap to be switched on, and it still fails with LRU eviction policy. So far I ended up with a named cache and default region configured as below:
```
IgniteConfiguration igniteConfiguration = new IgniteConfiguration();
igniteConfiguration.setDataStorageConfiguration(new DataStorageConfiguration()
    .setDefaultDataRegionConfiguration(new DataRegionConfiguration()
        .setPersistenceEnabled(false)
        .setInitialSize(256L * 1024 * 1024)
        .setMaxSize(512L * 1024 * 1024)
        .setPageEvictionMode(DataPageEvictionMode.RANDOM_LRU)
        .setSwapPath(null)
        .setEvictionThreshold(0.75)
    )
    .setPageSize(DataStorageConfiguration.MAX_PAGE_SIZE)
);
CacheConfiguration<SessionKey, ExpensiveObject> expensiveCache = new CacheConfiguration<>()
    .setName(CACHE_NAME)
    .setBackups(2)
    .setAtomicityMode(CacheAtomicityMode.ATOMIC)
    .setCacheStoreFactory(cacheJdbcBlobStoreFactory)
    .setWriteThrough(true)
    .setOnheapCacheEnabled(true)
    .setEvictionPolicyFactory(new LruEvictionPolicyFactory<>(1024))
    .setReadThrough(true);
igniteConfiguration.setCacheConfiguration(
    expensiveCache
);
```
What I observe is following - the cache keeps writing data into database, but it does not remove old entries fast enough to prevent crash.
JVM parameters I use are fairly basic:
-Xms1g -Xmx1g -XX:+AlwaysPreTouch -XX:+UseG1GC -XX:+ScavengeBeforeFullGC -XX:+DisableExplicitGC The store mechanism is jdbc blob store. Exceptions I get happen to occur in Ignite itself, processing (application code writing cache) or communication thread used to feed cache. I collected one case here:
https://gist.github.com/splatch/b5ec9134cd9df19bc62f007dd17a19a1
The error message in linked gist advice to enable persistence (which I did via cache store!), increase memory limit (which I don't want to do), or enable eviction/expiry policy (which somehow miss behave). To me it looks like self defense mechanisms Ignite has are being tricked   leading whole application to crash. Can you please advise me which settings to tune and how in order to get Ignite more stable under such load?
Kind regards,
Łukasz

Reply via email to