Hi Jason,

This error is unexpected. I don't think its caused by insufficient memory.
I'm including Stefan into the conversation, he's the RocksDB expert :)

On Thu, May 25, 2017 at 4:15 PM, Jason Brelloch <jb.bc....@gmail.com> wrote:

> Hey guys,
>
> We are running into a JVM crash on checkpointing when our rocksDB state
> reaches a certain size on a taskmanager (about 2GB).  The issue happens
> with both a hadoop backend and just writing to a local file.
>
> We are running on Flink 1.2.1.
>
> #
> # A fatal error has been detected by the Java Runtime Environment:
> #
> #  SIGSEGV (0xb) at pc=0x00007febf4261b42, pid=1, tid=0x00007fead135f700
> #
> # JRE version: Java(TM) SE Runtime Environment (8.0_131-b11) (build
> 1.8.0_131-b11)
> # Java VM: Java HotSpot(TM) 64-Bit Server VM (25.131-b11 mixed mode
> linux-amd64 compressed oops)
> # Problematic frame:
> # V  [libjvm.so+0x6d1b42]  jni_SetByteArrayRegion+0xc2
> #
> # Core dump written. Default location: //core or core.1
> #
> # An error report file with more information is saved as:
> # /tmp/hs_err_pid1.log
> #
> # If you would like to submit a bug report, please visit:
> #   http://bugreport.java.com/bugreport/crash.jsp
> #
>
> Is this an issue with not enough memory?  Or maybe not enough allocated to
> rocksDB?
>
> I have attached the taskmanager logs, and the core dump.  The jobmanager
> logs just say taskmanger lost/killed.
>
> Thanks!
>
> --
> *Jason Brelloch* | Product Developer
> 3405 Piedmont Rd. NE, Suite 325, Atlanta, GA 30305
> <http://www.bettercloud.com/>
> Subscribe to the BetterCloud Monitor
> <https://www.bettercloud.com/monitor?utm_source=bettercloud_email&utm_medium=email_signature&utm_campaign=monitor_launch>
>  -
> Get IT delivered to your inbox
>

Reply via email to