[ https://issues.apache.org/jira/browse/FLINK-18489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Ori Popowski updated FLINK-18489: --------------------------------- Environment: {code:java} OS current user: yarn Current Hadoop/Kerberos user: hadoop JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.141-b15 Maximum heap size: 28960 MiBytes JAVA_HOME: /usr/java/jdk1.8.0_141/jre Hadoop version: 2.8.5-amzn-6 JVM Options: -Xmx30360049728 -Xms30360049728 -XX:MaxDirectMemorySize=4429185024 -XX:MaxMetaspaceSize=1073741824 -XX:+UseG1GC -XX:+UnlockDiagnosticVMOptions -XX:+G1SummarizeConcMark -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -XX:+UnlockCommercialFeatures -XX:+FlightRecorder -XX:+DebugNonSafepoints -XX:FlightRecorderOptions=defaultrecording=true,settings=/home/hadoop/heap.jfc,dumponexit=true,dumponexitpath=/var/lib/hadoop-yarn/recording.jfr,loglevel=info -Dlog.file=/var/log/hadoop-yarn/containers/application_1593935560662_0002/container_1593935560662_0002_01_000002/taskmanager.log -Dlog4j.configuration=file:./log4j.properties Program Arguments: -Dtaskmanager.memory.framework.off-heap.size=134217728b -Dtaskmanager.memory.network.max=1073741824b -Dtaskmanager.memory.network.min=1073741824b -Dtaskmanager.memory.framework.heap.size=134217728b -Dtaskmanager.memory.managed.size=23192823744b -Dtaskmanager.cpu.cores=7.0 -Dtaskmanager.memory.task.heap.size=30225832000b -Dtaskmanager.memory.task.off-heap.size=3221225472b --configDir. -Djobmanager.rpc.address=ip-10-180-30-250.us-west-2.compute.internal-Dweb.port=0 -Dweb.tmpdir=/tmp/flink-web-64f613cf-bf04-4a09-8c14-75c31b619574 -Djobmanager.rpc.port=33739 -Drest.address=ip-10-180-30-250.us-west-2.compute.internal {code} was: OS current user: yarn Current Hadoop/Kerberos user: hadoop JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.141-b15 Maximum heap size: 18432 MiBytes JAVA_HOME: /usr/java/jdk1.8.0_141/jre Hadoop version: 2.8.5-amzn-6 JVM Options: -Xms18432m -Xmx18432m -XX:+UseG1GC -Dlog.file=/var/log/hadoop-yarn/containers/application_1593935560662_0002/container_1593935560662_0002_01_000001/jobmanager.log -Dlog4j.configuration=file:log4j.properties env.java.opts.jobmanager, -XX:+UseG1GC env.yarn.conf.dir, /etc/hadoop/conf env.java.home, /usr/java/jdk1.8.0_141/jre pipeline.default-kryo-serializers, class:com.fasterxml.jackson.databind.JsonNode,serializer:walkme.flink.JsonNodeKryoSerializer pipeline.time-characteristic, EventTime restart-strategy.failure-rate.max-failures-per-interval, 3 high-availability.cluster-id, application_1593935560662_0002 taskmanager.cpu.cores, 7 taskmanager.memory.jvm-metaspace.size, 1 gb taskmanager.memory.task.off-heap.size, 3 gb state.savepoints.dir, s3://... execution.savepoint.ignore-unclaimed-state, false parallelism.default, 189 env.hadoop.conf.dir, /etc/hadoop/conf taskmanager.numberOfTaskSlots, 1 akka.ask.timeout, 1 min restart-strategy.failure-rate.failure-rate-interval, 5 min metrics.reporter.prom.class, org.apache.flink.metrics.prometheus.PrometheusReporter jobmanager.heap.size, 24g execution.checkpointing.mode, AT_LEAST_ONCE taskmanager.memory.process.size, 56 gb restart-strategy.failure-rate.delay, 10 sec flink.partition-discovery.interval-millis, 60000 state.backend.incremental, true execution.target, yarn-per-job execution.checkpointing.interval, 10 min execution.checkpointing.timeout, 10 min execution.attached, true internal.cluster.execution-mode, NORMAL execution.shutdown-on-attached-exit, false pipeline.jars, [file:/home/hadoop/flink-session-playback-assembly-0.1.jar] execution.checkpointing.min-pause, 5 min restart-strategy, none state.backend, rocksdb env.java.opts.taskmanager, -XX:+UseG1GC -XX:+UnlockDiagnosticVMOptions -XX:+G1SummarizeConcMark -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -XX:+UnlockCommercialFeatures -XX:+FlightRecorder -XX:+DebugNonSafepoints -XX:FlightRecorderOptions=defaultrecording=true,settings=/home/hadoop/heap.jfc,dumponexit=true,dumponexitpath=/var/lib/hadoop-yarn/recording.jfr,loglevel=info pipeline.registered-kryo-types, java.lang.Number;java.lang.Object $internal.yarn.log-config-file, /etc/flink/conf/log4j.properties state.checkpoints.dir, s3://... > java.lang.ArrayIndexOutOfBoundsException > ---------------------------------------- > > Key: FLINK-18489 > URL: https://issues.apache.org/jira/browse/FLINK-18489 > Project: Flink > Issue Type: Bug > Affects Versions: 1.10.0 > Environment: {code:java} > OS current user: yarn > Current Hadoop/Kerberos user: hadoop > JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.141-b15 > Maximum heap size: 28960 MiBytes > JAVA_HOME: /usr/java/jdk1.8.0_141/jre > Hadoop version: 2.8.5-amzn-6 > JVM Options: > -Xmx30360049728 > -Xms30360049728 > -XX:MaxDirectMemorySize=4429185024 > -XX:MaxMetaspaceSize=1073741824 > -XX:+UseG1GC > -XX:+UnlockDiagnosticVMOptions > -XX:+G1SummarizeConcMark > -verbose:gc > -XX:+PrintGCDetails > -XX:+PrintGCDateStamps > -XX:+UnlockCommercialFeatures > -XX:+FlightRecorder > -XX:+DebugNonSafepoints > > -XX:FlightRecorderOptions=defaultrecording=true,settings=/home/hadoop/heap.jfc,dumponexit=true,dumponexitpath=/var/lib/hadoop-yarn/recording.jfr,loglevel=info > > -Dlog.file=/var/log/hadoop-yarn/containers/application_1593935560662_0002/container_1593935560662_0002_01_000002/taskmanager.log > -Dlog4j.configuration=file:./log4j.properties > Program Arguments: > -Dtaskmanager.memory.framework.off-heap.size=134217728b > -Dtaskmanager.memory.network.max=1073741824b > -Dtaskmanager.memory.network.min=1073741824b > -Dtaskmanager.memory.framework.heap.size=134217728b > -Dtaskmanager.memory.managed.size=23192823744b > -Dtaskmanager.cpu.cores=7.0 > -Dtaskmanager.memory.task.heap.size=30225832000b > -Dtaskmanager.memory.task.off-heap.size=3221225472b > --configDir. > > -Djobmanager.rpc.address=ip-10-180-30-250.us-west-2.compute.internal-Dweb.port=0 > -Dweb.tmpdir=/tmp/flink-web-64f613cf-bf04-4a09-8c14-75c31b619574 > -Djobmanager.rpc.port=33739 > -Drest.address=ip-10-180-30-250.us-west-2.compute.internal > {code} > Reporter: Ori Popowski > Priority: Major > > Getting {{java.lang.ArrayIndexOutOfBoundsException}} with the following > stacktrace: > {code:java} > 2020-07-05 18:25:04org.apache.flink.runtime.JobException: Recovery is > suppressed by NoRestartBackoffTimeStrategy at > org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:110) > at > org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:76) > at > org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:192) > at > org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:186) > at > org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:180) > at > org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:484) > at > org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:380) > at sun.reflect.GeneratedMethodAccessor85.invoke(Unknown Source) at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:279) > at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:194) > at > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74) > at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152) > at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) at > akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) at > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) at > akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170) at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) at > akka.actor.Actor$class.aroundReceive(Actor.scala:517) at > akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) at > akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) at > akka.actor.ActorCell.invoke(ActorCell.scala:561) at > akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) at > akka.dispatch.Mailbox.run(Mailbox.scala:225) at > akka.dispatch.Mailbox.exec(Mailbox.scala:235) at > akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at > akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)Caused > by: java.lang.RuntimeException at > org.apache.flink.streaming.runtime.io.RecordWriterOutput.pushToRecordWriter(RecordWriterOutput.java:110) > at > org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:89) > at > org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:45) > at > org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:730) > at > org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:708) > at > org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:641) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:616) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:596) > at > org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:730) > at > org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:708) > at > org.apache.flink.streaming.api.operators.TimestampedCollector.collect(TimestampedCollector.java:53) > at > walkme.flink.SessionWindowFunction$$anonfun$process$1.apply(SessionWindowFunction.scala:38) > at > walkme.flink.SessionWindowFunction$$anonfun$process$1.apply(SessionWindowFunction.scala:29) > at scala.Option.foreach(Option.scala:257) at > walkme.flink.SessionWindowFunction.process(SessionWindowFunction.scala:29) > at walkme.flink.SessionWindowFunction.process(SessionWindowFunction.scala:16) > at > org.apache.flink.streaming.api.scala.function.util.ScalaProcessWindowFunctionWrapper.process(ScalaProcessWindowFunctionWrapper.scala:63) > at > org.apache.flink.streaming.runtime.operators.windowing.functions.InternalIterableProcessWindowFunction.process(InternalIterableProcessWindowFunction.java:50) > at > org.apache.flink.streaming.runtime.operators.windowing.functions.InternalIterableProcessWindowFunction.process(InternalIterableProcessWindowFunction.java:32) > at > org.apache.flink.streaming.runtime.operators.windowing.WindowOperator.emitWindowContents(WindowOperator.java:549) > at > org.apache.flink.streaming.runtime.operators.windowing.WindowOperator.onEventTime(WindowOperator.java:457) > at > org.apache.flink.streaming.api.operators.InternalTimerServiceImpl.advanceWatermark(InternalTimerServiceImpl.java:276) > at > org.apache.flink.streaming.api.operators.InternalTimeServiceManager.advanceWatermark(InternalTimeServiceManager.java:128) > at > org.apache.flink.streaming.api.operators.AbstractStreamOperator.processWatermark(AbstractStreamOperator.java:787) > at > org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitWatermark(OneInputStreamTask.java:181) > at > org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.findAndOutputNewMinWatermarkAcrossAlignedChannels(StatusWatermarkValve.java:179) > at > org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.inputWatermark(StatusWatermarkValve.java:101) > at > org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.processElement(StreamTaskNetworkInput.java:153) > at > org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:128) > at > org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:69) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:311) > at > org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:187) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:487) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:470) > at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707) at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:532) at > java.lang.Thread.run(Thread.java:748)Caused by: > java.lang.ArrayIndexOutOfBoundsException at > org.apache.flink.core.memory.DataOutputSerializer.write(DataOutputSerializer.java:154) > at > org.apache.flink.core.memory.DataOutputSerializer.write(DataOutputSerializer.java:143) > at > org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.serialize(BytePrimitiveArraySerializer.java:78) > at > org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.serialize(BytePrimitiveArraySerializer.java:33) > at > org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:109) > at > org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:32) > at > org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:109) > at > org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:32) > at > org.apache.flink.streaming.runtime.streamrecord.StreamElementSerializer.serialize(StreamElementSerializer.java:175) > at > org.apache.flink.streaming.runtime.streamrecord.StreamElementSerializer.serialize(StreamElementSerializer.java:46) > at > org.apache.flink.runtime.plugable.SerializationDelegate.write(SerializationDelegate.java:54) > at > org.apache.flink.runtime.io.network.api.serialization.SpanningRecordSerializer.serializeRecord(SpanningRecordSerializer.java:71) > at > org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:113) > at > org.apache.flink.runtime.io.network.api.writer.ChannelSelectorRecordWriter.emit(ChannelSelectorRecordWriter.java:60) > at > org.apache.flink.streaming.runtime.io.RecordWriterOutput.pushToRecordWriter(RecordWriterOutput.java:107) > ... 37 more > {code} > -- This message was sent by Atlassian Jira (v8.3.4#803005)