[ 
https://issues.apache.org/jira/browse/FLINK-18489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ori Popowski updated FLINK-18489:
---------------------------------
    Environment: 
{code:java}
OS current user: yarn
Current Hadoop/Kerberos user: hadoop
JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.141-b15
Maximum heap size: 28960 MiBytes
JAVA_HOME: /usr/java/jdk1.8.0_141/jre
Hadoop version: 2.8.5-amzn-6
JVM Options:
   -Xmx30360049728
   -Xms30360049728
   -XX:MaxDirectMemorySize=4429185024
   -XX:MaxMetaspaceSize=1073741824
   -XX:+UseG1GC
   -XX:+UnlockDiagnosticVMOptions
   -XX:+G1SummarizeConcMark
   -verbose:gc
   -XX:+PrintGCDetails
   -XX:+PrintGCDateStamps
   -XX:+UnlockCommercialFeatures
   -XX:+FlightRecorder
   -XX:+DebugNonSafepoints
   
-XX:FlightRecorderOptions=defaultrecording=true,settings=/home/hadoop/heap.jfc,dumponexit=true,dumponexitpath=/var/lib/hadoop-yarn/recording.jfr,loglevel=info
   
-Dlog.file=/var/log/hadoop-yarn/containers/application_1593935560662_0002/container_1593935560662_0002_01_000002/taskmanager.log
   -Dlog4j.configuration=file:./log4j.properties
Program Arguments:
   -Dtaskmanager.memory.framework.off-heap.size=134217728b
   -Dtaskmanager.memory.network.max=1073741824b
   -Dtaskmanager.memory.network.min=1073741824b
   -Dtaskmanager.memory.framework.heap.size=134217728b
   -Dtaskmanager.memory.managed.size=23192823744b
   -Dtaskmanager.cpu.cores=7.0
   -Dtaskmanager.memory.task.heap.size=30225832000b
   -Dtaskmanager.memory.task.off-heap.size=3221225472b
   --configDir.
   
-Djobmanager.rpc.address=ip-10-180-30-250.us-west-2.compute.internal-Dweb.port=0
   -Dweb.tmpdir=/tmp/flink-web-64f613cf-bf04-4a09-8c14-75c31b619574
   -Djobmanager.rpc.port=33739
   -Drest.address=ip-10-180-30-250.us-west-2.compute.internal

{code}

  was:
OS current user: yarn
Current Hadoop/Kerberos user: hadoop
JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.141-b15
Maximum heap size: 18432 MiBytes
JAVA_HOME: /usr/java/jdk1.8.0_141/jre
Hadoop version: 2.8.5-amzn-6
JVM Options:
 -Xms18432m
 -Xmx18432m
 -XX:+UseG1GC
 
-Dlog.file=/var/log/hadoop-yarn/containers/application_1593935560662_0002/container_1593935560662_0002_01_000001/jobmanager.log
 -Dlog4j.configuration=file:log4j.properties
env.java.opts.jobmanager, -XX:+UseG1GC
env.yarn.conf.dir, /etc/hadoop/conf
env.java.home, /usr/java/jdk1.8.0_141/jre
pipeline.default-kryo-serializers, 
class:com.fasterxml.jackson.databind.JsonNode,serializer:walkme.flink.JsonNodeKryoSerializer
pipeline.time-characteristic, EventTime
restart-strategy.failure-rate.max-failures-per-interval, 3
high-availability.cluster-id, application_1593935560662_0002
taskmanager.cpu.cores, 7
taskmanager.memory.jvm-metaspace.size, 1 gb
taskmanager.memory.task.off-heap.size, 3 gb
state.savepoints.dir, s3://...
execution.savepoint.ignore-unclaimed-state, false
parallelism.default, 189
env.hadoop.conf.dir, /etc/hadoop/conf
taskmanager.numberOfTaskSlots, 1
akka.ask.timeout, 1 min
restart-strategy.failure-rate.failure-rate-interval, 5 min
metrics.reporter.prom.class, 
org.apache.flink.metrics.prometheus.PrometheusReporter
jobmanager.heap.size, 24g
execution.checkpointing.mode, AT_LEAST_ONCE
taskmanager.memory.process.size, 56 gb
restart-strategy.failure-rate.delay, 10 sec
flink.partition-discovery.interval-millis, 60000
state.backend.incremental, true
execution.target, yarn-per-job
execution.checkpointing.interval, 10 min
execution.checkpointing.timeout, 10 min
execution.attached, true
internal.cluster.execution-mode, NORMAL
execution.shutdown-on-attached-exit, false
pipeline.jars, [file:/home/hadoop/flink-session-playback-assembly-0.1.jar]
execution.checkpointing.min-pause, 5 min
restart-strategy, none
state.backend, rocksdb
env.java.opts.taskmanager, -XX:+UseG1GC -XX:+UnlockDiagnosticVMOptions 
-XX:+G1SummarizeConcMark -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps 
-XX:+UnlockCommercialFeatures -XX:+FlightRecorder -XX:+DebugNonSafepoints 
-XX:FlightRecorderOptions=defaultrecording=true,settings=/home/hadoop/heap.jfc,dumponexit=true,dumponexitpath=/var/lib/hadoop-yarn/recording.jfr,loglevel=info
pipeline.registered-kryo-types, java.lang.Number;java.lang.Object
$internal.yarn.log-config-file, /etc/flink/conf/log4j.properties
state.checkpoints.dir, s3://...


> java.lang.ArrayIndexOutOfBoundsException
> ----------------------------------------
>
>                 Key: FLINK-18489
>                 URL: https://issues.apache.org/jira/browse/FLINK-18489
>             Project: Flink
>          Issue Type: Bug
>    Affects Versions: 1.10.0
>         Environment: {code:java}
> OS current user: yarn
> Current Hadoop/Kerberos user: hadoop
> JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.141-b15
> Maximum heap size: 28960 MiBytes
> JAVA_HOME: /usr/java/jdk1.8.0_141/jre
> Hadoop version: 2.8.5-amzn-6
> JVM Options:
>    -Xmx30360049728
>    -Xms30360049728
>    -XX:MaxDirectMemorySize=4429185024
>    -XX:MaxMetaspaceSize=1073741824
>    -XX:+UseG1GC
>    -XX:+UnlockDiagnosticVMOptions
>    -XX:+G1SummarizeConcMark
>    -verbose:gc
>    -XX:+PrintGCDetails
>    -XX:+PrintGCDateStamps
>    -XX:+UnlockCommercialFeatures
>    -XX:+FlightRecorder
>    -XX:+DebugNonSafepoints
>    
> -XX:FlightRecorderOptions=defaultrecording=true,settings=/home/hadoop/heap.jfc,dumponexit=true,dumponexitpath=/var/lib/hadoop-yarn/recording.jfr,loglevel=info
>    
> -Dlog.file=/var/log/hadoop-yarn/containers/application_1593935560662_0002/container_1593935560662_0002_01_000002/taskmanager.log
>    -Dlog4j.configuration=file:./log4j.properties
> Program Arguments:
>    -Dtaskmanager.memory.framework.off-heap.size=134217728b
>    -Dtaskmanager.memory.network.max=1073741824b
>    -Dtaskmanager.memory.network.min=1073741824b
>    -Dtaskmanager.memory.framework.heap.size=134217728b
>    -Dtaskmanager.memory.managed.size=23192823744b
>    -Dtaskmanager.cpu.cores=7.0
>    -Dtaskmanager.memory.task.heap.size=30225832000b
>    -Dtaskmanager.memory.task.off-heap.size=3221225472b
>    --configDir.
>    
> -Djobmanager.rpc.address=ip-10-180-30-250.us-west-2.compute.internal-Dweb.port=0
>    -Dweb.tmpdir=/tmp/flink-web-64f613cf-bf04-4a09-8c14-75c31b619574
>    -Djobmanager.rpc.port=33739
>    -Drest.address=ip-10-180-30-250.us-west-2.compute.internal
> {code}
>            Reporter: Ori Popowski
>            Priority: Major
>
> Getting {{java.lang.ArrayIndexOutOfBoundsException}} with the following 
> stacktrace:
> {code:java}
> 2020-07-05 18:25:04org.apache.flink.runtime.JobException: Recovery is 
> suppressed by NoRestartBackoffTimeStrategy    at 
> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:110)
>     at 
> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:76)
>     at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:192)
>     at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:186)
>     at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:180)
>     at 
> org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:484)
>     at 
> org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:380)
>     at sun.reflect.GeneratedMethodAccessor85.invoke(Unknown Source)    at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:498)    at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:279)
>     at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:194)
>     at 
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
>     at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)    at 
> akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)    at 
> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)    at 
> akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)    at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)    at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)    at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)    at 
> akka.actor.Actor$class.aroundReceive(Actor.scala:517)    at 
> akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)    at 
> akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)    at 
> akka.actor.ActorCell.invoke(ActorCell.scala:561)    at 
> akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)    at 
> akka.dispatch.Mailbox.run(Mailbox.scala:225)    at 
> akka.dispatch.Mailbox.exec(Mailbox.scala:235)    at 
> akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)    at 
> akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) 
>    at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)   
>  at 
> akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)Caused
>  by: java.lang.RuntimeException    at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.pushToRecordWriter(RecordWriterOutput.java:110)
>     at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:89)
>     at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:45)
>     at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:730)
>     at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:708)
>     at 
> org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41)
>     at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:641)
>     at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:616)
>     at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:596)
>     at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:730)
>     at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:708)
>     at 
> org.apache.flink.streaming.api.operators.TimestampedCollector.collect(TimestampedCollector.java:53)
>     at 
> walkme.flink.SessionWindowFunction$$anonfun$process$1.apply(SessionWindowFunction.scala:38)
>     at 
> walkme.flink.SessionWindowFunction$$anonfun$process$1.apply(SessionWindowFunction.scala:29)
>     at scala.Option.foreach(Option.scala:257)    at 
> walkme.flink.SessionWindowFunction.process(SessionWindowFunction.scala:29)    
> at walkme.flink.SessionWindowFunction.process(SessionWindowFunction.scala:16) 
>    at 
> org.apache.flink.streaming.api.scala.function.util.ScalaProcessWindowFunctionWrapper.process(ScalaProcessWindowFunctionWrapper.scala:63)
>     at 
> org.apache.flink.streaming.runtime.operators.windowing.functions.InternalIterableProcessWindowFunction.process(InternalIterableProcessWindowFunction.java:50)
>     at 
> org.apache.flink.streaming.runtime.operators.windowing.functions.InternalIterableProcessWindowFunction.process(InternalIterableProcessWindowFunction.java:32)
>     at 
> org.apache.flink.streaming.runtime.operators.windowing.WindowOperator.emitWindowContents(WindowOperator.java:549)
>     at 
> org.apache.flink.streaming.runtime.operators.windowing.WindowOperator.onEventTime(WindowOperator.java:457)
>     at 
> org.apache.flink.streaming.api.operators.InternalTimerServiceImpl.advanceWatermark(InternalTimerServiceImpl.java:276)
>     at 
> org.apache.flink.streaming.api.operators.InternalTimeServiceManager.advanceWatermark(InternalTimeServiceManager.java:128)
>     at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator.processWatermark(AbstractStreamOperator.java:787)
>     at 
> org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitWatermark(OneInputStreamTask.java:181)
>     at 
> org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.findAndOutputNewMinWatermarkAcrossAlignedChannels(StatusWatermarkValve.java:179)
>     at 
> org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.inputWatermark(StatusWatermarkValve.java:101)
>     at 
> org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.processElement(StreamTaskNetworkInput.java:153)
>     at 
> org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:128)
>     at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:69)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:311)
>     at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:187)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:487)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:470)
>     at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)    at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)    at 
> java.lang.Thread.run(Thread.java:748)Caused by: 
> java.lang.ArrayIndexOutOfBoundsException    at 
> org.apache.flink.core.memory.DataOutputSerializer.write(DataOutputSerializer.java:154)
>     at 
> org.apache.flink.core.memory.DataOutputSerializer.write(DataOutputSerializer.java:143)
>     at 
> org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.serialize(BytePrimitiveArraySerializer.java:78)
>     at 
> org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.serialize(BytePrimitiveArraySerializer.java:33)
>     at 
> org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:109)
>     at 
> org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:32)
>     at 
> org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:109)
>     at 
> org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:32)
>     at 
> org.apache.flink.streaming.runtime.streamrecord.StreamElementSerializer.serialize(StreamElementSerializer.java:175)
>     at 
> org.apache.flink.streaming.runtime.streamrecord.StreamElementSerializer.serialize(StreamElementSerializer.java:46)
>     at 
> org.apache.flink.runtime.plugable.SerializationDelegate.write(SerializationDelegate.java:54)
>     at 
> org.apache.flink.runtime.io.network.api.serialization.SpanningRecordSerializer.serializeRecord(SpanningRecordSerializer.java:71)
>     at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:113)
>     at 
> org.apache.flink.runtime.io.network.api.writer.ChannelSelectorRecordWriter.emit(ChannelSelectorRecordWriter.java:60)
>     at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.pushToRecordWriter(RecordWriterOutput.java:107)
>     ... 37 more
> {code}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to