[ 
https://issues.apache.org/jira/browse/FLINK-29135?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin resolved FLINK-29135.
-------------------------------------
    Resolution: Duplicate

Close it in favor of https://issues.apache.org/jira/browse/FLINK-27995

> After updating to calcite 1.30.0 and janino 3.1.6 some tests started to fail 
> with compilation issues in generated code
> ----------------------------------------------------------------------------------------------------------------------
>
>                 Key: FLINK-29135
>                 URL: https://issues.apache.org/jira/browse/FLINK-29135
>             Project: Flink
>          Issue Type: Sub-task
>          Components: Table SQL / API, Table SQL / Planner
>            Reporter: Sergey Nuyanzin
>            Priority: Major
>
>  e.g. {{org.apache.flink.table.planner.functions.CastFunctionITCase}} for 
> {{cast(f0, ARRAY<BIGINT>) }}
> {noformat}
> /* 1 */
> /* 2 */      public class StreamExecCalc$4813 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4805 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4806 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4805, 3, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4813(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4806.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4806.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4806.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (true) {
> /* 33 */            writer$4806.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4806.writeInt(1, ((int) -1));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          
> /* 39 */          
> /* 40 */          if (false) {
> /* 41 */            writer$4806.setNullInt(2);
> /* 42 */          } else {
> /* 43 */            writer$4806.writeInt(2, ((int) 2));
> /* 44 */          }
> /* 45 */                    
> /* 46 */          writer$4806.complete();
> /* 47 */                   
> /* 48 */          this.setup(task, config, output);
> /* 49 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 50 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 51 */              .setProcessingTimeService(processingTimeService);
> /* 52 */          }
> /* 53 */        }
> /* 54 */
> /* 55 */        @Override
> /* 56 */        public void open() throws Exception {
> /* 57 */          super.open();
> /* 58 */          
> /* 59 */        }
> /* 60 */
> /* 61 */        @Override
> /* 62 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 63 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 64 */          
> /* 65 */          boolean isNull$4808;
> /* 66 */          org.apache.flink.table.data.ArrayData result$4809;
> /* 67 */          long result$4812;
> /* 68 */          
> /* 69 */          
> /* 70 */          
> /* 71 */          
> /* 72 */          out.setRowKind(in1.getRowKind());
> /* 73 */          
> /* 74 */          
> /* 75 */          
> /* 76 */          
> /* 77 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 78 */           isNull$4808 = false;
> /* 79 */          if (!isNull$4808) {
> /* 80 */          java.lang.Long[] objArray$4810 = new 
> java.lang.Long[array$4805.size()];
> /* 81 */          for (int i$4811 = 0; i$4811 < array$4805.size(); i$4811++) {
> /* 82 */          if (!array$4805.isNullAt(i$4811)) {
> /* 83 */          result$4812 = ((long)(array$4805.getInt(i$4811)));
> /* 84 */          objArray$4810[i$4811] = result$4812;
> /* 85 */          }
> /* 86 */          }
> /* 87 */          result$4809 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4810);
> /* 88 */          isNull$4808 = result$4809 == null;
> /* 89 */          } else {
> /* 90 */          result$4809 = null;
> /* 91 */          }
> /* 92 */          
> /* 93 */           // --- End cast section
> /* 94 */                         
> /* 95 */          if (isNull$4808) {
> /* 96 */            out.setNullAt(0);
> /* 97 */          } else {
> /* 98 */            out.setNonPrimitiveValue(0, result$4809);
> /* 99 */          }
> /* 100 */                    
> /* 101 */                  
> /* 102 */          output.collect(outElement.replace(out));
> /* 103 */          
> /* 104 */          
> /* 105 */        }
> /* 106 */
> /* 107 */        
> /* 108 */
> /* 109 */        @Override
> /* 110 */        public void close() throws Exception {
> /* 111 */           super.close();
> /* 112 */          
> /* 113 */        }
> /* 114 */
> /* 115 */        
> /* 116 */      }
> /* 117 */    
> /* 1 */
> /* 2 */      public class StreamExecCalc$4813 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4805 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4806 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4805, 3, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4813(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4806.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4806.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4806.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (true) {
> /* 33 */            writer$4806.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4806.writeInt(1, ((int) -1));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          
> /* 39 */          
> /* 40 */          if (false) {
> /* 41 */            writer$4806.setNullInt(2);
> /* 42 */          } else {
> /* 43 */            writer$4806.writeInt(2, ((int) 2));
> /* 44 */          }
> /* 45 */                    
> /* 46 */          writer$4806.complete();
> /* 47 */                   
> /* 48 */          this.setup(task, config, output);
> /* 49 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 50 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 51 */              .setProcessingTimeService(processingTimeService);
> /* 52 */          }
> /* 53 */        }
> /* 54 */
> /* 55 */        @Override
> /* 56 */        public void open() throws Exception {
> /* 57 */          super.open();
> /* 58 */          
> /* 59 */        }
> /* 60 */
> /* 61 */        @Override
> /* 62 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 63 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 64 */          
> /* 65 */          boolean isNull$4808;
> /* 66 */          org.apache.flink.table.data.ArrayData result$4809;
> /* 67 */          long result$4812;
> /* 68 */          
> /* 69 */          
> /* 70 */          
> /* 71 */          
> /* 72 */          out.setRowKind(in1.getRowKind());
> /* 73 */          
> /* 74 */          
> /* 75 */          
> /* 76 */          
> /* 77 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 78 */           isNull$4808 = false;
> /* 79 */          if (!isNull$4808) {
> /* 80 */          java.lang.Long[] objArray$4810 = new 
> java.lang.Long[array$4805.size()];
> /* 81 */          for (int i$4811 = 0; i$4811 < array$4805.size(); i$4811++) {
> /* 82 */          if (!array$4805.isNullAt(i$4811)) {
> /* 83 */          result$4812 = ((long)(array$4805.getInt(i$4811)));
> /* 84 */          objArray$4810[i$4811] = result$4812;
> /* 85 */          }
> /* 86 */          }
> /* 87 */          result$4809 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4810);
> /* 88 */          isNull$4808 = result$4809 == null;
> /* 89 */          } else {
> /* 90 */          result$4809 = null;
> /* 91 */          }
> /* 92 */          
> /* 93 */           // --- End cast section
> /* 94 */                         
> /* 95 */          if (isNull$4808) {
> /* 96 */            out.setNullAt(0);
> /* 97 */          } else {
> /* 98 */            out.setNonPrimitiveValue(0, result$4809);
> /* 99 */          }
> /* 100 */                    
> /* 101 */                  
> /* 102 */          output.collect(outElement.replace(out));
> /* 103 */          
> /* 104 */          
> /* 105 */        }
> /* 106 */
> /* 107 */        
> /* 108 */
> /* 109 */        @Override
> /* 110 */        public void close() throws Exception {
> /* 111 */           super.close();
> /* 112 */          
> /* 113 */        }
> /* 114 */
> /* 115 */        
> /* 116 */      }
> /* 117 */    
> /* 1 */
> /* 2 */      public class StreamExecCalc$4875 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4859 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4860 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4859, 2, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4875(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4860.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4860.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4860.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (false) {
> /* 33 */            writer$4860.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4860.writeInt(1, ((int) 2));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          writer$4860.complete();
> /* 39 */                   
> /* 40 */          this.setup(task, config, output);
> /* 41 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 42 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 43 */              .setProcessingTimeService(processingTimeService);
> /* 44 */          }
> /* 45 */        }
> /* 46 */
> /* 47 */        @Override
> /* 48 */        public void open() throws Exception {
> /* 49 */          super.open();
> /* 50 */          
> /* 51 */        }
> /* 52 */
> /* 53 */        @Override
> /* 54 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 55 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 56 */          
> /* 57 */          boolean isNull$4867;
> /* 58 */          org.apache.flink.table.data.ArrayData result$4870;
> /* 59 */          long result$4874;
> /* 60 */          
> /* 61 */          
> /* 62 */          
> /* 63 */          
> /* 64 */          out.setRowKind(in1.getRowKind());
> /* 65 */          
> /* 66 */          
> /* 67 */          
> /* 68 */          
> /* 69 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 70 */           isNull$4867 = false;
> /* 71 */          if (!isNull$4867) {
> /* 72 */          long[] objArray$4871 = new long[array$4859.size()];
> /* 73 */          for (int i$4872 = 0; i$4872 < array$4859.size(); i$4872++) {
> /* 74 */          result$4874 = ((long)(array$4859.getInt(i$4872)));
> /* 75 */          objArray$4871[i$4872] = result$4874;
> /* 76 */          }
> /* 77 */          result$4870 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4871);
> /* 78 */          isNull$4867 = result$4870 == null;
> /* 79 */          } else {
> /* 80 */          result$4870 = null;
> /* 81 */          }
> /* 82 */          
> /* 83 */           // --- End cast section
> /* 84 */                         
> /* 85 */          if (isNull$4867) {
> /* 86 */            out.setNullAt(0);
> /* 87 */          } else {
> /* 88 */            out.setNonPrimitiveValue(0, result$4870);
> /* 89 */          }
> /* 90 */                    
> /* 91 */                  
> /* 92 */          output.collect(outElement.replace(out));
> /* 93 */          
> /* 94 */          
> /* 95 */        }
> /* 96 */
> /* 97 */        
> /* 98 */
> /* 99 */        @Override
> /* 100 */        public void close() throws Exception {
> /* 101 */           super.close();
> /* 102 */          
> /* 103 */        }
> /* 104 */
> /* 105 */        
> /* 106 */      }
> /* 107 */    
> /* 1 */
> /* 2 */      public class StreamExecCalc$4875 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4859 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4860 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4859, 2, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4875(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4860.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4860.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4860.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (false) {
> /* 33 */            writer$4860.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4860.writeInt(1, ((int) 2));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          writer$4860.complete();
> /* 39 */                   
> /* 40 */          this.setup(task, config, output);
> /* 41 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 42 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 43 */              .setProcessingTimeService(processingTimeService);
> /* 44 */          }
> /* 45 */        }
> /* 46 */
> /* 47 */        @Override
> /* 48 */        public void open() throws Exception {
> /* 49 */          super.open();
> /* 50 */          
> /* 51 */        }
> /* 52 */
> /* 53 */        @Override
> /* 54 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 55 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 56 */          
> /* 57 */          boolean isNull$4867;
> /* 58 */          org.apache.flink.table.data.ArrayData result$4870;
> /* 59 */          long result$4874;
> /* 60 */          
> /* 61 */          
> /* 62 */          
> /* 63 */          
> /* 64 */          out.setRowKind(in1.getRowKind());
> /* 65 */          
> /* 66 */          
> /* 67 */          
> /* 68 */          
> /* 69 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 70 */           isNull$4867 = false;
> /* 71 */          if (!isNull$4867) {
> /* 72 */          long[] objArray$4871 = new long[array$4859.size()];
> /* 73 */          for (int i$4872 = 0; i$4872 < array$4859.size(); i$4872++) {
> /* 74 */          result$4874 = ((long)(array$4859.getInt(i$4872)));
> /* 75 */          objArray$4871[i$4872] = result$4874;
> /* 76 */          }
> /* 77 */          result$4870 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4871);
> /* 78 */          isNull$4867 = result$4870 == null;
> /* 79 */          } else {
> /* 80 */          result$4870 = null;
> /* 81 */          }
> /* 82 */          
> /* 83 */           // --- End cast section
> /* 84 */                         
> /* 85 */          if (isNull$4867) {
> /* 86 */            out.setNullAt(0);
> /* 87 */          } else {
> /* 88 */            out.setNonPrimitiveValue(0, result$4870);
> /* 89 */          }
> /* 90 */                    
> /* 91 */                  
> /* 92 */          output.collect(outElement.replace(out));
> /* 93 */          
> /* 94 */          
> /* 95 */        }
> /* 96 */
> /* 97 */        
> /* 98 */
> /* 99 */        @Override
> /* 100 */        public void close() throws Exception {
> /* 101 */           super.close();
> /* 102 */          
> /* 103 */        }
> /* 104 */
> /* 105 */        
> /* 106 */      }
> /* 107 */    
> /* 1 */
> /* 2 */      public class StreamExecCalc$4838 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4823 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4824 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4823, 3, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4838(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4824.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4824.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4824.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (true) {
> /* 33 */            writer$4824.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4824.writeInt(1, ((int) -1));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          
> /* 39 */          
> /* 40 */          if (false) {
> /* 41 */            writer$4824.setNullInt(2);
> /* 42 */          } else {
> /* 43 */            writer$4824.writeInt(2, ((int) 2));
> /* 44 */          }
> /* 45 */                    
> /* 46 */          writer$4824.complete();
> /* 47 */                   
> /* 48 */          this.setup(task, config, output);
> /* 49 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 50 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 51 */              .setProcessingTimeService(processingTimeService);
> /* 52 */          }
> /* 53 */        }
> /* 54 */
> /* 55 */        @Override
> /* 56 */        public void open() throws Exception {
> /* 57 */          super.open();
> /* 58 */          
> /* 59 */        }
> /* 60 */
> /* 61 */        @Override
> /* 62 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 63 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 64 */          
> /* 65 */          boolean isNull$4827;
> /* 66 */          org.apache.flink.table.data.ArrayData result$4828;
> /* 67 */          long result$4831;
> /* 68 */          
> /* 69 */          
> /* 70 */          
> /* 71 */          
> /* 72 */          out.setRowKind(in1.getRowKind());
> /* 73 */          
> /* 74 */          
> /* 75 */          
> /* 76 */          
> /* 77 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 78 */           isNull$4827 = false;
> /* 79 */          if (!isNull$4827) {
> /* 80 */          java.lang.Long[] objArray$4829 = new 
> java.lang.Long[array$4823.size()];
> /* 81 */          for (int i$4830 = 0; i$4830 < array$4823.size(); i$4830++) {
> /* 82 */          if (!array$4823.isNullAt(i$4830)) {
> /* 83 */          result$4831 = ((long)(array$4823.getInt(i$4830)));
> /* 84 */          objArray$4829[i$4830] = result$4831;
> /* 85 */          }
> /* 86 */          }
> /* 87 */          result$4828 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4829);
> /* 88 */          isNull$4827 = result$4828 == null;
> /* 89 */          } else {
> /* 90 */          result$4828 = null;
> /* 91 */          }
> /* 92 */          
> /* 93 */           // --- End cast section
> /* 94 */                         
> /* 95 */          if (isNull$4827) {
> /* 96 */            out.setNullAt(0);
> /* 97 */          } else {
> /* 98 */            out.setNonPrimitiveValue(0, result$4828);
> /* 99 */          }
> /* 100 */                    
> /* 101 */                  
> /* 102 */          output.collect(outElement.replace(out));
> /* 103 */          
> /* 104 */          
> /* 105 */        }
> /* 106 */
> /* 107 */        
> /* 108 */
> /* 109 */        @Override
> /* 110 */        public void close() throws Exception {
> /* 111 */           super.close();
> /* 112 */          
> /* 113 */        }
> /* 114 */
> /* 115 */        
> /* 116 */      }
> /* 117 */    
> /* 1 */
> /* 2 */      public class StreamExecCalc$4838 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4823 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4824 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4823, 3, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4838(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4824.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4824.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4824.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (true) {
> /* 33 */            writer$4824.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4824.writeInt(1, ((int) -1));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          
> /* 39 */          
> /* 40 */          if (false) {
> /* 41 */            writer$4824.setNullInt(2);
> /* 42 */          } else {
> /* 43 */            writer$4824.writeInt(2, ((int) 2));
> /* 44 */          }
> /* 45 */                    
> /* 46 */          writer$4824.complete();
> /* 47 */                   
> /* 48 */          this.setup(task, config, output);
> /* 49 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 50 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 51 */              .setProcessingTimeService(processingTimeService);
> /* 52 */          }
> /* 53 */        }
> /* 54 */
> /* 55 */        @Override
> /* 56 */        public void open() throws Exception {
> /* 57 */          super.open();
> /* 58 */          
> /* 59 */        }
> /* 60 */
> /* 61 */        @Override
> /* 62 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 63 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 64 */          
> /* 65 */          boolean isNull$4827;
> /* 66 */          org.apache.flink.table.data.ArrayData result$4828;
> /* 67 */          long result$4831;
> /* 68 */          
> /* 69 */          
> /* 70 */          
> /* 71 */          
> /* 72 */          out.setRowKind(in1.getRowKind());
> /* 73 */          
> /* 74 */          
> /* 75 */          
> /* 76 */          
> /* 77 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 78 */           isNull$4827 = false;
> /* 79 */          if (!isNull$4827) {
> /* 80 */          java.lang.Long[] objArray$4829 = new 
> java.lang.Long[array$4823.size()];
> /* 81 */          for (int i$4830 = 0; i$4830 < array$4823.size(); i$4830++) {
> /* 82 */          if (!array$4823.isNullAt(i$4830)) {
> /* 83 */          result$4831 = ((long)(array$4823.getInt(i$4830)));
> /* 84 */          objArray$4829[i$4830] = result$4831;
> /* 85 */          }
> /* 86 */          }
> /* 87 */          result$4828 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4829);
> /* 88 */          isNull$4827 = result$4828 == null;
> /* 89 */          } else {
> /* 90 */          result$4828 = null;
> /* 91 */          }
> /* 92 */          
> /* 93 */           // --- End cast section
> /* 94 */                         
> /* 95 */          if (isNull$4827) {
> /* 96 */            out.setNullAt(0);
> /* 97 */          } else {
> /* 98 */            out.setNonPrimitiveValue(0, result$4828);
> /* 99 */          }
> /* 100 */                    
> /* 101 */                  
> /* 102 */          output.collect(outElement.replace(out));
> /* 103 */          
> /* 104 */          
> /* 105 */        }
> /* 106 */
> /* 107 */        
> /* 108 */
> /* 109 */        @Override
> /* 110 */        public void close() throws Exception {
> /* 111 */           super.close();
> /* 112 */          
> /* 113 */        }
> /* 114 */
> /* 115 */        
> /* 116 */      }
> /* 117 */    
> /* 1 */
> /* 2 */      public class StreamExecCalc$4821 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4804 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4807 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4804, 2, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4821(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4807.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4807.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4807.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (false) {
> /* 33 */            writer$4807.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4807.writeInt(1, ((int) 2));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          writer$4807.complete();
> /* 39 */                   
> /* 40 */          this.setup(task, config, output);
> /* 41 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 42 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 43 */              .setProcessingTimeService(processingTimeService);
> /* 44 */          }
> /* 45 */        }
> /* 46 */
> /* 47 */        @Override
> /* 48 */        public void open() throws Exception {
> /* 49 */          super.open();
> /* 50 */          
> /* 51 */        }
> /* 52 */
> /* 53 */        @Override
> /* 54 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 55 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 56 */          
> /* 57 */          boolean isNull$4814;
> /* 58 */          org.apache.flink.table.data.ArrayData result$4815;
> /* 59 */          long result$4818;
> /* 60 */          
> /* 61 */          
> /* 62 */          
> /* 63 */          
> /* 64 */          out.setRowKind(in1.getRowKind());
> /* 65 */          
> /* 66 */          
> /* 67 */          
> /* 68 */          
> /* 69 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 70 */           isNull$4814 = false;
> /* 71 */          if (!isNull$4814) {
> /* 72 */          long[] objArray$4816 = new long[array$4804.size()];
> /* 73 */          for (int i$4817 = 0; i$4817 < array$4804.size(); i$4817++) {
> /* 74 */          result$4818 = ((long)(array$4804.getInt(i$4817)));
> /* 75 */          objArray$4816[i$4817] = result$4818;
> /* 76 */          }
> /* 77 */          result$4815 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4816);
> /* 78 */          isNull$4814 = result$4815 == null;
> /* 79 */          } else {
> /* 80 */          result$4815 = null;
> /* 81 */          }
> /* 82 */          
> /* 83 */           // --- End cast section
> /* 84 */                         
> /* 85 */          if (isNull$4814) {
> /* 86 */            out.setNullAt(0);
> /* 87 */          } else {
> /* 88 */            out.setNonPrimitiveValue(0, result$4815);
> /* 89 */          }
> /* 90 */                    
> /* 91 */                  
> /* 92 */          output.collect(outElement.replace(out));
> /* 93 */          
> /* 94 */          
> /* 95 */        }
> /* 96 */
> /* 97 */        
> /* 98 */
> /* 99 */        @Override
> /* 100 */        public void close() throws Exception {
> /* 101 */           super.close();
> /* 102 */          
> /* 103 */        }
> /* 104 */
> /* 105 */        
> /* 106 */      }
> /* 107 */    
> /* 1 */
> /* 2 */      public class StreamExecCalc$4821 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */        private final Object[] references;
> /* 6 */        
> /* 7 */        org.apache.flink.table.data.binary.BinaryArrayData array$4804 
> = new org.apache.flink.table.data.binary.BinaryArrayData();
> /* 8 */        org.apache.flink.table.data.writer.BinaryArrayWriter 
> writer$4807 = new 
> org.apache.flink.table.data.writer.BinaryArrayWriter(array$4804, 2, 4);
> /* 9 */               
> /* 10 */        org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(1);
> /* 11 */        private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 12 */
> /* 13 */        public StreamExecCalc$4821(
> /* 14 */            Object[] references,
> /* 15 */            org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 16 */            org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 17 */            org.apache.flink.streaming.api.operators.Output output,
> /* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 19 */          this.references = references;
> /* 20 */          
> /* 21 */          writer$4807.reset();
> /* 22 */          
> /* 23 */          
> /* 24 */          if (false) {
> /* 25 */            writer$4807.setNullInt(0);
> /* 26 */          } else {
> /* 27 */            writer$4807.writeInt(0, ((int) 1));
> /* 28 */          }
> /* 29 */                    
> /* 30 */          
> /* 31 */          
> /* 32 */          if (false) {
> /* 33 */            writer$4807.setNullInt(1);
> /* 34 */          } else {
> /* 35 */            writer$4807.writeInt(1, ((int) 2));
> /* 36 */          }
> /* 37 */                    
> /* 38 */          writer$4807.complete();
> /* 39 */                   
> /* 40 */          this.setup(task, config, output);
> /* 41 */          if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 42 */            
> ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) this)
> /* 43 */              .setProcessingTimeService(processingTimeService);
> /* 44 */          }
> /* 45 */        }
> /* 46 */
> /* 47 */        @Override
> /* 48 */        public void open() throws Exception {
> /* 49 */          super.open();
> /* 50 */          
> /* 51 */        }
> /* 52 */
> /* 53 */        @Override
> /* 54 */        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 55 */          org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 56 */          
> /* 57 */          boolean isNull$4814;
> /* 58 */          org.apache.flink.table.data.ArrayData result$4815;
> /* 59 */          long result$4818;
> /* 60 */          
> /* 61 */          
> /* 62 */          
> /* 63 */          
> /* 64 */          out.setRowKind(in1.getRowKind());
> /* 65 */          
> /* 66 */          
> /* 67 */          
> /* 68 */          
> /* 69 */           // --- Cast section generated by 
> org.apache.flink.table.planner.functions.casting.ArrayToArrayCastRule
> /* 70 */           isNull$4814 = false;
> /* 71 */          if (!isNull$4814) {
> /* 72 */          long[] objArray$4816 = new long[array$4804.size()];
> /* 73 */          for (int i$4817 = 0; i$4817 < array$4804.size(); i$4817++) {
> /* 74 */          result$4818 = ((long)(array$4804.getInt(i$4817)));
> /* 75 */          objArray$4816[i$4817] = result$4818;
> /* 76 */          }
> /* 77 */          result$4815 = new 
> org.apache.flink.table.data.GenericArrayData(objArray$4816);
> /* 78 */          isNull$4814 = result$4815 == null;
> /* 79 */          } else {
> /* 80 */          result$4815 = null;
> /* 81 */          }
> /* 82 */          
> /* 83 */           // --- End cast section
> /* 84 */                         
> /* 85 */          if (isNull$4814) {
> /* 86 */            out.setNullAt(0);
> /* 87 */          } else {
> /* 88 */            out.setNonPrimitiveValue(0, result$4815);
> /* 89 */          }
> /* 90 */                    
> /* 91 */                  
> /* 92 */          output.collect(outElement.replace(out));
> /* 93 */          
> /* 94 */          
> /* 95 */        }
> /* 96 */
> /* 97 */        
> /* 98 */
> /* 99 */        @Override
> /* 100 */        public void close() throws Exception {
> /* 101 */           super.close();
> /* 102 */          
> /* 103 */        }
> /* 104 */
> /* 105 */        
> /* 106 */      }
> /* 107 */    
> java.lang.RuntimeException: Failed to fetch next result
>       at 
> org.apache.flink.streaming.api.operators.collect.CollectResultIterator.nextResultFromFetcher(CollectResultIterator.java:109)
>       at 
> org.apache.flink.streaming.api.operators.collect.CollectResultIterator.hasNext(CollectResultIterator.java:80)
>       at 
> org.apache.flink.table.planner.connectors.CollectDynamicSink$CloseableRowIteratorWrapper.hasNext(CollectDynamicSink.java:222)
>       at org.assertj.core.internal.Iterators.assertHasNext(Iterators.java:49)
>       at 
> org.assertj.core.api.AbstractIteratorAssert.hasNext(AbstractIteratorAssert.java:60)
>       at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase$ResultTestItem.test(BuiltInFunctionTestBase.java:356)
>       at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase$TestSetSpec.lambda$getTestCase$4(BuiltInFunctionTestBase.java:320)
>       at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase$TestCase.execute(BuiltInFunctionTestBase.java:113)
>       at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase.test(BuiltInFunctionTestBase.java:93)
>       at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:498)
>       at 
> org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725)
>       at 
> org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
>       at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
>       at 
> org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
>       at 
> org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
>       at 
> org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestTemplateMethod(TimeoutExtension.java:92)
>       at 
> org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
>       at 
> org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
>       at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
>       at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
>       at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
>       at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
>       at 
> org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
>       at 
> org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
>       at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
>       at 
> org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
>       at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
>       at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
>       at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
>       at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
>       at 
> org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
>       at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
>       at 
> org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
>       at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
>       at 
> org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
>       at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
>       at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
>       at 
> org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService$ExclusiveTask.compute(ForkJoinPoolHierarchicalTestExecutorService.java:185)
>       at java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189)
>       at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
>       at 
> java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
>       at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
>       at 
> java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175)
> Caused by: java.io.IOException: Failed to fetch job execution result
>       at 
> org.apache.flink.streaming.api.operators.collect.CollectResultFetcher.getAccumulatorResults(CollectResultFetcher.java:184)
>       at 
> org.apache.flink.streaming.api.operators.collect.CollectResultFetcher.next(CollectResultFetcher.java:121)
>       at 
> org.apache.flink.streaming.api.operators.collect.CollectResultIterator.nextResultFromFetcher(CollectResultIterator.java:106)
>       ... 44 more
> Caused by: java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
>       at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
>       at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
>       at 
> org.apache.flink.streaming.api.operators.collect.CollectResultFetcher.getAccumulatorResults(CollectResultFetcher.java:182)
>       ... 46 more
> Caused by: org.apache.flink.runtime.client.JobExecutionException: Job 
> execution failed.
>       at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:144)
>       at 
> org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$3(MiniClusterJobClient.java:141)
>       at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
>       at 
> java.util.concurrent.CompletableFuture.uniApplyStage(CompletableFuture.java:628)
>       at 
> java.util.concurrent.CompletableFuture.thenApply(CompletableFuture.java:1996)
>       at 
> org.apache.flink.runtime.minicluster.MiniClusterJobClient.getJobExecutionResult(MiniClusterJobClient.java:138)
>       at 
> org.apache.flink.streaming.api.operators.collect.CollectResultFetcher.getAccumulatorResults(CollectResultFetcher.java:181)
>       ... 46 more
> Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by 
> NoRestartBackoffTimeStrategy
>       at 
> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:139)
>       at 
> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:83)
>       at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.recordTaskFailure(DefaultScheduler.java:256)
>       at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:247)
>       at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.onTaskFailed(DefaultScheduler.java:240)
>       at 
> org.apache.flink.runtime.scheduler.SchedulerBase.onTaskExecutionStateUpdate(SchedulerBase.java:738)
>       at 
> org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:715)
>       at 
> org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:78)
>       at 
> org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:477)
>       at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:498)
>       at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.lambda$handleRpcInvocation$1(AkkaRpcActor.java:309)
>       at 
> org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:83)
>       at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:307)
>       at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:222)
>       at 
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:84)
>       at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:168)
>       at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:24)
>       at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:20)
>       at scala.PartialFunction.applyOrElse(PartialFunction.scala:123)
>       at scala.PartialFunction.applyOrElse$(PartialFunction.scala:122)
>       at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:20)
>       at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>       at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172)
>       at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172)
>       at akka.actor.Actor.aroundReceive(Actor.scala:537)
>       at akka.actor.Actor.aroundReceive$(Actor.scala:535)
>       at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:220)
>       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:580)
>       at akka.actor.ActorCell.invoke(ActorCell.scala:548)
>       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:270)
>       at akka.dispatch.Mailbox.run(Mailbox.scala:231)
>       at akka.dispatch.Mailbox.exec(Mailbox.scala:243)
>       ... 4 more
> Caused by: java.lang.RuntimeException: Could not instantiate generated class 
> 'StreamExecCalc$4838'
>       at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:84)
>       at 
> org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
>       at 
> org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:81)
>       at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperator(OperatorChain.java:762)
>       at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:736)
>       at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:676)
>       at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:195)
>       at 
> org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.<init>(RegularOperatorChain.java:60)
>       at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:678)
>       at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:666)
>       at 
> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:935)
>       at 
> org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:904)
>       at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:728)
>       at org.apache.flink.runtime.taskmanager.Task.run(Task.java:550)
>       at java.lang.Thread.run(Thread.java:750)
> Caused by: org.apache.flink.util.FlinkRuntimeException: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.
>       at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:94)
>       at 
> org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:101)
>       at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:82)
>       ... 14 more
> Caused by: 
> org.apache.flink.shaded.guava30.com.google.common.util.concurrent.UncheckedExecutionException:
>  org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2051)
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache.get(LocalCache.java:3962)
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4859)
>       at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:92)
>       ... 16 more
> Caused by: org.apache.flink.api.common.InvalidProgramException: Table program 
> cannot be compiled. This is a bug. Please file an issue.
>       at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:107)
>       at 
> org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$0(CompileUtils.java:92)
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4864)
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3529)
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2278)
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2155)
>       at 
> org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2045)
>       ... 19 more
> Caused by: org.codehaus.commons.compiler.InternalCompilerException: Compiling 
> "StreamExecCalc$4838" in Line 2, Column 14: Line 62, Column 21: Compiling 
> "processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element)": Line 81, Column 32: Compiling "i$4830": Invalid local variable 
> index 8
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:369)
>       at org.codehaus.janino.UnitCompiler.access$000(UnitCompiler.java:231)
>       at 
> org.codehaus.janino.UnitCompiler$1.visitCompilationUnit(UnitCompiler.java:333)
>       at 
> org.codehaus.janino.UnitCompiler$1.visitCompilationUnit(UnitCompiler.java:330)
>       at org.codehaus.janino.Java$CompilationUnit.accept(Java.java:367)
>       at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:330)
>       at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:245)
>       at 
> org.codehaus.janino.SimpleCompiler.compileToClassLoader(SimpleCompiler.java:473)
>       at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:223)
>       at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:209)
>       at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:82)
>       at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:77)
>       at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:104)
>       ... 25 more
> Caused by: org.codehaus.commons.compiler.InternalCompilerException: Line 62, 
> Column 21: Compiling 
> "processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element)": Line 81, Column 32: Compiling "i$4830": Invalid local variable 
> index 8
>       at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:3222)
>       at 
> org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1379)
>       at 
> org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1352)
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:800)
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:412)
>       at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:231)
>       at 
> org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:391)
>       at 
> org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:386)
>       at 
> org.codehaus.janino.Java$PackageMemberClassDeclaration.accept(Java.java:1692)
>       at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:386)
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:359)
>       ... 37 more
> Caused by: org.codehaus.commons.compiler.InternalCompilerException: Line 81, 
> Column 32: Compiling "i$4830": Invalid local variable index 8
>       at 
> org.codehaus.janino.UnitCompiler.compileGetValue(UnitCompiler.java:5731)
>       at 
> org.codehaus.janino.UnitCompiler.compileBoolean2(UnitCompiler.java:4369)
>       at org.codehaus.janino.UnitCompiler.access$6500(UnitCompiler.java:231)
>       at 
> org.codehaus.janino.UnitCompiler$13.visitBinaryOperation(UnitCompiler.java:4166)
>       at 
> org.codehaus.janino.UnitCompiler$13.visitBinaryOperation(UnitCompiler.java:4144)
>       at org.codehaus.janino.Java$BinaryOperation.accept(Java.java:5222)
>       at 
> org.codehaus.janino.UnitCompiler.compileBoolean(UnitCompiler.java:4144)
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:1714)
>       at org.codehaus.janino.UnitCompiler.access$2000(UnitCompiler.java:231)
>       at 
> org.codehaus.janino.UnitCompiler$6.visitForStatement(UnitCompiler.java:1532)
>       at 
> org.codehaus.janino.UnitCompiler$6.visitForStatement(UnitCompiler.java:1523)
>       at org.codehaus.janino.Java$ForStatement.accept(Java.java:3321)
>       at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1523)
>       at 
> org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1607)
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:1592)
>       at org.codehaus.janino.UnitCompiler.access$1700(UnitCompiler.java:231)
>       at org.codehaus.janino.UnitCompiler$6.visitBlock(UnitCompiler.java:1529)
>       at org.codehaus.janino.UnitCompiler$6.visitBlock(UnitCompiler.java:1523)
>       at org.codehaus.janino.Java$Block.accept(Java.java:3103)
>       at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1523)
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2509)
>       at org.codehaus.janino.UnitCompiler.access$1900(UnitCompiler.java:231)
>       at 
> org.codehaus.janino.UnitCompiler$6.visitIfStatement(UnitCompiler.java:1531)
>       at 
> org.codehaus.janino.UnitCompiler$6.visitIfStatement(UnitCompiler.java:1523)
>       at org.codehaus.janino.Java$IfStatement.accept(Java.java:3274)
>       at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1523)
>       at 
> org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1607)
>       at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3531)
>       at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:3218)
>       ... 47 more
> Caused by: org.codehaus.commons.compiler.InternalCompilerException: Invalid 
> local variable index 8
>       at 
> org.codehaus.janino.UnitCompiler.getLocalVariableTypeInfo(UnitCompiler.java:13214)
>       at org.codehaus.janino.UnitCompiler.load(UnitCompiler.java:12232)
>       at org.codehaus.janino.UnitCompiler.load(UnitCompiler.java:12209)
>       at org.codehaus.janino.UnitCompiler.compileGet2(UnitCompiler.java:4679)
>       at org.codehaus.janino.UnitCompiler.access$7900(UnitCompiler.java:231)
>       at 
> org.codehaus.janino.UnitCompiler$15$1.visitLocalVariableAccess(UnitCompiler.java:4616)
>       at 
> org.codehaus.janino.UnitCompiler$15$1.visitLocalVariableAccess(UnitCompiler.java:4608)
>       at org.codehaus.janino.Java$LocalVariableAccess.accept(Java.java:4632)
>       at 
> org.codehaus.janino.UnitCompiler$15.visitLvalue(UnitCompiler.java:4608)
>       at 
> org.codehaus.janino.UnitCompiler$15.visitLvalue(UnitCompiler.java:4604)
>       at org.codehaus.janino.Java$Lvalue.accept(Java.java:4498)
>       at org.codehaus.janino.UnitCompiler.compileGet(UnitCompiler.java:4604)
>       at org.codehaus.janino.UnitCompiler.compileGet2(UnitCompiler.java:4675)
>       at org.codehaus.janino.UnitCompiler.access$7400(UnitCompiler.java:231)
>       at 
> org.codehaus.janino.UnitCompiler$15$1.visitAmbiguousName(UnitCompiler.java:4611)
>       at 
> org.codehaus.janino.UnitCompiler$15$1.visitAmbiguousName(UnitCompiler.java:4608)
>       at org.codehaus.janino.Java$AmbiguousName.accept(Java.java:4574)
>       at 
> org.codehaus.janino.UnitCompiler$15.visitLvalue(UnitCompiler.java:4608)
>       at 
> org.codehaus.janino.UnitCompiler$15.visitLvalue(UnitCompiler.java:4604)
>       at org.codehaus.janino.Java$Lvalue.accept(Java.java:4498)
>       at org.codehaus.janino.UnitCompiler.compileGet(UnitCompiler.java:4604)
>       at 
> org.codehaus.janino.UnitCompiler.compileGetValue(UnitCompiler.java:5729)
>       ... 75 more
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to