[ https://issues.apache.org/jira/browse/FLINK-33168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17773893#comment-17773893 ]
macdoor615 commented on FLINK-33168: ------------------------------------ [~luoyuxia] table-planner-loader.jar works. Docs should be updated. > An error occurred when executing sql, java.lang.NoSuchFieldError: operands > -------------------------------------------------------------------------- > > Key: FLINK-33168 > URL: https://issues.apache.org/jira/browse/FLINK-33168 > Project: Flink > Issue Type: Bug > Components: Table SQL / Client > Affects Versions: 1.18.0 > Reporter: macdoor615 > Assignee: Zheng yunhong > Priority: Major > > Environment: > > {code:java} > Linux hb3-prod-hadoop-006 4.18.0-477.27.1.el8_8.x86_64 #1 SMP Thu Sep 21 > 06:49:25 EDT 2023 x86_64 x86_64 x86_64 GNU/Linux > openjdk version "1.8.0_382" > OpenJDK Runtime Environment (build 1.8.0_382-b05) > OpenJDK 64-Bit Server VM (build 25.382-b05, mixed mode) > flink-1.18.0-RC1 , > https://github.com/apache/flink/releases/tag/release-1.18.0-rc1 > {code} > > I execute the following sql in sql-client.sh. > > {code:java} > insert into svc1_paimon_prod.cq.b_customer_ecus > select > rcus.id id, > if(cus.id is not null, cus.id, try_cast(NULL as string)) cus_id, > if(cus.id is null and cus_rownum = 1, rcus.id, try_cast(NULL as string)) > newCus_id, > companyID, > customerProvinceNumber, > mobilePhone, > oprCode, > customerNum, > staffName, > location, > staffNumber, > extendInfo, > customerName, > case when companyID='000' then '名称1' > when companyID='002' then '名称2' > else '新名称' > end prov, > row ( > accessToken, > busType, > cutOffDay, > domain, > envFlag, > routeType, > routeValue, > sessionID, > sign, > signMethod, > org_timeStamp, > transIDO, > userPartyID, > version > ) raw_message, > named_struct( > 'id', cus.id, > 'name', cus.name, > 'code', cus.code, > 'customerlevel', cus.customerlevel, > 'prov', cus.prov, > 'container', cus.container, > 'crtime', cus.crtime, > 'updtime', cus.updtime > ) existing_cus, > cus_rownum, > to_timestamp(org_timeStamp, 'yyyyMMddHHmmss') as org_timeStamp, > raw_rowtime, > localtimestamp as raw_rowtime1, > dt > from svc1_paimon_prod.raw_data.abscustinfoserv_content_append_cq > /*+ OPTIONS('consumer-id' = '创建新客户id') */ > rcus > left join svc1_mysql_test.gem_svc1_vpn.bv_customer > FOR SYSTEM_TIME AS OF rcus.proctime AS cus on rcus.customerNum=cus.code > {code} > There are the following jar files in the flink/lib directory. > {code:java} > commons-cli-1.5.0.jar > flink-cep-1.18.0.jar > flink-connector-files-1.18.0.jar > flink-connector-jdbc-3.1.1-1.17.jar > flink-csv-1.18.0.jar > flink-dist-1.18.0.jar > flink-json-1.18.0.jar > flink-orc-1.18.0.jar > flink-parquet-1.18.0.jar > flink-scala_2.12-1.18.0.jar > flink-sql-avro-1.18.0.jar > flink-sql-avro-confluent-registry-1.18.0.jar > flink-sql-connector-elasticsearch7-3.0.0-1.16.jar > flink-sql-connector-hive-3.1.3_2.12-1.18.0.jar > flink-sql-connector-kafka-3.0.0-1.17.jar > flink-sql-orc-1.18.0.jar > flink-sql-parquet-1.18.0.jar > flink-table-api-java-uber-1.18.0.jar > flink-table-api-scala_2.12-1.18.0.jar > flink-table-api-scala-bridge_2.12-1.18.0.jar > flink-table-planner_2.12-1.18.0.jar > flink-table-runtime-1.18.0.jar > jline-reader-3.23.0.jar > jline-terminal-3.23.0.jar > kafka-clients-3.5.1.jar > log4j-1.2-api-2.17.1.jar > log4j-api-2.17.1.jar > log4j-core-2.17.1.jar > log4j-slf4j-impl-2.17.1.jar > mysql-connector-j-8.1.0.jar > paimon-flink-1.18-0.6-20230929.002044-11.jar{code} > Works correctly in version 1.17.1, but produces the following error in > 1.18.0-RC1 > > {code:java} > 2023-09-29 14:04:11,438 ERROR > org.apache.flink.table.gateway.service.operation.OperationManager [] - Failed > to execute the operation fe1b0a58-b822-49c0-b1ae-ce73d16f92da. > java.lang.NoSuchFieldError: operands > at org.apache.calcite.plan.RelOptRule.operand(RelOptRule.java:129) > ~[flink-sql-connector-hive-3.1.3_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule.<init>(SimplifyFilterConditionRule.scala:36) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<init>(SimplifyFilterConditionRule.scala:94) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<clinit>(SimplifyFilterConditionRule.scala) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<init>(FlinkStreamRuleSets.scala:35) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<clinit>(FlinkStreamRuleSets.scala) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.program.FlinkStreamProgram$.buildProgram(FlinkStreamProgram.scala:57) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.$anonfun$optimizeTree$1(StreamCommonSubGraphBasedOptimizer.scala:169) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at scala.Option.getOrElse(Option.scala:189) > ~[flink-scala_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.optimizeTree(StreamCommonSubGraphBasedOptimizer.scala:169) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.doOptimize(StreamCommonSubGraphBasedOptimizer.scala:83) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:87) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:324) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:182) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1277) > ~[flink-table-api-java-uber-1.18.0.jar:1.18.0] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:862) > ~[flink-table-api-java-uber-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.callModifyOperations(OperationExecutor.java:513) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeOperation(OperationExecutor.java:426) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeStatement(OperationExecutor.java:207) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.SqlGatewayServiceImpl.lambda$executeStatement$1(SqlGatewayServiceImpl.java:212) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager.lambda$submitOperation$1(OperationManager.java:119) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.lambda$run$0(OperationManager.java:258) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > [?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_382] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > [?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > [?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > [?:1.8.0_382] > at java.lang.Thread.run(Thread.java:750) [?:1.8.0_382] > 2023-09-29 14:04:11,505 ERROR > org.apache.flink.table.gateway.service.SqlGatewayServiceImpl [] - Failed to > fetchResults. > org.apache.flink.table.gateway.service.utils.SqlExecutionException: Failed to > execute the operation fe1b0a58-b822-49c0-b1ae-ce73d16f92da. > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.processThrowable(OperationManager.java:414) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.lambda$run$0(OperationManager.java:267) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > ~[?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_382] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > ~[?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > ~[?:1.8.0_382] > at java.lang.Thread.run(Thread.java:750) [?:1.8.0_382] > Caused by: java.lang.NoSuchFieldError: operands > at org.apache.calcite.plan.RelOptRule.operand(RelOptRule.java:129) > ~[flink-sql-connector-hive-3.1.3_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule.<init>(SimplifyFilterConditionRule.scala:36) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<init>(SimplifyFilterConditionRule.scala:94) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<clinit>(SimplifyFilterConditionRule.scala) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<init>(FlinkStreamRuleSets.scala:35) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<clinit>(FlinkStreamRuleSets.scala) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.program.FlinkStreamProgram$.buildProgram(FlinkStreamProgram.scala:57) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.$anonfun$optimizeTree$1(StreamCommonSubGraphBasedOptimizer.scala:169) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at scala.Option.getOrElse(Option.scala:189) > ~[flink-scala_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.optimizeTree(StreamCommonSubGraphBasedOptimizer.scala:169) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.doOptimize(StreamCommonSubGraphBasedOptimizer.scala:83) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:87) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:324) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:182) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1277) > ~[flink-table-api-java-uber-1.18.0.jar:1.18.0] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:862) > ~[flink-table-api-java-uber-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.callModifyOperations(OperationExecutor.java:513) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeOperation(OperationExecutor.java:426) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeStatement(OperationExecutor.java:207) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.SqlGatewayServiceImpl.lambda$executeStatement$1(SqlGatewayServiceImpl.java:212) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager.lambda$submitOperation$1(OperationManager.java:119) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.lambda$run$0(OperationManager.java:258) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > ... 7 more > 2023-09-29 14:04:11,508 ERROR > org.apache.flink.table.gateway.rest.handler.statement.FetchResultsHandler [] > - Unhandled exception. > org.apache.flink.table.gateway.api.utils.SqlGatewayException: > org.apache.flink.table.gateway.api.utils.SqlGatewayException: Failed to > fetchResults. > at > org.apache.flink.table.gateway.rest.handler.statement.FetchResultsHandler.handleRequest(FetchResultsHandler.java:85) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.rest.handler.AbstractSqlGatewayRestHandler.respondToRequest(AbstractSqlGatewayRestHandler.java:84) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.rest.handler.AbstractSqlGatewayRestHandler.respondToRequest(AbstractSqlGatewayRestHandler.java:52) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.handler.AbstractHandler.respondAsLeader(AbstractHandler.java:196) > ~[flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.lambda$channelRead0$0(LeaderRetrievalHandler.java:83) > ~[flink-dist-1.18.0.jar:1.18.0] > at java.util.Optional.ifPresent(Optional.java:159) [?:1.8.0_382] > at org.apache.flink.util.OptionalConsumer.ifPresent(OptionalConsumer.java:45) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:80) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:49) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.handler.router.RouterHandler.routed(RouterHandler.java:115) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:94) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:55) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:208) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:69) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:436) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:346) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:318) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler.channelRead(CombinedChannelDuplexHandler.java:251) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) > [flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > [flink-dist-1.18.0.jar:1.18.0] > at java.lang.Thread.run(Thread.java:750) [?:1.8.0_382] > Caused by: org.apache.flink.table.gateway.api.utils.SqlGatewayException: > Failed to fetchResults. > at > org.apache.flink.table.gateway.service.SqlGatewayServiceImpl.fetchResults(SqlGatewayServiceImpl.java:229) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.rest.handler.statement.FetchResultsHandler.handleRequest(FetchResultsHandler.java:83) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > ... 48 more > Caused by: > org.apache.flink.table.gateway.service.utils.SqlExecutionException: Failed to > execute the operation fe1b0a58-b822-49c0-b1ae-ce73d16f92da. > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.processThrowable(OperationManager.java:414) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.lambda$run$0(OperationManager.java:267) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > ~[?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_382] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > ~[?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > ~[?:1.8.0_382] > ... 1 more > Caused by: java.lang.NoSuchFieldError: operands > at org.apache.calcite.plan.RelOptRule.operand(RelOptRule.java:129) > ~[flink-sql-connector-hive-3.1.3_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule.<init>(SimplifyFilterConditionRule.scala:36) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<init>(SimplifyFilterConditionRule.scala:94) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<clinit>(SimplifyFilterConditionRule.scala) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<init>(FlinkStreamRuleSets.scala:35) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<clinit>(FlinkStreamRuleSets.scala) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.program.FlinkStreamProgram$.buildProgram(FlinkStreamProgram.scala:57) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.$anonfun$optimizeTree$1(StreamCommonSubGraphBasedOptimizer.scala:169) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at scala.Option.getOrElse(Option.scala:189) > ~[flink-scala_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.optimizeTree(StreamCommonSubGraphBasedOptimizer.scala:169) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.doOptimize(StreamCommonSubGraphBasedOptimizer.scala:83) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:87) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:324) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:182) > ~[flink-table-planner_2.12-1.18.0.jar:1.18.0] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1277) > ~[flink-table-api-java-uber-1.18.0.jar:1.18.0] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:862) > ~[flink-table-api-java-uber-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.callModifyOperations(OperationExecutor.java:513) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeOperation(OperationExecutor.java:426) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeStatement(OperationExecutor.java:207) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.SqlGatewayServiceImpl.lambda$executeStatement$1(SqlGatewayServiceImpl.java:212) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager.lambda$submitOperation$1(OperationManager.java:119) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.lambda$run$0(OperationManager.java:258) > ~[flink-sql-gateway-1.18.0.jar:1.18.0] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > ~[?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_382] > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > ~[?:1.8.0_382] > at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > ~[?:1.8.0_382] > ... 1 more > 2023-09-29 14:04:11,517 WARN org.apache.flink.table.client.cli.CliClient > [] - Could not execute SQL statement. > org.apache.flink.table.client.gateway.SqlExecutionException: Failed to get > response for the operation fe1b0a58-b822-49c0-b1ae-ce73d16f92da. > at > org.apache.flink.table.client.gateway.ExecutorImpl.getFetchResultResponse(ExecutorImpl.java:488) > ~[flink-sql-client-1.18.0.jar:1.18.0] > at > org.apache.flink.table.client.gateway.ExecutorImpl.fetchUtilResultsReady(ExecutorImpl.java:448) > ~[flink-sql-client-1.18.0.jar:1.18.0] > at > org.apache.flink.table.client.gateway.ExecutorImpl.executeStatement(ExecutorImpl.java:309) > ~[flink-sql-client-1.18.0.jar:1.18.0] > at > org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:113) > ~[flink-sql-client-1.18.0.jar:1.18.0] > at org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2994) > ~[jline-reader-3.23.0.jar:?] > at org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3812) > ~[jline-reader-3.23.0.jar:?] > at org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:689) > ~[jline-reader-3.23.0.jar:?] > at > org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:194) > [flink-sql-client-1.18.0.jar:1.18.0] > at > org.apache.flink.table.client.cli.CliClient.executeFile(CliClient.java:243) > [flink-sql-client-1.18.0.jar:1.18.0] > at > org.apache.flink.table.client.cli.CliClient.executeInNonInteractiveMode(CliClient.java:131) > [flink-sql-client-1.18.0.jar:1.18.0] > at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:171) > [flink-sql-client-1.18.0.jar:1.18.0] > at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118) > [flink-sql-client-1.18.0.jar:1.18.0] > at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228) > [flink-sql-client-1.18.0.jar:1.18.0] > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179) > [flink-sql-client-1.18.0.jar:1.18.0] > Caused by: org.apache.flink.runtime.rest.util.RestClientException: [Internal > server error., <Exception on server side: > org.apache.flink.table.gateway.api.utils.SqlGatewayException: > org.apache.flink.table.gateway.api.utils.SqlGatewayException: Failed to > fetchResults. > at > org.apache.flink.table.gateway.rest.handler.statement.FetchResultsHandler.handleRequest(FetchResultsHandler.java:85) > at > org.apache.flink.table.gateway.rest.handler.AbstractSqlGatewayRestHandler.respondToRequest(AbstractSqlGatewayRestHandler.java:84) > at > org.apache.flink.table.gateway.rest.handler.AbstractSqlGatewayRestHandler.respondToRequest(AbstractSqlGatewayRestHandler.java:52) > at > org.apache.flink.runtime.rest.handler.AbstractHandler.respondAsLeader(AbstractHandler.java:196) > at > org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.lambda$channelRead0$0(LeaderRetrievalHandler.java:83) > at java.util.Optional.ifPresent(Optional.java:159) > at org.apache.flink.util.OptionalConsumer.ifPresent(OptionalConsumer.java:45) > at > org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:80) > at > org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:49) > at > org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > at > org.apache.flink.runtime.rest.handler.router.RouterHandler.routed(RouterHandler.java:115) > at > org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:94) > at > org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:55) > at > org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > at > org.apache.flink.shaded.netty4.io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > at > org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:208) > at > org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:69) > at > org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > at > org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:436) > at > org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:346) > at > org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:318) > at > org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler.channelRead(CombinedChannelDuplexHandler.java:251) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) > at > org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440) > at > org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) > at > org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) > at > org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) > at > org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) > at > org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > at java.lang.Thread.run(Thread.java:750) > Caused by: org.apache.flink.table.gateway.api.utils.SqlGatewayException: > Failed to fetchResults. > at > org.apache.flink.table.gateway.service.SqlGatewayServiceImpl.fetchResults(SqlGatewayServiceImpl.java:229) > at > org.apache.flink.table.gateway.rest.handler.statement.FetchResultsHandler.handleRequest(FetchResultsHandler.java:83) > ... 48 more > Caused by: > org.apache.flink.table.gateway.service.utils.SqlExecutionException: Failed to > execute the operation fe1b0a58-b822-49c0-b1ae-ce73d16f92da. > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.processThrowable(OperationManager.java:414) > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.lambda$run$0(OperationManager.java:267) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > ... 1 more > Caused by: java.lang.NoSuchFieldError: operands > at org.apache.calcite.plan.RelOptRule.operand(RelOptRule.java:129) > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule.<init>(SimplifyFilterConditionRule.scala:36) > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<init>(SimplifyFilterConditionRule.scala:94) > at > org.apache.flink.table.planner.plan.rules.logical.SimplifyFilterConditionRule$.<clinit>(SimplifyFilterConditionRule.scala) > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<init>(FlinkStreamRuleSets.scala:35) > at > org.apache.flink.table.planner.plan.rules.FlinkStreamRuleSets$.<clinit>(FlinkStreamRuleSets.scala) > at > org.apache.flink.table.planner.plan.optimize.program.FlinkStreamProgram$.buildProgram(FlinkStreamProgram.scala:57) > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.$anonfun$optimizeTree$1(StreamCommonSubGraphBasedOptimizer.scala:169) > at scala.Option.getOrElse(Option.scala:189) > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.optimizeTree(StreamCommonSubGraphBasedOptimizer.scala:169) > at > org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.doOptimize(StreamCommonSubGraphBasedOptimizer.scala:83) > at > org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:87) > at > org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:324) > at > org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:182) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1277) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:862) > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.callModifyOperations(OperationExecutor.java:513) > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeOperation(OperationExecutor.java:426) > at > org.apache.flink.table.gateway.service.operation.OperationExecutor.executeStatement(OperationExecutor.java:207) > at > org.apache.flink.table.gateway.service.SqlGatewayServiceImpl.lambda$executeStatement$1(SqlGatewayServiceImpl.java:212) > at > org.apache.flink.table.gateway.service.operation.OperationManager.lambda$submitOperation$1(OperationManager.java:119) > at > org.apache.flink.table.gateway.service.operation.OperationManager$Operation.lambda$run$0(OperationManager.java:258) > ... 7 more > > End of exception on server side>] > at > org.apache.flink.runtime.rest.RestClient.parseResponse(RestClient.java:646) > ~[flink-dist-1.18.0.jar:1.18.0] > at > org.apache.flink.runtime.rest.RestClient.lambda$submitRequest$6(RestClient.java:626) > ~[flink-dist-1.18.0.jar:1.18.0] > at > java.util.concurrent.CompletableFuture.uniCompose(CompletableFuture.java:966) > ~[?:1.8.0_382] > at > java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:940) > ~[?:1.8.0_382] > at > java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:456) > ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > ~[?:1.8.0_382] > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > ~[?:1.8.0_382] > at java.lang.Thread.run(Thread.java:750) ~[?:1.8.0_382] > 2023-09-29 14:04:11,528 INFO > org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog [] - Catalog > mysql_bnpmp closing > 2023-09-29 14:04:11,528 INFO > org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog [] - Catalog > mysql_service1 closing > 2023-09-29 14:04:11,528 INFO > org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog [] - Catalog > svc1_mysql_test closing > 2023-09-29 14:04:11,528 INFO > org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog [] - Catalog > bnpmp_mysql_prod closing > 2023-09-29 14:04:11,528 INFO > org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog [] - Catalog > mysql_test closing > 2023-09-29 14:04:11,528 INFO > org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog [] - Catalog > bnpmp_mysql_test closing > 2023-09-29 14:04:11,528 INFO > org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog [] - Catalog > svc1_mysql_prod closing > 2023-09-29 14:04:11,536 INFO org.apache.flink.table.catalog.hive.HiveCatalog > [] - Close connection to Hive metastore > 2023-09-29 14:04:11,546 INFO > org.apache.flink.table.gateway.rest.SqlGatewayRestEndpoint [] - Shutting > down rest endpoint. > 2023-09-29 14:04:11,555 INFO > org.apache.flink.table.gateway.rest.SqlGatewayRestEndpoint [] - Shut down > complete. > 2023-09-29 14:04:12,547 INFO > org.apache.flink.table.gateway.rest.SqlGatewayRestEndpoint [] - Shutting > down rest endpoint. > {code} > > > > -- This message was sent by Atlassian Jira (v8.20.10#820010)