Caizhi Weng created FLINK-31659: ----------------------------------- Summary: java.lang.ClassNotFoundException: org.apache.flink.table.planner.delegation.DialectFactory when bundled Hive connector jar is in classpath Key: FLINK-31659 URL: https://issues.apache.org/jira/browse/FLINK-31659 Project: Flink Issue Type: Bug Components: Connectors / Hive Affects Versions: 1.15.4, 1.16.1, 1.17.0 Reporter: Caizhi Weng
Steps to reproduce this bug: 1. Download a fresh 1.16.1 Flink distribution. 2. Add both {{flink-sql-connector-hive-2.3.9_2.12-1.16.0.jar}} and {{flink-shaded-hadoop-2-uber-2.8.3-10.0.jar}} into {{lib}} directory. 3. Start a standalone cluster with {{bin/start-cluster.sh}}. 4. Start SQL client and run the following SQL. {code:sql} create table T ( a int, b string ) with ( 'connector' = 'filesystem', 'format' = 'csv', 'path' = '/tmp/gao.csv' ); create table S ( a int, b string ) with ( 'connector' = 'print' ); insert into S select * from T; {code} The following exception will occur. {code} org.apache.flink.table.client.gateway.SqlExecutionException: Failed to parse statement: insert into S select * from T; at org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:174) ~[flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.cli.SqlCommandParserImpl.parseCommand(SqlCommandParserImpl.java:45) ~[flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.cli.SqlMultiLineParser.parse(SqlMultiLineParser.java:71) ~[flink-sql-client-1.16.1.jar:1.16.1] at org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964) ~[flink-sql-client-1.16.1.jar:1.16.1] at org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778) ~[flink-sql-client-1.16.1.jar:1.16.1] at org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:679) ~[flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:295) [flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:280) [flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:228) [flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:151) [flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.SqlClient.start(SqlClient.java:95) [flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:187) [flink-sql-client-1.16.1.jar:1.16.1] at org.apache.flink.table.client.SqlClient.main(SqlClient.java:161) [flink-sql-client-1.16.1.jar:1.16.1] Caused by: org.apache.flink.table.api.ValidationException: Unable to create a source for reading table 'default_catalog.default_database.T'. Table options are: 'connector'='filesystem' 'format'='json' 'path'='/tmp/gao.json' at org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:166) ~[flink-table-api-java-uber-1.16.1.jar:1.16.1] at org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:191) ~[flink-table-api-java-uber-1.16.1.jar:1.16.1] at org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:175) ~[?:?] at org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:115) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3619) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2559) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2175) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2095) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2038) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:669) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:657) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3462) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570) ~[?:?] at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215) ~[?:?] at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:1253) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:374) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNodeOrFail(SqlToOperationConverter.java:384) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlInsert(SqlToOperationConverter.java:828) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:351) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:262) ~[?:?] at org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:106) ~[?:?] at org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:172) ~[flink-sql-client-1.16.1.jar:1.16.1] ... 12 more Caused by: java.lang.NoClassDefFoundError: org/apache/flink/table/planner/delegation/DialectFactory at java.lang.ClassLoader.defineClass1(Native Method) ~[?:1.8.0_361] at java.lang.ClassLoader.defineClass(ClassLoader.java:756) ~[?:1.8.0_361] at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) ~[?:1.8.0_361] at java.net.URLClassLoader.defineClass(URLClassLoader.java:473) ~[?:1.8.0_361] at java.net.URLClassLoader.access$100(URLClassLoader.java:74) ~[?:1.8.0_361] at java.net.URLClassLoader$1.run(URLClassLoader.java:369) ~[?:1.8.0_361] at java.net.URLClassLoader$1.run(URLClassLoader.java:363) ~[?:1.8.0_361] at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_361] at java.net.URLClassLoader.findClass(URLClassLoader.java:362) ~[?:1.8.0_361] at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_361] at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355) ~[?:1.8.0_361] at java.lang.ClassLoader.loadClass(ClassLoader.java:405) ~[?:1.8.0_361] at org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67) ~[flink-dist-1.16.1.jar:1.16.1] at org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65) ~[flink-dist-1.16.1.jar:1.16.1] at org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51) ~[flink-dist-1.16.1.jar:1.16.1] at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_361] at org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:192) ~[flink-dist-1.16.1.jar:1.16.1] at java.lang.Class.forName0(Native Method) ~[?:1.8.0_361] at java.lang.Class.forName(Class.java:348) ~[?:1.8.0_361] at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370) ~[?:1.8.0_361] at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404) ~[?:1.8.0_361] at java.util.ServiceLoader$1.next(ServiceLoader.java:480) ~[?:1.8.0_361] at java.util.Iterator.forEachRemaining(Iterator.java:116) ~[?:1.8.0_361] at org.apache.flink.connector.file.table.FileSystemTableFactory.formatFactoryExists(FileSystemTableFactory.java:206) ~[flink-connector-files-1.16.1.jar:1.16.1] at org.apache.flink.connector.file.table.FileSystemTableFactory.discoverDecodingFormat(FileSystemTableFactory.java:172) ~[flink-connector-files-1.16.1.jar:1.16.1] at org.apache.flink.connector.file.table.FileSystemTableFactory.createDynamicTableSource(FileSystemTableFactory.java:77) ~[flink-connector-files-1.16.1.jar:1.16.1] at org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:163) ~[flink-table-api-java-uber-1.16.1.jar:1.16.1] at org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:191) ~[flink-table-api-java-uber-1.16.1.jar:1.16.1] at org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:175) ~[?:?] at org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:115) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3619) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2559) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2175) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2095) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2038) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:669) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:657) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3462) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570) ~[?:?] at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215) ~[?:?] at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:1253) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:374) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNodeOrFail(SqlToOperationConverter.java:384) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlInsert(SqlToOperationConverter.java:828) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:351) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:262) ~[?:?] at org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:106) ~[?:?] at org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:172) ~[flink-sql-client-1.16.1.jar:1.16.1] ... 12 more Caused by: java.lang.ClassNotFoundException: org.apache.flink.table.planner.delegation.DialectFactory at java.net.URLClassLoader.findClass(URLClassLoader.java:387) ~[?:1.8.0_361] at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_361] at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355) ~[?:1.8.0_361] at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_361] at java.lang.ClassLoader.defineClass1(Native Method) ~[?:1.8.0_361] at java.lang.ClassLoader.defineClass(ClassLoader.java:756) ~[?:1.8.0_361] at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) ~[?:1.8.0_361] at java.net.URLClassLoader.defineClass(URLClassLoader.java:473) ~[?:1.8.0_361] at java.net.URLClassLoader.access$100(URLClassLoader.java:74) ~[?:1.8.0_361] at java.net.URLClassLoader$1.run(URLClassLoader.java:369) ~[?:1.8.0_361] at java.net.URLClassLoader$1.run(URLClassLoader.java:363) ~[?:1.8.0_361] at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_361] at java.net.URLClassLoader.findClass(URLClassLoader.java:362) ~[?:1.8.0_361] at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_361] at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355) ~[?:1.8.0_361] at java.lang.ClassLoader.loadClass(ClassLoader.java:405) ~[?:1.8.0_361] at org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67) ~[flink-dist-1.16.1.jar:1.16.1] at org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65) ~[flink-dist-1.16.1.jar:1.16.1] at org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51) ~[flink-dist-1.16.1.jar:1.16.1] at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_361] at org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:192) ~[flink-dist-1.16.1.jar:1.16.1] at java.lang.Class.forName0(Native Method) ~[?:1.8.0_361] at java.lang.Class.forName(Class.java:348) ~[?:1.8.0_361] at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370) ~[?:1.8.0_361] at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404) ~[?:1.8.0_361] at java.util.ServiceLoader$1.next(ServiceLoader.java:480) ~[?:1.8.0_361] at java.util.Iterator.forEachRemaining(Iterator.java:116) ~[?:1.8.0_361] at org.apache.flink.connector.file.table.FileSystemTableFactory.formatFactoryExists(FileSystemTableFactory.java:206) ~[flink-connector-files-1.16.1.jar:1.16.1] at org.apache.flink.connector.file.table.FileSystemTableFactory.discoverDecodingFormat(FileSystemTableFactory.java:172) ~[flink-connector-files-1.16.1.jar:1.16.1] at org.apache.flink.connector.file.table.FileSystemTableFactory.createDynamicTableSource(FileSystemTableFactory.java:77) ~[flink-connector-files-1.16.1.jar:1.16.1] at org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:163) ~[flink-table-api-java-uber-1.16.1.jar:1.16.1] at org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:191) ~[flink-table-api-java-uber-1.16.1.jar:1.16.1] at org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:175) ~[?:?] at org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:115) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3619) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2559) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2175) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2095) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2038) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:669) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:657) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3462) ~[?:?] at org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570) ~[?:?] at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215) ~[?:?] at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:1253) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:374) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNodeOrFail(SqlToOperationConverter.java:384) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlInsert(SqlToOperationConverter.java:828) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:351) ~[?:?] at org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:262) ~[?:?] at org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:106) ~[?:?] at org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:172) ~[flink-sql-client-1.16.1.jar:1.16.1] ... 12 more {code} The [document|https://nightlies.apache.org/flink/flink-docs-release-1.16/docs/connectors/table/hive/overview/#moving-the-planner-jar] states that if we're not using Hive dialect, we don't have to move the planner jar. However we clearly did not use Hive dialect in the above SQL, but the exception still happens. -- This message was sent by Atlassian Jira (v8.20.10#820010)