Thanks Caizhi for the quick validation. Let's cancel this RC and wait for your fix.
Best, Jingsong On Thu, Apr 28, 2022 at 2:55 PM Caizhi Weng <tsreape...@gmail.com> wrote: > > Hi all! > > -1 for this release. Currently we cannot write array and map types into > table store due to this commit > <https://github.com/apache/flink-table-store/commit/3bf8cde932b0f8512b348c86eb412a483039804c>. > Run the following SQL in SQL client and an exception will be thrown: > > create table if not exists tstore2 ( > a ARRAY<STRING>, > b MAP<BIGINT, STRING> > ) with ( > 'path' = 'hdfs:///tstore2', > 'bucket' = '4' > ); > > insert into tstore2 values (array['hi','hello',cast(null as > string),'test'], map[1,'A',2,'BB',100,'CCC']), (cast(null as > array<string>), cast(null as map<bigint, string>)); > > The exception stack is > > Exception in thread "main" > org.apache.flink.table.client.SqlClientException: Unexpected exception. > This is a bug. Please consider filing an issue. > > at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:201) > > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:161) > > Caused by: java.lang.NoClassDefFoundError: > org/apache/flink/table/planner/plan/utils/SortUtil$ > > at > org.apache.flink.table.planner.codegen.GenerateUtils$.generateCompare(GenerateUtils.scala:675) > > at > org.apache.flink.table.planner.codegen.GenerateUtils$.$anonfun$generateRowCompare$1(GenerateUtils.scala:828) > > at > scala.collection.IndexedSeqOptimized.foreach(IndexedSeqOptimized.scala:32) > > at > scala.collection.IndexedSeqOptimized.foreach$(IndexedSeqOptimized.scala:29) > > at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:194) > > at > org.apache.flink.table.planner.codegen.GenerateUtils$.generateRowCompare(GenerateUtils.scala:802) > > at > org.apache.flink.table.planner.codegen.sort.ComparatorCodeGenerator$.gen(ComparatorCodeGenerator.scala:53) > > at > org.apache.flink.table.planner.codegen.sort.ComparatorCodeGenerator.gen(ComparatorCodeGenerator.scala) > > at > org.apache.flink.table.store.codegen.CodeGeneratorImpl.generateRecordComparator(CodeGeneratorImpl.java:60) > > at > org.apache.flink.table.store.codegen.CodeGenUtils.generateRecordComparator(CodeGenUtils.java:67) > > at > org.apache.flink.table.store.file.FileStoreImpl.<init>(FileStoreImpl.java:67) > > at > org.apache.flink.table.store.connector.TableStore.buildFileStore(TableStore.java:220) > > at > org.apache.flink.table.store.connector.TableStore.access$100(TableStore.java:81) > > at > org.apache.flink.table.store.connector.TableStore$SinkBuilder.build(TableStore.java:415) > > at > org.apache.flink.table.store.connector.sink.TableStoreSink.lambda$getSinkRuntimeProvider$0(TableStoreSink.java:143) > > at > org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecSink.applySinkProvider(CommonExecSink.java:446) > > at > org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecSink.createSinkTransformation(CommonExecSink.java:192) > > at > org.apache.flink.table.planner.plan.nodes.exec.batch.BatchExecSink.translateToPlanInternal(BatchExecSink.java:67) > > at > org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:148) > > at > org.apache.flink.table.planner.delegation.BatchPlanner.$anonfun$translateToPlan$1(BatchPlanner.scala:82) > > at > scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:233) > > at scala.collection.Iterator.foreach(Iterator.scala:937) > > at scala.collection.Iterator.foreach$(Iterator.scala:937) > > at scala.collection.AbstractIterator.foreach(Iterator.scala:1425) > > at scala.collection.IterableLike.foreach(IterableLike.scala:70) > > at scala.collection.IterableLike.foreach$(IterableLike.scala:69) > > at scala.collection.AbstractIterable.foreach(Iterable.scala:54) > > at scala.collection.TraversableLike.map(TraversableLike.scala:233) > > at scala.collection.TraversableLike.map$(TraversableLike.scala:226) > > at scala.collection.AbstractTraversable.map(Traversable.scala:104) > > at > org.apache.flink.table.planner.delegation.BatchPlanner.translateToPlan(BatchPlanner.scala:81) > > at > org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:181) > > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1656) > > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:782) > > at > org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$executeModifyOperations$4(LocalExecutor.java:222) > > at > org.apache.flink.table.client.gateway.context.ExecutionContext.wrapClassLoader(ExecutionContext.java:88) > > at > org.apache.flink.table.client.gateway.local.LocalExecutor.executeModifyOperations(LocalExecutor.java:222) > > at > org.apache.flink.table.client.cli.CliClient.callInserts(CliClient.java:600) > > at > org.apache.flink.table.client.cli.CliClient.callInsert(CliClient.java:589) > > at > org.apache.flink.table.client.cli.CliClient.callOperation(CliClient.java:443) > > at > org.apache.flink.table.client.cli.CliClient.executeOperation(CliClient.java:373) > > at > org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:330) > > at > org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:281) > > at > org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:229) > > at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:151) > > at org.apache.flink.table.client.SqlClient.start(SqlClient.java:95) > > at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:187) > > ... 1 more > > Caused by: java.lang.ClassNotFoundException: > org.apache.flink.table.planner.plan.utils.SortUtil$ > > at java.net.URLClassLoader.findClass(URLClassLoader.java:382) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:418) > > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:351) > > at > org.apache.flink.core.classloading.ComponentClassLoader.loadClassFromOwnerOnly(ComponentClassLoader.java:163) > > at > org.apache.flink.core.classloading.ComponentClassLoader.loadClassFromComponentFirst(ComponentClassLoader.java:157) > > at > org.apache.flink.core.classloading.ComponentClassLoader.loadClass(ComponentClassLoader.java:103) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:351) > > ... 48 more > > Sorry for introducing this bug. I plan to fix it very soon and we also need > an e2e test case for all supported data types. > > > Jingsong Li <jingsongl...@gmail.com> 于2022年4月28日周四 14:50写道: > > > Hi everyone, > > > > Please review and vote on the release candidate #1 for the version 0.1.0 of > > Apache Flink Table Store, as follows: > > > > [ ] +1, Approve the release > > [ ] -1, Do not approve the release (please provide specific comments) > > > > **Release Overview** > > > > As an overview, the release consists of the following: > > a) Table Store canonical source distribution, to be deployed to the > > release repository at dist.apache.org > > b) Maven artifacts to be deployed to the Maven Central Repository > > > > **Staging Areas to Review** > > > > The staging areas containing the above mentioned artifacts are as follows, > > for your review: > > * All artifacts for a) and b) can be found in the corresponding dev > > repository at dist.apache.org [2] > > * All artifacts for c) can be found at the Apache Nexus Repository [3] > > * Pre Bundled Binaries Jar can work fine with quick start [4][5] > > > > All artifacts are signed with the key > > 2C2B6A653B07086B65E4369F7C76245E0A318150 [6] > > > > Other links for your review: > > * JIRA release notes [7] > > * source code tag "release-0.1.0-rc1" [8] > > * PR to update the website Downloads page to include Table Store > > links [9] > > > > **Vote Duration** > > > > The voting time will run for at least 72 hours. > > It is adopted by majority approval, with at least 3 PMC affirmative votes. > > > > Best, > > Jingsong Lee > > > > [1] > > https://cwiki.apache.org/confluence/display/FLINK/Verifying+a+Flink+Table+Store+Release > > [2] > > https://dist.apache.org/repos/dist/dev/flink/flink-table-store-0.1.0-rc1/ > > [3] > > https://repository.apache.org/content/repositories/orgapacheflink-1501/ > > [4] > > https://repository.apache.org/content/repositories/orgapacheflink-1501/org/apache/flink/flink-table-store-dist/0.1.0/flink-table-store-dist-0.1.0.jar > > [5] > > https://nightlies.apache.org/flink/flink-table-store-docs-release-0.1/docs/try-table-store/quick-start/ > > [6] https://dist.apache.org/repos/dist/release/flink/KEYS > > [7] > > https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12351234 > > [8] https://github.com/apache/flink-table-store/tree/release-0.1.0-rc1 > > [9] https://github.com/apache/flink-web/pull/531 > >