Hi all!

-1 for this release. Currently we cannot write array and map types into
table store due to this commit
<https://github.com/apache/flink-table-store/commit/3bf8cde932b0f8512b348c86eb412a483039804c>.
Run the following SQL in SQL client and an exception will be thrown:

create table if not exists tstore2 (
a ARRAY<STRING>,
b MAP<BIGINT, STRING>
) with (
'path' = 'hdfs:///tstore2',
'bucket' = '4'
);

insert into tstore2 values (array['hi','hello',cast(null as
string),'test'], map[1,'A',2,'BB',100,'CCC']), (cast(null as
array<string>), cast(null as map<bigint, string>));

The exception stack is

Exception in thread "main"
org.apache.flink.table.client.SqlClientException: Unexpected exception.
This is a bug. Please consider filing an issue.

at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:201)

at org.apache.flink.table.client.SqlClient.main(SqlClient.java:161)

Caused by: java.lang.NoClassDefFoundError:
org/apache/flink/table/planner/plan/utils/SortUtil$

at
org.apache.flink.table.planner.codegen.GenerateUtils$.generateCompare(GenerateUtils.scala:675)

at
org.apache.flink.table.planner.codegen.GenerateUtils$.$anonfun$generateRowCompare$1(GenerateUtils.scala:828)

at
scala.collection.IndexedSeqOptimized.foreach(IndexedSeqOptimized.scala:32)

at
scala.collection.IndexedSeqOptimized.foreach$(IndexedSeqOptimized.scala:29)

at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:194)

at
org.apache.flink.table.planner.codegen.GenerateUtils$.generateRowCompare(GenerateUtils.scala:802)

at
org.apache.flink.table.planner.codegen.sort.ComparatorCodeGenerator$.gen(ComparatorCodeGenerator.scala:53)

at
org.apache.flink.table.planner.codegen.sort.ComparatorCodeGenerator.gen(ComparatorCodeGenerator.scala)

at
org.apache.flink.table.store.codegen.CodeGeneratorImpl.generateRecordComparator(CodeGeneratorImpl.java:60)

at
org.apache.flink.table.store.codegen.CodeGenUtils.generateRecordComparator(CodeGenUtils.java:67)

at
org.apache.flink.table.store.file.FileStoreImpl.<init>(FileStoreImpl.java:67)

at
org.apache.flink.table.store.connector.TableStore.buildFileStore(TableStore.java:220)

at
org.apache.flink.table.store.connector.TableStore.access$100(TableStore.java:81)

at
org.apache.flink.table.store.connector.TableStore$SinkBuilder.build(TableStore.java:415)

at
org.apache.flink.table.store.connector.sink.TableStoreSink.lambda$getSinkRuntimeProvider$0(TableStoreSink.java:143)

at
org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecSink.applySinkProvider(CommonExecSink.java:446)

at
org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecSink.createSinkTransformation(CommonExecSink.java:192)

at
org.apache.flink.table.planner.plan.nodes.exec.batch.BatchExecSink.translateToPlanInternal(BatchExecSink.java:67)

at
org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:148)

at
org.apache.flink.table.planner.delegation.BatchPlanner.$anonfun$translateToPlan$1(BatchPlanner.scala:82)

at
scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:233)

at scala.collection.Iterator.foreach(Iterator.scala:937)

at scala.collection.Iterator.foreach$(Iterator.scala:937)

at scala.collection.AbstractIterator.foreach(Iterator.scala:1425)

at scala.collection.IterableLike.foreach(IterableLike.scala:70)

at scala.collection.IterableLike.foreach$(IterableLike.scala:69)

at scala.collection.AbstractIterable.foreach(Iterable.scala:54)

at scala.collection.TraversableLike.map(TraversableLike.scala:233)

at scala.collection.TraversableLike.map$(TraversableLike.scala:226)

at scala.collection.AbstractTraversable.map(Traversable.scala:104)

at
org.apache.flink.table.planner.delegation.BatchPlanner.translateToPlan(BatchPlanner.scala:81)

at
org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:181)

at
org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1656)

at
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:782)

at
org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$executeModifyOperations$4(LocalExecutor.java:222)

at
org.apache.flink.table.client.gateway.context.ExecutionContext.wrapClassLoader(ExecutionContext.java:88)

at
org.apache.flink.table.client.gateway.local.LocalExecutor.executeModifyOperations(LocalExecutor.java:222)

at
org.apache.flink.table.client.cli.CliClient.callInserts(CliClient.java:600)

at
org.apache.flink.table.client.cli.CliClient.callInsert(CliClient.java:589)

at
org.apache.flink.table.client.cli.CliClient.callOperation(CliClient.java:443)

at
org.apache.flink.table.client.cli.CliClient.executeOperation(CliClient.java:373)

at
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:330)

at
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:281)

at
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:229)

at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:151)

at org.apache.flink.table.client.SqlClient.start(SqlClient.java:95)

at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:187)

... 1 more

Caused by: java.lang.ClassNotFoundException:
org.apache.flink.table.planner.plan.utils.SortUtil$

at java.net.URLClassLoader.findClass(URLClassLoader.java:382)

at java.lang.ClassLoader.loadClass(ClassLoader.java:418)

at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)

at java.lang.ClassLoader.loadClass(ClassLoader.java:351)

at
org.apache.flink.core.classloading.ComponentClassLoader.loadClassFromOwnerOnly(ComponentClassLoader.java:163)

at
org.apache.flink.core.classloading.ComponentClassLoader.loadClassFromComponentFirst(ComponentClassLoader.java:157)

at
org.apache.flink.core.classloading.ComponentClassLoader.loadClass(ComponentClassLoader.java:103)

at java.lang.ClassLoader.loadClass(ClassLoader.java:351)

... 48 more

Sorry for introducing this bug. I plan to fix it very soon and we also need
an e2e test case for all supported data types.


Jingsong Li <jingsongl...@gmail.com> 于2022年4月28日周四 14:50写道:

> Hi everyone,
>
> Please review and vote on the release candidate #1 for the version 0.1.0 of
> Apache Flink Table Store, as follows:
>
> [ ] +1, Approve the release
> [ ] -1, Do not approve the release (please provide specific comments)
>
> **Release Overview**
>
> As an overview, the release consists of the following:
> a) Table Store canonical source distribution, to be deployed to the
> release repository at dist.apache.org
> b) Maven artifacts to be deployed to the Maven Central Repository
>
> **Staging Areas to Review**
>
> The staging areas containing the above mentioned artifacts are as follows,
> for your review:
> * All artifacts for a) and b) can be found in the corresponding dev
> repository at dist.apache.org [2]
> * All artifacts for c) can be found at the Apache Nexus Repository [3]
> * Pre Bundled Binaries Jar can work fine with quick start [4][5]
>
> All artifacts are signed with the key
> 2C2B6A653B07086B65E4369F7C76245E0A318150 [6]
>
> Other links for your review:
> * JIRA release notes [7]
> * source code tag "release-0.1.0-rc1" [8]
> * PR to update the website Downloads page to include Table Store
> links [9]
>
> **Vote Duration**
>
> The voting time will run for at least 72 hours.
> It is adopted by majority approval, with at least 3 PMC affirmative votes.
>
> Best,
> Jingsong Lee
>
> [1]
> https://cwiki.apache.org/confluence/display/FLINK/Verifying+a+Flink+Table+Store+Release
> [2]
> https://dist.apache.org/repos/dist/dev/flink/flink-table-store-0.1.0-rc1/
> [3]
> https://repository.apache.org/content/repositories/orgapacheflink-1501/
> [4]
> https://repository.apache.org/content/repositories/orgapacheflink-1501/org/apache/flink/flink-table-store-dist/0.1.0/flink-table-store-dist-0.1.0.jar
> [5]
> https://nightlies.apache.org/flink/flink-table-store-docs-release-0.1/docs/try-table-store/quick-start/
> [6] https://dist.apache.org/repos/dist/release/flink/KEYS
> [7]
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12351234
> [8] https://github.com/apache/flink-table-store/tree/release-0.1.0-rc1
> [9] https://github.com/apache/flink-web/pull/531
>

Reply via email to