[jira] [Created] (FLINK-22448) FlinkRexUtil create Sarg String array elemet supplement space

2021-04-25 Thread Junning Liang (Jira)
Junning Liang created FLINK-22448:
-

 Summary: FlinkRexUtil create Sarg String array elemet supplement 
space
 Key: FLINK-22448
 URL: https://issues.apache.org/jira/browse/FLINK-22448
 Project: Flink
  Issue Type: Bug
Reporter: Junning Liang
 Fix For: 1.13.0


As we know, the new version of Calcite introduces the {{SEARCH}} rex call to 
express range conditions. But when i used string array to express range in the 
StreamSQL, i found that some string in the array had problems with the 
completion length by using space.

the following query:

 
{code:java}
create view tempView as
  select * from sourceTable where action in ('systemnotifyv2', 'session_auth', 
'create_session', 'close_single_chat'){code}
after Sarg operator created, the result is :

 

 
{code:java}
create view tempView as
 select * from sourceTable where action in ('systemnotifyv2', 'session_auth  ', 
'create_session', 'close_single_chat')
{code}
I debuged to see why dose the happans. After calling  rexBuilder.makeLiteral in

FlinkRexUtil#expandSearchOperands, the string 'session_auth' became 
'session_auth  '.And i also found that the type and length of the string array 
were determined by the first string in the array.Just like my example above, 
the type of the array was Char and the length of the array was 14.the length of 
'session_auth' string was 12 so that calcite would supplement  2 space to make 
it meet the length of 14.

Now, i All I can think of is adding trim parameter to remove the space。do you 
have a better way to fix or avoid the problem happens?

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22449) Casting an invalid constant string to int throws exception from SinkNotNullEnforcer

2021-04-25 Thread Caizhi Weng (Jira)
Caizhi Weng created FLINK-22449:
---

 Summary: Casting an invalid constant string to int throws 
exception from SinkNotNullEnforcer
 Key: FLINK-22449
 URL: https://issues.apache.org/jira/browse/FLINK-22449
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Planner
Affects Versions: 1.13.0
Reporter: Caizhi Weng


Add the following test case to {{CalcITCase}} to reproduce this bug:

{code:scala}
@Test
def myTest(): Unit = {
  checkResult("SELECT CAST('haha' AS INT)", Seq(row(null)))
}
{code}

The exception stack is
{code}
Caused by: org.apache.flink.table.api.TableException: Column 'EXPR$0' is NOT 
NULL, however, a null value is being written into it. You can set job 
configuration 'table.exec.sink.not-null-enforcer'='drop' to suppress this 
exception and drop such records silently.
at 
org.apache.flink.table.runtime.operators.sink.SinkNotNullEnforcer.filter(SinkNotNullEnforcer.java:56)
at 
org.apache.flink.table.runtime.operators.sink.SinkNotNullEnforcer.filter(SinkNotNullEnforcer.java:30)
at 
org.apache.flink.streaming.api.operators.StreamFilter.processElement(StreamFilter.java:38)
at 
org.apache.flink.streaming.runtime.tasks.ChainingOutput.pushToOperator(ChainingOutput.java:112)
at 
org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:93)
at 
org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:39)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28)
at BatchExecCalc$33.processElement(Unknown Source)
at 
org.apache.flink.streaming.runtime.tasks.ChainingOutput.pushToOperator(ChainingOutput.java:112)
at 
org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:93)
at 
org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:39)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28)
at 
org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollect(StreamSourceContexts.java:317)
at 
org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collect(StreamSourceContexts.java:411)
at 
org.apache.flink.streaming.api.functions.source.InputFormatSourceFunction.run(InputFormatSourceFunction.java:92)
at 
org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110)
at 
org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:66)
at 
org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:269)
{code}

This is because the result type of CAST is inferred as NOT NULL (see 
{{SqlCastFunction#inferReturnType}} and 
{{StandardConvertletTable#convertCast}}, the nullability is the same with the 
input argument), however parsing an invalid string to int will produce null 
values.

One way I could think of is to change the result type of CAST to always 
nullable (at least for some specific types of casting, for example casting from 
string to int), but as CAST is a very low-level function this might have a big 
impact (for example, if a rule adds casting, the resulting row type might not 
be equal to the original row type due to mismatch in nullability).

So it seems that at the current stage we should set all columns in a select 
sink to be nullable. However this indicates that one cannot truly rely on the 
nullability of any result type.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22450) SQL parser fails to parse subquery containing INTERSECT in a view

2021-04-25 Thread Caizhi Weng (Jira)
Caizhi Weng created FLINK-22450:
---

 Summary: SQL parser fails to parse subquery containing INTERSECT 
in a view
 Key: FLINK-22450
 URL: https://issues.apache.org/jira/browse/FLINK-22450
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Planner
Affects Versions: 1.13.0
Reporter: Caizhi Weng


Add the following test case to {{TableEnvironmentITCase}} to reproduce this bug.

{code:scala}
@Test
def myTest(): Unit = {
  tEnv.executeSql(
"""
  |CREATE TABLE T1 (
  |  a INT,
  |  b BIGINT
  |) WITH (
  |  'connector'='values'
  |)
  |""".stripMargin)
  tEnv.executeSql(
"""
  |CREATE TABLE T2 (
  |  c INT,
  |  d BIGINT
  |) WITH (
  |  'connector'='values'
  |)
  |""".stripMargin)
  tEnv.executeSql(
"""
  |CREATE TABLE T3 (
  |  c INT,
  |  d BIGINT
  |) WITH (
  |  'connector'='values'
  |)
  |""".stripMargin)
  tEnv.executeSql("CREATE VIEW myView AS SELECT * FROM T1, (SELECT * FROM T2 
WHERE c > 0 INTERSECT SELECT * FROM T3 WHERE c > 0) WHERE a = c")
  System.out.println(tEnv.explainSql("SELECT * FROM myView"))
}
{code}

The exception stack is
{code}
org.apache.flink.table.api.SqlParserException: SQL parse failed. Encountered ", 
SELECT" at line 2, column 10.
Was expecting one of:
 
"AS" ...
"EXCEPT" ...
"EXTEND" ...
"FETCH" ...
"FOR" ...
"GROUP" ...
"HAVING" ...
"INTERSECT" ...
"LIMIT" ...
"MATCH_RECOGNIZE" ...
"OFFSET" ...
"ORDER" ...
"PIVOT" ...
"MINUS" ...
"TABLESAMPLE" ...
"UNION" ...
"WHERE" ...
"WINDOW" ...
"(" ...
 ...
 ...
 ...
 ...
 ...
 ...
"/*+" ...
"NATURAL" ...
"JOIN" ...
"INNER" ...
"LEFT" ...
"RIGHT" ...
"FULL" ...
"CROSS" ...
","  ...
","  ...
","  ...
","  ...
","  ...
","  ...
"," "LATERAL" ...
"," "(" ...
"," "UNNEST" ...
"," "TABLE" ...
"OUTER" ...
"." ...


at 
org.apache.flink.table.planner.parse.CalciteParser.parse(CalciteParser.java:56)
at 
org.apache.flink.table.planner.utils.Expander.expanded(Expander.java:83)
at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.convertViewQuery(SqlToOperationConverter.java:849)
at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.convertCreateView(SqlToOperationConverter.java:819)
at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:248)
at 
org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:99)
at 
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeSql(TableEnvironmentImpl.java:723)
at 
org.apache.flink.table.api.TableEnvironmentITCase.myTest(TableEnvironmentITCase.scala:116)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:239)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
at 
org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.

[jira] [Created] (FLINK-22451) Support (*) as parameter of table UserDefinedFunction

2021-04-25 Thread Yi Tang (Jira)
Yi Tang created FLINK-22451:
---

 Summary: Support (*) as parameter of table UserDefinedFunction 
 Key: FLINK-22451
 URL: https://issues.apache.org/jira/browse/FLINK-22451
 Project: Flink
  Issue Type: Improvement
Reporter: Yi Tang


For now, one can use star (*) to act as a wild card, selecting all of the 
columns in the table.
{code}
Table result = orders.select($("*"));
{code}

When one use a star (*) as parameter of an UDF, it will fail 
{{ReferenceResolverRule}} in on 
{code}
"Cannot resolve field [*], input field list:[...]."
{code}

The cause is that, the parameter of an UDF is not expanded in 
{{StarReferenceFlatteningRule}}

I think we can support to expand the star parameter to the real fields list if 
it is the only parameter of the UDF.

then the parameters can be received by 
{code}
eval(@DataTypeHint(inputGroup = InputGroup.ANY) Object... row)
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22452) Support specifying custom transactional.id prefix in FlinkKafkaProducer

2021-04-25 Thread Wenhao Ji (Jira)
Wenhao Ji created FLINK-22452:
-

 Summary: Support specifying custom transactional.id prefix in 
FlinkKafkaProducer
 Key: FLINK-22452
 URL: https://issues.apache.org/jira/browse/FLINK-22452
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Kafka
Affects Versions: 1.12.2
Reporter: Wenhao Ji


Currently, the "transactional.id"s of the Kafka producers in FlinkKafkaProducer 
are generated based on the task name. This mechanism has some limitations:
 * It will exceed Kafka's limitation if the task name is too long. (resolved in 
FLINK-17691)
 * They will very likely clash each other if the job topologies are similar. 
(discussed in FLINK-11654)
 * Only certain "transactional.id" may be authorized by [Prefixed 
ACLs|https://docs.confluent.io/platform/current/kafka/authorization.html#prefixed-acls]
 on the target Kafka cluster.

Besides, the spring community has introduced the 
[setTransactionIdPrefix|https://docs.spring.io/spring-kafka/api/org/springframework/kafka/core/DefaultKafkaProducerFactory.html#setTransactionIdPrefix(java.lang.String)]
 method to their Kafka client.

Therefore, I think it will be necessary to have this feature in the Flink Kafka 
connector. 

 

As discussed in FLINK-11654, the possible solution will be,
 * either introduce an additional method called 
setTransactionalIdPrefix(String) in the FlinkKafkaProducer,
 * or use the existing "transactional.id" properties as the prefix.

 And the behavior of the "transactional.id" generation will be
 * keep the behavior as it was if absent,
 * use the one if present as the prefix for the TransactionalIdsGenerator.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22453) Can not stop job when not ust "-m" option

2021-04-25 Thread Liu (Jira)
Liu created FLINK-22453:
---

 Summary: Can not stop job when not ust "-m" option
 Key: FLINK-22453
 URL: https://issues.apache.org/jira/browse/FLINK-22453
 Project: Flink
  Issue Type: Bug
  Components: API / Core
Affects Versions: 1.12.2
Reporter: Liu


flink version: 1.12.2

yarn version : 3.1.1 (hdp 3.1.5)
h3. Starting a Flink Session on YARN

when i use ' flink stop xxx',   comond line output:

 
{code:java}
//代码占位符
Setting HBASE_CONF_DIR=/etc/hbase/conf because no HBASE_CONF_DIR was set.SLF4J: 
Class path contains multiple SLF4J bindings.SLF4J: Found binding in 
[jar:file:/data/flink-1.12.2/lib/log4j-slf4j-impl-2.12.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]SLF4J:
 Found binding in 
[jar:file:/usr/hdp/3.1.5.0-152/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]SLF4J:
 See http://www.slf4j.org/codes.html#multiple_bindings for an 
explanation.SLF4J: Actual binding is of type 
[org.apache.logging.slf4j.Log4jLoggerFactory]2021-04-25 16:10:43,369 INFO  
org.apache.flink.yarn.cli.FlinkYarnSessionCli                [] - Found Yarn 
properties file under /tmp/.yarn-properties-flink.2021-04-25 16:10:43,369 INFO  
org.apache.flink.yarn.cli.FlinkYarnSessionCli                [] - Found Yarn 
properties file under /tmp/.yarn-properties-flink.Suspending job 
"a81c0fe295871ef278a119cd44206216" with a savepoint.2021-04-25 16:10:45,126 
INFO  org.apache.hadoop.yarn.client.AHSProxy                       [] - 
Connecting to Application History server at 
adt-bd-c1-nn03.internal/172.20.33.149:102002021-04-25 16:10:45,174 INFO  
org.apache.flink.yarn.YarnClusterDescriptor                  [] - No path for 
the flink jar passed. Using the location of class 
org.apache.flink.yarn.YarnClusterDescriptor to locate the jar2021-04-25 
16:10:45,520 INFO  org.apache.flink.yarn.YarnClusterDescriptor                  
[] - Found Web Interface adt-bd-c1-flink06.internal:43379 of application 
'application_1618023905026_0005'.
 The program 
finished with the following exception:
org.apache.flink.util.FlinkException: Could not stop with a savepoint job 
"a81c0fe295871ef278a119cd44206216". at 
org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:581) at 
org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:1002) 
at org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:569) at 
org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1069) at 
org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132) 
at java.security.AccessController.doPrivileged(Native Method) at 
javax.security.auth.Subject.doAs(Subject.java:422) at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
 at 
org.apache.flink.runtime.security.contexts.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
 at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132)Caused 
by: java.util.concurrent.ExecutionException: 
java.util.concurrent.TimeoutException at 
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1915) at 
org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:579) ... 
9 moreCaused by: java.util.concurrent.TimeoutException at 
org.apache.flink.runtime.concurrent.FutureUtils$Timeout.run(FutureUtils.java:1220)
 at 
org.apache.flink.runtime.concurrent.DirectExecutorService.execute(DirectExecutorService.java:217)
 at 
org.apache.flink.runtime.concurrent.FutureUtils.lambda$orTimeout$15(FutureUtils.java:582)
 at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at 
java.util.concurrent.FutureTask.run(FutureTask.java:266) at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
at java.lang.Thread.run(Thread.java:748)
{code}
 

 flink can't stop the job,  

but when i user  'flink stop -m jobmanager.server.host:port    x'  ,  it 
work well.

'-m' is an option args,  The old version does not have this problem

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22454) Ignore CAST condition in temporal join

2021-04-25 Thread Shengkai Fang (Jira)
Shengkai Fang created FLINK-22454:
-

 Summary: Ignore CAST condition in temporal join 
 Key: FLINK-22454
 URL: https://issues.apache.org/jira/browse/FLINK-22454
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / API
Affects Versions: 1.13.0
Reporter: Shengkai Fang


Please add test in {{LookupJoinTest}}
{code:java}
 def before(): Unit ={
util.addDataStream[(Int, String, Long)](
  "MyTable", 'a, 'b, 'c, 'proctime.proctime, 'rowtime.rowtime)

if (legacyTableSource) {
  TestTemporalTable.createTemporaryTable(util.tableEnv, "LookupTable")
} else {
  util.addTable(
"""
  |CREATE TABLE LookupTable (
  |  `id` DECIMAL(38, 10),
  |  `to_qty` DECIMAL(38, 10),
  |  `name` STRING,
  |  `age` INT,
  |  `id_int` as CAST(`id` AS INT)
  |) WITH (
  |  'connector' = 'values'
  |)
  |""".stripMargin)
  }
{code}
{code:java}
@Test
  def test(): Unit = {
val sql =
"""
|SELECT MyTable.b, LookupTable.`to_qty`
|FROM MyTable
|LEFT JOIN LookupTable FOR SYSTEM_TIME AS OF MyTable.`proctime`
|ON MyTable.a = CAST(LookupTable.`id` as INT)
|""".stripMargin

util.tableEnv.sqlQuery(sql).explain()
  }

{code}


The exception stack is 

{code}
org.apache.flink.table.api.TableException: Temporal table join requires 
equivalent condition of the same type, but the condition is 
a[INT]=id[DECIMAL(38, 10)]

at 
org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecLookupJoin.validateLookupKeyType(CommonExecLookupJoin.java:303)
at 
org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecLookupJoin.translateToPlanInternal(CommonExecLookupJoin.java:222)
at 
org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:134)
at 
org.apache.flink.table.planner.plan.nodes.exec.ExecEdge.translateToPlan(ExecEdge.java:247)
at 
org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecCalc.translateToPlanInternal(CommonExecCalc.java:88)
at 
org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:134)
at 
org.apache.flink.table.planner.delegation.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:70)
at 
org.apache.flink.table.planner.delegation.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:69)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.Iterator$class.foreach(Iterator.scala:891)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
at scala.collection.AbstractTraversable.map(Traversable.scala:104)
at 
org.apache.flink.table.planner.delegation.StreamPlanner.translateToPlan(StreamPlanner.scala:69)
at 
org.apache.flink.table.planner.delegation.StreamPlanner.explain(StreamPlanner.scala:104)
at 
org.apache.flink.table.planner.delegation.StreamPlanner.explain(StreamPlanner.scala:46)
at 
org.apache.flink.table.api.internal.TableEnvironmentImpl.explainInternal(TableEnvironmentImpl.java:691)
at 
org.apache.flink.table.api.internal.TableImpl.explain(TableImpl.java:582)
at 
org.apache.flink.table.planner.plan.stream.sql.join.LookupJoinTest.test(LookupJoinTest.scala:197)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:239)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)

[jira] [Created] (FLINK-22455) FlinkRelBuilder#windowAggregate will throw ClassCastException when function reuse

2021-04-25 Thread tartarus (Jira)
tartarus created FLINK-22455:


 Summary: FlinkRelBuilder#windowAggregate will throw 
ClassCastException when function reuse
 Key: FLINK-22455
 URL: https://issues.apache.org/jira/browse/FLINK-22455
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Planner
Reporter: tartarus
 Attachments: FlinkRelBuilderTest.scala

If the input parameter aggCalls of FlinkRelBuilder#windowAggregate contains the 
same aggregate function. Then it will throw ClassCastException, because of the 
optimization of aggregate function reuse. We did not judge the return value 
type, but direct type conversion;
{code:java}
val aggregate = super.transform(
  new UnaryOperator[RelBuilder.Config] {
override def apply(t: RelBuilder.Config)
  : RelBuilder.Config = t.withPruneInputOfAggregate(false)
  })
  .push(build())
  .aggregate(groupKey, aggCalls)
  .build()
  .asInstanceOf[LogicalAggregate]
{code}
I wrote a test that triggered this problem.
You can use the attached code to reproduce this problem.




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [VOTE] Release 1.13.0, release candidate #2

2021-04-25 Thread Jark Wu
Hi Rui,

Documentation is not a part of the release package.
The changes to the documentation will take effect in 24 hours
 and can be accessed on the web [1] then.
Docs can even be updated after 1.13 is released.
So we don't need to cancel the RC for this.

Best,
Jark

[1]:
https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/

On Sun, 25 Apr 2021 at 13:47, Rui Li  wrote:

> Hi Dawid,
>
> I have just merged the doc change for hive dialect [1], which talks about
> how to use hive dialect to run hive queries. Do you think we can have
> another RC to include it? Sorry for the trouble.
>
> [1] https://issues.apache.org/jira/browse/FLINK-22119
>
> On Sat, Apr 24, 2021 at 1:36 AM Dawid Wysakowicz 
> wrote:
>
> > Hi everyone,
> > Please review and vote on the release candidate #2 for the version
> 1.13.0,
> > as follows:
> > [ ] +1, Approve the release
> > [ ] -1, Do not approve the release (please provide specific comments)
> >
> >
> > The complete staging area is available for your review, which includes:
> > * JIRA release notes [1],
> > * the official Apache source release and binary convenience releases to
> be
> > deployed to dist.apache.org [2], which are signed with the key with
> > fingerprint 31D2DD10BFC15A2D [3],
> > * all artifacts to be deployed to the Maven Central Repository [4],
> > * source code tag "release-1.13.0-rc2" [5],
> > * website pull request listing the new release and adding announcement
> > blog post [6].
> >
> > The vote will be open for at least 72 hours. It is adopted by majority
> > approval, with at least 3 PMC affirmative votes.
> >
> > Thanks,
> > Dawid Wysakowicz
> >
> > [1]
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12349287
> > [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.13.0-rc2/
> > [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> > [4]
> > https://repository.apache.org/content/repositories/orgapacheflink-1420/
> > [5] https://github.com/apache/flink/tree/release-1.13.0-rc2
> > [6] https://github.com/apache/flink-web/pull/436
> >
>
>
> --
> Best regards!
> Rui Li
>


Re: [DISCUSS]FLIP-150: Introduce Hybrid Source

2021-04-25 Thread Thomas Weise
Hi Becket,

I agree and am not planning to hard wire a specific combination of
sources (like S3 + Kafka). That also wouldn't help for the use case I
want to address, because there are customized connectors that we need
to be able to plug in.

Rather, the suggested simplification would be for the flexibility of
switching mechanism.

The prototype already supports fixed start positions and checkpoint
conversion for any combination of sources; no need to undo that.

But for testing/example purposes, we will need to settle on a specific
combination.

Thomas

On Sat, Apr 24, 2021 at 8:20 PM Becket Qin  wrote:
>
> Sorry for the late reply. Starting from a specific connector sounds
> reasonable to me.
>
> That said, I would suggest to keep the possibility of future generalization
> as much as possible. We have already seen some variation of source
> combinations from different users, HDFS + Kafka, S3 + Kafka, S3 + SQL
> Binlog, etc. So it would be good if we can reuse some base abstraction in
> the future instead of having to write each combination from scratch.
>
> Thanks,
>
> Jiangjie (Becket) Qin
>
> On Sat, Apr 17, 2021 at 7:34 PM Stephan Ewen  wrote:
>
> > Thanks, Thomas!
> >
> > @Becket and @Nicholas - would you be ok with that approach?
> >
> >
> > On Thu, Apr 15, 2021 at 6:30 PM Thomas Weise  wrote:
> >
> > > Hi Stephan,
> > >
> > > Thanks for the feedback!
> > >
> > > I agree with the approach of starting with a simple implementation
> > > that can address a well understood, significant portion of use cases.
> > >
> > > I'm planning to continue work on the prototype that I had shared.
> > > There is production level usage waiting for it fairly soon. I expect
> > > to open a PR in the coming weeks.
> > >
> > > Thomas
> > >
> > >
> > >
> > >
> > >
> > > On Tue, Apr 13, 2021 at 12:15 PM Stephan Ewen  wrote:
> > > >
> > > > Thanks all for this discussion. Looks like there are lots of ideas and
> > > > folks that are eager to do things, so let's see how we can get this
> > > moving.
> > > >
> > > > My take on this is the following:
> > > >
> > > > There will probably not be one Hybrid source, but possibly multiple
> > ones,
> > > > because of different strategies/requirements.
> > > > - One may be very simple, with switching points known up-front.
> > Would
> > > > be good to have this in a very simple implementation.
> > > > - There may be one where the switch is dynamic and the readers need
> > > to
> > > > report back where they left off.
> > > > - There may be one that switches back and forth multiple times
> > > during a
> > > > job, for example Kakfa going to DFS whenever it falls behind retention,
> > > in
> > > > order to catch up again.
> > > >
> > > > This also seems hard to "design on paper"; I expect there are nuances
> > in
> > > a
> > > > production setup that affect some details of the design. So I'd feel
> > most
> > > > comfortable in adding a variant of the hybrid source to Flink that has
> > > been
> > > > used already in a real use case (not necessarily in production, but
> > maybe
> > > > in a testing/staging environment, so it seems to meet all
> > requirements).
> > > >
> > > >
> > > > What do you think about the following approach?
> > > >   - If there is a tested PoC, let's try to get it contributed to Flink
> > > > without trying to make it much more general.
> > > >   - When we see similar but a bit different requirements for another
> > > hybrid
> > > > source, then let's try to evolve the contributed one.
> > > >   - If we see new requirements that are so different that they don't
> > fit
> > > > well with the existing hybrid source, then let us look at building a
> > > second
> > > > hybrid source for those requirements.
> > > >
> > > > We need to make connector contributions in general more easy, and I
> > think
> > > > it is not a bad thing to end up with different approaches and see how
> > > these
> > > > play out against each other when being used by users. For example
> > > switching
> > > > with known boundaries, dynamic switching, back-and-forth-switching,
> > etc.
> > > > (I know some committers are planning to do some work on making
> > > > connector contributions easier, with standardized testing frameworks,
> > > > decoupled CI, etc.)
> > > >
> > > > Best,
> > > > Stephan
> > > >
> > > >
> > > > On Thu, Mar 25, 2021 at 4:41 AM Thomas Weise  wrote:
> > > >
> > > > > Hi,
> > > > >
> > > > > As mentioned in my previous email, I had been working on a prototype
> > > for
> > > > > the hybrid source.
> > > > >
> > > > > You can find it at https://github.com/tweise/flink/pull/1
> > > > >
> > > > > It contains:
> > > > > * Switching with configurable chain of sources
> > > > > * Fixed or dynamic start positions
> > > > > * Test with MockSource and FileSource
> > > > >
> > > > > The purpose of the above PR is to gather feedback and help drive
> > > consensus
> > > > > on the FLIP.
> > > > >
> > > > > * How to support a dynamic start position within the source chain

Re: [DISCUSS]FLIP-150: Introduce Hybrid Source

2021-04-25 Thread Becket Qin
Thanks for the clarification, Thomas. Yes, that makes sense to me.

Cheers,

Jiangjie (Becket) Qin

On Mon, Apr 26, 2021 at 1:03 AM Thomas Weise  wrote:

> Hi Becket,
>
> I agree and am not planning to hard wire a specific combination of
> sources (like S3 + Kafka). That also wouldn't help for the use case I
> want to address, because there are customized connectors that we need
> to be able to plug in.
>
> Rather, the suggested simplification would be for the flexibility of
> switching mechanism.
>
> The prototype already supports fixed start positions and checkpoint
> conversion for any combination of sources; no need to undo that.
>
> But for testing/example purposes, we will need to settle on a specific
> combination.
>
> Thomas
>
> On Sat, Apr 24, 2021 at 8:20 PM Becket Qin  wrote:
> >
> > Sorry for the late reply. Starting from a specific connector sounds
> > reasonable to me.
> >
> > That said, I would suggest to keep the possibility of future
> generalization
> > as much as possible. We have already seen some variation of source
> > combinations from different users, HDFS + Kafka, S3 + Kafka, S3 + SQL
> > Binlog, etc. So it would be good if we can reuse some base abstraction in
> > the future instead of having to write each combination from scratch.
> >
> > Thanks,
> >
> > Jiangjie (Becket) Qin
> >
> > On Sat, Apr 17, 2021 at 7:34 PM Stephan Ewen  wrote:
> >
> > > Thanks, Thomas!
> > >
> > > @Becket and @Nicholas - would you be ok with that approach?
> > >
> > >
> > > On Thu, Apr 15, 2021 at 6:30 PM Thomas Weise  wrote:
> > >
> > > > Hi Stephan,
> > > >
> > > > Thanks for the feedback!
> > > >
> > > > I agree with the approach of starting with a simple implementation
> > > > that can address a well understood, significant portion of use cases.
> > > >
> > > > I'm planning to continue work on the prototype that I had shared.
> > > > There is production level usage waiting for it fairly soon. I expect
> > > > to open a PR in the coming weeks.
> > > >
> > > > Thomas
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > On Tue, Apr 13, 2021 at 12:15 PM Stephan Ewen 
> wrote:
> > > > >
> > > > > Thanks all for this discussion. Looks like there are lots of ideas
> and
> > > > > folks that are eager to do things, so let's see how we can get this
> > > > moving.
> > > > >
> > > > > My take on this is the following:
> > > > >
> > > > > There will probably not be one Hybrid source, but possibly multiple
> > > ones,
> > > > > because of different strategies/requirements.
> > > > > - One may be very simple, with switching points known up-front.
> > > Would
> > > > > be good to have this in a very simple implementation.
> > > > > - There may be one where the switch is dynamic and the readers
> need
> > > > to
> > > > > report back where they left off.
> > > > > - There may be one that switches back and forth multiple times
> > > > during a
> > > > > job, for example Kakfa going to DFS whenever it falls behind
> retention,
> > > > in
> > > > > order to catch up again.
> > > > >
> > > > > This also seems hard to "design on paper"; I expect there are
> nuances
> > > in
> > > > a
> > > > > production setup that affect some details of the design. So I'd
> feel
> > > most
> > > > > comfortable in adding a variant of the hybrid source to Flink that
> has
> > > > been
> > > > > used already in a real use case (not necessarily in production, but
> > > maybe
> > > > > in a testing/staging environment, so it seems to meet all
> > > requirements).
> > > > >
> > > > >
> > > > > What do you think about the following approach?
> > > > >   - If there is a tested PoC, let's try to get it contributed to
> Flink
> > > > > without trying to make it much more general.
> > > > >   - When we see similar but a bit different requirements for
> another
> > > > hybrid
> > > > > source, then let's try to evolve the contributed one.
> > > > >   - If we see new requirements that are so different that they
> don't
> > > fit
> > > > > well with the existing hybrid source, then let us look at building
> a
> > > > second
> > > > > hybrid source for those requirements.
> > > > >
> > > > > We need to make connector contributions in general more easy, and I
> > > think
> > > > > it is not a bad thing to end up with different approaches and see
> how
> > > > these
> > > > > play out against each other when being used by users. For example
> > > > switching
> > > > > with known boundaries, dynamic switching, back-and-forth-switching,
> > > etc.
> > > > > (I know some committers are planning to do some work on making
> > > > > connector contributions easier, with standardized testing
> frameworks,
> > > > > decoupled CI, etc.)
> > > > >
> > > > > Best,
> > > > > Stephan
> > > > >
> > > > >
> > > > > On Thu, Mar 25, 2021 at 4:41 AM Thomas Weise 
> wrote:
> > > > >
> > > > > > Hi,
> > > > > >
> > > > > > As mentioned in my previous email, I had been working on a
> prototype
> > > > for
> > > > > > the hybrid source.
> > > > > >
> > > > > > You ca

[jira] [Created] (FLINK-22456) Support InitializeOnMaster and FinalizeOnMaster to be used in InputFormat

2021-04-25 Thread Li (Jira)
Li created FLINK-22456:
--

 Summary: Support InitializeOnMaster and FinalizeOnMaster to be 
used in InputFormat
 Key: FLINK-22456
 URL: https://issues.apache.org/jira/browse/FLINK-22456
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Task
Reporter: Li


        In _InputOutputFormatVertex_, _initializeGlobal_ and _finalizeGlobal_ 
are only called when the Format is _OutputFormat_, however _InputFormat_ is not 
be called.
        In FLINK-1722, its say _HadoopOutputFormats_ ues it to do something 
before and after the task. And they only support _initializeGlobal_ and 
_finalizeGlobal_ in _OutputFormat_.
        I don't know why _InputFormat_ doesn't support, anyone can tell me why?
        But I think _InitializeOnMaster_ and _FinalizeOnMaster_ should also be 
supported in _InputFormat_.
        For example, an offline task in _JdbcInputFormat_, user can use 
_initializeGlobal_ to query the total counts of this task, and then user can 
create InputSplits by total counts. While task running, user can add progress 
indicators metric by calculating the total number of records divided by the 
current number of reads, and even the remaining time of the task can be 
estimated. It is very helpful for users to view task progress and remaining 
time through external systems.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22457) KafkaSourceLegacyITCase.testMultipleSourcesOnePartition fails because of timeout

2021-04-25 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22457:
-

 Summary: KafkaSourceLegacyITCase.testMultipleSourcesOnePartition 
fails because of timeout
 Key: FLINK-22457
 URL: https://issues.apache.org/jira/browse/FLINK-22457
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kafka
Affects Versions: 1.14.0
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17140&view=logs&j=1fc6e7bf-633c-5081-c32a-9dea24b05730&t=80a658d1-f7f6-5d93-2758-53ac19fd5b19&l=7045


{code:java}
Apr 24 23:47:33 [ERROR] Tests run: 21, Failures: 0, Errors: 1, Skipped: 0, Time 
elapsed: 174.335 s <<< FAILURE! - in 
org.apache.flink.connector.kafka.source.KafkaSourceLegacyITCase
Apr 24 23:47:33 [ERROR] 
testMultipleSourcesOnePartition(org.apache.flink.connector.kafka.source.KafkaSourceLegacyITCase)
  Time elapsed: 60.019 s  <<< ERROR!
Apr 24 23:47:33 org.junit.runners.model.TestTimedOutException: test timed out 
after 6 milliseconds
Apr 24 23:47:33 at sun.misc.Unsafe.park(Native Method)
Apr 24 23:47:33 at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
Apr 24 23:47:33 at 
java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
Apr 24 23:47:33 at 
java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
Apr 24 23:47:33 at 
java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
Apr 24 23:47:33 at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
Apr 24 23:47:33 at 
org.apache.flink.test.util.TestUtils.tryExecute(TestUtils.java:49)
Apr 24 23:47:33 at 
org.apache.flink.streaming.connectors.kafka.KafkaConsumerTestBase.runMultipleSourcesOnePartitionExactlyOnceTest(KafkaConsumerTestBase.java:1112)
Apr 24 23:47:33 at 
org.apache.flink.connector.kafka.source.KafkaSourceLegacyITCase.testMultipleSourcesOnePartition(KafkaSourceLegacyITCase.java:87)
Apr 24 23:47:33 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
Apr 24 23:47:33 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
Apr 24 23:47:33 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
Apr 24 23:47:33 at java.lang.reflect.Method.invoke(Method.java:498)
Apr 24 23:47:33 at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
Apr 24 23:47:33 at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
Apr 24 23:47:33 at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
Apr 24 23:47:33 at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
Apr 24 23:47:33 at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
Apr 24 23:47:33 at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
Apr 24 23:47:33 at 
java.util.concurrent.FutureTask.run(FutureTask.java:266)
Apr 24 23:47:33 at java.lang.Thread.run(Thread.java:748)

{code}




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22458) Failed when operating empty catalog table

2021-04-25 Thread Spongebob (Jira)
Spongebob created FLINK-22458:
-

 Summary: Failed when operating empty catalog table
 Key: FLINK-22458
 URL: https://issues.apache.org/jira/browse/FLINK-22458
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / API
Affects Versions: 1.12.2
 Environment: Flink: 1.12.2
Reporter: Spongebob


The pipline might like: HiveTable -> FlinkCatalogTable(might be empty) -> 
HiveTable

It runs normally when the FlinkCatalogTable is not empty, But When 
FlinkCatalogTable is empty, Jobmanager throws this exception:
{code:java}
java.lang.Exception: Failed to finalize execution on master at 
org.apache.flink.runtime.executiongraph.ExecutionGraph.vertexFinished(ExecutionGraph.java:1373)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.executiongraph.ExecutionVertex.executionFinished(ExecutionVertex.java:877)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.executiongraph.Execution.markFinished(Execution.java:1241)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.executiongraph.ExecutionGraph.updateStateInternal(ExecutionGraph.java:1610)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.executiongraph.ExecutionGraph.updateState(ExecutionGraph.java:1584)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:663)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:89)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:447)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_251] at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
~[?:1.8.0_251] at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_251] at java.lang.reflect.Method.invoke(Method.java:498) 
~[?:1.8.0_251] at 
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:305)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:212)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:77)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:158)
 ~[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
scala.PartialFunction.applyOrElse(PartialFunction.scala:123) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
scala.PartialFunction.applyOrElse$(PartialFunction.scala:122) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.actor.Actor.aroundReceive(Actor.scala:517) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.actor.Actor.aroundReceive$(Actor.scala:515) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.actor.ActorCell.invoke(ActorCell.scala:561) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.dispatch.Mailbox.run(Mailbox.scala:225) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.dispatch.Mailbox.exec(Mailbox.scala:235) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) 
[flownData-1.0-jar-with-dependencies.jar:?] at 
akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) 
[flownData-1.0-jar-with-dependencies.jar:?] Caused by: 
o

[jira] [Created] (FLINK-22459) FlinkKafkaProducerITCase testScaleUpAfterScalingDown failed because of Timeout expired after 60000milliseconds while awaiting EndTxn(COMMIT)

2021-04-25 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22459:
-

 Summary: FlinkKafkaProducerITCase testScaleUpAfterScalingDown 
failed because of Timeout expired after 6milliseconds while awaiting 
EndTxn(COMMIT)
 Key: FLINK-22459
 URL: https://issues.apache.org/jira/browse/FLINK-22459
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kafka
Affects Versions: 1.12.2
Reporter: Guowei Ma



https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17180&view=logs&j=1fc6e7bf-633c-5081-c32a-9dea24b05730&t=80a658d1-f7f6-5d93-2758-53ac19fd5b19&l=6694
{code:java}

[ERROR] Tests run: 12, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 
261.699 s <<< FAILURE! - in 
org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase
[ERROR] 
testScaleUpAfterScalingDown(org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase)
  Time elapsed: 67.768 s  <<< ERROR!
org.apache.kafka.common.errors.TimeoutException: Timeout expired after 
6milliseconds while awaiting EndTxn(COMMIT)

[INFO] 
[INFO] Results:
[INFO] 
[ERROR] Errors: 
[ERROR]   FlinkKafkaProducerITCase.testScaleUpAfterScalingDown » Timeout 
Timeout expired...
[INFO] 
[ERROR] Tests run: 136, Failures: 0, Errors: 1, Skipped: 0

{code}




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22460) Conversion to relational algebra failed caused by ''

2021-04-25 Thread Haiwei Zhou (Jira)
Haiwei Zhou created FLINK-22460:
---

 Summary: Conversion to relational algebra failed caused by ''
 Key: FLINK-22460
 URL: https://issues.apache.org/jira/browse/FLINK-22460
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / API
Affects Versions: 1.12.1
Reporter: Haiwei Zhou


Flink complains that an insert sql doesn't match the table schema. The 
validated type is missing a "NOT NULL" modifier.

 

 
{code:java}
py4j.protocol.Py4JJavaError: An error occurred while calling o18.executeSql.
: java.lang.AssertionError: Conversion to relational algebra failed to preserve 
datatypes:
validated type:
RecordType(VARCHAR(2147483647) CHARACTER SET "UTF-16LE" request, CHAR(7) 
CHARACTER SET "UTF-16LE" NOT NULL EXPR$1, BIGINT number, TIMESTAMP(3) 
start_time, TIMESTAMP(3) end_time) NOT NULL
converted type:
RecordType(VARCHAR(2147483647) CHARACTER SET "UTF-16LE" request, CHAR(7) 
CHARACTER SET "UTF-16LE" NOT NULL EXPR$1, BIGINT NOT NULL number, TIMESTAMP(3) 
start_time, TIMESTAMP(3) end_time) NOT
 NULL{code}
 

 
{code:java}
table_env.execute_sql('''
CREATE TABLE preload_stats (
 lineitems STRING,
 itype STRING,
 number BIGINT NOT NULL,
 start_time TIMESTAMP(3),
 end_time TIMESTAMP(3)
)'''
 
table_env.execute_sql(
 "SELECT request, 'request', number, start_time, end_time "
 "FROM result_1 ").execute_insert('preload_stats')
{code}
 

 

 

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22461) Resuming Savepoint (file, async, scale down) end-to-end test failed because of binding port fail

2021-04-25 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22461:
-

 Summary: Resuming Savepoint (file, async, scale down) end-to-end 
test failed because of binding port fail
 Key: FLINK-22461
 URL: https://issues.apache.org/jira/browse/FLINK-22461
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Checkpointing, Tests
Affects Versions: 1.11.3
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17179&view=logs&j=c88eea3b-64a0-564d-0031-9fdcd7b8abee&t=2b7514ee-e706-5046-657b-3430666e7bd9&l=1495


{code:java}
2021-04-25 20:27:11,375 INFO  org.apache.flink.core.fs.FileSystem   
   [] - Hadoop is not in the classpath/dependencies. The extended set 
of supported File Systems via Hadoop is not available.
2021-04-25 20:27:11,420 INFO  
org.apache.flink.runtime.entrypoint.ClusterEntrypoint[] - Install 
security context.
2021-04-25 20:27:11,441 INFO  
org.apache.flink.runtime.security.modules.HadoopModuleFactory [] - Cannot 
create Hadoop Security Module because Hadoop cannot be found in the Classpath.
2021-04-25 20:27:11,444 INFO  
org.apache.flink.runtime.security.modules.JaasModule [] - Jaas file 
will be created as /tmp/jaas-1367939375368529920.conf.
2021-04-25 20:27:11,450 INFO  
org.apache.flink.runtime.security.contexts.HadoopSecurityContextFactory [] - 
Cannot install HadoopSecurityContext because Hadoop cannot be found in the 
Classpath.
2021-04-25 20:27:11,454 INFO  
org.apache.flink.runtime.entrypoint.ClusterEntrypoint[] - Initializing 
cluster services.
2021-04-25 20:27:11,483 INFO  
org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils[] - Trying to 
start actor system, external address localhost:6123, bind address 0.0.0.0:6123.
2021-04-25 20:27:12,014 INFO  akka.event.slf4j.Slf4jLogger  
   [] - Slf4jLogger started
2021-04-25 20:27:12,040 INFO  akka.remote.Remoting  
   [] - Starting remoting
2021-04-25 20:27:12,127 ERROR akka.remote.transport.netty.NettyTransport
   [] - failed to bind to /0.0.0.0:6123, shutting down Netty transport
2021-04-25 20:27:12,138 INFO  
org.apache.flink.runtime.entrypoint.ClusterEntrypoint[] - Shutting 
StandaloneSessionClusterEntrypoint down with application status FAILED. 
Diagnostics java.net.BindException: Could not start actor system on any port in 
port range 6123
at 
org.apache.flink.runtime.clusterframework.BootstrapTools.startRemoteActorSystem(BootstrapTools.java:173)
at 
org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils$AkkaRpcServiceBuilder.createAndStart(AkkaRpcServiceUtils.java:363)
at 
org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils.createRemoteRpcService(AkkaRpcServiceUtils.java:91)
at 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint.initializeServices(ClusterEntrypoint.java:278)
at 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runCluster(ClusterEntrypoint.java:223)
at 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint.lambda$startCluster$1(ClusterEntrypoint.java:177)
at 
org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
at 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:174)
at 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runClusterEntrypoint(ClusterEntrypoint.java:577)
at 
org.apache.flink.runtime.entrypoint.StandaloneSessionClusterEntrypoint.main(StandaloneSessionClusterEntrypoint.java:67)
.
2021-04-25 20:27:12,139 INFO  
akka.remote.RemoteActorRefProvider$RemotingTerminator[] - Shutting down 
remote daemon.
2021-04-25 20:27:12,141 INFO  
akka.remote.RemoteActorRefProvider$RemotingTerminator[] - Remote daemon 
shut down; proceeding with flushing remote transports.
2021-04-25 20:27:12,143 ERROR akka.remote.Remoting  
   [] - Remoting system has been terminated abrubtly. Attempting to 
shut down transports
2021-04-25 20:27:12,143 INFO  
akka.remote.RemoteActorRefProvider$RemotingTerminator[] - Remoting shut 
down.
2021-04-25 20:27:12,187 ERROR 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint[] - Could not 
start cluster entrypoint StandaloneSessionClusterEntrypoint.
org.apache.flink.runtime.entrypoint.ClusterEntrypointException: Failed to 
initialize the cluster entrypoint StandaloneSessionClusterEntrypoint.
at 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:200)
 ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
at 
org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runClusterEntrypoint(ClusterEntrypoint.java:577)
 [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
at 
org.apache.flink.runtime.entrypoint.StandaloneSessionClusterEntrypo

Re: [VOTE] Release 1.12.3, release candidate #1

2021-04-25 Thread Dian Fu
+1 (binding)

- Verified the checksum and signature
- Reviewed the website PR and have left a few comments
- Installed PyFlink with the source package on Mac OS and everything works well
- Ran a PyFlink example which uses both Python Table API and Python DataStream 
API

Regards,
Dian

> 2021年4月24日 上午5:52,Arvid Heise  写道:
> 
> Hi everyone,
> Please review and vote on the release candidate #1 for the version 1.12.3,
> as follows:
> [ ] +1, Approve the release
> [ ] -1, Do not approve the release (please provide specific comments)
> 
> 
> The complete staging area is available for your review, which includes:
> * JIRA release notes [1],
> * the official Apache source release and binary convenience releases to be
> deployed to dist.apache.org [2], which are signed with the key with
> fingerprint 476DAA5D1FF08189 [3],
> * all artifacts to be deployed to the Maven Central Repository [4],
> * source code tag "release-1.2.3-rc3" [5],
> * website pull request listing the new release and adding announcement blog
> post [6].
> 
> The vote will be open for at least 72 hours. It is adopted by majority
> approval, with at least 3 PMC affirmative votes.
> 
> Thanks,
> Your friendly release manager Arvid
> 
> [1]
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12349691
> [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.12.3-rc1/
> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> [4] https://repository.apache.org/content/repositories/orgapacheflink-1419
> [5] https://github.com/apache/flink/releases/tag/release-1.12.3-rc1
> [6] https://github.com/apache/flink-web/pull/437



[jira] [Created] (FLINK-22462) JdbcExactlyOnceSinkE2eTest.testInsert failed because of too many clients.

2021-04-25 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22462:
-

 Summary: JdbcExactlyOnceSinkE2eTest.testInsert failed because of 
too many clients.
 Key: FLINK-22462
 URL: https://issues.apache.org/jira/browse/FLINK-22462
 Project: Flink
  Issue Type: Bug
  Components: Connectors / JDBC
Affects Versions: 1.14.0
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17178&view=logs&j=ba53eb01-1462-56a3-8e98-0dd97fbcaab5&t=bfbc6239-57a0-5db0-63f3-41551b4f7d51&l=13514


{code:java}
Apr 25 23:05:31 [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time 
elapsed: 138.743 s <<< FAILURE! - in 
org.apache.flink.connector.jdbc.xa.JdbcExactlyOnceSinkE2eTest
Apr 25 23:05:31 [ERROR] 
testInsert(org.apache.flink.connector.jdbc.xa.JdbcExactlyOnceSinkE2eTest)  Time 
elapsed: 137.267 s  <<< ERROR!
Apr 25 23:05:31 org.postgresql.util.PSQLException: FATAL: sorry, too many 
clients already
Apr 25 23:05:31 at 
org.postgresql.core.v3.ConnectionFactoryImpl.doAuthentication(ConnectionFactoryImpl.java:524)
Apr 25 23:05:31 at 
org.postgresql.core.v3.ConnectionFactoryImpl.tryConnect(ConnectionFactoryImpl.java:145)
Apr 25 23:05:31 at 
org.postgresql.core.v3.ConnectionFactoryImpl.openConnectionImpl(ConnectionFactoryImpl.java:196)
Apr 25 23:05:31 at 
org.postgresql.core.ConnectionFactory.openConnection(ConnectionFactory.java:49)
Apr 25 23:05:31 at 
org.postgresql.jdbc.PgConnection.(PgConnection.java:211)
Apr 25 23:05:31 at org.postgresql.Driver.makeConnection(Driver.java:459)
Apr 25 23:05:31 at org.postgresql.Driver.connect(Driver.java:261)
Apr 25 23:05:31 at 
java.sql.DriverManager.getConnection(DriverManager.java:664)
Apr 25 23:05:31 at 
java.sql.DriverManager.getConnection(DriverManager.java:247)
Apr 25 23:05:31 at 
org.apache.flink.connector.jdbc.xa.JdbcXaFacadeTestHelper.getInsertedIds(JdbcXaFacadeTestHelper.java:81)
Apr 25 23:05:31 at 
org.apache.flink.connector.jdbc.xa.JdbcExactlyOnceSinkE2eTest.testInsert(JdbcExactlyOnceSinkE2eTest.java:119)
Apr 25 23:05:31 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)

{code}




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22463) IllegalArgumentException is thrown in WindowAttachedWindowingStrategy when two phase is enabled for distinct agg

2021-04-25 Thread godfrey he (Jira)
godfrey he created FLINK-22463:
--

 Summary: IllegalArgumentException is thrown in 
WindowAttachedWindowingStrategy when two phase is enabled for distinct agg
 Key: FLINK-22463
 URL: https://issues.apache.org/jira/browse/FLINK-22463
 Project: Flink
  Issue Type: Bug
Reporter: godfrey he


Caused by: java.lang.IllegalArgumentException
at 
org.apache.flink.util.Preconditions.checkArgument(Preconditions.java:122)
at 
org.apache.flink.table.planner.plan.logical.WindowAttachedWindowingStrategy.(WindowAttachedWindowingStrategy.java:51)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)


The reason is the {{windowStart}} may be {{-1}} when two phase is enabled for 
distinct agg, see 
[TwoStageOptimizedWindowAggregateRule.java#L143|https://github.com/apache/flink/blob/a3363b91b144edfbae5ab114984ded622d3f8fbc/flink-table/flink-table-planner-blink/src/main/java/org/apache/flink/table/planner/plan/rules/physical/stream/TwoStageOptimizedWindowAggregateRule.java#L143]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22464) OperatorEventSendingCheckpointITCase.testOperatorEventLostWithReaderFailure hangs with `AdaptiveScheduler`

2021-04-25 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22464:
-

 Summary: 
OperatorEventSendingCheckpointITCase.testOperatorEventLostWithReaderFailure 
hangs with `AdaptiveScheduler`
 Key: FLINK-22464
 URL: https://issues.apache.org/jira/browse/FLINK-22464
 Project: Flink
  Issue Type: Bug
  Components: Tests
Affects Versions: 1.14.0
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17178&view=logs&j=8fd9202e-fd17-5b26-353c-ac1ff76c8f28&t=a0a633b8-47ef-5c5a-2806-3c13b9e48228&l=8171


{code:java}

at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)

{code}




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22465) KafkaSourceITCase.testValueOnlyDeserializer hangs

2021-04-25 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22465:
-

 Summary: KafkaSourceITCase.testValueOnlyDeserializer hangs
 Key: FLINK-22465
 URL: https://issues.apache.org/jira/browse/FLINK-22465
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kafka
Affects Versions: 1.14.0
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17104&view=logs&j=c5f0071e-1851-543e-9a45-9ac140befc32&t=1fb1a56f-e8b5-5a82-00a0-a2db7757b4f5&l=28977


{code:java}

at 
org.apache.flink.streaming.api.operators.collect.CollectResultIterator.nextResultFromFetcher(CollectResultIterator.java:106)
at 
org.apache.flink.streaming.api.operators.collect.CollectResultIterator.hasNext(CollectResultIterator.java:80)
at java.util.Iterator.forEachRemaining(Iterator.java:115)
at 
org.apache.flink.connector.kafka.source.KafkaSourceITCase.testValueOnlyDeserializer(KafkaSourceITCase.java:155)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.jav
{code}




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22466) KafkaSourceLegacyITCase.testOneToOneSources fail because the OperatorEvent lost

2021-04-25 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22466:
-

 Summary: KafkaSourceLegacyITCase.testOneToOneSources fail because 
the OperatorEvent lost
 Key: FLINK-22466
 URL: https://issues.apache.org/jira/browse/FLINK-22466
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kafka
Affects Versions: 1.13.0
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17110&view=logs&j=c5f0071e-1851-543e-9a45-9ac140befc32&t=1fb1a56f-e8b5-5a82-00a0-a2db7757b4f5&l=7010

{code:java}
2021-04-23T14:31:37.1620668Z Apr 23 14:31:37 [INFO] Running 
org.apache.flink.connector.kafka.source.KafkaSourceLegacyITCase
2021-04-23T14:32:27.0398155Z java.util.concurrent.ExecutionException: 
org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
2021-04-23T14:32:27.0400673Zat 
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
2021-04-23T14:32:27.0401550Zat 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
2021-04-23T14:32:27.0402365Zat 
org.apache.flink.test.util.TestUtils.tryExecute(TestUtils.java:49)
2021-04-23T14:32:27.0403227Zat 
org.apache.flink.streaming.connectors.kafka.KafkaConsumerTestBase.runOneToOneExactlyOnceTest(KafkaConsumerTestBase.java:1009)
2021-04-23T14:32:27.0403937Zat 
org.apache.flink.connector.kafka.source.KafkaSourceLegacyITCase.testOneToOneSources(KafkaSourceLegacyITCase.java:77)
2021-04-23T14:32:27.0404881Zat 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2021-04-23T14:32:27.0405293Zat 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2021-04-23T14:32:27.0406792Zat 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2021-04-23T14:32:27.0407333Zat 
java.lang.reflect.Method.invoke(Method.java:498)
2021-04-23T14:32:27.0407743Zat 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
2021-04-23T14:32:27.0408318Zat 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
2021-04-23T14:32:27.0408784Zat 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
2021-04-23T14:32:27.0409246Zat 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
2021-04-23T14:32:27.0409742Zat 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
2021-04-23T14:32:27.0410251Zat 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
2021-04-23T14:32:27.0410727Zat 
java.util.concurrent.FutureTask.run(FutureTask.java:266)
2021-04-23T14:32:27.0411065Zat java.lang.Thread.run(Thread.java:748)
2021-04-23T14:32:27.0411430Z Caused by: 
org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
2021-04-23T14:32:27.0411931Zat 
org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:144)
2021-04-23T14:32:27.0412631Zat 
org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$3(MiniClusterJobClient.java:137)
2021-04-23T14:32:27.0413144Zat 
java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
2021-04-23T14:32:27.0413605Zat 
java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
2021-04-23T14:32:27.0414063Zat 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
2021-04-23T14:32:27.0414497Zat 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
2021-04-23T14:32:27.0415002Zat 
org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:237)
2021-04-23T14:32:27.0415526Zat 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
2021-04-23T14:32:27.0416026Zat 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
2021-04-23T14:32:27.0416498Zat 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
2021-04-23T14:32:27.0417403Zat 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
2021-04-23T14:32:27.0417945Zat 
org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:1081)
2021-04-23T14:32:27.0418481Zat 
akka.dispatch.OnComplete.internal(Future.scala:264)
2021-04-23T14:32:27.0418820Zat 
akka.dispatch.OnComplete.internal(Future.scala:261)
2021-04-23T14:32:27.0419188Zat 
akka.dispatch.japi$CallbackBridge.apply(Future.scala:191)
2021-04-23T14:32:27.0419574Zat 
akka.dispatch.japi$CallbackBridge.apply(Future.scala:188)
2021-04-23T14:32:27.0419956Zat 
scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
2021-04-23T14:32:27.0420414Zat 
org.apache.flink.runtime.concurrent.Executors$

[jira] [Created] (FLINK-22467) OverAggregateITCase.testRowNumberOnOver is unstable

2021-04-25 Thread Jingsong Lee (Jira)
Jingsong Lee created FLINK-22467:


 Summary: OverAggregateITCase.testRowNumberOnOver is unstable
 Key: FLINK-22467
 URL: https://issues.apache.org/jira/browse/FLINK-22467
 Project: Flink
  Issue Type: Test
  Components: Table SQL / Planner
Reporter: Jingsong Lee
 Fix For: 1.14.0


assertEquals(expected, sink.getAppendResults)

Results should be sorted.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22468) Make StateFun build with Java 11

2021-04-25 Thread Tzu-Li (Gordon) Tai (Jira)
Tzu-Li (Gordon) Tai created FLINK-22468:
---

 Summary: Make StateFun build with Java 11
 Key: FLINK-22468
 URL: https://issues.apache.org/jira/browse/FLINK-22468
 Project: Flink
  Issue Type: Task
  Components: Stateful Functions
Reporter: Tzu-Li (Gordon) Tai
Assignee: Tzu-Li (Gordon) Tai
 Fix For: statefun-3.1.0, statefun-3.0.1


StateFun is currently not building with Java 11 due to the removal of 
javax.annotations in JDK 11. This should be fixable by manually adding the 
dependency.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22469) HistoryServer starts with NoSuchFileException

2021-04-25 Thread Gabor Somogyi (Jira)
Gabor Somogyi created FLINK-22469:
-

 Summary: HistoryServer starts with NoSuchFileException
 Key: FLINK-22469
 URL: https://issues.apache.org/jira/browse/FLINK-22469
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Web Frontend
Affects Versions: 1.12.2
Reporter: Gabor Somogyi


When history server started initially it throws the following exception:
{code:java}
2021-04-23 23:25:17,487 ERROR 
org.apache.flink.runtime.webmonitor.history.HistoryServerArchiveFetcher [] - 
Failed to update job overview.
java.nio.file.NoSuchFileException: 
/var/folders/jd/35_sh46s7zq0qc6khfw8hc80gn/T/flink-web-history-35a77053-0e52-4cb3-8d22-626e3ce3cbd7/jobs/overview.json
at 
sun.nio.fs.UnixException.translateToIOException(UnixException.java:86) 
~[?:1.8.0_282]
at 
sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102) 
~[?:1.8.0_282]
at 
sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107) 
~[?:1.8.0_282]
at 
sun.nio.fs.UnixFileSystemProvider.newByteChannel(UnixFileSystemProvider.java:214)
 ~[?:1.8.0_282]
at java.nio.file.Files.newByteChannel(Files.java:361) ~[?:1.8.0_282]
at java.nio.file.Files.createFile(Files.java:632) ~[?:1.8.0_282]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServer.createOrGetFile(HistoryServer.java:324)
 ~[flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServerArchiveFetcher.updateJobOverview(HistoryServerArchiveFetcher.java:464)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServerArchiveFetcher.access$000(HistoryServerArchiveFetcher.java:74)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServerArchiveFetcher$JobArchiveFetcherTask.(HistoryServerArchiveFetcher.java:199)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServerArchiveFetcher.(HistoryServerArchiveFetcher.java:124)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServer.(HistoryServer.java:230)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServer.(HistoryServer.java:146)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServer$1.call(HistoryServer.java:130)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServer$1.call(HistoryServer.java:127)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
at 
org.apache.flink.runtime.webmonitor.history.HistoryServer.main(HistoryServer.java:126)
 [flink-dist_2.11-1.14-SNAPSHOT.jar:1.14-SNAPSHOT]
{code}
The issue is that "webDi"r not yet created by "HistoryServer" when 
"HistoryServerArchiveFetcher" tries to reach "jobs/overview.json" file in it.




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [VOTE] Release 1.13.0, release candidate #2

2021-04-25 Thread Rui Li
I see. Thanks Jark for the clarification.

On Sun, Apr 25, 2021 at 10:36 PM Jark Wu  wrote:

> Hi Rui,
>
> Documentation is not a part of the release package.
> The changes to the documentation will take effect in 24 hours
>  and can be accessed on the web [1] then.
> Docs can even be updated after 1.13 is released.
> So we don't need to cancel the RC for this.
>
> Best,
> Jark
>
> [1]:
>
> https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/
>
> On Sun, 25 Apr 2021 at 13:47, Rui Li  wrote:
>
> > Hi Dawid,
> >
> > I have just merged the doc change for hive dialect [1], which talks about
> > how to use hive dialect to run hive queries. Do you think we can have
> > another RC to include it? Sorry for the trouble.
> >
> > [1] https://issues.apache.org/jira/browse/FLINK-22119
> >
> > On Sat, Apr 24, 2021 at 1:36 AM Dawid Wysakowicz  >
> > wrote:
> >
> > > Hi everyone,
> > > Please review and vote on the release candidate #2 for the version
> > 1.13.0,
> > > as follows:
> > > [ ] +1, Approve the release
> > > [ ] -1, Do not approve the release (please provide specific comments)
> > >
> > >
> > > The complete staging area is available for your review, which includes:
> > > * JIRA release notes [1],
> > > * the official Apache source release and binary convenience releases to
> > be
> > > deployed to dist.apache.org [2], which are signed with the key with
> > > fingerprint 31D2DD10BFC15A2D [3],
> > > * all artifacts to be deployed to the Maven Central Repository [4],
> > > * source code tag "release-1.13.0-rc2" [5],
> > > * website pull request listing the new release and adding announcement
> > > blog post [6].
> > >
> > > The vote will be open for at least 72 hours. It is adopted by majority
> > > approval, with at least 3 PMC affirmative votes.
> > >
> > > Thanks,
> > > Dawid Wysakowicz
> > >
> > > [1]
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12349287
> > > [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.13.0-rc2/
> > > [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> > > [4]
> > >
> https://repository.apache.org/content/repositories/orgapacheflink-1420/
> > > [5] https://github.com/apache/flink/tree/release-1.13.0-rc2
> > > [6] https://github.com/apache/flink-web/pull/436
> > >
> >
> >
> > --
> > Best regards!
> > Rui Li
> >
>


-- 
Best regards!
Rui Li


Re: when should `FlinkYarnSessionCli` be included for parsing CLI arguments?

2021-04-25 Thread Yangze Guo
Hi, Tony.

What is the version of your flink-dist. AFAIK, this issue should be
addressed in FLINK-15852[1]. Could you give the client log of case
2(set the log level to DEBUG would be better).

[1] https://issues.apache.org/jira/browse/FLINK-15852

Best,
Yangze Guo

On Sun, Apr 25, 2021 at 11:33 AM Tony Wei  wrote:
>
> Hi Experts,
>
> I recently tried to run yarn-application mode on my yarn cluster, and I had a 
> problem related to configuring `execution.target`.
> After reading the source code and doing some experiments, I found that there 
> should be some room of improvement for `FlinkYarnSessionCli` or 
> `AbstractYarnCli`.
>
> My experiments are:
>
> setting `execution.target: yarn-application` in flink-conf.yaml and run 
> `flink run-application -t yarn-application`: run job successfully.
>
> `FlinkYarnSessionCli` is not active
> `GenericCLI` is active
>
> setting `execution.target: yarn-per-job` in flink-conf.yaml and run `flink 
> run-application -t yarn-application`: run job failed
>
> failed due to `ClusterDeploymentException` [1]
> `FlinkYarnSessionCli` is active
>
> setting `execution.target: yarn-application` in flink-conf.yaml and run 
> `flink run -t yarn-per-job`: run job successfully.
>
> `FlinkYarnSessionCli` is not active
> `GenericCLI` is active
>
> setting `execution.target: yarn-per-job` in flink-conf.yaml and run `flink 
> run -t yarn-per-job`: run job successfully.
>
> `FlinkYarnSessionCli` is active
>
> From `AbstractYarnCli#isActive` [2] and `FlinkYarnSessionCli#isActive` [3], 
> `FlinkYarnSessionCli` will be active when `execution.target` is specified 
> with `yarn-per-job` or `yarn-session`.
>
> According to the flink official document [4], I thought the 2nd experiment 
> should also work well, but it didn't.
>>
>> The --target will overwrite the execution.target specified in the 
>> config/flink-config.yaml.
>
>
> The root cause is that `FlinkYarnSessionCli` only overwrite the 
> `execution.target` with `yarn-session` or `yarn-per-job` [5], but no 
> `yarn-application`.
> So, my question is
>
> should we use `FlinkYarnSessionCli` in case 2?
> if we should, how we can improve `FlinkYarnSessionCli` so that we can 
> overwrite `execution.target` via `--target`?
>
> and one more improvement, the config description for `execution.target` [6] 
> should include `yarn-application` as well.
>
> [1] 
> https://github.com/apache/flink/blob/master/flink-yarn/src/main/java/org/apache/flink/yarn/YarnClusterDescriptor.java#L439-L447
> [2] 
> https://github.com/apache/flink/blob/master/flink-yarn/src/main/java/org/apache/flink/yarn/cli/AbstractYarnCli.java#L54-L66
> [3] 
> https://github.com/apache/flink/blob/master/flink-yarn/src/main/java/org/apache/flink/yarn/cli/FlinkYarnSessionCli.java#L373-L377
> [4] 
> https://ci.apache.org/projects/flink/flink-docs-stable/deployment/cli.html#selecting-deployment-targets
> [5] 
> https://github.com/apache/flink/blob/master/flink-yarn/src/main/java/org/apache/flink/yarn/cli/FlinkYarnSessionCli.java#L397-L413
> [6] 
> https://github.com/apache/flink/blob/master/flink-core/src/main/java/org/apache/flink/configuration/DeploymentOptions.java#L41-L46
>
> best regards,
>