This is an automated email from the ASF dual-hosted git repository.

danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 1e95f14b6c [HUDI-4644] Change default flink profile to 1.15.x (#6445)
1e95f14b6c is described below

commit 1e95f14b6ca144b2569b20b15840c29b00009b95
Author: Danny Chan <[email protected]>
AuthorDate: Sat Aug 20 17:32:26 2022 +0800

    [HUDI-4644] Change default flink profile to 1.15.x (#6445)
---
 hudi-examples/hudi-examples-flink/pom.xml          |  7 ----
 hudi-flink-datasource/hudi-flink/pom.xml           |  5 ---
 .../apache/hudi/sink/ITTestDataStreamWrite.java    |  1 -
 pom.xml                                            | 39 ++++++----------------
 4 files changed, 11 insertions(+), 41 deletions(-)

diff --git a/hudi-examples/hudi-examples-flink/pom.xml 
b/hudi-examples/hudi-examples-flink/pom.xml
index a9092411d9..7038cd9d46 100644
--- a/hudi-examples/hudi-examples-flink/pom.xml
+++ b/hudi-examples/hudi-examples-flink/pom.xml
@@ -333,13 +333,6 @@
             <scope>test</scope>
             <type>test-jar</type>
         </dependency>
-        <dependency>
-            <groupId>org.apache.flink</groupId>
-            <artifactId>flink-json</artifactId>
-            <version>${flink.version}</version>
-            <scope>test</scope>
-            <type>test-jar</type>
-        </dependency>
         <dependency>
             <groupId>org.apache.flink</groupId>
             <artifactId>flink-csv</artifactId>
diff --git a/hudi-flink-datasource/hudi-flink/pom.xml 
b/hudi-flink-datasource/hudi-flink/pom.xml
index f6bd3a00ac..396b10b11a 100644
--- a/hudi-flink-datasource/hudi-flink/pom.xml
+++ b/hudi-flink-datasource/hudi-flink/pom.xml
@@ -239,11 +239,6 @@
             <artifactId>jcommander</artifactId>
             <scope>compile</scope>
         </dependency>
-        <dependency>
-            <groupId>com.twitter</groupId>
-            <artifactId>bijection-avro_${scala.binary.version}</artifactId>
-            <version>0.9.7</version>
-        </dependency>
         <dependency>
             <groupId>joda-time</groupId>
             <artifactId>joda-time</artifactId>
diff --git 
a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
 
b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
index aa420a433d..4862cda07a 100644
--- 
a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
+++ 
b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
@@ -246,7 +246,6 @@ public class ITTestDataStreamWrite extends TestLogger {
     execEnv.addOperator(pipeline.getTransformation());
 
     if (isMor) {
-      Pipelines.clean(conf, pipeline);
       Pipelines.compact(conf, pipeline);
     }
 
diff --git a/pom.xml b/pom.xml
index ff1f1e26ea..6489e632b4 100644
--- a/pom.xml
+++ b/pom.xml
@@ -131,20 +131,20 @@
     <flink1.15.version>1.15.1</flink1.15.version>
     <flink1.14.version>1.14.5</flink1.14.version>
     <flink1.13.version>1.13.6</flink1.13.version>
-    <flink.version>${flink1.13.version}</flink.version>
-    <hudi.flink.module>hudi-flink1.13.x</hudi.flink.module>
-    <flink.bundle.version>1.13</flink.bundle.version>
+    <flink.version>${flink1.15.version}</flink.version>
+    <hudi.flink.module>hudi-flink1.15.x</hudi.flink.module>
+    <flink.bundle.version>1.15</flink.bundle.version>
     <flink.format.parquet.version>1.12.2</flink.format.parquet.version>
     <flink.runtime.artifactId>flink-runtime</flink.runtime.artifactId>
-    
<flink.table.runtime.artifactId>flink-table-runtime_${scala.binary.version}</flink.table.runtime.artifactId>
-    
<flink.table.planner.artifactId>flink-table-planner_${scala.binary.version}</flink.table.planner.artifactId>
+    
<flink.table.runtime.artifactId>flink-table-runtime</flink.table.runtime.artifactId>
+    
<flink.table.planner.artifactId>flink-table-planner_2.12</flink.table.planner.artifactId>
     <flink.parquet.artifactId>flink-parquet</flink.parquet.artifactId>
     
<flink.statebackend.rocksdb.artifactId>flink-statebackend-rocksdb</flink.statebackend.rocksdb.artifactId>
     <flink.test.utils.artifactId>flink-test-utils</flink.test.utils.artifactId>
     
<flink.streaming.java.artifactId>flink-streaming-java</flink.streaming.java.artifactId>
     <flink.clients.artifactId>flink-clients</flink.clients.artifactId>
     
<flink.connector.kafka.artifactId>flink-connector-kafka</flink.connector.kafka.artifactId>
-    
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_${scala.binary.version}</flink.hadoop.compatibility.artifactId>
+    
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_2.12</flink.hadoop.compatibility.artifactId>
     <spark31.version>3.1.3</spark31.version>
     <spark32.version>3.2.1</spark32.version>
     <spark33.version>3.3.0</spark33.version>
@@ -1828,20 +1828,6 @@
 
     <profile>
       <id>flink1.15</id>
-      <properties>
-        <flink.version>${flink1.15.version}</flink.version>
-        
<flink.table.runtime.artifactId>flink-table-runtime</flink.table.runtime.artifactId>
-        <flink.parquet.artifactId>flink-parquet</flink.parquet.artifactId>
-        
<flink.statebackend.rocksdb.artifactId>flink-statebackend-rocksdb</flink.statebackend.rocksdb.artifactId>
-        
<flink.test.utils.artifactId>flink-test-utils</flink.test.utils.artifactId>
-        
<flink.streaming.java.artifactId>flink-streaming-java</flink.streaming.java.artifactId>
-        <flink.clients.artifactId>flink-clients</flink.clients.artifactId>
-        
<flink.connector.kafka.artifactId>flink-connector-kafka</flink.connector.kafka.artifactId>
-        <!-- 1.15 only supports scala2.12 -->
-        
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_2.12</flink.hadoop.compatibility.artifactId>
-        <hudi.flink.module>hudi-flink1.15.x</hudi.flink.module>
-        <flink.bundle.version>1.15</flink.bundle.version>
-      </properties>
       <activation>
         <property>
           <name>flink1.15</name>
@@ -1852,7 +1838,10 @@
       <id>flink1.14</id>
       <properties>
         <flink.version>${flink1.14.version}</flink.version>
+        <hudi.flink.module>hudi-flink1.14.x</hudi.flink.module>
+        <flink.bundle.version>1.14</flink.bundle.version>
         
<flink.table.runtime.artifactId>flink-table-runtime_${scala.binary.version}</flink.table.runtime.artifactId>
+        
<flink.table.planner.artifactId>flink-table-planner_${scala.binary.version}</flink.table.planner.artifactId>
         
<flink.parquet.artifactId>flink-parquet_${scala.binary.version}</flink.parquet.artifactId>
         
<flink.statebackend.rocksdb.artifactId>flink-statebackend-rocksdb_${scala.binary.version}</flink.statebackend.rocksdb.artifactId>
         
<flink.test.utils.artifactId>flink-test-utils_${scala.binary.version}</flink.test.utils.artifactId>
@@ -1860,23 +1849,19 @@
         
<flink.clients.artifactId>flink-clients_${scala.binary.version}</flink.clients.artifactId>
         
<flink.connector.kafka.artifactId>flink-connector-kafka_${scala.binary.version}</flink.connector.kafka.artifactId>
         
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_${scala.binary.version}</flink.hadoop.compatibility.artifactId>
-        <hudi.flink.module>hudi-flink1.14.x</hudi.flink.module>
-        <flink.bundle.version>1.14</flink.bundle.version>
       </properties>
       <activation>
-        <activeByDefault>true</activeByDefault>
         <property>
           <name>flink1.14</name>
-          <!-- add flink1.14 module to all profile -->
-          <value>!disabled</value>
         </property>
       </activation>
     </profile>
     <profile>
       <id>flink1.13</id>
       <properties>
-        <flink.scala.bnary.version>2.11</flink.scala.bnary.version>
         <flink.version>${flink1.13.version}</flink.version>
+        <hudi.flink.module>hudi-flink1.13.x</hudi.flink.module>
+        <flink.bundle.version>1.13</flink.bundle.version>
         
<flink.runtime.artifactId>flink-runtime_${scala.binary.version}</flink.runtime.artifactId>
         
<flink.table.runtime.artifactId>flink-table-runtime-blink_${scala.binary.version}</flink.table.runtime.artifactId>
         
<flink.table.planner.artifactId>flink-table-planner-blink_${scala.binary.version}</flink.table.planner.artifactId>
@@ -1887,8 +1872,6 @@
         
<flink.clients.artifactId>flink-clients_${scala.binary.version}</flink.clients.artifactId>
         
<flink.connector.kafka.artifactId>flink-connector-kafka_${scala.binary.version}</flink.connector.kafka.artifactId>
         
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_${scala.binary.version}</flink.hadoop.compatibility.artifactId>
-        <hudi.flink.module>hudi-flink1.13.x</hudi.flink.module>
-        <flink.bundle.version>1.13</flink.bundle.version>
         <skipITs>true</skipITs>
       </properties>
       <activation>

Reply via email to