This is an automated email from the ASF dual-hosted git repository.
danny0405 pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/asf-site by this push:
new e4d28aa [HUDI-2729] Fix the options of flink quick start page (#3958)
e4d28aa is described below
commit e4d28aa366560ab68136821f3a1297fd9a9fd18c
Author: Danny Chan <[email protected]>
AuthorDate: Wed Nov 10 15:45:24 2021 +0800
[HUDI-2729] Fix the options of flink quick start page (#3958)
---
website/docs/flink-quick-start-guide.md | 36 +++++++++++----------------------
1 file changed, 12 insertions(+), 24 deletions(-)
diff --git a/website/docs/flink-quick-start-guide.md
b/website/docs/flink-quick-start-guide.md
index 3f4d4a9..7ba0423 100644
--- a/website/docs/flink-quick-start-guide.md
+++ b/website/docs/flink-quick-start-guide.md
@@ -76,7 +76,7 @@ Creates a Flink Hudi table first and insert data into the
Hudi table using SQL `
set execution.result-mode=tableau;
CREATE TABLE t1(
- uuid VARCHAR(20),
+ uuid VARCHAR(20) PRIMARY KEY NOT ENFORCED,
name VARCHAR(10),
age INT,
ts TIMESTAMP(3),
@@ -85,7 +85,7 @@ CREATE TABLE t1(
PARTITIONED BY (`partition`)
WITH (
'connector' = 'hudi',
- 'path' = 'schema://base-path',
+ 'path' = '${path}',
'table.type' = 'MERGE_ON_READ' -- this creates a MERGE_ON_READ table, by
default is COPY_ON_WRITE
);
@@ -135,7 +135,7 @@ We do not need to specify endTime, if we want all changes
after the given commit
```sql
CREATE TABLE t1(
- uuid VARCHAR(20),
+ uuid VARCHAR(20) PRIMARY KEY NOT ENFORCED,
name VARCHAR(10),
age INT,
ts TIMESTAMP(3),
@@ -144,10 +144,10 @@ CREATE TABLE t1(
PARTITIONED BY (`partition`)
WITH (
'connector' = 'hudi',
- 'path' = 'oss://vvr-daily/hudi/t1',
+ 'path' = '${path}',
'table.type' = 'MERGE_ON_READ',
'read.streaming.enabled' = 'true', -- this option enable the streaming read
- 'read.streaming.start-commit' = '20210316134557' -- specifies the start
commit instant time
+ 'read.start-commit' = '20210316134557' -- specifies the start commit instant
time
'read.streaming.check-interval' = '4' -- specifies the check interval for
finding new source commits, default 60s.
);
@@ -157,7 +157,6 @@ select * from t1;
This will give all changes that happened after the
`read.streaming.start-commit` commit. The unique thing about this
feature is that it now lets you author streaming pipelines on streaming or
batch data source.
-{: .notice--info}
### Delete Data {#deletes}
@@ -460,11 +459,11 @@ CREATE TABLE t1(
PARTITIONED BY (`partition`)
WITH (
'connector' = 'hudi',
- 'path' = 'oss://vvr-daily/hudi/t1',
+ 'path' = '${db_path}/t1',
'table.type' = 'COPY_ON_WRITE', --If MERGE_ON_READ, hive query will not
have output until the parquet file is generated
'hive_sync.enable' = 'true', -- Required. To enable hive synchronization
'hive_sync.mode' = 'hms' -- Required. Setting hive sync mode to hms,
default jdbc
- 'hive_sync.metastore.uris' = 'thrift://ip:9083' -- Required. The port need
set on hive-site.xml
+ 'hive_sync.metastore.uris' = 'thrift://${ip}:9083' -- Required. The port
need set on hive-site.xml
);
@@ -479,16 +478,16 @@ CREATE TABLE t1(
PARTITIONED BY (`partition`)
WITH (
'connector' = 'hudi',
- 'path' = 'oss://vvr-daily/hudi/t1',
+ 'path' = '${db_path}/t1',
'table.type' = 'COPY_ON_WRITE', --If MERGE_ON_READ, hive query will not
have output until the parquet file is generated
'hive_sync.enable' = 'true', -- Required. To enable hive synchronization
'hive_sync.mode' = 'hms' -- Required. Setting hive sync mode to hms,
default jdbc
- 'hive_sync.metastore.uris' = 'thrift://ip:9083' -- Required. The port need
set on hive-site.xml
- 'hive_sync.jdbc_url'='jdbc:hive2://ip:10000', -- required, hiveServer port
+ 'hive_sync.metastore.uris' = 'thrift://${ip}:9083' -- Required. The port
need set on hive-site.xml
+ 'hive_sync.jdbc_url'='jdbc:hive2://${ip}:10000', -- required, hiveServer
port
'hive_sync.table'='t1', -- required, hive table name
'hive_sync.db'='testDB', -- required, hive database
name
- 'hive_sync.username'='root', -- required, HMS username
- 'hive_sync.password'='your password' -- required, HMS password
+ 'hive_sync.username'='${user_name}', -- required, HMS
username
+ 'hive_sync.password'='${password}' -- required, HMS password
);
```
@@ -499,17 +498,6 @@ While using hive beeline query, you need to enter settings:
set hive.input.format =
org.apache.hudi.hadoop.hive.HoodieCombineHiveInputFormat;
```
-### Conflict
-
-When there is a `flink-sql-connector-hive-xxx.jar` under Flink lib/, there
will be a jar conflicts between `flink-sql-connector-hive-xxx.jar`
-and `hudi-flink-bundle_2.11.xxx.jar`. The solution is to use another profile
`include-flink-sql-connector-hive` when install and delete
-the `flink-sql-connector-hive-xxx.jar` under Flink lib/. install command :
-
-```bash
-# Maven install command
-mvn install -DskipTests -Drat.skip=true -Pflink-bundle-shade-hive2
-Pinclude-flink-sql-connector-hive
-```
-
## Presto Query
### Hive Sync