This is an automated email from the ASF dual-hosted git repository.
hellostephen pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/master by this push:
new aaaf54561c2 [regression](s3) add default conf for s3 releated cases
(#37952)
aaaf54561c2 is described below
commit aaaf54561c2af53853a5c9843677ee18cda13141
Author: Dongyang Li <[email protected]>
AuthorDate: Mon Jul 29 11:00:35 2024 +0800
[regression](s3) add default conf for s3 releated cases (#37952)
## Proposed changes
replace COS with OSS in the TeamCity pipeline to improve stability
<!--Describe your changes.-->
---------
Co-authored-by: stephen <[email protected]>
---
regression-test/conf/regression-conf.groovy | 16 ++--
.../org/apache/doris/regression/Config.groovy | 92 +++++++++++++++++++++-
.../org/apache/doris/regression/suite/Suite.groovy | 2 +-
.../cloud_p0/conf/regression-conf-custom.groovy | 2 +-
regression-test/pipeline/cloud_p0/prepare.sh | 2 +-
regression-test/pipeline/cloud_p0/run.sh | 6 +-
.../cloud_p1/conf/regression-conf-custom.groovy | 2 +-
.../pipeline/external/conf/regression-conf.groovy | 5 +-
.../pipeline/p0/conf/regression-conf.groovy | 5 +-
.../pipeline/p1/conf/regression-conf.groovy | 5 +-
.../spark_connector/spark_connector.groovy | 3 +-
...test_export_table_with_materialized_view.groovy | 2 +-
.../hive/test_trino_different_parquet_types.groovy | 20 ++---
.../hive/test_trino_hive_orc.groovy | 20 ++---
.../hive/test_trino_hive_other.groovy | 20 ++---
.../hive/test_trino_hive_parquet.groovy | 22 +++---
.../hive/test_trino_hive_schema_evolution.groovy | 23 +++---
.../hive/test_trino_hive_serde_prop.groovy | 22 +++---
.../hive/test_trino_hive_tablesample_p0.groovy | 20 ++---
.../hive/test_trino_hive_tpch_sf1_orc.groovy | 20 ++---
.../hive/test_trino_hive_tpch_sf1_parquet.groovy | 20 ++---
.../test_trino_prepare_hive_data_in_case.groovy | 22 +++---
.../kafka/test_trino_kafka_base.groovy | 23 +++---
.../trino_connector/test_plugins_download.groovy | 21 ++---
regression-test/suites/variant_p0/load.groovy | 4 +-
25 files changed, 245 insertions(+), 154 deletions(-)
diff --git a/regression-test/conf/regression-conf.groovy
b/regression-test/conf/regression-conf.groovy
index 527b0231394..08d03632c37 100644
--- a/regression-test/conf/regression-conf.groovy
+++ b/regression-test/conf/regression-conf.groovy
@@ -104,8 +104,16 @@ brokerName = "broker_name"
// broker load test config
enableBrokerLoad=true
-ak=""
-sk=""
+
+// for s3 releated cases, "aliyun" or "aliyun-internal" or "tencent" or
"huawei" or "azure" or "gcp"
+// if s3Source is set, s3Endpoint s3BucketName s3Region s3Provider will be
filled with default value if not set
+s3Source="aliyun"
+// s3Endpoint = ""
+// s3BucketName = ""
+// s3Region = ""
+// s3Provider = ""
+ak="***********"
+sk="***********"
// jdbc connector test config
// To enable jdbc test, you need first start mysql/pg container.
@@ -194,10 +202,6 @@ aliYunSk="***********"
hwYunAk="***********"
hwYunSk="***********"
-s3Endpoint = "cos.ap-hongkong.myqcloud.com"
-s3BucketName = "doris-build-hk-1308700295"
-s3Region = "ap-hongkong"
-
//arrow flight sql test config
extArrowFlightSqlHost = "127.0.0.1"
extArrowFlightSqlPort = 8080
diff --git
a/regression-test/framework/src/main/groovy/org/apache/doris/regression/Config.groovy
b/regression-test/framework/src/main/groovy/org/apache/doris/regression/Config.groovy
index b901095eac1..b7c3090e0b8 100644
---
a/regression-test/framework/src/main/groovy/org/apache/doris/regression/Config.groovy
+++
b/regression-test/framework/src/main/groovy/org/apache/doris/regression/Config.groovy
@@ -144,9 +144,12 @@ class Config {
public String kafkaBrokerList
public String cloudVersion
+ public String s3Source
+
Config() {}
Config(
+ String s3Source,
String caseNamePrefix,
String defaultDb,
String jdbcUrl,
@@ -199,6 +202,7 @@ class Config {
String clusterDir,
String kafkaBrokerList,
String cloudVersion) {
+ this.s3Source = s3Source
this.caseNamePrefix = caseNamePrefix
this.defaultDb = defaultDb
this.jdbcUrl = jdbcUrl
@@ -449,7 +453,6 @@ class Config {
}
log.info("recycleAddr : $config.recycleServiceHttpAddress, socketAddr
: $config.recycleServiceHttpInetSocketAddress")
-
config.defaultDb = cmd.getOptionValue(defaultDbOpt, config.defaultDb)
config.jdbcUrl = cmd.getOptionValue(jdbcOpt, config.jdbcUrl)
config.jdbcUser = cmd.getOptionValue(userOpt, config.jdbcUser)
@@ -478,6 +481,16 @@ class Config {
log.info("withOutLoadData is ${config.withOutLoadData}".toString())
log.info("caseNamePrefix is ${config.caseNamePrefix}".toString())
log.info("dryRun is ${config.dryRun}".toString())
+ def s3SourceList = ["aliyun", "aliyun-internal", "tencent", "huawei",
"azure", "gcp"]
+ if (s3SourceList.contains(config.s3Source)) {
+ log.info("s3Source is ${config.s3Source}".toString())
+ log.info("s3Provider is
${config.otherConfigs.get("s3Provider")}".toString())
+ log.info("s3BucketName is
${config.otherConfigs.get("s3BucketName")}".toString())
+ log.info("s3Region is
${config.otherConfigs.get("s3Region")}".toString())
+ log.info("s3Endpoint is
${config.otherConfigs.get("s3Endpoint")}".toString())
+ } else {
+ throw new Exception("The s3Source '${config.s3Source}' is invalid,
optional values ${s3SourceList}")
+ }
Properties props = cmd.getOptionProperties("conf")
config.otherConfigs.putAll(props)
@@ -490,6 +503,7 @@ class Config {
static Config fromConfigObject(ConfigObject obj) {
def config = new Config(
+ configToString(obj.s3Source),
configToString(obj.caseNamePrefix),
configToString(obj.defaultDb),
configToString(obj.jdbcUrl),
@@ -602,6 +616,82 @@ class Config {
}
static void fillDefaultConfig(Config config) {
+ if (config.s3Source == null) {
+ config.s3Source = "aliyun"
+ log.info("Set s3Source to 'aliyun' because not
specify.".toString())
+ }
+
+ if (config.otherConfigs.get("s3Provider") == null) {
+ def s3Provider = "OSS"
+ if (config.s3Source == "aliyun" || config.s3Source ==
"aliyun-internal") {
+ s3Provider = "OSS"
+ } else if (config.s3Source == "tencent") {
+ s3Provider = "COS"
+ } else if (config.s3Source == "huawei") {
+ s3Provider = "OBS"
+ } else if (config.s3Source == "azure") {
+ s3Provider = "AZURE"
+ } else if (config.s3Source == "gcp") {
+ s3Provider = "GCP"
+ }
+ config.otherConfigs.put("s3Provider", "${s3Provider}")
+ log.info("Set s3Provider to '${s3Provider}' because not
specify.".toString())
+ }
+ if (config.otherConfigs.get("s3BucketName") == null) {
+ def s3BucketName = "doris-regression-hk"
+ if (config.s3Source == "aliyun") {
+ s3BucketName = "doris-regression-hk"
+ } else if (config.s3Source == "aliyun-internal") {
+ s3BucketName = "doris-regression"
+ } else if (config.s3Source == "tencent") {
+ s3BucketName = "doris-build-1308700295"
+ } else if (config.s3Source == "huawei") {
+ s3BucketName = "doris-build"
+ } else if (config.s3Source == "azure") {
+ s3BucketName = "qa-build"
+ } else if (config.s3Source == "gcp") {
+ s3BucketName = "doris-regression"
+ }
+ config.otherConfigs.put("s3BucketName", "${s3BucketName}")
+ log.info("Set s3BucketName to '${s3BucketName}' because not
specify.".toString())
+ }
+ if (config.otherConfigs.get("s3Region") == null) {
+ def s3Region = "oss-cn-hongkong"
+ if (config.s3Source == "aliyun") {
+ s3Region = "oss-cn-hongkong"
+ } else if (config.s3Source == "aliyun-internal") {
+ s3Region = "oss-cn-beijing"
+ } else if (config.s3Source == "tencent") {
+ s3Region = "ap-beijing"
+ } else if (config.s3Source == "huawei") {
+ s3Region = "cn-north-4"
+ } else if (config.s3Source == "azure") {
+ s3Region = "azure-region"
+ } else if (config.s3Source == "gcp") {
+ s3Region = "us-central1"
+ }
+ config.otherConfigs.put("s3Region", "${s3Region}")
+ log.info("Set s3Region to '${s3Region}' because not
specify.".toString())
+ }
+ if (config.otherConfigs.get("s3Endpoint") == null) {
+ def s3Endpoint = "oss-cn-hongkong.aliyuncs.com"
+ if (config.s3Source == "aliyun") {
+ s3Endpoint = "oss-cn-hongkong.aliyuncs.com"
+ } else if (config.s3Source == "aliyun-internal") {
+ s3Endpoint = "oss-cn-beijing-internal.aliyuncs.com"
+ } else if (config.s3Source == "tencent") {
+ s3Endpoint = "cos.ap-beijing.myqcloud.com"
+ } else if (config.s3Source == "huawei") {
+ s3Endpoint = "obs.cn-north-4.myhuaweicloud.com"
+ } else if (config.s3Source == "azure") {
+ s3Endpoint = "azure-endpoint"
+ } else if (config.s3Source == "gcp") {
+ s3Endpoint = "storage.googleapis.com"
+ }
+ config.otherConfigs.put("s3Endpoint", "${s3Endpoint}")
+ log.info("Set s3Endpoint to '${s3Endpoint}' because not
specify.".toString())
+ }
+
if (config.caseNamePrefix == null) {
config.caseNamePrefix = ""
log.info("set caseNamePrefix to '' because not
specify.".toString())
diff --git
a/regression-test/framework/src/main/groovy/org/apache/doris/regression/suite/Suite.groovy
b/regression-test/framework/src/main/groovy/org/apache/doris/regression/suite/Suite.groovy
index 4d5127b8644..2cd27b0968d 100644
---
a/regression-test/framework/src/main/groovy/org/apache/doris/regression/suite/Suite.groovy
+++
b/regression-test/framework/src/main/groovy/org/apache/doris/regression/suite/Suite.groovy
@@ -877,7 +877,7 @@ class Suite implements GroovyInterceptable {
if (exitcode != 0) {
staticLogger.info("exit code: ${exitcode}, output\n:
${proc.text}")
if (mustSuc == true) {
- Assert.assertEquals(0, exitCode)
+ Assert.assertEquals(0, exitcode)
}
}
} catch (IOException e) {
diff --git
a/regression-test/pipeline/cloud_p0/conf/regression-conf-custom.groovy
b/regression-test/pipeline/cloud_p0/conf/regression-conf-custom.groovy
index c04b3742bbf..4a79c5cc909 100644
--- a/regression-test/pipeline/cloud_p0/conf/regression-conf-custom.groovy
+++ b/regression-test/pipeline/cloud_p0/conf/regression-conf-custom.groovy
@@ -74,4 +74,4 @@ enableKafkaTest=true
// trino-connector catalog test config
enableTrinoConnectorTest = false
-s3Provider = "COS"
+s3Source = "aliyun"
diff --git a/regression-test/pipeline/cloud_p0/prepare.sh
b/regression-test/pipeline/cloud_p0/prepare.sh
index 211870d99d5..df3cbe9fb9e 100644
--- a/regression-test/pipeline/cloud_p0/prepare.sh
+++ b/regression-test/pipeline/cloud_p0/prepare.sh
@@ -47,7 +47,7 @@ if [[ -z "${pr_num_from_trigger}" ]]; then echo "ERROR: env
pr_num_from_trigger
if [[ -z "${commit_id_from_trigger}" ]]; then echo "ERROR: env
commit_id_from_trigger not set" && exit 1; fi
if [[ -z "${commit_id_from_checkout}" ]]; then echo "ERROR: env
commit_id_from_checkout not set" && exit 1; fi
if [[ -z "${target_branch}" ]]; then echo "ERROR: env target_branch not set"
&& exit 1; fi
-if [[ -z "${cos_ak}" || -z "${cos_sk}" ]]; then echo "ERROR: env cos_ak or
cos_sk not set" && exit 1; fi
+if [[ -z "${s3SourceAk}" || -z "${s3SourceSk}" ]]; then echo "ERROR: env
s3SourceAk or s3SourceSk not set" && exit 1; fi
if [[ -z "${oss_ak}" || -z "${oss_sk}" ]]; then echo "ERROR: env oss_ak or
oss_sk not set." && exit 1; fi
echo "#### 1. check if need run"
diff --git a/regression-test/pipeline/cloud_p0/run.sh
b/regression-test/pipeline/cloud_p0/run.sh
index 3c126460b56..2079d182ef3 100644
--- a/regression-test/pipeline/cloud_p0/run.sh
+++ b/regression-test/pipeline/cloud_p0/run.sh
@@ -33,7 +33,7 @@ echo "#### Check env"
if [[ -z "${teamcity_build_checkoutDir}" ]]; then echo "ERROR: env
teamcity_build_checkoutDir not set" && exit 1; fi
if [[ -z "${pr_num_from_trigger}" ]]; then echo "ERROR: env
pr_num_from_trigger not set" && exit 1; fi
if [[ -z "${commit_id_from_trigger}" ]]; then echo "ERROR: env
commit_id_from_trigger not set" && exit 1; fi
-if [[ -z "${cos_ak}" || -z "${cos_sk}" ]]; then echo "ERROR: env cos_ak or
cos_sk not set" && exit 1; fi
+if [[ -z "${s3SourceAk}" || -z "${s3SourceSk}" ]]; then echo "ERROR: env
s3SourceAk or s3SourceSk not set" && exit 1; fi
# shellcheck source=/dev/null
source "$(bash
"${teamcity_build_checkoutDir}"/regression-test/pipeline/common/get-or-set-tmp-env.sh
'get')"
@@ -53,8 +53,8 @@ run() {
cd "${teamcity_build_checkoutDir}" || return 1
{
echo # add a new line to prevent two config items from being combined,
which will cause the error "No signature of method"
- echo "ak='${cos_ak}'"
- echo "sk='${cos_sk}'"
+ echo "ak='${s3SourceAk}'"
+ echo "sk='${s3SourceSk}'"
}
>>"${teamcity_build_checkoutDir}"/regression-test/pipeline/cloud_p0/conf/regression-conf-custom.groovy
cp -f
"${teamcity_build_checkoutDir}"/regression-test/pipeline/cloud_p0/conf/regression-conf-custom.groovy
\
"${teamcity_build_checkoutDir}"/regression-test/conf/
diff --git
a/regression-test/pipeline/cloud_p1/conf/regression-conf-custom.groovy
b/regression-test/pipeline/cloud_p1/conf/regression-conf-custom.groovy
index 42a18b7f22e..2662090b401 100644
--- a/regression-test/pipeline/cloud_p1/conf/regression-conf-custom.groovy
+++ b/regression-test/pipeline/cloud_p1/conf/regression-conf-custom.groovy
@@ -21,4 +21,4 @@ excludeDirectories = "000_the_start_sentinel_do_not_touch," +
// keep this line
max_failure_num = 50
-s3Provider = "COS"
+s3Source = "aliyun"
diff --git a/regression-test/pipeline/external/conf/regression-conf.groovy
b/regression-test/pipeline/external/conf/regression-conf.groovy
index 28956568e58..93965b84219 100644
--- a/regression-test/pipeline/external/conf/regression-conf.groovy
+++ b/regression-test/pipeline/external/conf/regression-conf.groovy
@@ -138,10 +138,7 @@ es_8_port=39200
cacheDataPath = "/data/regression/"
-s3Endpoint = "cos.ap-hongkong.myqcloud.com"
-s3BucketName = "doris-build-hk-1308700295"
-s3Region = "ap-hongkong"
-s3Provider = "COS"
+s3Source="aliyun"
max_failure_num=50
diff --git a/regression-test/pipeline/p0/conf/regression-conf.groovy
b/regression-test/pipeline/p0/conf/regression-conf.groovy
index c30dd20481c..72368697f2c 100644
--- a/regression-test/pipeline/p0/conf/regression-conf.groovy
+++ b/regression-test/pipeline/p0/conf/regression-conf.groovy
@@ -137,10 +137,7 @@ es_8_port=39200
cacheDataPath = "/data/regression/"
-s3Endpoint = "cos.ap-hongkong.myqcloud.com"
-s3BucketName = "doris-build-hk-1308700295"
-s3Region = "ap-hongkong"
-s3Provider = "COS"
+s3Source="aliyun"
//arrow flight sql test config
extArrowFlightSqlHost = "127.0.0.1"
diff --git a/regression-test/pipeline/p1/conf/regression-conf.groovy
b/regression-test/pipeline/p1/conf/regression-conf.groovy
index d4ecd55d38f..49f52c1fa32 100644
--- a/regression-test/pipeline/p1/conf/regression-conf.groovy
+++ b/regression-test/pipeline/p1/conf/regression-conf.groovy
@@ -70,10 +70,7 @@ excludeDirectories = "000_the_start_sentinel_do_not_touch,"
+ // keep this line
cacheDataPath="/data/regression/"
-s3Endpoint = "cos.ap-hongkong.myqcloud.com"
-s3BucketName = "doris-build-hk-1308700295"
-s3Region = "ap-hongkong"
-s3Provider = "COS"
+s3Source="aliyun"
max_failure_num=0
diff --git
a/regression-test/suites/connector_p0/spark_connector/spark_connector.groovy
b/regression-test/suites/connector_p0/spark_connector/spark_connector.groovy
index 2bd618fcc3c..06699d7c8ff 100644
--- a/regression-test/suites/connector_p0/spark_connector/spark_connector.groovy
+++ b/regression-test/suites/connector_p0/spark_connector/spark_connector.groovy
@@ -25,7 +25,8 @@ suite("spark_connector", "connector") {
logger.info("start download spark doris demo ...")
logger.info("getS3Url ==== ${getS3Url()}")
def download_spark_jar = "/usr/bin/curl
${getS3Url()}/regression/spark-doris-connector-demo-jar-with-dependencies.jar
--output spark-doris-demo.jar".execute().getText()
- logger.info("finish download spark doris demo ...")
+ def out = "/usr/bin/ls -al spark-doris-demo.jar".execute().getText()
+ logger.info("finish download spark doris demo, out: ${out}")
def run_cmd = "java -jar spark-doris-demo.jar
$context.config.feHttpAddress $context.config.feHttpUser
regression_test_connector_p0_spark_connector.$tableName"
logger.info("run_cmd : $run_cmd")
def proc = run_cmd.execute()
diff --git
a/regression-test/suites/export_p0/test_export_table_with_materialized_view.groovy
b/regression-test/suites/export_p0/test_export_table_with_materialized_view.groovy
index bc51b99a991..adba998b16f 100644
---
a/regression-test/suites/export_p0/test_export_table_with_materialized_view.groovy
+++
b/regression-test/suites/export_p0/test_export_table_with_materialized_view.groovy
@@ -114,7 +114,7 @@ suite("test_export_table_with_materialized_view", "p0") {
def outfile_url = waiting_export.call(label)
qt_select_load1 """ select * from s3(
- "uri" = "http://${s3_endpoint}${outfile_url.substring(4,
outfile_url.length() - 1)}0.parquet",
+ "uri" =
"http://${bucket}.${s3_endpoint}${outfile_url.substring(5+bucket.length(),
outfile_url.length() - 1)}0.parquet",
"s3.access_key"= "${ak}",
"s3.secret_key" = "${sk}",
"format" = "parquet",
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_different_parquet_types.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_different_parquet_types.groovy
index 4a450b99901..fbf60ed2497 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_different_parquet_types.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_different_parquet_types.groovy
@@ -16,16 +16,6 @@
// under the License.
suite("test_trino_different_parquet_types",
"p0,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String hdfs_port = context.config.otherConfigs.get("hive2HdfsPort")
@@ -185,6 +175,16 @@ suite("test_trino_different_parquet_types",
"p0,external,hive,external_docker,ex
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
try {
String catalog_name = "test_trino_different_parquet_types"
sql """drop catalog if exists ${catalog_name}"""
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_orc.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_orc.groovy
index 89255cfbc56..d4cbcbe409b 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_orc.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_orc.groovy
@@ -16,16 +16,6 @@
// under the License.
suite("test_trino_hive_orc",
"all_types,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
// Ensure that all types are parsed correctly
def select_top50 = {
@@ -88,6 +78,16 @@ suite("test_trino_hive_orc",
"all_types,external,hive,external_docker,external_d
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
try {
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String catalog_name = "test_trino_hive_orc"
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_other.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_other.groovy
index 9d3430d1ad0..6d410b2cb9a 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_other.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_other.groovy
@@ -16,16 +16,6 @@
// under the License.
suite("test_trino_hive_other",
"external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
def q01 = {
qt_q24 """ select name, count(1) as c from student group by name order
by name desc;"""
@@ -62,6 +52,16 @@ suite("test_trino_hive_other",
"external,hive,external_docker,external_docker_hi
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String hdfs_port = context.config.otherConfigs.get("hive2HdfsPort")
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_parquet.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_parquet.groovy
index 658b342dff0..748ac02b651 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_parquet.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_parquet.groovy
@@ -16,17 +16,7 @@
// under the License.
suite("test_trino_hive_parquet",
"p0,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
-
+
def q01 = {
qt_q01 """
select * from partition_table order by l_orderkey, l_partkey,
l_suppkey;
@@ -182,6 +172,16 @@ suite("test_trino_hive_parquet",
"p0,external,hive,external_docker,external_dock
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
try {
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String catalog_name = "test_trino_hive_parquet"
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_schema_evolution.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_schema_evolution.groovy
index 1eb4a0f1b9e..c803a988574 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_schema_evolution.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_schema_evolution.groovy
@@ -16,17 +16,7 @@
// under the License.
suite("test_trino_hive_schema_evolution",
"p0,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
-
+
def q_text = {
qt_q01 """
select * from schema_evo_test_text order by id;
@@ -67,6 +57,17 @@ suite("test_trino_hive_schema_evolution",
"p0,external,hive,external_docker,exte
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
+
try {
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String catalog_name = "test_trino_hive_schema_evolution"
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_serde_prop.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_serde_prop.groovy
index 8479c14fcbc..b996d94f95d 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_serde_prop.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_serde_prop.groovy
@@ -16,19 +16,19 @@
// under the License.
suite("test_trino_hive_serde_prop",
"external_docker,hive,external_docker_hive,p0,external") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
-
+
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
String catalog_name = "test_trino_hive_serde_prop"
String ex_db_name = "`stats_test`"
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tablesample_p0.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tablesample_p0.groovy
index be760d381df..8752f2f4b41 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tablesample_p0.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tablesample_p0.groovy
@@ -16,19 +16,19 @@
// under the License.
suite("test_trino_hive_tablesample_p0",
"all_types,p0,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
try {
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String catalog_name = "test_trino_hive_tablesample_p0"
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_orc.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_orc.groovy
index cafd9301753..6ee38e0021d 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_orc.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_orc.groovy
@@ -16,16 +16,6 @@
// under the License.
suite("test_trino_hive_tpch_sf1_orc",
"p0,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
String enable_file_cache = "false"
def q01 = {
@@ -857,6 +847,16 @@ order by
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String catalog_name = "test_trino_hive_tpch_sf1_orc"
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_parquet.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_parquet.groovy
index 4034ac805ef..8bd86c7def1 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_parquet.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_hive_tpch_sf1_parquet.groovy
@@ -16,16 +16,6 @@
// under the License.
suite("test_trino_hive_tpch_sf1_parquet",
"p0,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
String enable_file_cache = "false"
def q01 = {
@@ -857,6 +847,16 @@ order by
String enabled = context.config.otherConfigs.get("enableHiveTest")
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
String catalog_name = "test_trino_hive_tpch_sf1_parquet"
String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
diff --git
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_prepare_hive_data_in_case.groovy
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_prepare_hive_data_in_case.groovy
index 84049ac1db6..a8b86014aa9 100644
---
a/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_prepare_hive_data_in_case.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/hive/test_trino_prepare_hive_data_in_case.groovy
@@ -16,20 +16,20 @@
// under the License.
suite("test_trino_prepare_hive_data_in_case",
"p0,external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
-
+
String enabled = context.config.otherConfigs.get("enableHiveTest")
def catalog_name = "test_trino_prepare_hive_data_in_case"
if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
try {
String externalEnvIp =
context.config.otherConfigs.get("externalEnvIp")
String hms_port = context.config.otherConfigs.get("hive2HmsPort")
diff --git
a/regression-test/suites/external_table_p0/trino_connector/kafka/test_trino_kafka_base.groovy
b/regression-test/suites/external_table_p0/trino_connector/kafka/test_trino_kafka_base.groovy
index b1da4641866..2f61764ec12 100644
---
a/regression-test/suites/external_table_p0/trino_connector/kafka/test_trino_kafka_base.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/kafka/test_trino_kafka_base.groovy
@@ -21,17 +21,6 @@ import org.apache.kafka.clients.producer.ProducerRecord
import org.apache.kafka.clients.producer.ProducerConfig
suite("test_trino_kafka_base",
"external,kafka,external_docker,external_docker_kafka") {
- // set up trino-connector plugins
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
- }
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
// Ensure that all types are parsed correctly
def select_top50 = {
@@ -43,6 +32,18 @@ suite("test_trino_kafka_base",
"external,kafka,external_docker,external_docker_k
String enabled_trino_connector =
context.config.otherConfigs.get("enableTrinoConnectorTest")
if (enabled != null && enabled.equalsIgnoreCase("true")
&& enabled_trino_connector!= null &&
enabled_trino_connector.equalsIgnoreCase("true")) {
+ // set up trino-connector plugins
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
+
def kafkaCsvTpoics = [
"trino_kafka_basic_data"
]
diff --git
a/regression-test/suites/external_table_p0/trino_connector/test_plugins_download.groovy
b/regression-test/suites/external_table_p0/trino_connector/test_plugins_download.groovy
index 3d28612cf62..028383b6609 100644
---
a/regression-test/suites/external_table_p0/trino_connector/test_plugins_download.groovy
+++
b/regression-test/suites/external_table_p0/trino_connector/test_plugins_download.groovy
@@ -16,14 +16,17 @@
// under the License.
suite("test_plugins_download",
"external,hive,external_docker,external_docker_hive") {
- def host_ips = new ArrayList()
- String[][] backends = sql """ show backends """
- for (def b in backends) {
- host_ips.add(b[1])
+ String enabled =
context.config.otherConfigs.get("enableTrinoConnectorTest")
+ if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ def host_ips = new ArrayList()
+ String[][] backends = sql """ show backends """
+ for (def b in backends) {
+ host_ips.add(b[1])
+ }
+ String [][] frontends = sql """ show frontends """
+ for (def f in frontends) {
+ host_ips.add(f[1])
+ }
+ dispatchTrinoConnectors(host_ips.unique())
}
- String [][] frontends = sql """ show frontends """
- for (def f in frontends) {
- host_ips.add(f[1])
- }
- dispatchTrinoConnectors(host_ips.unique())
}
\ No newline at end of file
diff --git a/regression-test/suites/variant_p0/load.groovy
b/regression-test/suites/variant_p0/load.groovy
index 2b9ec514031..d74155b182d 100644
--- a/regression-test/suites/variant_p0/load.groovy
+++ b/regression-test/suites/variant_p0/load.groovy
@@ -203,7 +203,7 @@ suite("regression_test_variant", "p0"){
// 7. gh data
table_name = "ghdata"
create_table table_name
- load_json_data.call(table_name, """${getS3Url() +
'/load/ghdata_sample.json'}""")
+ load_json_data.call(table_name, """${getS3Url() +
'/regression/load/ghdata_sample.json'}""")
qt_sql_26 "select count() from ${table_name}"
// 8. json empty string
@@ -218,7 +218,7 @@ suite("regression_test_variant", "p0"){
// // // 9. btc data
// // table_name = "btcdata"
// // create_table table_name
- // // load_json_data.call(table_name, """${getS3Url() +
'/load/btc_transactions.json'}""")
+ // // load_json_data.call(table_name, """${getS3Url() +
'/regression/load/btc_transactions.json'}""")
// // qt_sql_28 "select count() from ${table_name}"
// 10. alter add variant
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]