This is an automated email from the ASF dual-hosted git repository. morningman pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/master by this push: new cd8d707e120 [test](external)move hive cases from p2 to p0 (#36855) cd8d707e120 is described below commit cd8d707e120060071047957d81c5f4d997b5f789 Author: zy-kkk <zhongy...@gmail.com> AuthorDate: Tue Jul 2 12:56:19 2024 +0800 [test](external)move hive cases from p2 to p0 (#36855) test_hive_same_db_table_name test_hive_special_char_partition test_complex_types test_wide_table --- .../test_complex_types/create_table.hql | 138 +++++ .../data/multi_catalog/test_complex_types/run.sh | 22 + .../test_hive_same_db_table_name/create_table.hql | 22 + .../test_hive_same_db_table_name/data.tar.gz | Bin 0 -> 183 bytes .../test_hive_same_db_table_name/run.sh | 12 + .../create_table.hql | 22 + .../test_hive_special_char_partition/data.tar.gz | Bin 0 -> 894 bytes .../test_hive_special_char_partition/run.sh | 12 + .../multi_catalog/test_wide_table/create_table.hql | 654 +++++++++++++++++++++ .../data/multi_catalog/test_wide_table/run.sh | 22 + .../hive/test_complex_types.out | 54 +- .../hive/test_hive_same_db_table_name.out | 15 + .../hive/test_hive_special_char_partition.out | 50 ++ .../external_table_p0/hive/test_wide_table.out | 45 ++ .../external_table_p2/hive/test_wide_table.out | 23 - .../hive/test_complex_types.groovy | 39 +- .../hive/test_hive_same_db_table_name.groovy | 31 +- .../hive/test_hive_special_char_partition.groovy | 31 +- .../hive/test_wide_table.groovy | 33 +- 19 files changed, 1132 insertions(+), 93 deletions(-) diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/create_table.hql b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/create_table.hql new file mode 100644 index 00000000000..f22d7ff246b --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/create_table.hql @@ -0,0 +1,138 @@ +create database if not exists multi_catalog; + +use multi_catalog; + +CREATE TABLE `byd`( + `id` int, + `capacity` array<double>, + `singles` map<string,double>, + `favor` struct<name:string,age:int,tip:timestamp>) +ROW FORMAT SERDE + 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe' +STORED AS INPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' +LOCATION + '/user/doris/suites/multi_catalog/byd' +TBLPROPERTIES ( + 'transient_lastDdlTime'='1690356922'); + +msck repair table byd; + +CREATE TABLE `complex_offsets_check`( + `id` int, + `array1` array<int>, + `array2` array<array<int>>, + `map1` map<string,int>, + `struct1` struct<s1:string,s2:int>) +ROW FORMAT SERDE + 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe' +STORED AS INPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' +LOCATION + '/user/doris/suites/multi_catalog/complex_offsets_check' +TBLPROPERTIES ( + 'transient_lastDdlTime'='1690974653'); + +msck repair table complex_offsets_check; + +CREATE TABLE `parquet_all_types`( + `t_null_string` string, + `t_null_varchar` varchar(65535), + `t_null_char` char(10), + `t_null_decimal_precision_2` decimal(2,1), + `t_null_decimal_precision_4` decimal(4,2), + `t_null_decimal_precision_8` decimal(8,4), + `t_null_decimal_precision_17` decimal(17,8), + `t_null_decimal_precision_18` decimal(18,8), + `t_null_decimal_precision_38` decimal(38,16), + `t_empty_string` string, + `t_string` string, + `t_empty_varchar` varchar(65535), + `t_varchar` varchar(65535), + `t_varchar_max_length` varchar(65535), + `t_char` char(10), + `t_int` int, + `t_bigint` bigint, + `t_float` float, + `t_double` double, + `t_boolean_true` boolean, + `t_boolean_false` boolean, + `t_decimal_precision_2` decimal(2,1), + `t_decimal_precision_4` decimal(4,2), + `t_decimal_precision_8` decimal(8,4), + `t_decimal_precision_17` decimal(17,8), + `t_decimal_precision_18` decimal(18,8), + `t_decimal_precision_38` decimal(38,16), + `t_binary` binary, + `t_map_string` map<string,string>, + `t_map_varchar` map<varchar(65535),varchar(65535)>, + `t_map_char` map<char(10),char(10)>, + `t_map_int` map<int,int>, + `t_map_bigint` map<bigint,bigint>, + `t_map_float` map<float,float>, + `t_map_double` map<double,double>, + `t_map_boolean` map<boolean,boolean>, + `t_map_decimal_precision_2` map<decimal(2,1),decimal(2,1)>, + `t_map_decimal_precision_4` map<decimal(4,2),decimal(4,2)>, + `t_map_decimal_precision_8` map<decimal(8,4),decimal(8,4)>, + `t_map_decimal_precision_17` map<decimal(17,8),decimal(17,8)>, + `t_map_decimal_precision_18` map<decimal(18,8),decimal(18,8)>, + `t_map_decimal_precision_38` map<decimal(38,16),decimal(38,16)>, + `t_array_string` array<string>, + `t_array_int` array<int>, + `t_array_bigint` array<bigint>, + `t_array_float` array<float>, + `t_array_double` array<double>, + `t_array_boolean` array<boolean>, + `t_array_varchar` array<varchar(65535)>, + `t_array_char` array<char(10)>, + `t_array_decimal_precision_2` array<decimal(2,1)>, + `t_array_decimal_precision_4` array<decimal(4,2)>, + `t_array_decimal_precision_8` array<decimal(8,4)>, + `t_array_decimal_precision_17` array<decimal(17,8)>, + `t_array_decimal_precision_18` array<decimal(18,8)>, + `t_array_decimal_precision_38` array<decimal(38,16)>, + `t_struct_bigint` struct<s_bigint:bigint>, + `t_complex` map<string,array<struct<s_int:int>>>, + `t_struct_nested` struct<struct_field:array<string>>, + `t_struct_null` struct<struct_field_null:string,struct_field_null2:string>, + `t_struct_non_nulls_after_nulls` struct<struct_non_nulls_after_nulls1:int,struct_non_nulls_after_nulls2:string>, + `t_nested_struct_non_nulls_after_nulls` struct<struct_field1:int,struct_field2:string,strict_field3:struct<nested_struct_field1:int,nested_struct_field2:string>>, + `t_map_null_value` map<string,string>, + `t_array_string_starting_with_nulls` array<string>, + `t_array_string_with_nulls_in_between` array<string>, + `t_array_string_ending_with_nulls` array<string>, + `t_array_string_all_nulls` array<string>) +ROW FORMAT SERDE + 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe' +STORED AS INPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' +LOCATION + '/user/doris/suites/multi_catalog/parquet_all_types' +TBLPROPERTIES ( + 'transient_lastDdlTime'='1692347490'); + +msck repair table parquet_all_types; + +CREATE TABLE `date_dict`( + `date1` date, + `date2` date, + `date3` date) +ROW FORMAT SERDE + 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe' +STORED AS INPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' +LOCATION + '/user/doris/suites/multi_catalog/date_dict' +TBLPROPERTIES ( + 'transient_lastDdlTime'='1693396885'); + +msck repair table date_dict; diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh new file mode 100755 index 00000000000..947fb796eeb --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh @@ -0,0 +1,22 @@ +#!/bin/bash +set -x + +CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)" + +if [[ ! -d "${CUR_DIR}/data" ]]; then + echo "${CUR_DIR}/data does not exist" + cd "${CUR_DIR}" && rm -f data.tar.gz \ + && curl -O https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_complex_types/data.tar.gz \ + && tar xzf data.tar.gz + cd - +else + echo "${CUR_DIR}/data exist, continue !" +fi + +## mkdir and put data to hdfs +hadoop fs -mkdir -p /user/doris/suites/multi_catalog/ +hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/ + +# create table +hive -f "${CUR_DIR}"/create_table.hql + diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/create_table.hql b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/create_table.hql new file mode 100644 index 00000000000..3d672596230 --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/create_table.hql @@ -0,0 +1,22 @@ +create database if not exists multi_catalog; + +use multi_catalog; + +CREATE TABLE `region`( + `r_regionkey` int, + `r_name` char(25)) +ROW FORMAT SERDE + 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +WITH SERDEPROPERTIES ( + 'field.delim'='|', + 'serialization.format'='|') +STORED AS INPUTFORMAT + 'org.apache.hadoop.mapred.TextInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +LOCATION + '/user/doris/suites/multi_catalog/region' +TBLPROPERTIES ( + 'transient_lastDdlTime'='1670483235'); + +msck repair table region; diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/data.tar.gz b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/data.tar.gz new file mode 100644 index 00000000000..c5367912ddf Binary files /dev/null and b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/data.tar.gz differ diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/run.sh b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/run.sh new file mode 100644 index 00000000000..a52ee0579a2 --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/run.sh @@ -0,0 +1,12 @@ +#!/bin/bash +set -x + +CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)" + +## mkdir and put data to hdfs +cd "${CUR_DIR}" && rm -rf data/ && tar xzf data.tar.gz +hadoop fs -mkdir -p /user/doris/suites/multi_catalog/ +hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/ + +# create table +hive -f "${CUR_DIR}"/create_table.hql diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/create_table.hql b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/create_table.hql new file mode 100644 index 00000000000..2631d1360c9 --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/create_table.hql @@ -0,0 +1,22 @@ +create database if not exists multi_catalog; + +use multi_catalog; + +CREATE TABLE `special_character_1_partition`( + `name` string) +PARTITIONED BY ( + `part` string) +ROW FORMAT SERDE + 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe' +STORED AS INPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' +LOCATION + '/user/doris/suites/multi_catalog/special_character_1_partition' +TBLPROPERTIES ( + 'transient_lastDdlTime'='1689575322'); + +set hive.msck.path.validation=ignore; + +msck repair table special_character_1_partition; diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/data.tar.gz b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/data.tar.gz new file mode 100644 index 00000000000..bce0219357e Binary files /dev/null and b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/data.tar.gz differ diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/run.sh b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/run.sh new file mode 100644 index 00000000000..a52ee0579a2 --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/run.sh @@ -0,0 +1,12 @@ +#!/bin/bash +set -x + +CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)" + +## mkdir and put data to hdfs +cd "${CUR_DIR}" && rm -rf data/ && tar xzf data.tar.gz +hadoop fs -mkdir -p /user/doris/suites/multi_catalog/ +hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/ + +# create table +hive -f "${CUR_DIR}"/create_table.hql diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/create_table.hql b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/create_table.hql new file mode 100644 index 00000000000..e57a9148383 --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/create_table.hql @@ -0,0 +1,654 @@ +create database if not exists multi_catalog; + +use multi_catalog; + +CREATE TABLE `wide_table1_orc`( + `col1` decimal(16,0), + `col2` string, + `col3` string, + `col4` string, + `col5` string, + `col6` string, + `col7` string, + `col8` string, + `col9` string, + `col10` string, + `col11` string, + `col12` string, + `col13` string, + `col14` string, + `col15` decimal(2,0), + `col16` decimal(2,0), + `col17` decimal(2,0), + `col18` decimal(2,0), + `col19` decimal(2,0), + `col20` decimal(10,0), + `col21` string, + `col22` decimal(10,0), + `col23` string, + `col24` decimal(12,0), + `col25` string, + `col26` decimal(8,0), + `col27` decimal(8,0), + `col28` decimal(8,0), + `col29` decimal(8,0), + `col30` decimal(2,0), + `col31` string, + `col32` decimal(12,0), + `col33` string, + `col34` decimal(8,0), + `col35` decimal(12,0), + `col36` string, + `col37` decimal(12,0), + `col38` string, + `col39` decimal(12,0), + `col40` string, + `col41` int, + `col42` int, + `col43` string, + `col44` decimal(12,0), + `col45` string, + `col46` int, + `col47` decimal(12,0), + `col48` string, + `col49` decimal(8,0), + `col50` string, + `col51` string, + `col52` string, + `col53` string, + `col54` string, + `col55` string, + `col56` string, + `col57` string, + `col58` string, + `col59` string, + `col60` string, + `col61` string, + `col62` string, + `col63` string, + `col64` decimal(12,0), + `col65` string, + `col66` string, + `col67` string, + `col68` decimal(12,0), + `col69` string, + `col70` decimal(16,2), + `col71` decimal(8,0), + `col72` string, + `col73` decimal(16,2), + `col74` decimal(16,2), + `col75` decimal(16,2), + `col76` decimal(16,2), + `col77` string, + `col78` string, + `col79` string, + `col80` decimal(16,2), + `col81` decimal(16,2), + `col82` decimal(16,2), + `col83` decimal(16,2), + `col84` decimal(16,2), + `col85` decimal(16,2), + `col86` decimal(16,2), + `col87` decimal(16,2), + `col88` decimal(16,2), + `col89` decimal(16,2), + `col90` decimal(16,2), + `col91` decimal(16,2), + `col92` decimal(16,2), + `col93` decimal(16,2), + `col94` decimal(16,2), + `col95` decimal(16,2), + `col96` decimal(16,2), + `col97` decimal(19,6), + `col98` decimal(12,5), + `col99` decimal(12,5), + `col100` decimal(12,5), + `col101` decimal(16,2), + `col102` decimal(16,2), + `col103` decimal(16,2), + `col104` decimal(16,2), + `col105` decimal(16,2), + `col106` decimal(16,2), + `col107` decimal(16,2), + `col108` decimal(16,2), + `col109` decimal(16,2), + `col110` decimal(16,2), + `col111` decimal(16,2), + `col112` decimal(16,2), + `col113` decimal(16,2), + `col114` decimal(16,2), + `col115` decimal(16,2), + `col116` decimal(16,2), + `col117` decimal(8,0), + `col118` decimal(8,0), + `col119` decimal(8,0), + `col120` decimal(8,0), + `col121` decimal(8,0), + `col122` decimal(8,0), + `col123` decimal(8,0), + `col124` decimal(8,0), + `col125` decimal(8,0), + `col126` decimal(8,0), + `col127` decimal(8,0), + `col128` decimal(8,0), + `col129` decimal(8,0), + `col130` decimal(8,0), + `col131` decimal(8,0), + `col132` decimal(16,2), + `col133` decimal(16,2), + `col134` decimal(16,2), + `col135` decimal(16,2), + `col136` decimal(16,2), + `col137` decimal(16,2), + `col138` decimal(16,2), + `col139` decimal(16,2), + `col140` decimal(16,2), + `col141` decimal(16,2), + `col142` decimal(16,2), + `col143` decimal(16,2), + `col144` decimal(16,2), + `col145` decimal(16,2), + `col146` decimal(16,2), + `col147` decimal(16,2), + `col148` decimal(16,2), + `col149` decimal(16,2), + `col150` decimal(16,2), + `col151` decimal(16,2), + `col152` decimal(16,2), + `col153` decimal(16,2), + `col154` decimal(16,2), + `col155` decimal(16,2), + `col156` decimal(16,2), + `col157` decimal(16,2), + `col158` decimal(16,2), + `col159` decimal(16,2), + `col160` decimal(16,2), + `col161` decimal(16,2), + `col162` decimal(16,2), + `col163` decimal(16,2), + `col164` decimal(16,2), + `col165` decimal(16,2), + `col166` decimal(16,2), + `col167` decimal(16,2), + `col168` decimal(16,2), + `col169` decimal(16,2), + `col170` decimal(16,2), + `col171` decimal(16,2), + `col172` decimal(16,2), + `col173` decimal(16,2), + `col174` decimal(16,2), + `col175` decimal(16,2), + `col176` decimal(16,2), + `col177` decimal(16,2), + `col178` decimal(16,2), + `col179` decimal(16,2), + `col180` decimal(16,2), + `col181` decimal(16,2), + `col182` decimal(16,2), + `col183` decimal(16,2), + `col184` decimal(16,2), + `col185` decimal(16,2), + `col186` decimal(16,2), + `col187` decimal(16,2), + `col188` decimal(16,2), + `col189` decimal(16,2), + `col190` decimal(16,2), + `col191` decimal(16,2), + `col192` decimal(16,2), + `col193` decimal(16,2), + `col194` decimal(16,2), + `col195` decimal(16,2), + `col196` decimal(16,2), + `col197` decimal(16,2), + `col198` decimal(16,2), + `col199` decimal(16,2), + `col200` decimal(16,2), + `col201` decimal(16,2), + `col202` decimal(16,2), + `col203` decimal(16,2), + `col204` decimal(16,2), + `col205` decimal(16,2), + `col206` decimal(16,2), + `col207` decimal(16,2), + `col208` decimal(16,2), + `col209` decimal(16,2), + `col210` decimal(16,2), + `col211` decimal(16,2), + `col212` decimal(16,2), + `col213` decimal(16,2), + `col214` decimal(16,2), + `col215` decimal(16,2), + `col216` decimal(16,2), + `col217` decimal(16,2), + `col218` decimal(16,2), + `col219` decimal(16,2), + `col220` decimal(16,2), + `col221` decimal(16,2), + `col222` decimal(16,2), + `col223` decimal(16,2), + `col224` decimal(16,2), + `col225` decimal(16,2), + `col226` decimal(16,2), + `col227` decimal(16,2), + `col228` decimal(16,2), + `col229` decimal(16,2), + `col230` decimal(16,2), + `col231` decimal(22,6), + `col232` decimal(22,6), + `col233` decimal(22,6), + `col234` decimal(22,6), + `col235` decimal(22,6), + `col236` decimal(22,6), + `col237` decimal(22,6), + `col238` decimal(22,6), + `col239` decimal(22,6), + `col240` decimal(22,6), + `col241` decimal(22,6), + `col242` decimal(22,6), + `col243` decimal(22,6), + `col244` decimal(22,6), + `col245` decimal(22,6), + `col246` decimal(22,6), + `col247` decimal(22,6), + `col248` decimal(22,6), + `col249` decimal(22,6), + `col250` decimal(22,6), + `col251` decimal(22,6), + `col252` decimal(16,2), + `col253` decimal(16,2), + `col254` decimal(16,2), + `col255` decimal(16,2), + `col256` decimal(16,2), + `col257` decimal(16,2), + `col258` decimal(16,2), + `col259` decimal(16,2), + `col260` decimal(16,2), + `col261` decimal(16,2), + `col262` decimal(16,2), + `col263` decimal(16,2), + `col264` decimal(16,2), + `col265` decimal(16,2), + `col266` decimal(16,2), + `col267` decimal(16,2), + `col268` decimal(16,2), + `col269` decimal(16,2), + `col270` decimal(16,2), + `col271` decimal(16,2), + `col272` decimal(16,2), + `col273` decimal(16,2), + `col274` decimal(16,2), + `col275` decimal(16,2), + `col276` decimal(16,2), + `col277` decimal(16,2), + `col278` decimal(16,2), + `col279` decimal(16,2), + `col280` decimal(16,2), + `col281` decimal(16,2), + `col282` decimal(16,2), + `col283` decimal(16,2), + `col284` decimal(16,2), + `col285` decimal(16,2), + `col286` decimal(16,2), + `col287` decimal(16,2), + `col288` decimal(16,2), + `col289` decimal(16,2), + `col290` decimal(16,2), + `col291` decimal(16,2), + `col292` decimal(16,2), + `col293` decimal(16,2), + `col294` decimal(16,2), + `col295` decimal(16,2), + `col296` decimal(16,2), + `col297` decimal(16,2), + `col298` decimal(16,2), + `col299` decimal(16,2), + `col300` decimal(16,2), + `col301` decimal(16,2), + `col302` decimal(16,2), + `col303` decimal(16,2), + `col304` decimal(16,2), + `col305` decimal(16,2), + `col306` decimal(16,2), + `col307` decimal(16,2), + `col308` decimal(16,2), + `col309` decimal(16,2), + `col310` decimal(16,2), + `col311` decimal(16,2), + `col312` decimal(16,2), + `col313` decimal(16,2), + `col314` decimal(16,2), + `col315` decimal(16,2), + `col316` decimal(16,2), + `col317` decimal(16,2), + `col318` decimal(16,2), + `col319` decimal(16,2), + `col320` decimal(16,2), + `col321` decimal(16,2), + `col322` decimal(16,2), + `col323` decimal(16,2), + `col324` decimal(16,2), + `col325` decimal(16,2), + `col326` decimal(16,2), + `col327` decimal(16,2), + `col328` decimal(16,2), + `col329` decimal(16,2), + `col330` decimal(16,2), + `col331` decimal(16,2), + `col332` decimal(16,2), + `col333` decimal(16,2), + `col334` decimal(16,2), + `col335` decimal(16,2), + `col336` decimal(16,2), + `col337` decimal(16,2), + `col338` decimal(16,2), + `col339` decimal(22,6), + `col340` decimal(22,6), + `col341` decimal(22,6), + `col342` decimal(22,6), + `col343` decimal(22,6), + `col344` decimal(22,6), + `col345` decimal(22,6), + `col346` decimal(22,6), + `col347` decimal(22,6), + `col348` decimal(22,6), + `col349` decimal(22,6), + `col350` decimal(22,6), + `col351` decimal(22,6), + `col352` decimal(22,6), + `col353` decimal(16,2), + `col354` decimal(16,2), + `col355` decimal(16,2), + `col356` decimal(16,2), + `col357` decimal(16,2), + `col358` decimal(16,2), + `col359` decimal(16,2), + `col360` decimal(16,2), + `col361` decimal(16,2), + `col362` decimal(16,2), + `col363` decimal(16,2), + `col364` decimal(16,2), + `col365` decimal(16,2), + `col366` decimal(16,2), + `col367` decimal(16,2), + `col368` decimal(16,2), + `col369` decimal(16,2), + `col370` decimal(16,2), + `col371` decimal(16,2), + `col372` decimal(16,2), + `col373` decimal(16,2), + `col374` decimal(16,2), + `col375` decimal(16,2), + `col376` decimal(16,2), + `col377` decimal(16,2), + `col378` decimal(16,2), + `col379` decimal(16,2), + `col380` decimal(16,2), + `col381` decimal(16,2), + `col382` decimal(16,2), + `col383` decimal(16,2), + `col384` decimal(16,2), + `col385` decimal(16,2), + `col386` decimal(16,2), + `col387` decimal(16,2), + `col388` decimal(16,2), + `col389` decimal(16,2), + `col390` decimal(16,2), + `col391` decimal(16,2), + `col392` decimal(16,2), + `col393` decimal(16,2), + `col394` decimal(16,2), + `col395` decimal(8,0), + `col396` decimal(8,0), + `col397` string, + `col398` string, + `col399` decimal(16,2), + `col400` decimal(16,2), + `col401` decimal(16,2), + `col402` decimal(16,2), + `col403` decimal(16,2), + `col404` decimal(16,2), + `col405` decimal(16,2), + `col406` decimal(16,2), + `col407` decimal(16,2), + `col408` decimal(16,2), + `col409` decimal(16,2), + `col410` decimal(16,2), + `col411` decimal(16,2), + `col412` decimal(16,2), + `col413` decimal(16,2), + `col414` decimal(16,2), + `col415` decimal(16,2), + `col416` decimal(16,2), + `col417` decimal(16,2), + `col418` decimal(16,2), + `col419` decimal(16,2), + `col420` int, + `col421` int, + `col422` int, + `col423` decimal(16,2), + `col424` decimal(16,2), + `col425` decimal(16,2), + `col426` decimal(16,2), + `col427` decimal(16,2), + `col428` decimal(16,2), + `col429` string, + `col430` string, + `col431` string, + `col432` decimal(16,2), + `col433` decimal(16,2), + `col434` decimal(16,2), + `col435` decimal(16,2), + `col436` decimal(16,2), + `col437` decimal(16,2), + `col438` decimal(16,2), + `col439` decimal(16,2), + `col440` decimal(22,6), + `col441` decimal(22,6), + `col442` decimal(16,3), + `col443` decimal(16,3), + `col444` decimal(16,3), + `col445` decimal(16,2), + `col446` decimal(16,2), + `col447` decimal(9,6), + `col448` decimal(9,6), + `col449` decimal(16,2), + `col450` decimal(16,2), + `col451` decimal(16,2), + `col452` decimal(16,2), + `col453` decimal(16,2), + `col454` int, + `col455` decimal(16,2), + `col456` decimal(16,2), + `col457` decimal(16,2), + `col458` decimal(16,2), + `col459` decimal(16,2), + `col460` decimal(16,2), + `col461` decimal(16,2), + `col462` decimal(16,2), + `col463` decimal(22,6), + `col464` decimal(22,6), + `col465` decimal(22,6), + `col466` decimal(22,6), + `col467` decimal(22,6), + `col468` decimal(22,6), + `col469` decimal(22,6), + `col470` decimal(22,6), + `col471` decimal(16,2), + `col472` decimal(16,2), + `col473` decimal(16,2), + `col474` decimal(16,2), + `col475` decimal(16,2), + `col476` decimal(16,2), + `col477` decimal(16,2), + `col478` decimal(16,2), + `col479` decimal(16,2), + `col480` decimal(16,2), + `col481` decimal(16,2), + `col482` decimal(16,2), + `col483` decimal(16,2), + `col484` decimal(16,2), + `col485` decimal(16,2), + `col486` decimal(16,2), + `col487` decimal(16,2), + `col488` decimal(16,2), + `col489` decimal(16,2), + `col490` decimal(16,2), + `col491` decimal(16,2), + `col492` decimal(16,2), + `col493` decimal(16,2), + `col494` decimal(16,2), + `col495` decimal(16,2), + `col496` decimal(16,2), + `col497` decimal(16,2), + `col498` decimal(16,2), + `col499` decimal(16,2), + `col500` decimal(16,2), + `col501` decimal(16,2), + `col502` decimal(16,2), + `col503` decimal(16,2), + `col504` decimal(16,2), + `col505` decimal(16,2), + `col506` decimal(16,2), + `col507` decimal(16,2), + `col508` decimal(8,0), + `col509` decimal(8,0), + `col510` decimal(8,0), + `col511` decimal(8,0), + `col512` decimal(8,0), + `col513` decimal(8,0), + `col514` decimal(8,0), + `col515` decimal(8,0), + `col516` decimal(8,0), + `col517` decimal(8,0), + `col518` decimal(8,0), + `col519` decimal(8,0), + `col520` decimal(8,0), + `col521` decimal(8,0), + `col522` decimal(8,0), + `col523` decimal(8,0), + `col524` decimal(8,0), + `col525` decimal(8,0), + `col526` decimal(8,0), + `col527` decimal(8,0), + `col528` decimal(8,0), + `col529` decimal(8,0), + `col530` decimal(16,2), + `col531` decimal(16,2), + `col532` decimal(16,2), + `col533` decimal(16,2), + `col534` decimal(22,6), + `col535` decimal(16,2), + `col536` decimal(16,2), + `col537` decimal(16,2), + `col538` decimal(16,2), + `col539` decimal(22,6), + `col540` decimal(22,6), + `col541` decimal(22,6), + `col542` string, + `col543` decimal(16,2), + `col544` decimal(16,2), + `col545` decimal(16,2), + `col546` decimal(16,2), + `col547` decimal(16,2), + `col548` decimal(16,2), + `col549` decimal(16,2), + `col550` decimal(16,2), + `col551` decimal(16,2), + `col552` decimal(16,2), + `col553` decimal(16,2), + `col554` decimal(16,2), + `col555` decimal(16,2), + `col556` decimal(16,2), + `col557` string, + `col558` string, + `col559` string, + `col560` string, + `col561` string, + `col562` string, + `col563` string, + `col564` string, + `col565` decimal(16,2), + `col566` decimal(22,6), + `col567` decimal(22,6), + `col568` decimal(22,6), + `col569` string, + `col570` string, + `col571` decimal(16,2), + `col572` decimal(12,0), + `col573` string, + `col574` string, + `col575` decimal(22,6), + `col576` decimal(22,6), + `col577` decimal(22,6), + `col578` decimal(16,2), + `col579` decimal(22,2), + `col580` decimal(22,2), + `col581` decimal(22,2), + `col582` decimal(16,2), + `col583` decimal(16,2), + `col584` decimal(16,2), + `col585` decimal(16,2), + `col586` decimal(16,2), + `col587` decimal(16,2), + `col588` decimal(16,2), + `col589` decimal(16,2), + `col590` decimal(16,2), + `col591` decimal(16,2), + `col592` decimal(16,2), + `col593` decimal(16,2), + `col594` decimal(16,2), + `col595` decimal(16,2), + `col596` decimal(16,2), + `col597` string, + `col598` string, + `col599` decimal(16,2), + `col600` decimal(16,2), + `col601` decimal(16,2), + `col602` decimal(16,2), + `col603` decimal(16,2), + `col604` decimal(16,2), + `col605` decimal(16,2), + `col606` decimal(16,2), + `col607` int, + `col608` string, + `col609` string, + `col610` string, + `col611` decimal(22,2), + `col612` string, + `col613` string, + `col614` string, + `col615` string, + `col616` string, + `col617` string, + `col618` string, + `col619` string, + `col620` string, + `col621` string, + `col622` string, + `col623` string, + `col624` string, + `col625` string, + `col626` string, + `col627` string, + `col628` string, + `col629` decimal(16,2), + `col630` decimal(16,2), + `col631` string, + `col632` string, + `col633` string, + `col634` string, + `col635` string, + `col636` string, + `col637` decimal(16,2)) +ROW FORMAT SERDE + 'org.apache.hadoop.hive.ql.io.orc.OrcSerde' +STORED AS INPUTFORMAT + 'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat' +LOCATION + '/user/doris/suites/multi_catalog/wide_table1_orc' +TBLPROPERTIES ( + 'transient_lastDdlTime'='1680503244'); + +msck repair table wide_table1_orc; diff --git a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh new file mode 100755 index 00000000000..bfaeadfad06 --- /dev/null +++ b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh @@ -0,0 +1,22 @@ +#!/bin/bash +set -x + +CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)" + +if [[ ! -d "${CUR_DIR}/data" ]]; then + echo "${CUR_DIR}/data does not exist" + cd "${CUR_DIR}" && rm -f data.tar.gz \ + && curl -O https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_wide_table/data.tar.gz \ + && tar xzf data.tar.gz + cd - +else + echo "${CUR_DIR}/data exist, continue !" +fi + +## mkdir and put data to hdfs +hadoop fs -mkdir -p /user/doris/suites/multi_catalog/ +hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/ + +# create table +hive -f "${CUR_DIR}"/create_table.hql + diff --git a/regression-test/data/external_table_p2/hive/test_complex_types.out b/regression-test/data/external_table_p0/hive/test_complex_types.out similarity index 50% rename from regression-test/data/external_table_p2/hive/test_complex_types.out rename to regression-test/data/external_table_p0/hive/test_complex_types.out index 31cff0e45b2..4b94a4f1511 100644 --- a/regression-test/data/external_table_p2/hive/test_complex_types.out +++ b/regression-test/data/external_table_p0/hive/test_complex_types.out @@ -17,12 +17,6 @@ -- !array_max -- 11028 --- !array_filter -- -11028 - --- !array_last -- -0.9899828598260161 - -- !null_struct_element_orc -- 0 @@ -41,11 +35,53 @@ -- !array_max_orc -- 11028 --- !array_filter_orc -- +-- !offsets_check -- +0 [1, 2] [[], [3], null] {"a":1, "b":2} {"s1": "e", "s2": null} +1 [] [] {} \N +2 \N \N \N {"s1": "h", "s2": 10} +3 [5, null] [[6, 7], [8, null], null] {"f":1, "g":null} {"s1": null, "s2": 9} + +-- !map_with_nullable_key -- +\N \N \N \N \N \N \N \N \N test test aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa [...] + +-- !date_dict -- +2036-12-28 1898-12-28 2539-12-28 + +-- !null_struct_element -- +0 + +-- !map_key_select -- +38111 0.770169659057425 + +-- !map_keys -- +["9wXr9n-TBm9Wyt-r8H-SkAq", "CPDH4G-ZXGPkku-3wY-ktaQ", "RvNlMt-HHjHN5M-VjP-xHAI", "qKIhKy-Ws344os-haX-2pmT", "DOJJ5l-UEkwVMs-x9F-HifD", "m871g8-1eFi7jt-oBq-S0yc", "wXugVP-v2fc6IF-DeU-On3T", "B0mXFX-QvgUgo7-Dih-6rDu", "E9zv3F-xMqSbMa-il4-FuDg", "msuFIN-ZkKO8TY-tu4-veH0", "0rSUyl-Un07aIW-KAx-WHnX", "XvbmO8-WA6oAqc-ihc-s8IL", "G6B6RD-AicAlZb-16u-Pn1I", "coDK0Q-tMg1294-JMQ-ZWQu", "4c0aWh-yhL6BOX-rRu-1n0r", "G4iUcG-ZhWw62v-VLt-n6lH", "IIB7qD-WQistwT-Vux-0c9B", "7cTyuR-5ssXm2S-sJR-JTIZ", "3KPh [...] + +-- !map_values -- +[0.9805502029231666, 0.5330291595754054, 0.30024744873379805, 0.48563601750302665, 0.7687106425158624, 0.6993506644925102, 0.2849354808825807, 0.3473417455186141, 0.13500129443045072, 0.9708132103700939, 0.18583042639943448, 0.4886337264552073, 0.36354741695157655, 0.5640845268971175, 0.1374134087807577, 0.7766547647451623, 0.5835323296668318, 0.3654459547110349, 0.5479776709993764, 0.8379932542117192, 0.15665046278350814, 0.03371222042250388, 0.1699781825927229, 0.3579630495075078, 0.02 [...] + +-- !map_contains_key -- +1077 [0.7805560995873845, 0.9303489002269559, 0.2529522997521877, 0.662270811026298, 0.664725297532439, 0.1019441091764477, 0.9614059300688174, 0.5278126009983843, 0.5287505841216708, 0.426116738236779, 0.42300502393871175, 0.5327026330053651, 0.6025481777942603, 0.2710733647257627, 0.613792118138183, 0.0021003027835629906, 0.3200675048728582, 0.5485611014660204, 0.5121510581313707, 0.5145136652805358] {"9wXr9n-TBm9Wyt-r8H-SkAq":0.9338329010480995, "CPDH4G-ZXGPkku-3wY-ktaQ":0.43552569633 [...] + +-- !array_max -- 11028 --- !array_last_orc -- -0.9899828598260161 +-- !null_struct_element_orc -- +0 + +-- !map_key_select_orc -- +38111 0.770169659057425 + +-- !map_keys_orc -- +["9wXr9n-TBm9Wyt-r8H-SkAq", "CPDH4G-ZXGPkku-3wY-ktaQ", "RvNlMt-HHjHN5M-VjP-xHAI", "qKIhKy-Ws344os-haX-2pmT", "DOJJ5l-UEkwVMs-x9F-HifD", "m871g8-1eFi7jt-oBq-S0yc", "wXugVP-v2fc6IF-DeU-On3T", "B0mXFX-QvgUgo7-Dih-6rDu", "E9zv3F-xMqSbMa-il4-FuDg", "msuFIN-ZkKO8TY-tu4-veH0", "0rSUyl-Un07aIW-KAx-WHnX", "XvbmO8-WA6oAqc-ihc-s8IL", "G6B6RD-AicAlZb-16u-Pn1I", "coDK0Q-tMg1294-JMQ-ZWQu", "4c0aWh-yhL6BOX-rRu-1n0r", "G4iUcG-ZhWw62v-VLt-n6lH", "IIB7qD-WQistwT-Vux-0c9B", "7cTyuR-5ssXm2S-sJR-JTIZ", "3KPh [...] + +-- !map_values_orc -- +[0.9805502029231666, 0.5330291595754054, 0.30024744873379805, 0.48563601750302665, 0.7687106425158624, 0.6993506644925102, 0.2849354808825807, 0.3473417455186141, 0.13500129443045072, 0.9708132103700939, 0.18583042639943448, 0.4886337264552073, 0.36354741695157655, 0.5640845268971175, 0.1374134087807577, 0.7766547647451623, 0.5835323296668318, 0.3654459547110349, 0.5479776709993764, 0.8379932542117192, 0.15665046278350814, 0.03371222042250388, 0.1699781825927229, 0.3579630495075078, 0.02 [...] + +-- !map_contains_key_orc -- +1077 [0.7805560995873845, 0.9303489002269559, 0.2529522997521877, 0.662270811026298, 0.664725297532439, 0.1019441091764477, 0.9614059300688174, 0.5278126009983843, 0.5287505841216708, 0.426116738236779, 0.42300502393871175, 0.5327026330053651, 0.6025481777942603, 0.2710733647257627, 0.613792118138183, 0.0021003027835629906, 0.3200675048728582, 0.5485611014660204, 0.5121510581313707, 0.5145136652805358] {"9wXr9n-TBm9Wyt-r8H-SkAq":0.9338329010480995, "CPDH4G-ZXGPkku-3wY-ktaQ":0.43552569633 [...] + +-- !array_max_orc -- +11028 -- !offsets_check -- 0 [1, 2] [[], [3], null] {"a":1, "b":2} {"s1": "e", "s2": null} diff --git a/regression-test/data/external_table_p2/hive/test_hive_same_db_table_name.out b/regression-test/data/external_table_p0/hive/test_hive_same_db_table_name.out similarity index 62% rename from regression-test/data/external_table_p2/hive/test_hive_same_db_table_name.out rename to regression-test/data/external_table_p0/hive/test_hive_same_db_table_name.out index 03e8be6475d..b3ff544066f 100644 --- a/regression-test/data/external_table_p2/hive/test_hive_same_db_table_name.out +++ b/regression-test/data/external_table_p0/hive/test_hive_same_db_table_name.out @@ -14,3 +14,18 @@ -- !6 -- 1 name1 +-- !1 -- + +-- !2 -- + +-- !3 -- + +-- !4 -- +1 name1 + +-- !5 -- +1 name1 + +-- !6 -- +1 name1 + diff --git a/regression-test/data/external_table_p2/hive/test_hive_special_char_partition.out b/regression-test/data/external_table_p0/hive/test_hive_special_char_partition.out similarity index 54% rename from regression-test/data/external_table_p2/hive/test_hive_special_char_partition.out rename to regression-test/data/external_table_p0/hive/test_hive_special_char_partition.out index 0bd26b12760..f81719d2d0e 100644 --- a/regression-test/data/external_table_p2/hive/test_hive_special_char_partition.out +++ b/regression-test/data/external_table_p0/hive/test_hive_special_char_partition.out @@ -49,3 +49,53 @@ name6 2023%01%01 -- !13 -- name# 2023#01#01 +-- !1 -- +name# 2023#01#01 +name1 2023/01/01 +name10 2023<01><01> +name11 2023\\01\\01 +name12 2023.01.01 +name2 2023 01 01 +name3 2023:01:01 +name4 2023?01?01 +name5 2023=01=01 +name6 2023%01%01 +name8 2023"01"01 +name9 2023'01'01 + +-- !2 -- +name2 + +-- !3 -- +name1 + +-- !4 -- +name4 2023?01?01 + +-- !5 -- +name12 2023.01.01 + +-- !6 -- +name10 2023<01><01> + +-- !7 -- +name3 2023:01:01 + +-- !8 -- +name5 2023=01=01 + +-- !9 -- +name8 2023"01"01 + +-- !10 -- +name9 2023'01'01 + +-- !11 -- +name11 2023\\01\\01 + +-- !12 -- +name6 2023%01%01 + +-- !13 -- +name# 2023#01#01 + diff --git a/regression-test/data/external_table_p0/hive/test_wide_table.out b/regression-test/data/external_table_p0/hive/test_wide_table.out new file mode 100644 index 00000000000..9b5b21a8d98 --- /dev/null +++ b/regression-test/data/external_table_p0/hive/test_wide_table.out @@ -0,0 +1,45 @@ +-- This file is automatically generated. You should know what you did if you want to edit this +-- !01 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !02 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !03 -- + +-- !04 -- + +-- !05 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !06 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !07 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !08 -- +9999999541515682 99999218685068.86 99999869 221095586.11 27.54254 61077635638.763621 + +-- !01 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !02 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !03 -- + +-- !04 -- + +-- !05 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !06 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !07 -- +6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 + +-- !08 -- +9999999541515682 99999218685068.86 99999869 221095586.11 27.54254 61077635638.763621 + diff --git a/regression-test/data/external_table_p2/hive/test_wide_table.out b/regression-test/data/external_table_p2/hive/test_wide_table.out deleted file mode 100644 index 143aeb9bf50..00000000000 --- a/regression-test/data/external_table_p2/hive/test_wide_table.out +++ /dev/null @@ -1,23 +0,0 @@ --- This file is automatically generated. You should know what you did if you want to edit this --- !01 -- -6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 - --- !02 -- -6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 - --- !03 -- - --- !04 -- - --- !05 -- -6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 - --- !06 -- -6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 - --- !07 -- -6117920261 28156890937818.64 11058113 84788841307158.93 9988065.83660 8116313253956313.527443 - --- !08 -- -9999999541515682.000000000 99999218685068.860000000 99999869.000000000 221095586.110000000 27.542540000 61077635638.763621000 - diff --git a/regression-test/suites/external_table_p2/hive/test_complex_types.groovy b/regression-test/suites/external_table_p0/hive/test_complex_types.groovy similarity index 64% rename from regression-test/suites/external_table_p2/hive/test_complex_types.groovy rename to regression-test/suites/external_table_p0/hive/test_complex_types.groovy index 1a90570ff42..e8cba5df36d 100644 --- a/regression-test/suites/external_table_p2/hive/test_complex_types.groovy +++ b/regression-test/suites/external_table_p0/hive/test_complex_types.groovy @@ -15,20 +15,23 @@ // specific language governing permissions and limitations // under the License. -suite("test_complex_types", "p2,external,hive,external_remote,external_remote_hive") { - String enabled = context.config.otherConfigs.get("enableExternalHiveTest") - if (enabled != null && enabled.equalsIgnoreCase("true")) { - String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost") - String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort") - String catalog_name = "test_complex_types" - sql """drop catalog if exists ${catalog_name};""" - sql """ - create catalog if not exists ${catalog_name} properties ( - 'type'='hms', - 'hadoop.username' = 'hadoop', - 'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}' - ); - """ +suite("test_complex_types", "p0,external,hive,external_docker,external_docker_hive") { + String enabled = context.config.otherConfigs.get("enableHiveTest") + if (enabled == null || !enabled.equalsIgnoreCase("true")) { + logger.info("diable Hive test.") + return; + } + + for (String hivePrefix : ["hive2", "hive3"]) { + String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort") + String catalog_name = "${hivePrefix}_test_complex_types" + String externalEnvIp = context.config.otherConfigs.get("externalEnvIp") + + sql """drop catalog if exists ${catalog_name}""" + sql """create catalog if not exists ${catalog_name} properties ( + "type"="hms", + 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}' + );""" logger.info("catalog " + catalog_name + " created") sql """switch ${catalog_name};""" logger.info("switched to catalog " + catalog_name) @@ -47,9 +50,9 @@ suite("test_complex_types", "p2,external,hive,external_remote,external_remote_hi qt_array_max """select count(array_max(capacity)) from byd where array_max(capacity) > 0.99""" - qt_array_filter """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))""" + // qt_array_filter """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))""" - qt_array_last """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99""" + // qt_array_last """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99""" qt_null_struct_element_orc """select count(struct_element(favor, 'tip')) from byd where id % 13 = 0""" @@ -63,9 +66,9 @@ suite("test_complex_types", "p2,external,hive,external_remote,external_remote_hi qt_array_max_orc """select count(array_max(capacity)) from byd where array_max(capacity) > 0.99""" - qt_array_filter_orc """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))""" + // qt_array_filter_orc """select count(array_size(array_filter(i -> (i > 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99), capacity))""" - qt_array_last_orc """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99""" + // qt_array_last_orc """select max(array_last(i -> i > 0, capacity)) from byd where array_last(i -> i > 0, capacity) < 0.99""" qt_offsets_check """select * from complex_offsets_check order by id""" diff --git a/regression-test/suites/external_table_p2/hive/test_hive_same_db_table_name.groovy b/regression-test/suites/external_table_p0/hive/test_hive_same_db_table_name.groovy similarity index 66% rename from regression-test/suites/external_table_p2/hive/test_hive_same_db_table_name.groovy rename to regression-test/suites/external_table_p0/hive/test_hive_same_db_table_name.groovy index 1ae209f53f1..d881facbba2 100644 --- a/regression-test/suites/external_table_p2/hive/test_hive_same_db_table_name.groovy +++ b/regression-test/suites/external_table_p0/hive/test_hive_same_db_table_name.groovy @@ -15,20 +15,23 @@ // specific language governing permissions and limitations // under the License. -suite("test_hive_same_db_table_name", "p2,external,hive,external_remote,external_remote_hive") { - String enabled = context.config.otherConfigs.get("enableExternalHiveTest") - if (enabled != null && enabled.equalsIgnoreCase("true")) { - String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost") - String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort") - String catalog_name = "test_hive_same_db_table_name" - sql """drop catalog if exists ${catalog_name};""" - sql """ - create catalog if not exists ${catalog_name} properties ( - 'type'='hms', - 'hadoop.username' = 'hadoop', - 'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}' - ); - """ +suite("test_hive_same_db_table_name", "p0,external,hive,external_docker,external_docker_hive") { + String enabled = context.config.otherConfigs.get("enableHiveTest") + if (enabled == null || !enabled.equalsIgnoreCase("true")) { + logger.info("diable Hive test.") + return; + } + + for (String hivePrefix : ["hive2", "hive3"]) { + String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort") + String catalog_name = "${hivePrefix}_test_hive_same_db_table_name" + String externalEnvIp = context.config.otherConfigs.get("externalEnvIp") + + sql """drop catalog if exists ${catalog_name}""" + sql """create catalog if not exists ${catalog_name} properties ( + "type"="hms", + 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}' + );""" sql """switch internal;""" sql """create database if not exists multi_catalog;""" sql """use multi_catalog;""" diff --git a/regression-test/suites/external_table_p2/hive/test_hive_special_char_partition.groovy b/regression-test/suites/external_table_p0/hive/test_hive_special_char_partition.groovy similarity index 71% rename from regression-test/suites/external_table_p2/hive/test_hive_special_char_partition.groovy rename to regression-test/suites/external_table_p0/hive/test_hive_special_char_partition.groovy index 563a3a68c94..8b78ab2e3ce 100644 --- a/regression-test/suites/external_table_p2/hive/test_hive_special_char_partition.groovy +++ b/regression-test/suites/external_table_p0/hive/test_hive_special_char_partition.groovy @@ -15,20 +15,23 @@ // specific language governing permissions and limitations // under the License. -suite("test_hive_special_char_partition", "p2,external,hive,external_remote,external_remote_hive") { - String enabled = context.config.otherConfigs.get("enableExternalHiveTest") - if (enabled != null && enabled.equalsIgnoreCase("true")) { - String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost") - String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort") - String catalog_name = "test_hive_special_char_partition" - sql """drop catalog if exists ${catalog_name};""" - sql """ - create catalog if not exists ${catalog_name} properties ( - 'type'='hms', - 'hadoop.username' = 'hadoop', - 'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}' - ); - """ +suite("test_hive_special_char_partition", "p0,external,hive,external_docker,external_docker_hive") { + String enabled = context.config.otherConfigs.get("enableHiveTest") + if (enabled == null || !enabled.equalsIgnoreCase("true")) { + logger.info("diable Hive test.") + return; + } + + for (String hivePrefix : ["hive2", "hive3"]) { + String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort") + String catalog_name = "${hivePrefix}_test_hive_special_char_partition" + String externalEnvIp = context.config.otherConfigs.get("externalEnvIp") + + sql """drop catalog if exists ${catalog_name}""" + sql """create catalog if not exists ${catalog_name} properties ( + "type"="hms", + 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}' + );""" logger.info("catalog " + catalog_name + " created") sql """switch ${catalog_name};""" logger.info("switched to catalog " + catalog_name) diff --git a/regression-test/suites/external_table_p2/hive/test_wide_table.groovy b/regression-test/suites/external_table_p0/hive/test_wide_table.groovy similarity index 77% rename from regression-test/suites/external_table_p2/hive/test_wide_table.groovy rename to regression-test/suites/external_table_p0/hive/test_wide_table.groovy index e8de2380eb5..1c266d22c67 100644 --- a/regression-test/suites/external_table_p2/hive/test_wide_table.groovy +++ b/regression-test/suites/external_table_p0/hive/test_wide_table.groovy @@ -15,7 +15,7 @@ // specific language governing permissions and limitations // under the License. -suite("test_wide_table", "p2,external,hive,external_remote,external_remote_hive") { +suite("test_wide_table", "p0,external,hive,external_docker,external_docker_hive") { def formats = ["_orc"] def decimal_test1 = """select col1, col70, col71, col81, col100, col534 from wide_table1SUFFIX where col1 is not null order by col1 limit 1;""" @@ -39,24 +39,27 @@ suite("test_wide_table", "p2,external,hive,external_remote,external_remote_hive" """ def decimal_test8 = """select max(col1), max(col70), max(col71), min(col81), min(col100), min(col534) from wide_table1SUFFIX;""" - String enabled = context.config.otherConfigs.get("enableExternalHiveTest") - if (enabled != null && enabled.equalsIgnoreCase("true")) { - String extHiveHmsHost = context.config.otherConfigs.get("extHiveHmsHost") - String extHiveHmsPort = context.config.otherConfigs.get("extHiveHmsPort") - String catalog_name = "external_wide_table" + String enabled = context.config.otherConfigs.get("enableHiveTest") + if (enabled == null || !enabled.equalsIgnoreCase("true")) { + logger.info("diable Hive test.") + return; + } + + for (String hivePrefix : ["hive2", "hive3"]) { + String hms_port = context.config.otherConfigs.get(hivePrefix + "HmsPort") + String catalog_name = "${hivePrefix}_test_complex_types" + String externalEnvIp = context.config.otherConfigs.get("externalEnvIp") - sql """drop catalog if exists ${catalog_name};""" - sql """ - create catalog if not exists ${catalog_name} properties ( - 'type'='hms', - 'hive.metastore.uris' = 'thrift://${extHiveHmsHost}:${extHiveHmsPort}' - ); - """ + sql """drop catalog if exists ${catalog_name}""" + sql """create catalog if not exists ${catalog_name} properties ( + "type"="hms", + 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}' + );""" logger.info("catalog " + catalog_name + " created") sql """switch ${catalog_name};""" logger.info("switched to catalog " + catalog_name) - sql """use wide_tables;""" - logger.info("use wide_tables") + sql """use multi_catalog;""" + logger.info("use multi_catalog") for (String format in formats) { logger.info("Process format " + format) --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org For additional commands, e-mail: commits-h...@doris.apache.org