github-actions[bot] commented on PR #50405:
URL: https://github.com/apache/doris/pull/50405#issuecomment-2829265452
#### `sh-checker report`
To get the full details, please check in the
[job]("https://github.com/apache/doris/actions/runs/14655946192") output.
<details>
<summary>shellcheck errors</summary>
```
'shellcheck ' returned error 1 finding the following syntactical issues:
----------
In build.sh line 895:
if [[ "${TARGET_SYSTEM}" == 'Linux' ]] && [[ "$TARGET_ARCH" == 'x86_64'
]]; then
^----------^ SC2250
(style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
if [[ "${TARGET_SYSTEM}" == 'Linux' ]] && [[ "${TARGET_ARCH}" ==
'x86_64' ]]; then
In build.sh line 899:
elif [[ "${TARGET_SYSTEM}" == 'Linux' ]] && [[ "$TARGET_ARCH" ==
'aarch64' ]]; then
^----------^ SC2250
(style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
elif [[ "${TARGET_SYSTEM}" == 'Linux' ]] && [[ "${TARGET_ARCH}" ==
'aarch64' ]]; then
In docker/thirdparties/docker-compose/common/hive-configure.sh line 22:
export CORE_CONF_fs_defaultFS=${CORE_CONF_fs_defaultFS:-hdfs://`hostname
-f`:8020}
^-----------^
SC2006 (style): Use $(...) notation instead of legacy backticks `...`.
Did you mean:
export CORE_CONF_fs_defaultFS=${CORE_CONF_fs_defaultFS:-hdfs://$(hostname
-f):8020}
In docker/thirdparties/docker-compose/common/hive-configure.sh line 29:
local
entry="<property><name>$name</name><value>${value}</value></property>"
^---^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
Did you mean:
local
entry="<property><name>${name}</name><value>${value}</value></property>"
In docker/thirdparties/docker-compose/common/hive-configure.sh line 30:
local escapedEntry=$(echo $entry | sed 's/\//\\\//g')
^----------^ SC2155 (warning): Declare and assign separately to
avoid masking return values.
^----^ SC2086 (info): Double quote to prevent
globbing and word splitting.
^----^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
Did you mean:
local escapedEntry=$(echo "${entry}" | sed 's/\//\\\//g')
In docker/thirdparties/docker-compose/common/hive-configure.sh line 31:
sed -i "/<\/configuration>/ s/.*/${escapedEntry}\n&/" $path
^---^ SC2086 (info):
Double quote to prevent globbing and word splitting.
^---^ SC2250
(style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
sed -i "/<\/configuration>/ s/.*/${escapedEntry}\n&/" "${path}"
In docker/thirdparties/docker-compose/common/hive-configure.sh line 42:
echo "Configuring $module"
^-----^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
echo "Configuring ${module}"
In docker/thirdparties/docker-compose/common/hive-configure.sh line 43:
for c in `printenv | perl -sne 'print "$1 " if
m/^${envPrefix}_(.+?)=.*/' -- -envPrefix=$envPrefix`; do
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^--------^ SC2086 (info): Double quote to prevent globbing and
word splitting.
^--------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
for c in $(printenv | perl -sne 'print "$1 " if
m/^${envPrefix}_(.+?)=.*/' -- -envPrefix="${envPrefix}"); do
In docker/thirdparties/docker-compose/common/hive-configure.sh line 44:
name=`echo ${c} | perl -pe 's/___/-/g; s/__/_/g; s/_/./g'`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^--^ SC2086 (info): Double quote to prevent globbing and
word splitting.
Did you mean:
name=$(echo "${c}" | perl -pe 's/___/-/g; s/__/_/g; s/_/./g')
In docker/thirdparties/docker-compose/common/hive-configure.sh line 47:
echo " - Setting $name=$ "
^---^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
echo " - Setting ${name}=$ "
In docker/thirdparties/docker-compose/common/hive-configure.sh line 48:
addProperty $path $name "$value"
^---^ SC2086 (info): Double quote to prevent globbing
and word splitting.
^---^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
^---^ SC2086 (info): Double quote to prevent
globbing and word splitting.
^---^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
^----^ SC2250 (style): Prefer putting
braces around variable references even when not strictly required.
Did you mean:
addProperty "${path}" "${name}" "${value}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
26:
// clear output file
^-- SC1127 (error): Was this intended as a comment? Use # in sh.
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
27:
echo "" > "$output_file"
^----------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
echo "" > "${output_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
31:
if [ -n "$type_value" ]; then
^------------------^ SC2292 (style): Prefer [[ ]] over [ ] for
tests in Bash/Ksh.
^---------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
if [[ -n "${type_value}" ]]; then
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
32:
echo "{\"index\": {\"_index\": \"$index_name\", \"_type\":
\"$type_value\", \"_id\": \"${id_prefix}${id}\"}}" >> "$output_file"
^---------^ SC2250 (style):
Prefer putting braces around variable references even when not strictly
required.
^---------^ SC2250 (style): Prefer putting braces around variable references
even when not strictly required.
^----------^ SC2250 (style):
Prefer putting braces around variable references even when not strictly
required.
Did you mean:
echo "{\"index\": {\"_index\": \"${index_name}\", \"_type\":
\"${type_value}\", \"_id\": \"${id_prefix}${id}\"}}" >> "${output_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
34:
echo "{\"index\": {\"_index\": \"$index_name\", \"_id\":
\"${id_prefix}${id}\"}}" >> "$output_file"
^---------^ SC2250 (style):
Prefer putting braces around variable references even when not strictly
required.
^----------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
echo "{\"index\": {\"_index\": \"${index_name}\", \"_id\":
\"${id_prefix}${id}\"}}" >> "${output_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
36:
echo "$line" >> "$output_file"
^---^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
^----------^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
Did you mean:
echo "${line}" >> "${output_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
38:
done < "$data_file"
^--------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
done < "${data_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
79:
generate_bulk_request "composite_type_array" "doc" "item_"
"$array_data_file" "$bulk_request_file"
^--------------^
SC2250 (style): Prefer putting braces around variable references even when not
strictly required.
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
generate_bulk_request "composite_type_array" "doc" "item_"
"${array_data_file}" "${bulk_request_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
80:
curl -X POST "http://${ES_5_HOST}:9200/_bulk" --data-binary
"@$bulk_request_file" -H "Content-Type: application/json"
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
curl -X POST "http://${ES_5_HOST}:9200/_bulk" --data-binary
"@${bulk_request_file}" -H "Content-Type: application/json"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
119:
generate_bulk_request "composite_type_array" "doc" "item_"
"$array_data_file" "$bulk_request_file"
^--------------^
SC2250 (style): Prefer putting braces around variable references even when not
strictly required.
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
generate_bulk_request "composite_type_array" "doc" "item_"
"${array_data_file}" "${bulk_request_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
120:
curl -X POST "http://${ES_6_HOST}:9200/_bulk" --data-binary
"@$bulk_request_file" -H "Content-Type: application/json"
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
curl -X POST "http://${ES_6_HOST}:9200/_bulk" --data-binary
"@${bulk_request_file}" -H "Content-Type: application/json"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
126:
curl "http://${ES_7_HOST}:9200/test1" -H "Content-Type:application/json" -X
PUT -d "@/mnt/scripts/index/es7_test1.json"
^----------^ SC2154 (warning): ES_7_HOST is referenced but not
assigned.
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
164:
generate_bulk_request "composite_type_array" "_doc" "item_"
"$array_data_file" "$bulk_request_file"
^--------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
generate_bulk_request "composite_type_array" "_doc" "item_"
"${array_data_file}" "${bulk_request_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
165:
curl -X POST "http://${ES_7_HOST}:9200/_bulk" --data-binary
"@$bulk_request_file" -H "Content-Type: application/json"
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
curl -X POST "http://${ES_7_HOST}:9200/_bulk" --data-binary
"@${bulk_request_file}" -H "Content-Type: application/json"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
171:
curl "http://${ES_8_HOST}:9200/test1" -H "Content-Type:application/json" -X
PUT -d "@/mnt/scripts/index/es7_test1.json"
^----------^ SC2154 (warning): ES_8_HOST is referenced but not
assigned.
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
207:
generate_bulk_request "composite_type_array" "" "item_" "$array_data_file"
"$bulk_request_file"
^--------------^
SC2250 (style): Prefer putting braces around variable references even when not
strictly required.
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
generate_bulk_request "composite_type_array" "" "item_" "${array_data_file}"
"${bulk_request_file}"
In docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh line
208:
curl -X POST "http://${ES_8_HOST}:9200/_bulk" --data-binary
"@$bulk_request_file" -H "Content-Type: application/json"
^----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
curl -X POST "http://${ES_8_HOST}:9200/_bulk" --data-binary
"@${bulk_request_file}" -H "Content-Type: application/json"
In
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/logs1_parquet/run.sh
line 11:
cd -
^--^ SC2164 (warning): Use 'cd ... || exit' or 'cd ... || return' in
case cd fails.
Did you mean:
cd - || exit
In
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh
line 11:
cd -
^--^ SC2164 (warning): Use 'cd ... || exit' or 'cd ... || return' in
case cd fails.
Did you mean:
cd - || exit
In
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_compress_partitioned/run.sh
line 11:
cd -
^--^ SC2164 (warning): Use 'cd ... || exit' or 'cd ... || return' in
case cd fails.
Did you mean:
cd - || exit
In
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh
line 11:
cd -
^--^ SC2164 (warning): Use 'cd ... || exit' or 'cd ... || return' in
case cd fails.
Did you mean:
cd - || exit
In
docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_hdfs_tvf_compression/run.sh
line 9:
cd ${CUR_DIR}/ && rm -f test_data.tar.gz \
^--------^ SC2086 (info): Double quote to prevent globbing and word
splitting.
Did you mean:
cd "${CUR_DIR}"/ && rm -f test_data.tar.gz \
In
docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_hdfs_tvf_compression/run.sh
line 12:
cd -
^--^ SC2164 (warning): Use 'cd ... || exit' or 'cd ... || return' in
case cd fails.
Did you mean:
cd - || exit
In docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_tvf/run.sh
line 9:
cd ${CUR_DIR}/ && rm -f data.tar.gz \
^--------^ SC2086 (info): Double quote to prevent globbing and word
splitting.
Did you mean:
cd "${CUR_DIR}"/ && rm -f data.tar.gz \
In docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_tvf/run.sh
line 12:
cd -
^--^ SC2164 (warning): Use 'cd ... || exit' or 'cd ... || return' in
case cd fails.
Did you mean:
cd - || exit
In docker/thirdparties/docker-compose/hive/scripts/hive-metastore.sh line 33:
while ! $(nc -z localhost "${HMS_PORT:-9083}"); do
^-- SC2091 (warning): Remove surrounding $() to avoid executing
output (or use eval if intentional).
In docker/thirdparties/docker-compose/hive/scripts/hive-metastore.sh line 42:
find "${DATA_DIR}" -type f -name "run.sh" -print0 | xargs -0 -n 1 -P
"${parallel}" -I {} bash -ec '
^-- SC2016 (info): Expressions don't expand in single
quotes, use double quotes for that.
In docker/thirdparties/docker-compose/hive/scripts/hive-metastore.sh line
148:
ls /mnt/scripts/create_preinstalled_scripts/*.hql | xargs -n 1 -P
"${parallel}" -I {} bash -ec '
^-- SC2011 (warning): Use 'find .. -print0 | xargs -0 ..' or 'find .. -exec
.. +' to allow non-alphanumeric filenames.
^-- SC2016 (info): Expressions don't expand in single quotes,
use double quotes for that.
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 25:
if [ -z "${HADOOP_HOME}" ]; then
^---------------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests
in Bash/Ksh.
Did you mean:
if [[ -z "${HADOOP_HOME}" ]]; then
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 29:
if [ -z "${HIVE_HOME}" ]; then
^-------------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
Did you mean:
if [[ -z "${HIVE_HOME}" ]]; then
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 35:
HUDI_HIVE_UBER_JAR=`ls -c
$DIR/./hudi_docker_compose_attached_file/jar/hoodie-hive-sync-bundle.jar | grep
-v source | head -1`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^-- SC2010 (warning): Don't use ls | grep. Use a glob or
a for loop with a condition to allow non-alphanumeric filenames.
^--^ SC2086 (info): Double quote to prevent
globbing and word splitting.
^--^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
HUDI_HIVE_UBER_JAR=$(ls -c
"${DIR}"/./hudi_docker_compose_attached_file/jar/hoodie-hive-sync-bundle.jar |
grep -v source | head -1)
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 37:
if [ -z "$HADOOP_CONF_DIR" ]; then
^-----------------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests
in Bash/Ksh.
^--------------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
if [[ -z "${HADOOP_CONF_DIR}" ]]; then
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 43:
HIVE_EXEC=`ls ${HIVE_HOME}/lib/hive-exec-*.jar | tr '\n' ':'`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^-- SC2012 (info): Use find instead of ls to better handle
non-alphanumeric filenames.
^----------^ SC2086 (info): Double quote to prevent globbing
and word splitting.
Did you mean:
HIVE_EXEC=$(ls "${HIVE_HOME}"/lib/hive-exec-*.jar | tr '\n' ':')
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 44:
HIVE_SERVICE=`ls ${HIVE_HOME}/lib/hive-service-*.jar | grep -v rpc | tr '\n'
':'`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^-- SC2010 (warning): Don't use ls | grep. Use a glob or a for
loop with a condition to allow non-alphanumeric filenames.
^----------^ SC2086 (info): Double quote to prevent
globbing and word splitting.
Did you mean:
HIVE_SERVICE=$(ls "${HIVE_HOME}"/lib/hive-service-*.jar | grep -v rpc | tr
'\n' ':')
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 45:
HIVE_METASTORE=`ls ${HIVE_HOME}/lib/hive-metastore-*.jar | tr '\n' ':'`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^-- SC2012 (info): Use find instead of ls to better handle
non-alphanumeric filenames.
^----------^ SC2086 (info): Double quote to prevent
globbing and word splitting.
Did you mean:
HIVE_METASTORE=$(ls "${HIVE_HOME}"/lib/hive-metastore-*.jar | tr '\n' ':')
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 46:
HIVE_JDBC=`ls ${HIVE_HOME}/lib/hive-jdbc-*.jar | tr '\n' ':'`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^-- SC2012 (info): Use find instead of ls to better handle
non-alphanumeric filenames.
^----------^ SC2086 (info): Double quote to prevent globbing
and word splitting.
Did you mean:
HIVE_JDBC=$(ls "${HIVE_HOME}"/lib/hive-jdbc-*.jar | tr '\n' ':')
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 47:
if [ -z "${HIVE_JDBC}" ]; then
^-------------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
Did you mean:
if [[ -z "${HIVE_JDBC}" ]]; then
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 48:
HIVE_JDBC=`ls ${HIVE_HOME}/lib/hive-jdbc-*.jar | grep -v handler | tr '\n'
':'`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^-- SC2010 (warning): Don't use ls | grep. Use a glob or a for
loop with a condition to allow non-alphanumeric filenames.
^----------^ SC2086 (info): Double quote to prevent globbing
and word splitting.
Did you mean:
HIVE_JDBC=$(ls "${HIVE_HOME}"/lib/hive-jdbc-*.jar | grep -v handler | tr
'\n' ':')
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 50:
HIVE_JACKSON=`ls ${HIVE_HOME}/lib/jackson-*.jar | tr '\n' ':'`
^-- SC2006 (style): Use $(...) notation instead of legacy
backticks `...`.
^-- SC2012 (info): Use find instead of ls to better handle
non-alphanumeric filenames.
^----------^ SC2086 (info): Double quote to prevent
globbing and word splitting.
Did you mean:
HIVE_JACKSON=$(ls "${HIVE_HOME}"/lib/jackson-*.jar | tr '\n' ':')
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 51:
HIVE_JARS=$HIVE_METASTORE:$HIVE_SERVICE:$HIVE_EXEC:$HIVE_JDBC:$HIVE_JACKSON
^-------------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
^-----------^ SC2250 (style): Prefer putting
braces around variable references even when not strictly required.
^--------^ SC2250 (style): Prefer
putting braces around variable references even when not strictly required.
^--------^ SC2250
(style): Prefer putting braces around variable references even when not
strictly required.
^-----------^
SC2250 (style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
HIVE_JARS=${HIVE_METASTORE}:${HIVE_SERVICE}:${HIVE_EXEC}:${HIVE_JDBC}:${HIVE_JACKSON}
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 53:
HADOOP_HIVE_JARS=${HIVE_JARS}:${HADOOP_HOME}/share/hadoop/common/*:${HADOOP_HOME}/share/hadoop/mapreduce/*:${HADOOP_HOME}/share/hadoop/hdfs/*:${HADOOP_HOME}/share/hadoop/common/lib/*:${HADOOP_HOME}/share/hadoop/hdfs/lib/*
^-- SC2125 (warning): Brace expansions and globs are
literal in assignments. Quote it or use an array.
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 55:
echo "Running Command : java -cp
${HADOOP_HIVE_JARS}:${HADOOP_CONF_DIR}:$HUDI_HIVE_UBER_JAR
org.apache.hudi.hive.HiveSyncTool $@"
^-----------------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
^-- SC2145 (error): Argument
mixes string and array. Use * or separate argument.
Did you mean:
echo "Running Command : java -cp
${HADOOP_HIVE_JARS}:${HADOOP_CONF_DIR}:${HUDI_HIVE_UBER_JAR}
org.apache.hudi.hive.HiveSyncTool $@"
In docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh line 56:
java -cp $HUDI_HIVE_UBER_JAR:${HADOOP_HIVE_JARS}:${HADOOP_CONF_DIR}
org.apache.hudi.hive.HiveSyncTool "$@"
^-----------------^ SC2086 (info): Double quote to prevent globbing
and word splitting.
^-----------------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
^-----------------^ SC2086 (info): Double quote
to prevent globbing and word splitting.
^----------------^ SC2086
(info): Double quote to prevent globbing and word splitting.
Did you mean:
java -cp "${HUDI_HIVE_UBER_JAR}":"${HADOOP_HIVE_JARS}":"${HADOOP_CONF_DIR}"
org.apache.hudi.hive.HiveSyncTool "$@"
In
docker/thirdparties/docker-compose/hudi/scripts/setup_demo_container_adhoc_1.sh
line 20:
cp /var/scripts/config/spark-defaults.conf $SPARK_CONF_DIR/.
^-------------^ SC2154 (warning):
SPARK_CONF_DIR is referenced but not assigned.
^-------------^ SC2086 (info):
Double quote to prevent globbing and word splitting.
^-------------^ SC2250 (style):
Prefer putting braces around variable references even when not strictly
required.
Did you mean:
cp /var/scripts/config/spark-defaults.conf "${SPARK_CONF_DIR}"/.
In
docker/thirdparties/docker-compose/hudi/scripts/setup_demo_container_adhoc_1.sh
line 21:
cp /var/scripts/config/log4j2.properties $SPARK_CONF_DIR/.
^-------------^ SC2086 (info):
Double quote to prevent globbing and word splitting.
^-------------^ SC2250 (style):
Prefer putting braces around variable references even when not strictly
required.
Did you mean:
cp /var/scripts/config/log4j2.properties "${SPARK_CONF_DIR}"/.
In
docker/thirdparties/docker-compose/hudi/scripts/setup_demo_container_adhoc_2.sh
line 20:
cp /var/scripts/config/spark-defaults.conf $SPARK_CONF_DIR/.
^-------------^ SC2154 (warning):
SPARK_CONF_DIR is referenced but not assigned.
^-------------^ SC2086 (info):
Double quote to prevent globbing and word splitting.
^-------------^ SC2250 (style):
Prefer putting braces around variable references even when not strictly
required.
Did you mean:
cp /var/scripts/config/spark-defaults.conf "${SPARK_CONF_DIR}"/.
In
docker/thirdparties/docker-compose/hudi/scripts/setup_demo_container_adhoc_2.sh
line 21:
cp /var/scripts/config/log4j2.properties $SPARK_CONF_DIR/.
^-------------^ SC2086 (info):
Double quote to prevent globbing and word splitting.
^-------------^ SC2250 (style):
Prefer putting braces around variable references even when not strictly
required.
Did you mean:
cp /var/scripts/config/log4j2.properties "${SPARK_CONF_DIR}"/.
In docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
line 19:
source /usr/local/common/hive-configure.sh
^-- SC1091 (info): Not following:
/usr/local/common/hive-configure.sh: openBinaryFile: does not exist (No such
file or directory)
In docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
line 20:
source /usr/local/common/event-hook.sh
^-----------------------------^ SC1091 (info): Not following:
/usr/local/common/event-hook.sh: openBinaryFile: does not exist (No such file
or directory)
In docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
line 34:
if [ "$1" == "1" ]; then
^-------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
Did you mean:
if [[ "$1" == "1" ]]; then
In docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
line 36:
elif [ "$1" == "2" ]; then
^-------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
Did you mean:
elif [[ "$1" == "2" ]]; then
In docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
line 57:
if [ $i -eq 60 ]; then
^-----------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in Bash/Ksh.
^-- SC2086 (info): Double quote to prevent globbing and word splitting.
^-- SC2250 (style): Prefer putting braces around variable references
even when not strictly required.
Did you mean:
if [[ "${i}" -eq 60 ]]; then
In docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
line 64:
if [ "$1" == "1" ]; then
^-------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
Did you mean:
if [[ "$1" == "1" ]]; then
In docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
line 66:
elif [ "$1" == "2" ]; then
^-------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
Did you mean:
elif [[ "$1" == "2" ]]; then
In
docker/thirdparties/docker-compose/kerberos/health-checks/supervisorctl-check.sh
line 34:
if [ "$FAILED" == "" ]; then
^-----------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
^-----^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
if [[ "${FAILED}" == "" ]]; then
In
docker/thirdparties/docker-compose/ranger/ranger-admin/ranger-entrypoint.sh
line 24:
cd $RANGER_HOME
^----------^ SC2154 (warning): RANGER_HOME is referenced but not assigned.
^----------^ SC2086 (info): Double quote to prevent globbing and word
splitting.
^----------^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
cd "${RANGER_HOME}"
In
docker/thirdparties/docker-compose/ranger/script/install_doris_ranger_plugins.sh
line 16:
#!/bin/bash
^-- SC1128 (error): The shebang must be on the first line. Delete blanks and
move comments.
In
docker/thirdparties/docker-compose/ranger/script/install_doris_ranger_plugins.sh
line 19:
if [ ! -d "${RANGER_HOME}/ews/webapp/WEB-INF/classes/ranger-plugins/doris"
]; then
^-- SC2292 (style): Prefer [[ ]] over [ ] for tests in Bash/Ksh.
^------------^ SC2154 (warning): RANGER_HOME is referenced but
not assigned.
Did you mean:
if [[ ! -d "${RANGER_HOME}/ews/webapp/WEB-INF/classes/ranger-plugins/doris"
]]; then
In
docker/thirdparties/docker-compose/ranger/script/install_doris_service_def.sh
line 15:
#!/bin/bash
^-- SC1128 (error): The shebang must be on the first line. Delete blanks and
move comments.
In docker/thirdparties/run-thirdparties-docker.sh line 117:
echo ${COMPONENTS}
^-----------^ SC2086 (info): Double quote to prevent globbing and
word splitting.
Did you mean:
echo "${COMPONENTS}"
In docker/thirdparties/run-thirdparties-docker.sh line 150:
RUN_OCENABASE=0
^-----------^ SC2034 (warning): RUN_OCENABASE appears unused. Verify use (or
export if used externally).
In docker/thirdparties/run-thirdparties-docker.sh line 330:
IP_HOST=$(ifconfig "${eth_name}" | grep inet | grep -v 127.0.0.1 | grep
-v inet6 | awk '{print $2}' | tr -d "addr:" | head -n 1)
^-----^ SC2020 (info): tr replaces sets of
chars, not words (mentioned due to duplicates).
In docker/thirdparties/run-thirdparties-docker.sh line 339:
local backup_dir=/home/work/pipline/backup_center
^--------^ SC2034 (warning): backup_dir appears unused. Verify
use (or export if used externally).
In docker/thirdparties/run-thirdparties-docker.sh line 344:
echo "docker exec "${container_id}" bash -c echo
'/opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server
'${ip_host}:19193' --topic '${topic}'"
^-------------^ SC2027 (warning): The
surrounding quotes actually unquote this. Remove or escape them.
^-------------^ SC2086 (info): Double quote
to prevent globbing and word splitting.
Did you mean:
echo "docker exec ""${container_id}"" bash -c echo
'/opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server
'${ip_host}:19193' --topic '${topic}'"
In docker/thirdparties/run-thirdparties-docker.sh line 363:
IP_HOST=$(ifconfig "${eth_name}" | grep inet | grep -v 127.0.0.1 | grep
-v inet6 | awk '{print $2}' | tr -d "addr:" | head -n 1)
^-----^ SC2020 (info): tr replaces sets of
chars, not words (mentioned due to duplicates).
In docker/thirdparties/run-thirdparties-docker.sh line 365:
if [ "_${IP_HOST}" == "_" ]; then
^----------------------^ SC2292 (style): Prefer [[ ]] over [ ] for
tests in Bash/Ksh.
Did you mean:
if [[ "_${IP_HOST}" == "_" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 367:
exit -1
^-- SC2242 (error): Can only exit with status 0-255. Other data
should be written to stdout/stderr.
In docker/thirdparties/run-thirdparties-docker.sh line 375:
if [ -f "$file" ]; then
^------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
^---^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
if [[ -f "${file}" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 376:
echo "Processing $file"
^---^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
Did you mean:
echo "Processing ${file}"
In docker/thirdparties/run-thirdparties-docker.sh line 380:
echo "File not found: $file"
^---^ SC2250 (style): Prefer putting
braces around variable references even when not strictly required.
Did you mean:
echo "File not found: ${file}"
In docker/thirdparties/run-thirdparties-docker.sh line 391:
sudo docker compose -p ${CONTAINER_UID}hive2 -f
"${ROOT}"/docker-compose/hive/hive-2x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env down
^--------------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
Did you mean:
sudo docker compose -p "${CONTAINER_UID}"hive2 -f
"${ROOT}"/docker-compose/hive/hive-2x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env down
In docker/thirdparties/run-thirdparties-docker.sh line 393:
sudo docker compose -p ${CONTAINER_UID}hive2 -f
"${ROOT}"/docker-compose/hive/hive-2x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env up --build --remove-orphans -d
--wait
^--------------^ SC2248 (style): Prefer
double quoting even when variables don't contain special characters.
Did you mean:
sudo docker compose -p "${CONTAINER_UID}"hive2 -f
"${ROOT}"/docker-compose/hive/hive-2x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env up --build --remove-orphans -d
--wait
In docker/thirdparties/run-thirdparties-docker.sh line 403:
IP_HOST=$(ifconfig "${eth_name}" | grep inet | grep -v 127.0.0.1 | grep
-v inet6 | awk '{print $2}' | tr -d "addr:" | head -n 1)
^-----^ SC2020 (info): tr replaces sets of
chars, not words (mentioned due to duplicates).
In docker/thirdparties/run-thirdparties-docker.sh line 404:
if [ "_${IP_HOST}" == "_" ]; then
^----------------------^ SC2292 (style): Prefer [[ ]] over [ ] for
tests in Bash/Ksh.
Did you mean:
if [[ "_${IP_HOST}" == "_" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 406:
exit -1
^-- SC2242 (error): Can only exit with status 0-255. Other data
should be written to stdout/stderr.
In docker/thirdparties/run-thirdparties-docker.sh line 414:
if [ -f "$file" ]; then
^------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
^---^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
if [[ -f "${file}" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 415:
echo "Processing $file"
^---^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
Did you mean:
echo "Processing ${file}"
In docker/thirdparties/run-thirdparties-docker.sh line 419:
echo "File not found: $file"
^---^ SC2250 (style): Prefer putting
braces around variable references even when not strictly required.
Did you mean:
echo "File not found: ${file}"
In docker/thirdparties/run-thirdparties-docker.sh line 430:
sudo docker compose -p ${CONTAINER_UID}hive3 -f
"${ROOT}"/docker-compose/hive/hive-3x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env down
^--------------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
Did you mean:
sudo docker compose -p "${CONTAINER_UID}"hive3 -f
"${ROOT}"/docker-compose/hive/hive-3x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env down
In docker/thirdparties/run-thirdparties-docker.sh line 432:
sudo docker compose -p ${CONTAINER_UID}hive3 -f
"${ROOT}"/docker-compose/hive/hive-3x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env up --build --remove-orphans -d
--wait
^--------------^ SC2248 (style): Prefer
double quoting even when variables don't contain special characters.
Did you mean:
sudo docker compose -p "${CONTAINER_UID}"hive3 -f
"${ROOT}"/docker-compose/hive/hive-3x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env up --build --remove-orphans -d
--wait
In docker/thirdparties/run-thirdparties-docker.sh line 499:
cp "${trino_docker}/$file.tpl" "${trino_docker}/$file"
^---^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
^---^ SC2250
(style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
cp "${trino_docker}/${file}.tpl" "${trino_docker}/${file}"
In docker/thirdparties/run-thirdparties-docker.sh line 500:
if [[ $file != "hive.properties" ]]; then
^---^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
Did you mean:
if [[ ${file} != "hive.properties" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 501:
sed -i "s/doris--/${CONTAINER_UID}/g" "${trino_docker}/$file"
^---^
SC2250 (style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
sed -i "s/doris--/${CONTAINER_UID}/g" "${trino_docker}/${file}"
In docker/thirdparties/run-thirdparties-docker.sh line 510:
sudo echo "127.0.0.1 ${NAMENODE_CONTAINER_ID}" >>/etc/hosts
^-- SC2024 (warning):
sudo doesn't affect redirects. Use .. | sudo tee -a file
In docker/thirdparties/run-thirdparties-docker.sh line 512:
hive_metastore_ip=$(docker inspect --format='{{range
.NetworkSettings.Networks}}{{.IPAddress}}{{end}}'
${HIVE_METASTORE_CONTAINER_ID})
^----------------------------^ SC2248
(style): Prefer double quoting even when variables don't contain special
characters.
Did you mean:
hive_metastore_ip=$(docker inspect --format='{{range
.NetworkSettings.Networks}}{{.IPAddress}}{{end}}'
"${HIVE_METASTORE_CONTAINER_ID}")
In docker/thirdparties/run-thirdparties-docker.sh line 514:
if [ -z "$hive_metastore_ip" ]; then
^-------------------------^ SC2292 (style): Prefer [[ ]] over [ ]
for tests in Bash/Ksh.
^----------------^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
Did you mean:
if [[ -z "${hive_metastore_ip}" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 518:
echo "Hive Metastore IP address is: $hive_metastore_ip"
^----------------^ SC2250
(style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
echo "Hive Metastore IP address is: ${hive_metastore_ip}"
In docker/thirdparties/run-thirdparties-docker.sh line 533:
while [ $retries -lt $max_retries ]; do
^---------------------------^ SC2292 (style): Prefer [[ ]]
over [ ] for tests in Bash/Ksh.
^------^ SC2248 (style): Prefer double quoting even when
variables don't contain special characters.
^------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
^----------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
^----------^ SC2250 (style): Prefer putting
braces around variable references even when not strictly required.
Did you mean:
while [[ "${retries}" -lt "${max_retries}" ]]; do
In docker/thirdparties/run-thirdparties-docker.sh line 534:
status=$(docker inspect --format '{{.State.Running}}'
${TRINO_CONTAINER_ID})
^-------------------^ SC2248 (style): Prefer double quoting even when variables
don't contain special characters.
Did you mean:
status=$(docker inspect --format '{{.State.Running}}'
"${TRINO_CONTAINER_ID}")
In docker/thirdparties/run-thirdparties-docker.sh line 535:
if [ "${status}" == "${expected_status}" ]; then
^-- SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
Did you mean:
if [[ "${status}" == "${expected_status}" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 546:
if [ $retries -eq $max_retries ]; then
^---------------------------^ SC2292 (style): Prefer [[ ]]
over [ ] for tests in Bash/Ksh.
^------^ SC2248 (style): Prefer double quoting even when
variables don't contain special characters.
^------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
^----------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
^----------^ SC2250 (style): Prefer putting
braces around variable references even when not strictly required.
Did you mean:
if [[ "${retries}" -eq "${max_retries}" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 547:
echo "${operation} operation failed to complete after
$max_retries attempts."
^----------^ SC2250 (style): Prefer putting braces around variable references
even when not strictly required.
Did you mean:
echo "${operation} operation failed to complete after
${max_retries} attempts."
In docker/thirdparties/run-thirdparties-docker.sh line 552:
docker stop ${TRINO_CONTAINER_ID}
^-------------------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
Did you mean:
docker stop "${TRINO_CONTAINER_ID}"
In docker/thirdparties/run-thirdparties-docker.sh line 557:
docker start ${TRINO_CONTAINER_ID}
^-------------------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
Did you mean:
docker start "${TRINO_CONTAINER_ID}"
In docker/thirdparties/run-thirdparties-docker.sh line 563:
docker exec -it ${TRINO_CONTAINER_ID} /bin/bash -c 'trino -f
/scripts/create_trino_table.sql'
^-------------------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
Did you mean:
docker exec -it "${TRINO_CONTAINER_ID}" /bin/bash -c 'trino -f
/scripts/create_trino_table.sql'
In docker/thirdparties/run-thirdparties-docker.sh line 601:
mv *.tbl ../lakesoul/test_files/tpch/data
^-- SC2035 (info): Use ./*glob* or -- *glob* so names with dashes
won't become options.
In docker/thirdparties/run-thirdparties-docker.sh line 603:
export TPCH_DATA=$(realpath lakesoul/test_files/tpch/data)
^-------^ SC2155 (warning): Declare and assign separately to
avoid masking return values.
In docker/thirdparties/run-thirdparties-docker.sh line 615:
IP_HOST=$(ifconfig "${eth_name}" | grep inet | grep -v 127.0.0.1 | grep
-v inet6 | awk '{print $2}' | tr -d "addr:" | head -n 1)
^-----^ SC2020 (info): tr replaces sets of
chars, not words (mentioned due to duplicates).
In docker/thirdparties/run-thirdparties-docker.sh line 620:
. "${ROOT}"/docker-compose/kerberos/kerberos${i}_settings.env
^-- SC1090 (warning): ShellCheck can't follow non-constant source.
Use a directive to specify location.
^--^ SC2086 (info):
Double quote to prevent globbing and word splitting.
Did you mean:
. "${ROOT}"/docker-compose/kerberos/kerberos"${i}"_settings.env
In docker/thirdparties/run-thirdparties-docker.sh line 621:
envsubst <"${ROOT}"/docker-compose/kerberos/hadoop-hive.env.tpl
>"${ROOT}"/docker-compose/kerberos/hadoop-hive-${i}.env
^--^ SC2086 (info): Double quote to
prevent globbing and word splitting.
Did you mean:
envsubst <"${ROOT}"/docker-compose/kerberos/hadoop-hive.env.tpl
>"${ROOT}"/docker-compose/kerberos/hadoop-hive-"${i}".env
In docker/thirdparties/run-thirdparties-docker.sh line 622:
envsubst <"${ROOT}"/docker-compose/kerberos/conf/my.cnf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/my.cnf
^--^ SC2086 (info): Double quote to
prevent globbing and word splitting.
Did you mean:
envsubst <"${ROOT}"/docker-compose/kerberos/conf/my.cnf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos"${i}"/my.cnf
In docker/thirdparties/run-thirdparties-docker.sh line 623:
envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/kdc.conf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/kdc.conf
^--^ SC2086
(info): Double quote to prevent globbing and word splitting.
^--^ SC2086 (info):
Double quote to prevent globbing and word splitting.
Did you mean:
envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos"${i}"/kdc.conf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos"${i}"/kdc.conf
In docker/thirdparties/run-thirdparties-docker.sh line 624:
envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/krb5.conf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/krb5.conf
^--^ SC2086
(info): Double quote to prevent globbing and word splitting.
^--^ SC2086 (info):
Double quote to prevent globbing and word splitting.
Did you mean:
envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos"${i}"/krb5.conf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos"${i}"/krb5.conf
In docker/thirdparties/run-thirdparties-docker.sh line 778:
wait "${pids[$compose]}" || status=$?
^------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
Did you mean:
wait "${pids[${compose}]}" || status=$?
In docker/thirdparties/run-thirdparties-docker.sh line 779:
if [ $status -ne 0 ] && [ $compose != "db2" ]; then
^---------------^ SC2292 (style): Prefer [[ ]] over [ ] for tests in
Bash/Ksh.
^-----^ SC2248 (style): Prefer double quoting even when variables
don't contain special characters.
^-----^ SC2250 (style): Prefer putting braces around variable
references even when not strictly required.
^-------------------^ SC2292 (style): Prefer [[
]] over [ ] for tests in Bash/Ksh.
^------^ SC2086 (info): Double quote to
prevent globbing and word splitting.
^------^ SC2250 (style): Prefer putting braces
around variable references even when not strictly required.
Did you mean:
if [[ "${status}" -ne 0 ]] && [[ "${compose}" != "db2" ]]; then
In docker/thirdparties/run-thirdparties-docker.sh line 780:
echo "docker $compose started failed with status $status"
^------^ SC2250 (style): Prefer putting braces around
variable references even when not strictly required.
^-----^ SC2250
(style): Prefer putting braces around variable references even when not
strictly required.
Did you mean:
echo "docker ${compose} started failed with status ${status}"
In docker/thirdparties/run-thirdparties-docker.sh line 782:
cat start_${compose}.log || true
^--------^ SC2086 (info): Double quote to prevent globbing
and word splitting.
Did you mean:
cat start_"${compose}".log || true
In thirdparty/build-thirdparty.sh line 1863:
cp -r ${TP_SOURCE_DIR}/${JINDOFS_SOURCE}/*
"${TP_INSTALL_DIR}/jindofs_libs/"
^--------------^ SC2086 (info): Double quote to prevent globbing
and word splitting.
^---------------^ SC2248 (style): Prefer double
quoting even when variables don't contain special characters.
Did you mean:
cp -r "${TP_SOURCE_DIR}"/"${JINDOFS_SOURCE}"/*
"${TP_INSTALL_DIR}/jindofs_libs/"
In tools/lzo/build.sh line 1:
# Licensed to the Apache Software Foundation (ASF) under one
^-- SC2148 (error): Tips depend on target shell and yours is unknown. Add a
shebang or a 'shell' directive.
In tools/lzo/build.sh line 20:
g++ -o lzo_writer lzo_writer.cpp -I. -Isrc
-I${DORIS_THIRDPARTY}/installed/include -L${DORIS_THIRDPARTY}/installed/lib
-llzo2 -std=c++17
^-----------------^ SC2086
(info): Double quote to prevent globbing and word splitting.
^-----------------^ SC2086 (info): Double quote to prevent globbing and
word splitting.
Did you mean:
g++ -o lzo_writer lzo_writer.cpp -I. -Isrc
-I"${DORIS_THIRDPARTY}"/installed/include -L"${DORIS_THIRDPARTY}"/installed/lib
-llzo2 -std=c++17
For more information:
https://www.shellcheck.net/wiki/SC1127 -- Was this intended as a comment?
U...
https://www.shellcheck.net/wiki/SC1128 -- The shebang must be on the first
...
https://www.shellcheck.net/wiki/SC2145 -- Argument mixes string and array.
...
----------
You can address the above issues in one of three ways:
1. Manually correct the issue in the offending shell script;
2. Disable specific issues by adding the comment:
# shellcheck disable=NNNN
above the line that contains the issue, where NNNN is the error code;
3. Add '-e NNNN' to the SHELLCHECK_OPTS setting in your .yml action file.
```
</details>
<details>
<summary>shfmt errors</summary>
```
'shfmt ' returned error 1 finding the following formatting issues:
----------
--- docker/thirdparties/docker-compose/common/event-hook.sh.orig
+++ docker/thirdparties/docker-compose/common/event-hook.sh
--- docker/thirdparties/docker-compose/common/hive-configure.sh.orig
+++ docker/thirdparties/docker-compose/common/hive-configure.sh
@@ -19,16 +19,16 @@
# Referenced from
[docker-hive](https://github.com/big-data-europe/docker-hive)
# Set some sensible defaults
-export CORE_CONF_fs_defaultFS=${CORE_CONF_fs_defaultFS:-hdfs://`hostname
-f`:8020}
+export CORE_CONF_fs_defaultFS=${CORE_CONF_fs_defaultFS:-hdfs://$(hostname
-f):8020}
function addProperty() {
- local path=$1
- local name=$2
- local value=$3
+ local path=$1
+ local name=$2
+ local value=$3
- local
entry="<property><name>$name</name><value>${value}</value></property>"
- local escapedEntry=$(echo $entry | sed 's/\//\\\//g')
- sed -i "/<\/configuration>/ s/.*/${escapedEntry}\n&/" $path
+ local
entry="<property><name>$name</name><value>${value}</value></property>"
+ local escapedEntry=$(echo $entry | sed 's/\//\\\//g')
+ sed -i "/<\/configuration>/ s/.*/${escapedEntry}\n&/" $path
}
function configure() {
@@ -38,10 +38,10 @@
local var
local value
-
+
echo "Configuring $module"
- for c in `printenv | perl -sne 'print "$1 " if
m/^${envPrefix}_(.+?)=.*/' -- -envPrefix=$envPrefix`; do
- name=`echo ${c} | perl -pe 's/___/-/g; s/__/_/g; s/_/./g'`
+ for c in $(printenv | perl -sne 'print "$1 " if
m/^${envPrefix}_(.+?)=.*/' -- -envPrefix=$envPrefix); do
+ name=$(echo ${c} | perl -pe 's/___/-/g; s/__/_/g; s/_/./g')
var="${envPrefix}_${c}"
value=${!var}
echo " - Setting $name=$ "
--- docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh.orig
+++ docker/thirdparties/docker-compose/elasticsearch/scripts/es_init.sh
@@ -24,18 +24,18 @@
local output_file=$5
// clear output file
- echo "" > "$output_file"
+ echo "" >"$output_file"
local id=1
while IFS= read -r line; do
if [ -n "$type_value" ]; then
- echo "{\"index\": {\"_index\": \"$index_name\", \"_type\":
\"$type_value\", \"_id\": \"${id_prefix}${id}\"}}" >> "$output_file"
+ echo "{\"index\": {\"_index\": \"$index_name\", \"_type\":
\"$type_value\", \"_id\": \"${id_prefix}${id}\"}}" >>"$output_file"
else
- echo "{\"index\": {\"_index\": \"$index_name\", \"_id\":
\"${id_prefix}${id}\"}}" >> "$output_file"
+ echo "{\"index\": {\"_index\": \"$index_name\", \"_id\":
\"${id_prefix}${id}\"}}" >>"$output_file"
fi
- echo "$line" >> "$output_file"
+ echo "$line" >>"$output_file"
id=$((id + 1))
- done < "$data_file"
+ done <"$data_file"
}
array_data_file="/mnt/scripts/data/composite_type_array_bulk.json"
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/datev2_csv/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/datev2_csv/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}/create_table.hql"
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/datev2_orc/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/datev2_orc/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/datev2_parquet/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/datev2_parquet/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_config_test/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_config_test/run.sh
@@ -11,4 +11,3 @@
hive -f "${CUR_DIR}"/create_table.hql
hadoop fs -rm -r
/user/doris/suites/default/hive_ignore_absent_partitions_table/country=India
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_textfile_array_all_types/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_textfile_array_all_types/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_textfile_array_delimiter/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_textfile_array_delimiter/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_textfile_nestedarray/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_textfile_nestedarray/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_upper_case_orc/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_upper_case_orc/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_upper_case_parquet/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/hive_upper_case_parquet/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/logs1_parquet/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/logs1_parquet/run.sh
@@ -5,9 +5,9 @@
if [[ ! -d "${CUR_DIR}/data" ]]; then
echo "${CUR_DIR}/data does not exist"
- cd "${CUR_DIR}" && rm -f data.tar.gz \
- && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/logs1_parquet/data.tar.gz
\
- && tar xzf data.tar.gz
+ cd "${CUR_DIR}" && rm -f data.tar.gz &&
+ curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/logs1_parquet/data.tar.gz
&&
+ tar xzf data.tar.gz
cd -
else
echo "${CUR_DIR}/data exist, continue !"
@@ -19,4 +19,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/one_partition/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/one_partition/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}/create_table.hql"
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/orc_predicate/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/orc_predicate/run.sh
@@ -5,5 +5,3 @@
# create table
hive -f "${CUR_DIR}"/orc_predicate_table.hql
-
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/parquet_lz4_compression/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/parquet_lz4_compression/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/parquet_lzo_compression/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/parquet_lzo_compression/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/partition_manual_remove/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/partition_manual_remove/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh
@@ -5,9 +5,9 @@
if [[ ! -d "${CUR_DIR}/data" ]]; then
echo "${CUR_DIR}/data does not exist"
- cd "${CUR_DIR}" && rm -f data.tar.gz \
- && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_complex_types/data.tar.gz
\
- && tar xzf data.tar.gz
+ cd "${CUR_DIR}" && rm -f data.tar.gz &&
+ curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_complex_types/data.tar.gz
&&
+ tar xzf data.tar.gz
cd -
else
echo "${CUR_DIR}/data exist, continue !"
@@ -19,4 +19,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_compress_partitioned/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_compress_partitioned/run.sh
@@ -5,9 +5,9 @@
if [[ ! -d "${CUR_DIR}/data" ]]; then
echo "${CUR_DIR}/data does not exist"
- cd "${CUR_DIR}" && rm -f data.tar.gz \
- && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_compress_partitioned/data.tar.gz
\
- && tar xzf data.tar.gz
+ cd "${CUR_DIR}" && rm -f data.tar.gz &&
+ curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_compress_partitioned/data.tar.gz
&&
+ tar xzf data.tar.gz
cd -
else
echo "${CUR_DIR}/data exist, continue !"
@@ -19,4 +19,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_csv_format_error/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_csv_format_error/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}/create_table.hql"
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_date_string_partition/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_date_string_partition/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}/create_table.hql"
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh
@@ -5,9 +5,9 @@
if [[ ! -d "${CUR_DIR}/data" ]]; then
echo "${CUR_DIR}/data does not exist"
- cd "${CUR_DIR}" && rm -f data.tar.gz \
- && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_wide_table/data.tar.gz
\
- && tar xzf data.tar.gz
+ cd "${CUR_DIR}" && rm -f data.tar.gz &&
+ curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_wide_table/data.tar.gz
&&
+ tar xzf data.tar.gz
cd -
else
echo "${CUR_DIR}/data exist, continue !"
@@ -19,4 +19,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/two_partition/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/two_partition/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}/create_table.hql"
-
---
docker/thirdparties/docker-compose/hive/scripts/data/regression/crdmm_data/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/regression/crdmm_data/run.sh
@@ -10,5 +10,3 @@
# create table
hive -f "${CUR_DIR}"/create_table.hql
-
-
---
docker/thirdparties/docker-compose/hive/scripts/data/regression/serde_prop/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/regression/serde_prop/run.sh
@@ -5,5 +5,3 @@
# create table
hive -f "${CUR_DIR}"/some_serde_table.hql
-
-
---
docker/thirdparties/docker-compose/hive/scripts/data/statistics/statistics/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/statistics/statistics/run.sh
@@ -10,4 +10,3 @@
# create table
hive -f "${CUR_DIR}/create_table.hql"
-
---
docker/thirdparties/docker-compose/hive/scripts/data/tpch_1000_parquet/part/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/tpch_1000_parquet/part/run.sh
@@ -19,4 +19,3 @@
# # create table
# hive -f "${CUR_DIR}"/create_table.hql
-
---
docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_hdfs_tvf_compression/run.sh.orig
+++
docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_hdfs_tvf_compression/run.sh
@@ -6,9 +6,9 @@
if [[ ! -d "${CUR_DIR}/test_data" ]]; then
echo "${CUR_DIR}/test_data does not exist"
- cd ${CUR_DIR}/ && rm -f test_data.tar.gz \
- && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/test_hdfs_tvf_compression/test_data.tar.gz
\
- && tar xzf test_data.tar.gz
+ cd ${CUR_DIR}/ && rm -f test_data.tar.gz &&
+ curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/test_hdfs_tvf_compression/test_data.tar.gz
&&
+ tar xzf test_data.tar.gz
cd -
else
echo "${CUR_DIR}/test_data exist, continue !"
---
docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_tvf/run.sh.orig
+++ docker/thirdparties/docker-compose/hive/scripts/data/tvf/test_tvf/run.sh
@@ -6,9 +6,9 @@
if [[ ! -d "${CUR_DIR}/tvf" ]]; then
echo "${CUR_DIR}/tvf does not exist"
- cd ${CUR_DIR}/ && rm -f data.tar.gz \
- && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/test_tvf/data.tar.gz
\
- && tar xzf data.tar.gz
+ cd ${CUR_DIR}/ && rm -f data.tar.gz &&
+ curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/test_tvf/data.tar.gz
&&
+ tar xzf data.tar.gz
cd -
else
echo "${CUR_DIR}/tvf exist, continue !"
--- docker/thirdparties/docker-compose/hive/scripts/hive-metastore.sh.orig
+++ docker/thirdparties/docker-compose/hive/scripts/hive-metastore.sh
@@ -97,7 +97,6 @@
hadoop_put_pids=()
hadoop fs -mkdir -p /user/doris/
-
## put tpch1
if [[ -z "$(ls /mnt/scripts/tpch1.db)" ]]; then
echo "tpch1.db does not exist"
@@ -126,7 +125,6 @@
hadoop fs -copyFromLocal -f /mnt/scripts/preinstalled_data /user/doris/ &
hadoop_put_pids+=($!)
-
# wait put finish
set +e
wait "${hadoop_put_pids[@]}"
--- docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh.orig
+++ docker/thirdparties/docker-compose/hudi/scripts/run_sync_tool.sh
@@ -18,36 +18,36 @@
# under the License.
function error_exit {
- echo "$1" >&2 ## Send message to stderr. Exclude >&2 if you don't
want it that way.
- exit "${2:-1}" ## Return a code specified by $2 or 1 by default.
+ echo "$1" >&2 ## Send message to stderr. Exclude >&2 if you don't want
it that way.
+ exit "${2:-1}" ## Return a code specified by $2 or 1 by default.
}
if [ -z "${HADOOP_HOME}" ]; then
- error_exit "Please make sure the environment variable HADOOP_HOME is
setup"
+ error_exit "Please make sure the environment variable HADOOP_HOME is
setup"
fi
if [ -z "${HIVE_HOME}" ]; then
- error_exit "Please make sure the environment variable HIVE_HOME is setup"
+ error_exit "Please make sure the environment variable HIVE_HOME is
setup"
fi
-DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
+DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
#Ensure we pick the right jar even for hive11 builds
-HUDI_HIVE_UBER_JAR=`ls -c
$DIR/./hudi_docker_compose_attached_file/jar/hoodie-hive-sync-bundle.jar | grep
-v source | head -1`
+HUDI_HIVE_UBER_JAR=$(ls -c
$DIR/./hudi_docker_compose_attached_file/jar/hoodie-hive-sync-bundle.jar | grep
-v source | head -1)
if [ -z "$HADOOP_CONF_DIR" ]; then
- echo "setting hadoop conf dir"
- HADOOP_CONF_DIR="${HADOOP_HOME}/etc/hadoop"
+ echo "setting hadoop conf dir"
+ HADOOP_CONF_DIR="${HADOOP_HOME}/etc/hadoop"
fi
## Include only specific packages from HIVE_HOME/lib to avoid version
mismatches
-HIVE_EXEC=`ls ${HIVE_HOME}/lib/hive-exec-*.jar | tr '\n' ':'`
-HIVE_SERVICE=`ls ${HIVE_HOME}/lib/hive-service-*.jar | grep -v rpc | tr
'\n' ':'`
-HIVE_METASTORE=`ls ${HIVE_HOME}/lib/hive-metastore-*.jar | tr '\n' ':'`
-HIVE_JDBC=`ls ${HIVE_HOME}/lib/hive-jdbc-*.jar | tr '\n' ':'`
+HIVE_EXEC=$(ls ${HIVE_HOME}/lib/hive-exec-*.jar | tr '\n' ':')
+HIVE_SERVICE=$(ls ${HIVE_HOME}/lib/hive-service-*.jar | grep -v rpc | tr
'\n' ':')
+HIVE_METASTORE=$(ls ${HIVE_HOME}/lib/hive-metastore-*.jar | tr '\n' ':')
+HIVE_JDBC=$(ls ${HIVE_HOME}/lib/hive-jdbc-*.jar | tr '\n' ':')
if [ -z "${HIVE_JDBC}" ]; then
- HIVE_JDBC=`ls ${HIVE_HOME}/lib/hive-jdbc-*.jar | grep -v handler | tr
'\n' ':'`
+ HIVE_JDBC=$(ls ${HIVE_HOME}/lib/hive-jdbc-*.jar | grep -v handler | tr
'\n' ':')
fi
-HIVE_JACKSON=`ls ${HIVE_HOME}/lib/jackson-*.jar | tr '\n' ':'`
+HIVE_JACKSON=$(ls ${HIVE_HOME}/lib/jackson-*.jar | tr '\n' ':')
HIVE_JARS=$HIVE_METASTORE:$HIVE_SERVICE:$HIVE_EXEC:$HIVE_JDBC:$HIVE_JACKSON
HADOOP_HIVE_JARS=${HIVE_JARS}:${HADOOP_HOME}/share/hadoop/common/*:${HADOOP_HOME}/share/hadoop/mapreduce/*:${HADOOP_HOME}/share/hadoop/hdfs/*:${HADOOP_HOME}/share/hadoop/common/lib/*:${HADOOP_HOME}/share/hadoop/hdfs/lib/*
---
docker/thirdparties/docker-compose/hudi/scripts/setup_demo_container_adhoc_2.sh.orig
+++
docker/thirdparties/docker-compose/hudi/scripts/setup_demo_container_adhoc_2.sh
@@ -36,42 +36,42 @@
echo "Start synchronizing the stock_ticks_cow table"
/var/scripts/run_sync_tool.sh \
- --jdbc-url jdbc:hive2://hiveserver:10000 \
- --user hive \
- --pass hive \
- --partitioned-by date \
- --base-path /user/hive/warehouse/stock_ticks_cow \
- --database default \
- --table stock_ticks_cow \
- --partition-value-extractor
org.apache.hudi.hive.SlashEncodedDayPartitionValueExtractor
+ --jdbc-url jdbc:hive2://hiveserver:10000 \
+ --user hive \
+ --pass hive \
+ --partitioned-by date \
+ --base-path /user/hive/warehouse/stock_ticks_cow \
+ --database default \
+ --table stock_ticks_cow \
+ --partition-value-extractor
org.apache.hudi.hive.SlashEncodedDayPartitionValueExtractor
echo "Start synchronizing the stock_ticks_mor table"
/var/scripts/run_sync_tool.sh \
- --jdbc-url jdbc:hive2://hiveserver:10000 \
- --user hive \
- --pass hive \
- --partitioned-by date \
- --base-path /user/hive/warehouse/stock_ticks_mor \
- --database default \
- --table stock_ticks_mor \
- --partition-value-extractor
org.apache.hudi.hive.SlashEncodedDayPartitionValueExtractor
+ --jdbc-url jdbc:hive2://hiveserver:10000 \
+ --user hive \
+ --pass hive \
+ --partitioned-by date \
+ --base-path /user/hive/warehouse/stock_ticks_mor \
+ --database default \
+ --table stock_ticks_mor \
+ --partition-value-extractor
org.apache.hudi.hive.SlashEncodedDayPartitionValueExtractor
echo "Start synchronizing the hudi_cow_pt_tbl table"
/var/scripts/run_sync_tool.sh \
- --jdbc-url jdbc:hive2://hiveserver:10000 \
- --user hive \
- --pass hive \
- --partitioned-by dt \
- --base-path /user/hive/warehouse/hudi_cow_pt_tbl \
- --database default \
- --table hudi_cow_pt_tbl \
- --partition-value-extractor
org.apache.hudi.hive.HiveStylePartitionValueExtractor
+ --jdbc-url jdbc:hive2://hiveserver:10000 \
+ --user hive \
+ --pass hive \
+ --partitioned-by dt \
+ --base-path /user/hive/warehouse/hudi_cow_pt_tbl \
+ --database default \
+ --table hudi_cow_pt_tbl \
+ --partition-value-extractor
org.apache.hudi.hive.HiveStylePartitionValueExtractor
echo "Start synchronizing the hudi_non_part_cow table"
/var/scripts/run_sync_tool.sh \
- --jdbc-url jdbc:hive2://hiveserver:10000 \
- --user hive \
- --pass hive \
- --base-path /user/hive/warehouse/hudi_non_part_cow \
- --database default \
- --table hudi_non_part_cow \
+ --jdbc-url jdbc:hive2://hiveserver:10000 \
+ --user hive \
+ --pass hive \
+ --base-path /user/hive/warehouse/hudi_non_part_cow \
+ --database default \
+ --table hudi_non_part_cow
--- docker/thirdparties/docker-compose/iceberg/tools/save_docker.sh.orig
+++ docker/thirdparties/docker-compose/iceberg/tools/save_docker.sh
---
docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh.orig
+++ docker/thirdparties/docker-compose/kerberos/entrypoint-hive-master.sh
@@ -69,6 +69,6 @@
echo "Invalid index parameter. Exiting."
exit 1
fi
-hive -f /usr/local/sql/create_kerberos_hive_table.sql
+hive -f /usr/local/sql/create_kerberos_hive_table.sql
exec_success_hook
--- docker/thirdparties/docker-compose/kerberos/health-checks/health.sh.orig
+++ docker/thirdparties/docker-compose/kerberos/health-checks/health.sh
@@ -29,6 +29,6 @@
if test -d "${HEALTH_D}"; then
for health_script in "${HEALTH_D}"/*; do
- "${health_script}" &>> /var/log/container-health.log || exit 1
+ "${health_script}" &>>/var/log/container-health.log || exit 1
done
fi
---
docker/thirdparties/docker-compose/kerberos/health-checks/hive-health-check-2.sh.orig
+++
docker/thirdparties/docker-compose/kerberos/health-checks/hive-health-check-2.sh
---
docker/thirdparties/docker-compose/kerberos/health-checks/hive-health-check.sh.orig
+++
docker/thirdparties/docker-compose/kerberos/health-checks/hive-health-check.sh
---
docker/thirdparties/docker-compose/kerberos/health-checks/supervisorctl-check.sh.orig
+++
docker/thirdparties/docker-compose/kerberos/health-checks/supervisorctl-check.sh
@@ -32,9 +32,9 @@
FAILED=$(supervisorctl status | grep -v RUNNING || true)
if [ "$FAILED" == "" ]; then
- echo "All services are running"
- exit 0
+ echo "All services are running"
+ exit 0
else
- echo "Some of the services are failing: ${FAILED}"
- exit 1
+ echo "Some of the services are failing: ${FAILED}"
+ exit 1
fi
---
docker/thirdparties/docker-compose/ranger/script/install_doris_ranger_plugins.sh.orig
+++
docker/thirdparties/docker-compose/ranger/script/install_doris_ranger_plugins.sh
---
docker/thirdparties/docker-compose/ranger/script/install_doris_service_def.sh.orig
+++
docker/thirdparties/docker-compose/ranger/script/install_doris_service_def.sh
--- docker/thirdparties/run-thirdparties-docker.sh.orig
+++ docker/thirdparties/run-thirdparties-docker.sh
@@ -187,7 +187,7 @@
RUN_MARIADB=1
elif [[ "${element}"x == "db2"x ]]; then
RUN_DB2=1
- elif [[ "${element}"x == "oceanbase"x ]];then
+ elif [[ "${element}"x == "oceanbase"x ]]; then
RUN_OCEANBASE=1
elif [[ "${element}"x == "lakesoul"x ]]; then
RUN_LAKESOUL=1
@@ -387,7 +387,7 @@
. "${ROOT}"/docker-compose/hive/hive-2x_settings.env
envsubst <"${ROOT}"/docker-compose/hive/hive-2x.yaml.tpl
>"${ROOT}"/docker-compose/hive/hive-2x.yaml
envsubst <"${ROOT}"/docker-compose/hive/hadoop-hive.env.tpl
>"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env
- envsubst <"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env.tpl >>
"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env
+ envsubst <"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env.tpl
>>"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env
sudo docker compose -p ${CONTAINER_UID}hive2 -f
"${ROOT}"/docker-compose/hive/hive-2x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env down
if [[ "${STOP}" -ne 1 ]]; then
sudo docker compose -p ${CONTAINER_UID}hive2 -f
"${ROOT}"/docker-compose/hive/hive-2x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-2x.env up --build --remove-orphans -d
--wait
@@ -426,7 +426,7 @@
. "${ROOT}"/docker-compose/hive/hive-3x_settings.env
envsubst <"${ROOT}"/docker-compose/hive/hive-3x.yaml.tpl
>"${ROOT}"/docker-compose/hive/hive-3x.yaml
envsubst <"${ROOT}"/docker-compose/hive/hadoop-hive.env.tpl
>"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env
- envsubst <"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env.tpl >>
"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env
+ envsubst <"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env.tpl
>>"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env
sudo docker compose -p ${CONTAINER_UID}hive3 -f
"${ROOT}"/docker-compose/hive/hive-3x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env down
if [[ "${STOP}" -ne 1 ]]; then
sudo docker compose -p ${CONTAINER_UID}hive3 -f
"${ROOT}"/docker-compose/hive/hive-3x.yaml --env-file
"${ROOT}"/docker-compose/hive/hadoop-hive-3x.env up --build --remove-orphans -d
--wait
@@ -452,12 +452,12 @@
if [[ "${STOP}" -ne 1 ]]; then
if [[ ! -d "${ICEBERG_DIR}/data" ]]; then
echo "${ICEBERG_DIR}/data does not exist"
- cd "${ICEBERG_DIR}" \
- && rm -f iceberg_data.zip \
- && wget -P "${ROOT}"/docker-compose/iceberg
https://"${s3BucketName}.${s3Endpoint}"/regression/datalake/pipeline_data/iceberg_data.zip
\
- && sudo unzip iceberg_data.zip \
- && sudo mv iceberg_data data \
- && sudo rm -rf iceberg_data.zip
+ cd "${ICEBERG_DIR}" &&
+ rm -f iceberg_data.zip &&
+ wget -P "${ROOT}"/docker-compose/iceberg
https://"${s3BucketName}.${s3Endpoint}"/regression/datalake/pipeline_data/iceberg_data.zip
&&
+ sudo unzip iceberg_data.zip &&
+ sudo mv iceberg_data data &&
+ sudo rm -rf iceberg_data.zip
cd -
else
echo "${ICEBERG_DIR}/data exist, continue !"
@@ -619,9 +619,9 @@
for i in {1..2}; do
. "${ROOT}"/docker-compose/kerberos/kerberos${i}_settings.env
envsubst <"${ROOT}"/docker-compose/kerberos/hadoop-hive.env.tpl
>"${ROOT}"/docker-compose/kerberos/hadoop-hive-${i}.env
- envsubst <"${ROOT}"/docker-compose/kerberos/conf/my.cnf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/my.cnf
- envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/kdc.conf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/kdc.conf
- envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/krb5.conf.tpl >
"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/krb5.conf
+ envsubst <"${ROOT}"/docker-compose/kerberos/conf/my.cnf.tpl
>"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/my.cnf
+ envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/kdc.conf.tpl
>"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/kdc.conf
+ envsubst
<"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/krb5.conf.tpl
>"${ROOT}"/docker-compose/kerberos/conf/kerberos${i}/krb5.conf
done
sudo chmod a+w /etc/hosts
sudo sed -i "1i${IP_HOST} hadoop-master" /etc/hosts
@@ -672,101 +672,101 @@
declare -A pids
if [[ "${RUN_ES}" -eq 1 ]]; then
- start_es > start_es.log 2>&1 &
+ start_es >start_es.log 2>&1 &
pids["es"]=$!
fi
if [[ "${RUN_MYSQL}" -eq 1 ]]; then
- start_mysql > start_mysql.log 2>&1 &
+ start_mysql >start_mysql.log 2>&1 &
pids["mysql"]=$!
fi
if [[ "${RUN_PG}" -eq 1 ]]; then
- start_pg > start_pg.log 2>&1 &
+ start_pg >start_pg.log 2>&1 &
pids["pg"]=$!
fi
if [[ "${RUN_ORACLE}" -eq 1 ]]; then
- start_oracle > start_oracle.log 2>&1 &
+ start_oracle >start_oracle.log 2>&1 &
pids["oracle"]=$!
fi
if [[ "${RUN_DB2}" -eq 1 ]]; then
- start_db2 > start_db2.log 2>&1 &
+ start_db2 >start_db2.log 2>&1 &
pids["db2"]=$!
fi
if [[ "${RUN_OCEANBASE}" -eq 1 ]]; then
- start_oceanbase > start_oceanbase.log 2>&1 &
+ start_oceanbase >start_oceanbase.log 2>&1 &
pids["oceanbase"]=$!
fi
if [[ "${RUN_SQLSERVER}" -eq 1 ]]; then
- start_sqlserver > start_sqlserver.log 2>&1 &
+ start_sqlserver >start_sqlserver.log 2>&1 &
pids["sqlserver"]=$!
fi
if [[ "${RUN_CLICKHOUSE}" -eq 1 ]]; then
- start_clickhouse > start_clickhouse.log 2>&1 &
+ start_clickhouse >start_clickhouse.log 2>&1 &
pids["clickhouse"]=$!
fi
if [[ "${RUN_KAFKA}" -eq 1 ]]; then
- start_kafka > start_kafka.log 2>&1 &
+ start_kafka >start_kafka.log 2>&1 &
pids["kafka"]=$!
fi
if [[ "${RUN_HIVE2}" -eq 1 ]]; then
- start_hive2 > start_hive2.log 2>&1 &
+ start_hive2 >start_hive2.log 2>&1 &
pids["hive2"]=$!
fi
if [[ "${RUN_HIVE3}" -eq 1 ]]; then
- start_hive3 > start_hive3.log 2>&1 &
+ start_hive3 >start_hive3.log 2>&1 &
pids["hive3"]=$!
fi
if [[ "${RUN_SPARK}" -eq 1 ]]; then
- start_spark > start_spark.log 2>&1 &
+ start_spark >start_spark.log 2>&1 &
pids["spark"]=$!
fi
if [[ "${RUN_ICEBERG}" -eq 1 ]]; then
- start_iceberg > start_icerberg.log 2>&1 &
+ start_iceberg >start_icerberg.log 2>&1 &
pids["iceberg"]=$!
fi
if [[ "${RUN_HUDI}" -eq 1 ]]; then
- start_hudi > start_hudi.log 2>&1 &
+ start_hudi >start_hudi.log 2>&1 &
pids["hudi"]=$!
fi
if [[ "${RUN_TRINO}" -eq 1 ]]; then
- start_trino > start_trino.log 2>&1 &
+ start_trino >start_trino.log 2>&1 &
pids["trino"]=$!
fi
if [[ "${RUN_MARIADB}" -eq 1 ]]; then
- start_mariadb > start_mariadb.log 2>&1 &
+ start_mariadb >start_mariadb.log 2>&1 &
pids["mariadb"]=$!
fi
if [[ "${RUN_LAKESOUL}" -eq 1 ]]; then
- start_lakesoul > start_lakesoule.log 2>&1 &
+ start_lakesoul >start_lakesoule.log 2>&1 &
pids["lakesoul"]=$!
fi
if [[ "${RUN_MINIO}" -eq 1 ]]; then
- start_minio > start_minio.log 2>&1 &
+ start_minio >start_minio.log 2>&1 &
pids["minio"]=$!
fi
if [[ "${RUN_KERBEROS}" -eq 1 ]]; then
- start_kerberos > start_kerberos.log 2>&1 &
+ start_kerberos >start_kerberos.log 2>&1 &
pids["kerberos"]=$!
fi
if [[ "${RUN_RANGER}" -eq 1 ]]; then
- start_ranger > start_ranger.log 2>&1 &
+ start_ranger >start_ranger.log 2>&1 &
pids["ranger"]=$!
fi
----------
You can reformat the above files to meet shfmt's requirements by typing:
shfmt -w filename
```
</details>
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]