This is an automated email from the ASF dual-hosted git repository.

zichaowang pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/shardingsphere.git


The following commit(s) were added to refs/heads/master by this push:
     new a8174aed869 Fix code inspection for kernel module (#31209)
a8174aed869 is described below

commit a8174aed86929ef9a8204a8d8672c92426fa4938
Author: Liang Zhang <[email protected]>
AuthorDate: Sun May 12 04:34:50 2024 +0800

    Fix code inspection for kernel module (#31209)
---
 .../core/channel/memory/MemoryPipelineChannel.java |  2 +-
 .../ConsistencyCheckJobItemProgressContext.java    |  2 +-
 .../DataConsistencyCheckUtils.java                 |  2 +-
 .../result/TableDataConsistencyCheckResult.java    |  5 +--
 .../table/MatchingTableInventoryChecker.java       |  2 +-
 .../importer/sink/type/PipelineDataSourceSink.java |  2 +-
 .../type/pk/type/UnsupportedKeyIngestPosition.java |  2 +-
 .../pipeline/core/job/id/PipelineJobIdUtils.java   |  2 +-
 .../job/progress/TransmissionJobItemProgress.java  |  2 +-
 .../persist/PipelineJobProgressPersistContext.java |  2 +-
 .../persist/PipelineJobProgressPersistService.java |  6 +--
 .../YamlJobItemInventoryTasksProgressSwapper.java  |  2 +-
 .../core/job/service/TransmissionJobManager.java   |  4 +-
 .../StandardPipelineTableMetaDataLoader.java       |  2 +-
 .../datasource/PipelineJobDataSourcePreparer.java  |  2 +-
 .../inventory/InventoryRecordsCountCalculator.java |  2 +-
 .../preparer/inventory/InventoryTaskSplitter.java  |  6 +--
 .../job/PipelineJobCheckGovernanceRepository.java  |  2 +-
 .../sqlbuilder/sql/PipelineImportSQLBuilder.java   |  2 +-
 .../core/util/IntervalToRangeIterator.java         |  8 ++--
 .../core/util/ShardingColumnsExtractor.java        |  2 +-
 .../core/channel/MultiplexPipelineChannelTest.java |  2 +-
 .../ConsistencyCheckDataBuilder.java               |  8 ++--
 .../DataConsistencyCheckUtilsTest.java             |  4 +-
 .../CRC32SingleTableInventoryCalculatorTest.java   |  2 +-
 .../datanode/JobDataNodeLineConvertUtilsTest.java  |  6 +--
 .../type/IntegerPrimaryKeyIngestPositionTest.java  |  2 +-
 .../progress/PipelineJobProgressDetectorTest.java  | 50 ++++++----------------
 .../progress/TransmissionJobItemProgressTest.java  |  4 +-
 .../metadata/node/PipelineMetaDataNodeTest.java    |  3 +-
 .../mysql/ingest/GlobalTableMapEventMapping.java   |  2 +-
 .../pipeline/mysql/ingest/client/MySQLClient.java  |  2 +-
 .../value/impl/MySQLUnsignedBigintHandler.java     |  2 +-
 .../ingest/MySQLIngestPositionManagerTest.java     |  2 +-
 .../mysql/ingest/binlog/BinlogPositionTest.java    |  2 +-
 .../mysql/ingest/client/MySQLClientTest.java       |  2 +-
 .../opengauss/ingest/OpenGaussWALDumper.java       |  2 +-
 .../ddlgenerator/AbstractPostgreSQLDDLAdapter.java | 12 +++---
 .../PostgreSQLColumnPropertiesAppender.java        |  2 +-
 .../ddlgenerator/PostgreSQLColumnType.java         |  2 +-
 .../postgresql/ingest/PostgreSQLWALDumper.java     |  4 +-
 .../ingest/wal/WALEventConverterTest.java          |  2 +-
 .../query/ShowStreamingJobStatusExecutor.java      |  2 +-
 .../query/ShowMigrationJobStatusExecutor.java      |  2 +-
 .../cdc/client/util/ProtobufAnyValueConverter.java |  2 +-
 .../data/pipeline/cdc/CDCJobType.java              |  2 +-
 .../data/pipeline/cdc/api/CDCJobAPI.java           |  2 +-
 .../pipeline/cdc/context/CDCJobItemContext.java    |  4 +-
 .../pipeline/cdc/core/importer/CDCImporter.java    |  2 +-
 .../pipeline/cdc/handler/CDCBackendHandler.java    |  2 +-
 .../data/pipeline/cdc/util/CDCDataNodeUtils.java   |  2 +-
 .../pipeline/cdc/util/CDCSchemaTableUtils.java     |  6 +--
 .../pipeline/cdc/util/ColumnValueConvertUtils.java |  8 ++--
 .../pipeline/cdc/util/CDCSchemaTableUtilsTest.java |  7 ++-
 .../cdc/util/ColumnValueConvertUtilsTest.java      |  2 +-
 .../cdc/util/DataRecordResultConvertUtilsTest.java |  2 +-
 .../api/ConsistencyCheckJobAPI.java                |  8 ++--
 .../scenario/migration/api/MigrationJobAPI.java    |  6 +--
 .../migration/context/MigrationJobItemContext.java |  6 +--
 .../migration/preparer/MigrationJobPreparer.java   |  2 +-
 .../type/tso/provider/RedisTSOProvider.java        |  2 +-
 .../shardingsphere/logging/util/LoggingUtils.java  |  2 +-
 .../database/DatabaseMetaDataPersistService.java   |  4 +-
 .../single/datanode/SingleTableDataNodeLoader.java |  4 +-
 .../single/decider/SingleSQLFederationDecider.java |  2 +-
 .../SingleRuleConfigurationDecorator.java          |  8 ++--
 .../single/util/SingleTableLoadUtils.java          |  2 +-
 .../config/pojo/YamlSingleRuleConfiguration.java   |  2 +-
 .../handler/update/LoadSingleTableExecutor.java    |  2 +-
 .../resultset/SQLFederationResultSetMetaData.java  |  2 +-
 .../core/SQLFederationDistSQLStatementVisitor.java |  2 +-
 .../enumerable/EnumerableScanExecutor.java         |  2 +-
 .../executor/enumerator/JDBCRowEnumerator.java     |  2 +-
 .../expression/impl/RowExpressionConverter.java    |  2 +-
 .../optimizer/operator/logical/LogicalScan.java    |  2 +-
 .../planner/util/SQLFederationPlannerUtils.java    |  2 +-
 .../optimizer/it/SQLStatementCompilerIT.java       |  2 +-
 .../parser/rule/SQLParserRuleTest.java             |  2 +-
 .../rule/builder/SQLParserRuleBuilderTest.java     |  6 +--
 ...LParserCacheOptionConfigurationSwapperTest.java |  2 +-
 .../YamlSQLParserRuleConfigurationSwapperTest.java |  2 +-
 .../query/ShowSQLParserRuleExecutorTest.java       |  2 +-
 .../core/SQLParserDistSQLStatementVisitor.java     |  2 +-
 .../transaction/ConnectionTransaction.java         |  4 +-
 .../transaction/xa/fixture/DataSourceUtils.java    |  4 +-
 .../manager/switcher/ResourceSwitchManager.java    |  2 +-
 .../mode/metadata/MetaDataContexts.java            |  2 +-
 .../manager/cluster/ClusterModeContextManager.java |  2 +-
 .../compute/service/ComputeNodeStatusService.java  |  2 +-
 .../ResourceMetaDataChangedSubscriberTest.java     |  2 +-
 .../subscriber/StateChangedSubscriberTest.java     |  2 +-
 .../cluster/lock/holder/DistributedLockHolder.java |  3 +-
 .../listener/SessionConnectionListener.java        |  2 +-
 .../standalone/StandaloneModeContextManager.java   |  2 +-
 94 files changed, 151 insertions(+), 181 deletions(-)

diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
index a2a6212fc0e..7061d515ad2 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
@@ -57,7 +57,7 @@ public final class MemoryPipelineChannel implements 
PipelineChannel {
         long startMillis = System.currentTimeMillis();
         int recordsCount = 0;
         do {
-            List<Record> records = queue.poll(Math.max(0, timeoutMillis - 
(System.currentTimeMillis() - startMillis)), TimeUnit.MILLISECONDS);
+            List<Record> records = queue.poll(Math.max(0L, timeoutMillis - 
(System.currentTimeMillis() - startMillis)), TimeUnit.MILLISECONDS);
             if (null == records || records.isEmpty()) {
                 continue;
             }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
index c493dc9afb9..dba540c9f70 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
@@ -48,7 +48,7 @@ public final class ConsistencyCheckJobItemProgressContext 
implements PipelineJob
     
     private volatile long recordsCount;
     
-    private final AtomicLong checkedRecordsCount = new AtomicLong(0);
+    private final AtomicLong checkedRecordsCount = new AtomicLong(0L);
     
     private final long checkBeginTimeMillis = System.currentTimeMillis();
     
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
index 0aa0844db66..f78eb54f972 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
@@ -126,6 +126,6 @@ public final class DataConsistencyCheckUtils {
                 decimalTwo = another;
             }
         }
-        return decimalOne.equals(decimalTwo);
+        return 0 == decimalOne.compareTo(decimalTwo);
     }
 }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
index 8d73495befa..ebf462baea7 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
@@ -56,9 +56,6 @@ public final class TableDataConsistencyCheckResult {
      * @return matched or not
      */
     public boolean isMatched() {
-        if (null != ignoredType) {
-            return false;
-        }
-        return matched;
+        return null == ignoredType && matched;
     }
 }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
index abd1e1ecc04..f9faf20e2ce 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
@@ -62,7 +62,7 @@ public abstract class MatchingTableInventoryChecker 
implements TableInventoryChe
     @Override
     public TableDataConsistencyCheckResult checkSingleTableInventoryData() {
         ThreadFactory threadFactory = 
ExecutorThreadFactoryBuilder.build("job-" + getJobIdDigest(param.getJobId()) + 
"-matching-check-%d");
-        ThreadPoolExecutor executor = new ThreadPoolExecutor(2, 2, 60, 
TimeUnit.SECONDS, new ArrayBlockingQueue<>(2), threadFactory);
+        ThreadPoolExecutor executor = new ThreadPoolExecutor(2, 2, 60L, 
TimeUnit.SECONDS, new ArrayBlockingQueue<>(2), threadFactory);
         try {
             return checkSingleTableInventoryData(param, executor);
         } finally {
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
index 1cabadec3a8..8fdf6cfb94d 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
@@ -101,7 +101,7 @@ public final class PipelineDataSourceSink implements 
PipelineSink {
                 if (i == importerConfig.getRetryTimes()) {
                     throw new PipelineImporterJobWriteException(ex);
                 }
-                Thread.sleep(Math.min(5 * 60 * 1000L, 1000L << i));
+                Thread.sleep(Math.min(5L * 60L * 1000L, 1000L << i));
             }
         }
     }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
index e589193eb45..d8563ac1e41 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
@@ -36,7 +36,7 @@ public final class UnsupportedKeyIngestPosition implements 
PrimaryKeyIngestPosit
     
     @Override
     public Void convert(final String value) {
-        throw new UnsupportedOperationException();
+        throw new UnsupportedOperationException("");
     }
     
     @Override
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
index 8972531cca2..459193309f3 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
@@ -101,7 +101,7 @@ public final class PipelineJobIdUtils {
      * @return ElasticJob configuration POJO
      */
     public static JobConfigurationPOJO getElasticJobConfigurationPOJO(final 
String jobId) {
-        JobConfigurationPOJO result = 
PipelineAPIFactory.getJobConfigurationAPI(PipelineJobIdUtils.parseContextKey(jobId)).getJobConfiguration(jobId);
+        JobConfigurationPOJO result = 
PipelineAPIFactory.getJobConfigurationAPI(parseContextKey(jobId)).getJobConfiguration(jobId);
         ShardingSpherePreconditions.checkNotNull(result, () -> new 
PipelineJobNotFoundException(jobId));
         return result;
     }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
index 8d38f79fe0b..56c18fa144c 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
@@ -70,7 +70,7 @@ public final class TransmissionJobItemProgress implements 
PipelineJobItemProgres
     }
     
     private JobItemInventoryTasksProgress getInventoryTasksProgress(final 
Collection<PipelineTask> inventoryTasks) {
-        Map<String, InventoryTaskProgress> inventoryTaskProgressMap = new 
HashMap<>();
+        Map<String, InventoryTaskProgress> inventoryTaskProgressMap = new 
HashMap<>(inventoryTasks.size(), 1F);
         for (PipelineTask each : inventoryTasks) {
             inventoryTaskProgressMap.put(each.getTaskId(), 
(InventoryTaskProgress) each.getTaskProgress());
         }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
index beafba57ba3..839950933e0 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
@@ -34,7 +34,7 @@ public final class PipelineJobProgressPersistContext {
     
     private final int shardingItem;
     
-    private final AtomicLong unhandledEventCount = new AtomicLong(0);
+    private final AtomicLong unhandledEventCount = new AtomicLong(0L);
     
     private final AtomicBoolean firstExceptionLogged = new 
AtomicBoolean(false);
 }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
index e6bc68e0547..461efafcc9c 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
@@ -53,7 +53,7 @@ public final class PipelineJobProgressPersistService {
     private static final long DELAY_SECONDS = 1L;
     
     static {
-        JOB_PERSIST_EXECUTOR.scheduleWithFixedDelay(new 
PersistJobContextRunnable(), 0, DELAY_SECONDS, TimeUnit.SECONDS);
+        JOB_PERSIST_EXECUTOR.scheduleWithFixedDelay(new 
PersistJobContextRunnable(), 0L, DELAY_SECONDS, TimeUnit.SECONDS);
     }
     
     /**
@@ -126,8 +126,8 @@ public final class PipelineJobProgressPersistService {
         
         private static void persist0(final String jobId, final int 
shardingItem, final PipelineJobProgressPersistContext persistContext) {
             long currentUnhandledEventCount = 
persistContext.getUnhandledEventCount().get();
-            ShardingSpherePreconditions.checkState(currentUnhandledEventCount 
>= 0, () -> new IllegalStateException("Current unhandled event count must be 
greater than or equal to 0"));
-            if (0 == currentUnhandledEventCount) {
+            ShardingSpherePreconditions.checkState(currentUnhandledEventCount 
>= 0L, () -> new IllegalStateException("Current unhandled event count must be 
greater than or equal to 0"));
+            if (0L == currentUnhandledEventCount) {
                 return;
             }
             Optional<PipelineJobItemContext> jobItemContext = 
PipelineJobRegistry.getItemContext(jobId, shardingItem);
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
index 91486949172..e3c464b8f66 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
@@ -75,7 +75,7 @@ public final class YamlJobItemInventoryTasksProgressSwapper {
         if (null == yamlProgress) {
             return new JobItemInventoryTasksProgress(Collections.emptyMap());
         }
-        Map<String, InventoryTaskProgress> taskProgressMap = new 
LinkedHashMap<>();
+        Map<String, InventoryTaskProgress> taskProgressMap = new 
LinkedHashMap<>(yamlProgress.getFinished().length + 
yamlProgress.getUnfinished().size(), 1F);
         
taskProgressMap.putAll(Arrays.stream(yamlProgress.getFinished()).collect(Collectors.toMap(key
 -> key, value -> new InventoryTaskProgress(new IngestFinishedPosition()))));
         
taskProgressMap.putAll(yamlProgress.getUnfinished().entrySet().stream().collect(Collectors.toMap(Entry::getKey,
 getInventoryTaskProgressFunction())));
         return new JobItemInventoryTasksProgress(taskProgressMap);
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
index 9dcd2cd4676..df81d4312cd 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
@@ -81,8 +81,8 @@ public final class TransmissionJobManager {
         if (JobStatus.EXECUTE_INCREMENTAL_TASK == jobItemProgress.getStatus() 
|| JobStatus.FINISHED == jobItemProgress.getStatus()) {
             return 100;
         }
-        if (0 != jobItemProgress.getProcessedRecordsCount() && 0 != 
jobItemProgress.getInventoryRecordsCount()) {
-            return (int) Math.min(100, 
jobItemProgress.getProcessedRecordsCount() * 100 / 
jobItemProgress.getInventoryRecordsCount());
+        if (0L != jobItemProgress.getProcessedRecordsCount() && 0L != 
jobItemProgress.getInventoryRecordsCount()) {
+            return (int) Math.min(100L, 
jobItemProgress.getProcessedRecordsCount() * 100L / 
jobItemProgress.getInventoryRecordsCount());
         }
         return 0;
     }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
index 659777adb18..ac97df73d35 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
@@ -89,7 +89,7 @@ public final class StandardPipelineTableMetaDataLoader 
implements PipelineTableM
                 tableNames.add(tableName);
             }
         }
-        Map<CaseInsensitiveIdentifier, PipelineTableMetaData> result = new 
LinkedHashMap<>();
+        Map<CaseInsensitiveIdentifier, PipelineTableMetaData> result = new 
LinkedHashMap<>(tableNames.size(), 1F);
         for (String each : tableNames) {
             Set<String> primaryKeys = loadPrimaryKeys(connection, schemaName, 
each);
             Map<String, Collection<String>> uniqueKeys = 
loadUniqueIndexesOfTable(connection, schemaName, each);
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
index b62427ccfea..20f71add9c0 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
@@ -69,7 +69,7 @@ public final class PipelineJobDataSourcePreparer {
         }
         String defaultSchema = 
dialectDatabaseMetaData.getDefaultSchema().orElse(null);
         PipelinePrepareSQLBuilder pipelineSQLBuilder = new 
PipelinePrepareSQLBuilder(targetDatabaseType);
-        Collection<String> createdSchemaNames = new HashSet<>();
+        Collection<String> createdSchemaNames = new 
HashSet<>(param.getCreateTableConfigurations().size(), 1F);
         for (CreateTableConfiguration each : 
param.getCreateTableConfigurations()) {
             String targetSchemaName = 
each.getTargetName().getSchemaName().toString();
             if (null == targetSchemaName || 
targetSchemaName.equalsIgnoreCase(defaultSchema) || 
createdSchemaNames.contains(targetSchemaName)) {
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
index 86c461ad110..0544ce606e5 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
@@ -60,7 +60,7 @@ public final class InventoryRecordsCountCalculator {
             if (sql.isPresent()) {
                 DatabaseType databaseType = 
TypedSPILoader.getService(DatabaseType.class, 
dataSource.getDatabaseType().getType());
                 long result = getEstimatedCount(databaseType, dataSource, 
sql.get());
-                return result > 0 ? result : getCount(dataSource, 
pipelineSQLBuilder.buildCountSQL(schemaName, actualTableName));
+                return result > 0L ? result : getCount(dataSource, 
pipelineSQLBuilder.buildCountSQL(schemaName, actualTableName));
             }
             return getCount(dataSource, 
pipelineSQLBuilder.buildCountSQL(schemaName, actualTableName));
         } catch (final SQLException ex) {
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
index cfe3c9e87d6..88a2f5536a5 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
@@ -183,13 +183,13 @@ public final class InventoryTaskSplitter {
     
     private Collection<IngestPosition> 
getPositionByIntegerUniqueKeyRange(final InventoryDumperContext dumperContext, 
final long tableRecordsCount,
                                                                           
final TransmissionJobItemContext jobItemContext, final 
PipelineDataSourceWrapper dataSource) {
-        if (0 == tableRecordsCount) {
-            return Collections.singletonList(new 
IntegerPrimaryKeyIngestPosition(0, 0));
+        if (0L == tableRecordsCount) {
+            return Collections.singletonList(new 
IntegerPrimaryKeyIngestPosition(0L, 0L));
         }
         Collection<IngestPosition> result = new LinkedList<>();
         Range<Long> uniqueKeyValuesRange = 
getUniqueKeyValuesRange(jobItemContext, dataSource, dumperContext);
         int shardingSize = 
jobItemContext.getJobProcessContext().getProcessConfiguration().getRead().getShardingSize();
-        long splitCount = tableRecordsCount / shardingSize + 
(tableRecordsCount % shardingSize > 0 ? 1 : 0);
+        long splitCount = tableRecordsCount / shardingSize + 
(tableRecordsCount % shardingSize > 0L ? 1 : 0);
         long interval = (uniqueKeyValuesRange.getMaximum() - 
uniqueKeyValuesRange.getMinimum()) / splitCount;
         IntervalToRangeIterator rangeIterator = new 
IntervalToRangeIterator(uniqueKeyValuesRange.getMinimum(), 
uniqueKeyValuesRange.getMaximum(), interval);
         while (rangeIterator.hasNext()) {
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
index 4875f179837..acaf3779139 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
@@ -125,7 +125,7 @@ public final class PipelineJobCheckGovernanceRepository {
         if (null == checkResultMap) {
             return;
         }
-        Map<String, String> yamlCheckResultMap = new LinkedHashMap<>();
+        Map<String, String> yamlCheckResultMap = new 
LinkedHashMap<>(checkResultMap.size(), 1F);
         for (Entry<String, TableDataConsistencyCheckResult> entry : 
checkResultMap.entrySet()) {
             YamlTableDataConsistencyCheckResult yamlCheckResult = new 
YamlTableDataConsistencyCheckResultSwapper().swapToYamlConfiguration(entry.getValue());
             yamlCheckResultMap.put(entry.getKey(), 
YamlEngine.marshal(yamlCheckResult));
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
index 735d6d266d3..d7c7e7861de 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
@@ -51,7 +51,7 @@ public final class PipelineImportSQLBuilder {
     public PipelineImportSQLBuilder(final DatabaseType databaseType) {
         dialectSQLBuilder = 
DatabaseTypedSPILoader.getService(DialectPipelineSQLBuilder.class, 
databaseType);
         sqlSegmentBuilder = new PipelineSQLSegmentBuilder(databaseType);
-        sqlCache = 
Caffeine.newBuilder().initialCapacity(16).maximumSize(1024).build();
+        sqlCache = 
Caffeine.newBuilder().initialCapacity(16).maximumSize(1024L).build();
     }
     
     /**
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
index 2ac010aaaa7..ab8bbbfb7fa 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
@@ -40,12 +40,12 @@ public final class IntervalToRangeIterator implements 
Iterator<Range<Long>> {
         if (minimum > maximum) {
             throw new IllegalArgumentException("minimum greater than maximum");
         }
-        if (interval < 0) {
+        if (interval < 0L) {
             throw new IllegalArgumentException("interval is less than zero");
         }
         this.maximum = maximum;
         this.interval = interval;
-        this.current = minimum;
+        current = minimum;
     }
     
     @Override
@@ -56,11 +56,11 @@ public final class IntervalToRangeIterator implements 
Iterator<Range<Long>> {
     @Override
     public Range<Long> next() {
         if (!hasNext()) {
-            throw new NoSuchElementException();
+            throw new NoSuchElementException("");
         }
         long upperLimit = Math.min(maximum, current + interval);
         Range<Long> result = Range.between(current, upperLimit);
-        current = upperLimit + 1;
+        current = upperLimit + 1L;
         return result;
     }
 }
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
index 0d1773b5fd6..276b8581c70 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
@@ -55,7 +55,7 @@ public final class ShardingColumnsExtractor {
         }
         Set<String> defaultDatabaseShardingColumns = 
extractShardingColumns(shardingRuleConfig.get().getDefaultDatabaseShardingStrategy());
         Set<String> defaultTableShardingColumns = 
extractShardingColumns(shardingRuleConfig.get().getDefaultTableShardingStrategy());
-        Map<CaseInsensitiveIdentifier, Set<String>> result = new 
ConcurrentHashMap<>();
+        Map<CaseInsensitiveIdentifier, Set<String>> result = new 
ConcurrentHashMap<>(shardingRuleConfig.get().getTables().size(), 1F);
         for (ShardingTableRuleConfiguration each : 
shardingRuleConfig.get().getTables()) {
             CaseInsensitiveIdentifier logicTableName = new 
CaseInsensitiveIdentifier(each.getLogicTable());
             if (!logicTableNames.contains(logicTableName)) {
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
index 22e1292d10e..1d7f852b89b 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
@@ -84,7 +84,7 @@ class MultiplexPipelineChannelTest {
         MultiplexPipelineChannel channel = new 
MultiplexPipelineChannel(CHANNEL_NUMBER, 
TypedSPILoader.getService(PipelineChannelCreator.class, "MEMORY"), 10000, 
ackCallback);
         fetchWithMultiThreads(channel, countDownLatch);
         channel.push(Arrays.asList(records));
-        boolean awaitResult = countDownLatch.await(10, TimeUnit.SECONDS);
+        boolean awaitResult = countDownLatch.await(10L, TimeUnit.SECONDS);
         assertTrue(awaitResult, "await failed");
     }
     
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
index 9284bdff825..8174bc01255 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
@@ -33,7 +33,7 @@ public final class ConsistencyCheckDataBuilder {
      * @return built record
      */
     public static Map<String, Object> buildFixedFullTypeRecord(final int id) {
-        Map<String, Object> result = new LinkedHashMap<>();
+        Map<String, Object> result = new LinkedHashMap<>(15, 1F);
         result.put("id", id);
         result.put("c_bool", true);
         result.put("c_int1", Byte.MAX_VALUE);
@@ -97,13 +97,13 @@ public final class ConsistencyCheckDataBuilder {
             return value + "-";
         }
         if (value instanceof Time) {
-            return new Time(((Time) value).getTime() - 1);
+            return new Time(((Time) value).getTime() - 1L);
         }
         if (value instanceof Date) {
-            return new Date(((Date) value).getTime() - 1);
+            return new Date(((Date) value).getTime() - 1L);
         }
         if (value instanceof Timestamp) {
-            return new Timestamp(((Timestamp) value).getTime() - 1);
+            return new Timestamp(((Timestamp) value).getTime() - 1L);
         }
         if (value instanceof int[]) {
             int[] result = ((int[]) value).clone();
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
index 01676117059..7e2de6daa18 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
@@ -38,8 +38,8 @@ class DataConsistencyCheckUtilsTest {
     
     @Test
     void assertIsBigDecimalEquals() {
-        BigDecimal one = BigDecimal.valueOf(3322, 1);
-        BigDecimal another = BigDecimal.valueOf(33220, 2);
+        BigDecimal one = BigDecimal.valueOf(3322L, 1);
+        BigDecimal another = BigDecimal.valueOf(33220L, 2);
         assertTrue(DataConsistencyCheckUtils.isBigDecimalEquals(one, another));
     }
 }
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
index 04635829c2e..cf5b661cb60 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
@@ -93,7 +93,7 @@ class CRC32SingleTableInventoryCalculatorTest {
     
     @Test
     void assertCalculateFailed() throws SQLException {
-        when(connection.prepareStatement(anyString())).thenThrow(new 
SQLException());
+        when(connection.prepareStatement(anyString())).thenThrow(new 
SQLException(""));
         
assertThrows(PipelineTableDataConsistencyCheckLoadingFailedException.class, () 
-> new CRC32SingleTableInventoryCalculator().calculate(parameter));
     }
 }
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
index 33750199d46..2ae0b5792bb 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
@@ -35,7 +35,7 @@ class JobDataNodeLineConvertUtilsTest {
     
     @Test
     void assertConvertDataNodesToLines() {
-        Map<String, List<DataNode>> mockDataNodes = new LinkedHashMap<>();
+        Map<String, List<DataNode>> mockDataNodes = new LinkedHashMap<>(2, 1F);
         List<DataNode> dataNodes = Arrays.asList(new DataNode("ds_0", 
"t_order_0"), new DataNode("ds_0", "t_order_1"));
         List<DataNode> itemDataNodes = Collections.singletonList(new 
DataNode("ds_0", "t_order_item_0"));
         mockDataNodes.put("t_order", dataNodes);
@@ -51,10 +51,8 @@ class JobDataNodeLineConvertUtilsTest {
     
     @Test
     void assertConvertDataNodesToLinesWithMultipleDataSource() {
-        Map<String, List<DataNode>> mockDataNodes = new LinkedHashMap<>();
         List<DataNode> dataNodes = Arrays.asList(new DataNode("ds_0", 
"t_order_0"), new DataNode("ds_0", "t_order_2"), new DataNode("ds_1", 
"t_order_1"), new DataNode("ds_1", "t_order_3"));
-        mockDataNodes.put("t_order", dataNodes);
-        List<JobDataNodeLine> jobDataNodeLines = 
JobDataNodeLineConvertUtils.convertDataNodesToLines(mockDataNodes);
+        List<JobDataNodeLine> jobDataNodeLines = 
JobDataNodeLineConvertUtils.convertDataNodesToLines(Collections.singletonMap("t_order",
 dataNodes));
         assertThat(jobDataNodeLines.size(), is(2));
         JobDataNodeEntry jobDataNodeEntry = 
jobDataNodeLines.get(0).getEntries().iterator().next();
         
assertThat(jobDataNodeEntry.getDataNodes().stream().map(DataNode::getTableName).collect(Collectors.toList()),
 is(Arrays.asList("t_order_0", "t_order_2")));
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
index 327d472ea9f..9153d853bd2 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
@@ -34,6 +34,6 @@ class IntegerPrimaryKeyIngestPositionTest {
     
     @Test
     void assertToString() {
-        assertThat(new IntegerPrimaryKeyIngestPosition(1, 100).toString(), 
is("i,1,100"));
+        assertThat(new IntegerPrimaryKeyIngestPosition(1L, 100L).toString(), 
is("i,1,100"));
     }
 }
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
index fdbe2d2a88d..9a3bfdd5971 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
@@ -24,15 +24,11 @@ import 
org.apache.shardingsphere.data.pipeline.core.ingest.position.IngestPositi
 import 
org.apache.shardingsphere.data.pipeline.core.ingest.position.type.finished.IngestFinishedPosition;
 import 
org.apache.shardingsphere.data.pipeline.core.ingest.position.type.placeholder.IngestPlaceholderPosition;
 import org.apache.shardingsphere.data.pipeline.core.task.InventoryTask;
-import org.apache.shardingsphere.data.pipeline.core.task.PipelineTask;
 import 
org.apache.shardingsphere.data.pipeline.core.task.progress.InventoryTaskProgress;
 import org.junit.jupiter.api.Test;
 
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
+import java.util.Arrays;
+import java.util.Collections;
 import java.util.concurrent.atomic.AtomicReference;
 
 import static org.junit.jupiter.api.Assertions.assertFalse;
@@ -43,7 +39,7 @@ class PipelineJobProgressDetectorTest {
     
     @Test
     void assertAllInventoryTasksAreFinishedWhenCollectionIsEmpty() {
-        assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(new 
ArrayList<>()));
+        
assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(Collections.emptyList()));
     }
     
     @Test
@@ -52,10 +48,7 @@ class PipelineJobProgressDetectorTest {
         AtomicReference<IngestPosition> unfinishedPosition = new 
AtomicReference<>(new IngestPlaceholderPosition());
         InventoryTask actualTask1 = new InventoryTask("foo_id_1", 
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class), 
mock(Importer.class), finishedPosition);
         InventoryTask actualTask2 = new InventoryTask("foo_id_2", 
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class), 
mock(Importer.class), unfinishedPosition);
-        Collection<PipelineTask> inventoryTaskArrayList = new ArrayList<>();
-        inventoryTaskArrayList.add(actualTask1);
-        inventoryTaskArrayList.add(actualTask2);
-        
assertFalse(PipelineJobProgressDetector.isAllInventoryTasksFinished(inventoryTaskArrayList));
+        
assertFalse(PipelineJobProgressDetector.isAllInventoryTasksFinished(Arrays.asList(actualTask1,
 actualTask2)));
     }
     
     @Test
@@ -63,59 +56,42 @@ class PipelineJobProgressDetectorTest {
         AtomicReference<IngestPosition> finishedPosition = new 
AtomicReference<>(new IngestFinishedPosition());
         InventoryTask actualTask1 = new InventoryTask("foo_id_1", 
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class), 
mock(Importer.class), finishedPosition);
         InventoryTask actualTask2 = new InventoryTask("foo_id_2", 
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class), 
mock(Importer.class), finishedPosition);
-        Collection<PipelineTask> inventoryTaskArrayList = new ArrayList<>();
-        inventoryTaskArrayList.add(actualTask1);
-        inventoryTaskArrayList.add(actualTask2);
-        
assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(inventoryTaskArrayList));
+        
assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(Arrays.asList(actualTask1,
 actualTask2)));
     }
     
     @Test
     void assertIsInventoryFinishedWhenCollectionElementIsNull() {
         TransmissionJobItemProgress jobItemProgress = null;
-        Collection<TransmissionJobItemProgress> jobItemProgresses = new 
ArrayList<>();
-        jobItemProgresses.add(jobItemProgress);
-        assertFalse(PipelineJobProgressDetector.isInventoryFinished(1, 
jobItemProgresses));
+        assertFalse(PipelineJobProgressDetector.isInventoryFinished(1, 
Collections.singleton(jobItemProgress)));
     }
     
     @Test
     void assertIsInventoryFinishedWhenJobCountDoesNotMatchJobItemProgresses() {
         TransmissionJobItemProgress transmissionJobItemProgress = new 
TransmissionJobItemProgress();
-        List<TransmissionJobItemProgress> jobItemProgresses = new 
ArrayList<>();
-        jobItemProgresses.add(transmissionJobItemProgress);
-        assertFalse(PipelineJobProgressDetector.isInventoryFinished(2, 
jobItemProgresses));
+        assertFalse(PipelineJobProgressDetector.isInventoryFinished(2, 
Collections.singleton(transmissionJobItemProgress)));
     }
     
     @Test
     void assertIsInventoryFinishedWhenInventoryTaskProgressHasEmptyMap() {
-        JobItemInventoryTasksProgress jobItemInventoryTasksProgress = new 
JobItemInventoryTasksProgress(new HashMap<>());
+        JobItemInventoryTasksProgress jobItemInventoryTasksProgress = new 
JobItemInventoryTasksProgress(Collections.emptyMap());
         TransmissionJobItemProgress transmissionJobItemProgress = new 
TransmissionJobItemProgress();
         
transmissionJobItemProgress.setInventory(jobItemInventoryTasksProgress);
-        List<TransmissionJobItemProgress> jobItemProgresses = new 
ArrayList<>();
-        jobItemProgresses.add(transmissionJobItemProgress);
-        assertFalse(PipelineJobProgressDetector.isInventoryFinished(1, 
jobItemProgresses));
+        assertFalse(PipelineJobProgressDetector.isInventoryFinished(1, 
Collections.singleton(transmissionJobItemProgress)));
     }
     
     @Test
     void assertIsInventoryFinishedWhenNotAllInventoryTasksCompleted() {
-        Map<String, InventoryTaskProgress> progresses = new HashMap<>();
-        progresses.put("TEST", new InventoryTaskProgress(new 
IngestPlaceholderPosition()));
-        JobItemInventoryTasksProgress inventoryTasksProgress = new 
JobItemInventoryTasksProgress(progresses);
+        JobItemInventoryTasksProgress inventoryTasksProgress = new 
JobItemInventoryTasksProgress(Collections.singletonMap("TEST", new 
InventoryTaskProgress(new IngestPlaceholderPosition())));
         TransmissionJobItemProgress transmissionJobItemProgress = new 
TransmissionJobItemProgress();
         transmissionJobItemProgress.setInventory(inventoryTasksProgress);
-        List<TransmissionJobItemProgress> jobItemProgresses = new 
ArrayList<>();
-        jobItemProgresses.add(transmissionJobItemProgress);
-        assertFalse(PipelineJobProgressDetector.isInventoryFinished(1, 
jobItemProgresses));
+        assertFalse(PipelineJobProgressDetector.isInventoryFinished(1, 
Collections.singleton(transmissionJobItemProgress)));
     }
     
     @Test
     void assertIsInventoryFinished() {
-        Map<String, InventoryTaskProgress> progresses = new HashMap<>();
-        progresses.put("TEST", new InventoryTaskProgress(new 
IngestFinishedPosition()));
-        JobItemInventoryTasksProgress inventoryTasksProgress = new 
JobItemInventoryTasksProgress(progresses);
+        JobItemInventoryTasksProgress inventoryTasksProgress = new 
JobItemInventoryTasksProgress(Collections.singletonMap("TEST", new 
InventoryTaskProgress(new IngestFinishedPosition())));
         TransmissionJobItemProgress transmissionJobItemProgress = new 
TransmissionJobItemProgress();
         transmissionJobItemProgress.setInventory(inventoryTasksProgress);
-        List<TransmissionJobItemProgress> jobItemProgresses = new 
ArrayList<>();
-        jobItemProgresses.add(transmissionJobItemProgress);
-        assertTrue(PipelineJobProgressDetector.isInventoryFinished(1, 
jobItemProgresses));
+        assertTrue(PipelineJobProgressDetector.isInventoryFinished(1, 
Collections.singleton(transmissionJobItemProgress)));
     }
 }
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
index 0831af6e1d8..6588d0bf8de 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
@@ -83,8 +83,8 @@ class TransmissionJobItemProgressTest {
     
     @Test
     void assertGetProgressesCorrectly() {
-        Map<String, InventoryTaskProgress> progresses = new HashMap<>();
-        progresses.put("ds.order_item#0", new InventoryTaskProgress(new 
IntegerPrimaryKeyIngestPosition(1, 100)));
+        Map<String, InventoryTaskProgress> progresses = new HashMap<>(4, 1F);
+        progresses.put("ds.order_item#0", new InventoryTaskProgress(new 
IntegerPrimaryKeyIngestPosition(1L, 100L)));
         progresses.put("ds.order_item#1", new InventoryTaskProgress(new 
UnsupportedKeyIngestPosition()));
         progresses.put("ds.order#0", new InventoryTaskProgress(new 
IngestFinishedPosition()));
         progresses.put("ds.test_order#0", new InventoryTaskProgress(new 
StringPrimaryKeyIngestPosition("1", "100")));
diff --git 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
index 2f2dd47479b..7a0e9ce8467 100644
--- 
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
+++ 
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
@@ -17,7 +17,6 @@
 
 package org.apache.shardingsphere.data.pipeline.core.metadata.node;
 
-import org.hamcrest.MatcherAssert;
 import org.junit.jupiter.api.Test;
 
 import static org.hamcrest.CoreMatchers.is;
@@ -37,7 +36,7 @@ class PipelineMetaDataNodeTest {
     
     @Test
     void assertGetMetaDataDataSourcesPath() {
-        
MatcherAssert.assertThat(PipelineMetaDataNode.getMetaDataDataSourcesPath("FIXTURE"),
 is(migrationMetaDataRootPath + "/data_sources"));
+        assertThat(PipelineMetaDataNode.getMetaDataDataSourcesPath("FIXTURE"), 
is(migrationMetaDataRootPath + "/data_sources"));
     }
     
     @Test
diff --git 
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
 
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
index 095bcc999a6..78a215e80da 100644
--- 
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
+++ 
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
@@ -40,6 +40,6 @@ public final class GlobalTableMapEventMapping {
      * @return table map event map
      */
     public static Map<Long, MySQLBinlogTableMapEventPacket> 
getTableMapEventMap(final String databaseURL) {
-        return TABLE_MAP_EVENT_MAPPING.computeIfAbsent(databaseURL, k -> new 
ConcurrentHashMap<>());
+        return TABLE_MAP_EVENT_MAPPING.computeIfAbsent(databaseURL, key -> new 
ConcurrentHashMap<>());
     }
 }
diff --git 
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
 
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
index 9ea0b0809b0..df60c00de2f 100644
--- 
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
+++ 
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
@@ -361,7 +361,7 @@ public final class MySQLClient {
         private void reconnect() throws ExecutionException, 
InterruptedException, TimeoutException {
             Optional<ChannelFuture> future = closeChannel();
             if (future.isPresent()) {
-                future.get().get(1, TimeUnit.SECONDS);
+                future.get().get(1L, TimeUnit.SECONDS);
             }
             if (reconnectTimes.incrementAndGet() > 3) {
                 log.warn("Exceeds the maximum number of retry times, last 
binlog event:{}", lastBinlogEvent);
diff --git 
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
 
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
index 1a88f402bac..6691cc0207e 100644
--- 
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
+++ 
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
@@ -35,7 +35,7 @@ public final class MySQLUnsignedBigintHandler implements 
MySQLDataTypeHandler {
             return null;
         }
         long longValue = (long) value;
-        return 0 > longValue ? 
BIGINT_MODULO.add(BigInteger.valueOf(longValue)) : longValue;
+        return 0L > longValue ? 
BIGINT_MODULO.add(BigInteger.valueOf(longValue)) : longValue;
     }
     
     @Override
diff --git 
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
 
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
index eca2f990241..a6107d23c0b 100644
--- 
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
+++ 
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
@@ -42,7 +42,7 @@ class MySQLIngestPositionManagerTest {
     
     private static final long LOG_POSITION = 4L;
     
-    private static final long SERVER_ID = 555555;
+    private static final long SERVER_ID = 555555L;
     
     @Mock(extraInterfaces = AutoCloseable.class)
     private DataSource dataSource;
diff --git 
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
 
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
index 5d7a31939a8..31ae65d93bc 100644
--- 
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
+++ 
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
@@ -26,6 +26,6 @@ class BinlogPositionTest {
     
     @Test
     void assertToString() {
-        assertThat(new BinlogPosition("mysql-bin.000001", 4, 0L).toString(), 
is("mysql-bin.000001#4"));
+        assertThat(new BinlogPosition("mysql-bin.000001", 4L, 0L).toString(), 
is("mysql-bin.000001#4"));
     }
 }
diff --git 
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
 
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
index f117d8c680c..0485c7d65be 100644
--- 
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
+++ 
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
@@ -100,7 +100,7 @@ class MySQLClientTest {
     
     @Test
     void assertExecuteUpdate() throws ReflectiveOperationException {
-        MySQLOKPacket expected = new MySQLOKPacket(10, 0, 0);
+        MySQLOKPacket expected = new MySQLOKPacket(10L, 0L, 0);
         
Plugins.getMemberAccessor().set(MySQLOKPacket.class.getDeclaredField("affectedRows"),
 expected, 10L);
         mockChannelResponse(expected);
         
Plugins.getMemberAccessor().set(MySQLClient.class.getDeclaredField("channel"), 
mysqlClient, channel);
diff --git 
a/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
 
b/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
index bbed76a6b96..b50fcbc6eee 100644
--- 
a/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
+++ 
b/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
@@ -95,7 +95,7 @@ public final class OpenGaussWALDumper extends 
AbstractPipelineLifecycleRunnable
         this.channel = channel;
         walEventConverter = new WALEventConverter(dumperContext, 
metaDataLoader);
         logicalReplication = new OpenGaussLogicalReplication();
-        this.decodeWithTX = dumperContext.isDecodeWithTX();
+        decodeWithTX = dumperContext.isDecodeWithTX();
     }
     
     @SneakyThrows(InterruptedException.class)
diff --git 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
index 1c0e88a6ca0..de2ae2abc0b 100644
--- 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
+++ 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
@@ -65,8 +65,9 @@ public abstract class AbstractPostgreSQLDDLAdapter {
         ResultSetMetaData metaData = resultSet.getMetaData();
         Collection<Map<String, Object>> result = new LinkedList<>();
         while (resultSet.next()) {
-            Map<String, Object> row = new LinkedHashMap<>();
-            for (int i = 1; i <= metaData.getColumnCount(); i++) {
+            int columnCount = metaData.getColumnCount();
+            Map<String, Object> row = new LinkedHashMap<>(columnCount);
+            for (int i = 1; i <= columnCount; i++) {
                 row.put(metaData.getColumnName(i), resultSet.getObject(i));
             }
             result.add(row);
@@ -86,9 +87,10 @@ public abstract class AbstractPostgreSQLDDLAdapter {
     
     private Map<String, Object> getSingleRow(final ResultSet resultSet) throws 
SQLException {
         ResultSetMetaData metaData = resultSet.getMetaData();
-        Map<String, Object> result = new LinkedHashMap<>();
+        int columnCount = metaData.getColumnCount();
+        Map<String, Object> result = new LinkedHashMap<>(columnCount);
         if (resultSet.next()) {
-            for (int i = 1; i <= metaData.getColumnCount(); i++) {
+            for (int i = 1; i <= columnCount; i++) {
                 result.put(metaData.getColumnName(i), resultSet.getObject(i));
             }
         }
@@ -102,7 +104,7 @@ public abstract class AbstractPostgreSQLDDLAdapter {
         Collection<Map<String, String>> formatLabels = new LinkedList<>();
         Collection<String> securityLabels = Arrays.stream((String[]) ((Array) 
data.get("seclabels")).getArray()).collect(Collectors.toList());
         for (String each : securityLabels) {
-            Map<String, String> securityLabel = new LinkedHashMap<>();
+            Map<String, String> securityLabel = new LinkedHashMap<>(2, 1F);
             securityLabel.put("provider", each.substring(0, 
each.indexOf(SECURITY_LABEL_SPLIT)));
             securityLabel.put("label", 
each.substring(each.indexOf(SECURITY_LABEL_SPLIT) + 1));
             formatLabels.add(securityLabel);
diff --git 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
index ea99972f0bd..4be24d07ca3 100644
--- 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
+++ 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
@@ -221,7 +221,7 @@ public final class PostgreSQLColumnPropertiesAppender 
extends AbstractPostgreSQL
             name = name.substring(1, name.length() - 1);
         }
         Integer typmod = (Integer) column.get("atttypmod");
-        String length = -1 != typmod ? checkTypmod(typmod, name) : "";
+        String length = -1 == typmod ? "" : checkTypmod(typmod, name);
         return getFullTypeValue(name, schema, length, numdims == 1 ? "[]" : 
"");
     }
     
diff --git 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
index 649a20e900b..5fbf94c41e3 100644
--- 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
+++ 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
@@ -44,7 +44,7 @@ public enum PostgreSQLColumnType {
      * @return value of column type
      */
     public static PostgreSQLColumnType valueOf(final Long elemoid) {
-        if (0 == elemoid) {
+        if (0L == elemoid) {
             return UNKNOWN;
         }
         for (PostgreSQLColumnType each : values()) {
diff --git 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
index 5f7f2adac97..2d1c5d9dff6 100644
--- 
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
+++ 
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
@@ -85,7 +85,7 @@ public final class PostgreSQLWALDumper extends 
AbstractPipelineLifecycleRunnable
         this.channel = channel;
         walEventConverter = new WALEventConverter(dumperContext, 
metaDataLoader);
         logicalReplication = new PostgreSQLLogicalReplication();
-        this.decodeWithTX = dumperContext.isDecodeWithTX();
+        decodeWithTX = dumperContext.isDecodeWithTX();
     }
     
     @SneakyThrows(InterruptedException.class)
@@ -100,7 +100,7 @@ public final class PostgreSQLWALDumper extends 
AbstractPipelineLifecycleRunnable
                 int times = reconnectTimes.incrementAndGet();
                 log.error("Connect failed, reconnect times={}", times, ex);
                 if (isRunning()) {
-                    Thread.sleep(5000);
+                    Thread.sleep(5000L);
                 }
                 if (times >= 5) {
                     throw new IngestException(ex);
diff --git 
a/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
 
b/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
index fb1782907da..8492f2756c7 100644
--- 
a/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
+++ 
b/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
@@ -153,7 +153,7 @@ class WALEventConverterTest {
     
     @Test
     void assertConvertCommitTXEvent() {
-        CommitTXEvent commitTXEvent = new CommitTXEvent(1, 3468L);
+        CommitTXEvent commitTXEvent = new CommitTXEvent(1L, 3468L);
         commitTXEvent.setLogSequenceNumber(new 
PostgreSQLLogSequenceNumber(logSequenceNumber));
         Record record = walEventConverter.convert(commitTXEvent);
         assertInstanceOf(PlaceholderRecord.class, record);
diff --git 
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
 
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
index ca408b4e353..be7253f94fe 100644
--- 
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
+++ 
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
@@ -67,7 +67,7 @@ public final class ShowStreamingJobStatusExecutor implements 
DistSQLQueryExecuto
     }
     
     private static Optional<Long> getIncrementalIdleSeconds(final 
TransmissionJobItemProgress jobItemProgress, final TransmissionJobItemInfo 
transmissionJobItemInfo, final long currentTimeMillis) {
-        if 
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0) {
+        if 
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0L) {
             long latestActiveTimeMillis = 
Math.max(transmissionJobItemInfo.getStartTimeMillis(), 
jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis());
             return 
Optional.of(TimeUnit.MILLISECONDS.toSeconds(currentTimeMillis - 
latestActiveTimeMillis));
         }
diff --git 
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
 
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
index 83bf0bf39ee..bb6e3553924 100644
--- 
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
+++ 
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
@@ -60,7 +60,7 @@ public final class ShowMigrationJobStatusExecutor implements 
DistSQLQueryExecuto
     }
     
     private Optional<Long> getIncrementalIdleSeconds(final 
TransmissionJobItemProgress jobItemProgress, final TransmissionJobItemInfo 
jobItemInfo, final long currentTimeMillis) {
-        if 
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0) {
+        if 
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0L) {
             long latestActiveTimeMillis = 
Math.max(jobItemInfo.getStartTimeMillis(), 
jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis());
             return 
Optional.of(TimeUnit.MILLISECONDS.toSeconds(currentTimeMillis - 
latestActiveTimeMillis));
         }
diff --git 
a/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
 
b/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
index e1eebc88028..0a563cb3b01 100644
--- 
a/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
+++ 
b/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
@@ -98,7 +98,7 @@ public final class ProtobufAnyValueConverter {
     }
     
     private static Timestamp converProtobufTimestamp(final 
com.google.protobuf.Timestamp timestamp) {
-        Timestamp result = new Timestamp(timestamp.getSeconds() * 1000);
+        Timestamp result = new Timestamp(timestamp.getSeconds() * 1000L);
         result.setNanos(timestamp.getNanos());
         return result;
     }
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
index 47fc27d3e94..eb97ea51f0c 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
@@ -72,7 +72,7 @@ public final class CDCJobType implements PipelineJobType {
     @Override
     public PipelineDataConsistencyChecker buildDataConsistencyChecker(final 
PipelineJobConfiguration jobConfig, final TransmissionProcessContext 
processContext,
                                                                       final 
ConsistencyCheckJobItemProgressContext progressContext) {
-        throw new UnsupportedOperationException();
+        throw new UnsupportedOperationException("");
     }
     
     @Override
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
index 29e65fbfce1..b91a975f7a9 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
@@ -168,7 +168,7 @@ public final class CDCJobAPI implements TransmissionJobAPI {
     }
     
     private ShardingSpherePipelineDataSourceConfiguration 
getDataSourceConfiguration(final ShardingSphereDatabase database) {
-        Map<String, Map<String, Object>> dataSourcePoolProps = new HashMap<>();
+        Map<String, Map<String, Object>> dataSourcePoolProps = new 
HashMap<>(database.getResourceMetaData().getStorageUnits().size(), 1F);
         for (Entry<String, StorageUnit> entry : 
database.getResourceMetaData().getStorageUnits().entrySet()) {
             dataSourcePoolProps.put(entry.getKey(), 
dataSourceConfigSwapper.swapToMap(entry.getValue().getDataSourcePoolProperties()));
         }
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
index dfb8b23d5b0..b721cc44b7b 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
@@ -71,9 +71,9 @@ public final class CDCJobItemContext implements 
TransmissionJobItemContext {
     
     private final Collection<PipelineTask> incrementalTasks = new 
LinkedList<>();
     
-    private final AtomicLong processedRecordsCount = new AtomicLong(0);
+    private final AtomicLong processedRecordsCount = new AtomicLong(0L);
     
-    private final AtomicLong inventoryRecordsCount = new AtomicLong(0);
+    private final AtomicLong inventoryRecordsCount = new AtomicLong(0L);
     
     private final LazyInitializer<PipelineDataSourceWrapper> 
sourceDataSourceLazyInitializer = new 
LazyInitializer<PipelineDataSourceWrapper>() {
         
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
index ff130efc1a6..2357fb41384 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
@@ -71,7 +71,7 @@ public final class CDCImporter extends 
AbstractPipelineLifecycleRunnable impleme
     
     private final PriorityQueue<CSNRecords> csnRecordsQueue = new 
PriorityQueue<>(new CSNRecordsComparator());
     
-    private final Cache<String, List<Pair<CDCChannelProgressPair, 
CDCAckPosition>>> ackCache = 
Caffeine.newBuilder().maximumSize(10000).expireAfterAccess(5, 
TimeUnit.MINUTES).build();
+    private final Cache<String, List<Pair<CDCChannelProgressPair, 
CDCAckPosition>>> ackCache = 
Caffeine.newBuilder().maximumSize(10000L).expireAfterAccess(5L, 
TimeUnit.MINUTES).build();
     
     @Override
     protected void runBlocking() {
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
index 825e50df67a..4baf8622e45 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
@@ -110,7 +110,7 @@ public final class CDCBackendHandler {
             schemaTableNameMap = 
CDCSchemaTableUtils.parseTableExpressionWithSchema(database, 
requestBody.getSourceSchemaTableList());
             // TODO if different schema have same table names, table name may 
be overwritten, because the table name at sharding rule not contain schema.
             tableNames = 
schemaTableNameMap.values().stream().flatMap(Collection::stream).collect(Collectors.toList());
-            schemaTableNameMap.forEach((k, v) -> v.forEach(tableName -> 
schemaTableNames.add(k.isEmpty() ? tableName : String.join(".", k, 
tableName))));
+            schemaTableNameMap.forEach((key, value) -> value.forEach(tableName 
-> schemaTableNames.add(key.isEmpty() ? tableName : String.join(".", key, 
tableName))));
         } else {
             
schemaTableNames.addAll(CDCSchemaTableUtils.parseTableExpressionWithoutSchema(database,
 requestBody.getSourceSchemaTableList().stream().map(SchemaTable::getTable)
                     .collect(Collectors.toList())));
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
index a7fab4f593e..2fc2d55845d 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
@@ -54,7 +54,7 @@ public final class CDCDataNodeUtils {
         Optional<ShardingRule> shardingRule = 
database.getRuleMetaData().findSingleRule(ShardingRule.class);
         Optional<SingleRule> singleRule = 
database.getRuleMetaData().findSingleRule(SingleRule.class);
         Optional<BroadcastRule> broadcastRule = 
database.getRuleMetaData().findSingleRule(BroadcastRule.class);
-        Map<String, List<DataNode>> result = new HashMap<>();
+        Map<String, List<DataNode>> result = new HashMap<>(tableNames.size(), 
1F);
         // TODO support virtual data source name
         for (String each : tableNames) {
             if (singleRule.isPresent() && 
singleRule.get().getAttributes().getAttribute(DataNodeRuleAttribute.class).getAllDataNodes().containsKey(each))
 {
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
index aaed4565c58..be6fcb8150b 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
@@ -77,7 +77,7 @@ public final class CDCSchemaTableUtils {
     }
     
     private static Map<String, Set<String>> 
parseTableExpressionWithAllTables(final ShardingSphereDatabase database, final 
Collection<String> systemSchemas) {
-        Map<String, Set<String>> result = new 
HashMap<>(database.getSchemas().size(), 1);
+        Map<String, Set<String>> result = new 
HashMap<>(database.getSchemas().size(), 1F);
         for (Entry<String, ShardingSphereSchema> entry : 
database.getSchemas().entrySet()) {
             if (!systemSchemas.contains(entry.getKey())) {
                 entry.getValue().getAllTableNames().forEach(tableName -> 
result.computeIfAbsent(entry.getKey(), ignored -> new 
HashSet<>()).add(tableName));
@@ -88,7 +88,7 @@ public final class CDCSchemaTableUtils {
     }
     
     private static Map<String, Set<String>> 
parseTableExpressionWithAllSchema(final ShardingSphereDatabase database, final 
Collection<String> systemSchemas, final SchemaTable table) {
-        Map<String, Set<String>> result = new HashMap<>();
+        Map<String, Set<String>> result = new 
HashMap<>(database.getSchemas().size(), 1F);
         for (Entry<String, ShardingSphereSchema> entry : 
database.getSchemas().entrySet()) {
             if (!systemSchemas.contains(entry.getKey())) {
                 entry.getValue().getAllTableNames().stream().filter(tableName 
-> tableName.equals(table.getTable())).findFirst()
@@ -99,11 +99,11 @@ public final class CDCSchemaTableUtils {
     }
     
     private static Map<String, Set<String>> 
parseTableExpressionWithAllTable(final ShardingSphereDatabase database, final 
SchemaTable each) {
-        Map<String, Set<String>> result = new HashMap<>();
         DialectDatabaseMetaData dialectDatabaseMetaData = new 
DatabaseTypeRegistry(database.getProtocolType()).getDialectDatabaseMetaData();
         String schemaName = each.getSchema().isEmpty() ? 
dialectDatabaseMetaData.getDefaultSchema().orElseThrow(() -> new 
IllegalStateException("Default schema should exist.")) : each.getSchema();
         ShardingSphereSchema schema = database.getSchema(schemaName);
         ShardingSpherePreconditions.checkNotNull(schema, () -> new 
SchemaNotFoundException(each.getSchema()));
+        Map<String, Set<String>> result = new 
HashMap<>(schema.getAllTableNames().size(), 1F);
         schema.getAllTableNames().forEach(tableName -> 
result.computeIfAbsent(schemaName, ignored -> new HashSet<>()).add(tableName));
         return result;
     }
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
index 1f61502d80a..0c200c8fc2f 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
@@ -135,11 +135,11 @@ public final class ColumnValueConvertUtils {
         }
         if (object instanceof Clob) {
             Clob clob = (Clob) object;
-            return StringValue.of(clob.getSubString(1, (int) clob.length()));
+            return StringValue.of(clob.getSubString(1L, (int) clob.length()));
         }
         if (object instanceof Blob) {
             Blob blob = (Blob) object;
-            return BytesValue.of(ByteString.copyFrom(blob.getBytes(1, (int) 
blob.length())));
+            return BytesValue.of(ByteString.copyFrom(blob.getBytes(1L, (int) 
blob.length())));
         }
         return StringValue.newBuilder().setValue(object.toString()).build();
     }
@@ -147,9 +147,9 @@ public final class ColumnValueConvertUtils {
     private static com.google.protobuf.Timestamp 
converToProtobufTimestamp(final Date timestamp) {
         if (timestamp instanceof Timestamp) {
             Timestamp value = (Timestamp) timestamp;
-            return 
com.google.protobuf.Timestamp.newBuilder().setSeconds(value.getTime() / 
1000).setNanos(value.getNanos()).build();
+            return 
com.google.protobuf.Timestamp.newBuilder().setSeconds(value.getTime() / 
1000L).setNanos(value.getNanos()).build();
         }
         long millis = timestamp.getTime();
-        return com.google.protobuf.Timestamp.newBuilder().setSeconds(millis / 
1000).setNanos((int) ((millis % 1000) * 1000000)).build();
+        return com.google.protobuf.Timestamp.newBuilder().setSeconds(millis / 
1000L).setNanos((int) ((millis % 1000L) * 1000000L)).build();
     }
 }
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
index 220be845acc..09ae021e0d7 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
@@ -42,13 +42,13 @@ class CDCSchemaTableUtilsTest {
     
     @Test
     void assertParseTableExpression() {
-        Map<String, ShardingSphereSchema> schemas = new HashMap<>();
+        Map<String, ShardingSphereSchema> schemas = new HashMap<>(2, 1F);
         schemas.put("public", mockedPublicSchema());
         schemas.put("test", mockedTestSchema());
         ShardingSphereDatabase database = new 
ShardingSphereDatabase("sharding_db", 
TypedSPILoader.getService(DatabaseType.class, "openGauss"), null, null, 
schemas);
         List<SchemaTable> schemaTables = 
Arrays.asList(SchemaTable.newBuilder().setSchema("public").setTable("t_order").build(),
                 
SchemaTable.newBuilder().setSchema("test").setTable("*").build());
-        Map<String, Set<String>> expected = new HashMap<>();
+        Map<String, Set<String>> expected = new HashMap<>(2, 1F);
         expected.put("test", new HashSet<>(Arrays.asList("t_order_item", 
"t_order_item2")));
         expected.put("public", Collections.singleton("t_order"));
         Map<String, Set<String>> actual = 
CDCSchemaTableUtils.parseTableExpressionWithSchema(database, schemaTables);
@@ -59,8 +59,7 @@ class CDCSchemaTableUtilsTest {
         assertThat(actual, is(expected));
         schemaTables = 
Collections.singletonList(SchemaTable.newBuilder().setSchema("*").setTable("t_order").build());
         actual = CDCSchemaTableUtils.parseTableExpressionWithSchema(database, 
schemaTables);
-        expected = new HashMap<>();
-        expected.put("public", Collections.singleton("t_order"));
+        expected = Collections.singletonMap("public", 
Collections.singleton("t_order"));
         assertThat(actual, is(expected));
     }
     
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
index e9d33ef5d18..86974d9597b 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
@@ -110,7 +110,7 @@ class ColumnValueConvertUtilsTest {
     
     @Test
     void assertTimeConvert() {
-        Time time = new Time(-3600 * 1000 - 1234);
+        Time time = new Time(-3600L * 1000L - 1234L);
         int nanos = new Timestamp(time.getTime()).getNanos();
         Int64Value actualMessage = (Int64Value) 
ColumnValueConvertUtils.convertToProtobufMessage(time);
         assertThat(LocalTime.ofNanoOfDay(actualMessage.getValue()), 
is(time.toLocalTime().withNano(nanos)));
diff --git 
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
 
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
index ffbb09e938b..0f3c39146d5 100644
--- 
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
+++ 
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
@@ -55,7 +55,7 @@ class DataRecordResultConvertUtilsTest {
     
     @Test
     void assertConvertDataRecordToRecord() throws 
InvalidProtocolBufferException, SQLException {
-        DataRecord dataRecord = new 
DataRecord(PipelineSQLOperationType.INSERT, "t_order", new 
IntegerPrimaryKeyIngestPosition(0, 1), 2);
+        DataRecord dataRecord = new 
DataRecord(PipelineSQLOperationType.INSERT, "t_order", new 
IntegerPrimaryKeyIngestPosition(0L, 1L), 2);
         dataRecord.addColumn(new Column("order_id", BigInteger.ONE, false, 
true));
         dataRecord.addColumn(new Column("price", BigDecimal.valueOf(123), 
false, false));
         dataRecord.addColumn(new Column("user_id", Long.MAX_VALUE, false, 
false));
diff --git 
a/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
 
b/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
index c22d89843c2..5726948ca12 100644
--- 
a/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
+++ 
b/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
@@ -243,8 +243,8 @@ public final class ConsistencyCheckJobAPI {
             result.setDurationSeconds(duration.getSeconds());
             
result.setCheckEndTime(DateTimeFormatterFactory.getLongMillsFormatter().format(checkEndTime));
             result.setInventoryRemainingSeconds(0L);
-        } else if (0 != recordsCount && 0 != checkedRecordsCount) {
-            result.setInventoryFinishedPercentage((int) (checkedRecordsCount * 
100 / recordsCount));
+        } else if (0L != recordsCount && 0L != checkedRecordsCount) {
+            result.setInventoryFinishedPercentage((int) (checkedRecordsCount * 
100L / recordsCount));
             LocalDateTime stopTime = jobConfigPOJO.isDisabled() ? 
LocalDateTime.from(DateTimeFormatterFactory.getStandardFormatter().parse(jobConfigPOJO.getProps().getProperty("stop_time")))
                     : null;
             long durationMillis = (null != stopTime ? 
Timestamp.valueOf(stopTime).getTime() : System.currentTimeMillis()) - 
jobItemProgress.getCheckBeginTimeMillis();
@@ -252,8 +252,8 @@ public final class ConsistencyCheckJobAPI {
             if (null != stopTime) {
                 
result.setCheckEndTime(jobConfigPOJO.getProps().getProperty("stop_time"));
             }
-            long remainingMills = Math.max(0, (long) ((recordsCount - 
checkedRecordsCount) * 1.0D / checkedRecordsCount * durationMillis));
-            result.setInventoryRemainingSeconds(remainingMills / 1000);
+            long remainingMills = Math.max(0L, (long) ((recordsCount - 
checkedRecordsCount) * 1.0D / checkedRecordsCount * durationMillis));
+            result.setInventoryRemainingSeconds(remainingMills / 1000L);
         }
     }
     
diff --git 
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
 
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
index 5e2ff9914d0..ba4e11f69de 100644
--- 
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
+++ 
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
@@ -173,7 +173,7 @@ public final class MigrationJobAPI implements 
TransmissionJobAPI {
     }
     
     private PipelineDataSourceConfiguration 
buildTargetPipelineDataSourceConfiguration(final ShardingSphereDatabase 
targetDatabase) {
-        Map<String, Map<String, Object>> targetPoolProps = new HashMap<>();
+        Map<String, Map<String, Object>> targetPoolProps = new 
HashMap<>(targetDatabase.getResourceMetaData().getStorageUnits().size(), 1F);
         YamlDataSourceConfigurationSwapper dataSourceConfigSwapper = new 
YamlDataSourceConfigurationSwapper();
         for (Entry<String, StorageUnit> entry : 
targetDatabase.getResourceMetaData().getStorageUnits().entrySet()) {
             targetPoolProps.put(entry.getKey(), 
dataSourceConfigSwapper.swapToMap(entry.getValue().getDataSourcePoolProperties()));
@@ -192,7 +192,7 @@ public final class MigrationJobAPI implements 
TransmissionJobAPI {
     }
     
     private Map<String, String> buildTargetTableSchemaMap(final Map<String, 
List<DataNode>> sourceDataNodes) {
-        Map<String, String> result = new LinkedHashMap<>();
+        Map<String, String> result = new 
LinkedHashMap<>(sourceDataNodes.size(), 1F);
         sourceDataNodes.forEach((tableName, dataNodes) -> 
result.put(tableName, dataNodes.get(0).getSchemaName()));
         return result;
     }
@@ -285,7 +285,7 @@ public final class MigrationJobAPI implements 
TransmissionJobAPI {
     }
     
     private void refreshTableMetadata(final String jobId, final String 
databaseName) {
-        // TODO use origin database name now, wait reloadDatabaseMetaData fix 
case-sensitive probelm
+        // TODO use origin database name now, wait reloadDatabaseMetaData fix 
case-sensitive problem
         ContextManager contextManager = 
PipelineContextManager.getContext(PipelineJobIdUtils.parseContextKey(jobId)).getContextManager();
         ShardingSphereDatabase database = 
contextManager.getMetaDataContexts().getMetaData().getDatabase(databaseName);
         contextManager.refreshTableMetaData(database);
diff --git 
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
 
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
index a7caa8aa922..743ca8f47e4 100644
--- 
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
+++ 
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
@@ -67,9 +67,9 @@ public final class MigrationJobItemContext implements 
TransmissionJobItemContext
     
     private final Collection<PipelineTask> incrementalTasks = new 
LinkedList<>();
     
-    private final AtomicLong processedRecordsCount = new AtomicLong(0);
+    private final AtomicLong processedRecordsCount = new AtomicLong(0L);
     
-    private final AtomicLong inventoryRecordsCount = new AtomicLong(0);
+    private final AtomicLong inventoryRecordsCount = new AtomicLong(0L);
     
     private final MigrationJobConfiguration jobConfig;
     
@@ -98,7 +98,7 @@ public final class MigrationJobItemContext implements 
TransmissionJobItemContext
         this.jobConfig = jobConfig;
         jobId = jobConfig.getJobId();
         this.shardingItem = shardingItem;
-        this.dataSourceName = taskConfig.getDataSourceName();
+        dataSourceName = taskConfig.getDataSourceName();
         this.initProgress = initProgress;
         if (null != initProgress) {
             processedRecordsCount.set(initProgress.getProcessedRecordsCount());
diff --git 
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
 
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
index b0c695fc383..a893e286c24 100644
--- 
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
+++ 
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
@@ -130,7 +130,7 @@ public final class MigrationJobPreparer {
         }
         LockDefinition lockDefinition = new 
GlobalLockDefinition(String.format(GlobalLockNames.PREPARE.getLockName(), 
jobConfig.getJobId()));
         long startTimeMillis = System.currentTimeMillis();
-        if (lockContext.tryLock(lockDefinition, 600000)) {
+        if (lockContext.tryLock(lockDefinition, 600000L)) {
             log.info("try lock success, jobId={}, shardingItem={}, cost {} 
ms", jobId, jobItemContext.getShardingItem(), System.currentTimeMillis() - 
startTimeMillis);
             try {
                 JobOffsetInfo offsetInfo = 
PipelineAPIFactory.getPipelineGovernanceFacade(PipelineJobIdUtils.parseContextKey(jobId)).getJobFacade().getOffset().load(jobId);
diff --git 
a/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
 
b/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
index 48b69f7d39a..67838b40bd9 100644
--- 
a/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
+++ 
b/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
@@ -32,7 +32,7 @@ public final class RedisTSOProvider implements TSOProvider {
     
     private static final String CSN_KEY = "csn";
     
-    private static final long ERROR_CSN = 0;
+    private static final long ERROR_CSN = 0L;
     
     private static final long INIT_CSN = Integer.MAX_VALUE;
     
diff --git 
a/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
 
b/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
index 206e713952a..440283a7bc1 100644
--- 
a/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
+++ 
b/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
@@ -64,7 +64,7 @@ public final class LoggingUtils {
      * @param props configuration properties
      */
     public static void syncLoggingConfig(final LoggingRuleConfiguration 
loggingRuleConfig, final ConfigurationProperties props) {
-        LoggingUtils.getSQLLogger(loggingRuleConfig).ifPresent(option -> {
+        getSQLLogger(loggingRuleConfig).ifPresent(option -> {
             Properties loggerProps = option.getProps();
             syncPropsToLoggingRule(loggerProps, props);
             syncLoggingRuleToProps(loggerProps, props);
diff --git 
a/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
 
b/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
index 351629379eb..d3034bea2ee 100644
--- 
a/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
+++ 
b/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
@@ -48,8 +48,8 @@ public final class DatabaseMetaDataPersistService implements 
DatabaseMetaDataBas
     public DatabaseMetaDataPersistService(final PersistRepository repository, 
final MetaDataVersionPersistService metaDataVersionPersistService) {
         this.repository = repository;
         this.metaDataVersionPersistService = metaDataVersionPersistService;
-        this.tableMetaDataPersistService = new 
TableMetaDataPersistService(repository, metaDataVersionPersistService);
-        this.viewMetaDataPersistService = new 
ViewMetaDataPersistService(repository, metaDataVersionPersistService);
+        tableMetaDataPersistService = new 
TableMetaDataPersistService(repository, metaDataVersionPersistService);
+        viewMetaDataPersistService = new 
ViewMetaDataPersistService(repository, metaDataVersionPersistService);
     }
     
     @Override
diff --git 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
index 42500e70f37..d8443e5d7ca 100644
--- 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
+++ 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
@@ -114,7 +114,7 @@ public final class SingleTableDataNodeLoader {
     
     private static Map<String, Collection<DataNode>> 
loadSpecifiedDataNodes(final Map<String, Collection<DataNode>> actualDataNodes, 
final Collection<String> featureRequiredSingleTables,
                                                                             
final Map<String, Map<String, Collection<String>>> configuredTableMap) {
-        Map<String, Collection<DataNode>> result = new ConcurrentHashMap<>();
+        Map<String, Collection<DataNode>> result = new 
ConcurrentHashMap<>(actualDataNodes.size(), 1F);
         for (Entry<String, Collection<DataNode>> entry : 
actualDataNodes.entrySet()) {
             Collection<DataNode> singleNode = 
loadSpecifiedDataNode(entry.getValue(), featureRequiredSingleTables, 
configuredTableMap);
             if (!singleNode.isEmpty()) {
@@ -158,8 +158,8 @@ public final class SingleTableDataNodeLoader {
         if (configuredTables.isEmpty()) {
             return Collections.emptyMap();
         }
-        Map<String, Map<String, Collection<String>>> result = new 
LinkedHashMap<>();
         Collection<DataNode> dataNodes = 
SingleTableLoadUtils.convertToDataNodes(databaseName, protocolType, 
configuredTables);
+        Map<String, Map<String, Collection<String>>> result = new 
LinkedHashMap<>(dataNodes.size(), 1F);
         for (DataNode each : dataNodes) {
             Map<String, Collection<String>> schemaTables = 
result.getOrDefault(each.getDataSourceName(), new LinkedHashMap<>());
             Collection<String> tables = 
schemaTables.getOrDefault(each.getSchemaName(), new LinkedList<>());
diff --git 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
index 0799b25da5f..89963073110 100644
--- 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
+++ 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
@@ -66,7 +66,7 @@ public final class SingleSQLFederationDecider implements 
SQLFederationDecider<Si
     }
     
     private Collection<DataNode> getTableDataNodes(final SingleRule rule, 
final Collection<QualifiedTable> singleTables) {
-        Collection<DataNode> result = new HashSet<>();
+        Collection<DataNode> result = new HashSet<>(singleTables.size(), 1F);
         for (QualifiedTable each : singleTables) {
             
rule.getAttributes().getAttribute(MutableDataNodeRuleAttribute.class).findTableDataNode(each.getSchemaName(),
 each.getTableName()).ifPresent(result::add);
         }
diff --git 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
index 34182d0f86b..e9b2c81838f 100644
--- 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
+++ 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
@@ -82,8 +82,8 @@ public final class SingleRuleConfigurationDecorator 
implements RuleConfiguration
     
     private Collection<String> loadSpecifiedTables(final boolean 
isSchemaSupportedDatabaseType, final Map<String, Collection<DataNode>> 
actualDataNodes,
                                                    final 
Collection<ShardingSphereRule> builtRules, final Collection<DataNode> 
configuredDataNodes) {
-        Collection<String> expandRequiredDataSources = new LinkedHashSet<>();
-        Map<String, DataNode> expectedDataNodes = new LinkedHashMap<>();
+        Collection<String> expandRequiredDataSources = new 
LinkedHashSet<>(configuredDataNodes.size(), 1F);
+        Map<String, DataNode> expectedDataNodes = new 
LinkedHashMap<>(configuredDataNodes.size(), 1F);
         for (DataNode each : configuredDataNodes) {
             if (SingleTableConstants.ASTERISK.equals(each.getTableName())) {
                 expandRequiredDataSources.add(each.getDataSourceName());
@@ -101,7 +101,7 @@ public final class SingleRuleConfigurationDecorator 
implements RuleConfiguration
     private Collection<String> loadSpecifiedTablesWithExpand(final boolean 
isSchemaSupportedDatabaseType, final Map<String, Collection<DataNode>> 
actualDataNodes,
                                                              final 
Collection<String> featureRequiredSingleTables, final Collection<String> 
expandRequiredDataSources,
                                                              final Map<String, 
DataNode> expectedDataNodes) {
-        Collection<String> result = new LinkedHashSet<>();
+        Collection<String> result = new 
LinkedHashSet<>(actualDataNodes.size(), 1F);
         for (Entry<String, Collection<DataNode>> entry : 
actualDataNodes.entrySet()) {
             if (featureRequiredSingleTables.contains(entry.getKey())) {
                 continue;
@@ -125,7 +125,7 @@ public final class SingleRuleConfigurationDecorator 
implements RuleConfiguration
     
     private Collection<String> loadSpecifiedTablesWithoutExpand(final boolean 
isSchemaSupportedDatabaseType, final Map<String, Collection<DataNode>> 
actualDataNodes,
                                                                 final 
Collection<DataNode> configuredDataNodes) {
-        Collection<String> result = new LinkedHashSet<>();
+        Collection<String> result = new 
LinkedHashSet<>(configuredDataNodes.size(), 1F);
         for (DataNode each : configuredDataNodes) {
             ShardingSpherePreconditions.checkContainsKey(actualDataNodes, 
each.getTableName(), () -> new 
SingleTableNotFoundException(getTableNodeString(isSchemaSupportedDatabaseType, 
each)));
             DataNode actualDataNode = 
actualDataNodes.get(each.getTableName()).iterator().next();
diff --git 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
index dc0309e4570..3a14dd9b0ed 100644
--- 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
+++ 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
@@ -143,7 +143,7 @@ public final class SingleTableLoadUtils {
      * @return data nodes
      */
     public static Collection<DataNode> convertToDataNodes(final String 
databaseName, final DatabaseType databaseType, final Collection<String> tables) 
{
-        Collection<DataNode> result = new LinkedHashSet<>();
+        Collection<DataNode> result = new LinkedHashSet<>(tables.size(), 1F);
         for (String each : tables) {
             result.add(new DataNode(databaseName, databaseType, each));
         }
diff --git 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
index 3ce2a9ab09a..96e765ea2a6 100644
--- 
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
+++ 
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
@@ -30,7 +30,7 @@ import java.util.LinkedList;
 /**
  * Single rule configuration for YAML.
  */
-@RepositoryTupleEntity(value = "single")
+@RepositoryTupleEntity("single")
 @Getter
 @Setter
 public final class YamlSingleRuleConfiguration implements 
YamlRuleConfiguration {
diff --git 
a/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
 
b/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
index 5e9263ba320..216839ac96d 100644
--- 
a/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
+++ 
b/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
@@ -131,7 +131,7 @@ public final class LoadSingleTableExecutor implements 
DatabaseRuleCreateExecutor
     }
     
     private Map<String, Map<String, Collection<String>>> 
getActualTableNodes(final Collection<String> requiredDataSources, final 
Map<String, DataSource> aggregateDataSourceMap) {
-        Map<String, Map<String, Collection<String>>> result = new 
LinkedHashMap<>();
+        Map<String, Map<String, Collection<String>>> result = new 
LinkedHashMap<>(requiredDataSources.size(), 1F);
         for (String each : requiredDataSources) {
             DataSource dataSource = aggregateDataSourceMap.get(each);
             Map<String, Collection<String>> schemaTableNames = 
SingleTableDataNodeLoader.loadSchemaTableNames(database.getName(), 
DatabaseTypeEngine.getStorageType(dataSource), dataSource, each);
diff --git 
a/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
 
b/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
index d6a847fc057..0330f32d4dd 100644
--- 
a/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
+++ 
b/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
@@ -54,7 +54,7 @@ public final class SQLFederationResultSetMetaData extends 
WrapperAdapter impleme
     public SQLFederationResultSetMetaData(final Schema sqlFederationSchema, 
final SelectStatementContext selectStatementContext, final RelDataType 
resultColumnType,
                                           final Map<Integer, String> 
indexAndColumnLabels, final SQLFederationColumnTypeConverter 
columnTypeConverter) {
         this.sqlFederationSchema = sqlFederationSchema;
-        this.relDataTypeFactory = new JavaTypeFactoryImpl();
+        relDataTypeFactory = new JavaTypeFactoryImpl();
         this.selectStatementContext = selectStatementContext;
         this.resultColumnType = resultColumnType;
         this.indexAndColumnLabels = indexAndColumnLabels;
diff --git 
a/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
 
b/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
index 5909767f1cf..633ed53c809 100644
--- 
a/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
+++ 
b/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
@@ -42,7 +42,7 @@ public final class SQLFederationDistSQLStatementVisitor 
extends SQLFederationDis
     
     @Override
     public ASTNode visitAlterSQLFederationRule(final 
AlterSQLFederationRuleContext ctx) {
-        return super.visit(ctx.sqlFederationRuleDefinition());
+        return visit(ctx.sqlFederationRuleDefinition());
     }
     
     @Override
diff --git 
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
 
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
index cce7fc4e64d..f0921d47722 100644
--- 
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
+++ 
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
@@ -217,7 +217,7 @@ public final class EnumerableScanExecutor implements 
ScanExecutor {
         if (null == paramIndexes) {
             return Collections.emptyList();
         }
-        List<Object> result = new ArrayList<>();
+        List<Object> result = new ArrayList<>(paramIndexes.length);
         for (int each : paramIndexes) {
             
result.add(federationContext.getQueryContext().getParameters().get(each));
         }
diff --git 
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
 
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
index 033d1979648..bb26814c93b 100644
--- 
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
+++ 
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
@@ -69,7 +69,7 @@ public final class JDBCRowEnumerator implements 
Enumerator<Object> {
         for (int i = 0; i < metaData.getColumnCount(); i++) {
             rowValues[i] = getValue(i);
         }
-        this.currentRow = 1 == metaData.getColumnCount() ? rowValues[0] : 
rowValues;
+        currentRow = 1 == metaData.getColumnCount() ? rowValues[0] : rowValues;
     }
     
     private Object getValue(final int index) throws SQLException {
diff --git 
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
 
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
index b2859fb3201..2134bce1fc0 100644
--- 
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
+++ 
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
@@ -43,7 +43,7 @@ public final class RowExpressionConverter {
      * @return sql node
      */
     public static Optional<SqlNode> convert(final RowExpression segment) {
-        List<SqlNode> sqlNodes = new ArrayList<>();
+        List<SqlNode> sqlNodes = new ArrayList<>(segment.getItems().size());
         for (ExpressionSegment each : segment.getItems()) {
             ExpressionConverter.convert(each).ifPresent(sqlNodes::add);
         }
diff --git 
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
 
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
index 05b24844403..eaf4c188dc6 100644
--- 
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
+++ 
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
@@ -48,7 +48,7 @@ public final class LogicalScan extends TableScan {
     }
     
     private void resetRowType(final RelNode relNode) {
-        this.rowType = relNode.getRowType();
+        rowType = relNode.getRowType();
     }
     
     /**
diff --git 
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
 
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
index 6a7596ab65e..8ffc3f3bab3 100644
--- 
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
+++ 
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
@@ -259,6 +259,6 @@ public final class SQLFederationPlannerUtils {
      * @return rel opt cluster
      */
     public static RelOptCluster createRelOptCluster(final RelDataTypeFactory 
relDataTypeFactory) {
-        return 
RelOptCluster.create(SQLFederationPlannerUtils.createVolcanoPlanner(), new 
RexBuilder(relDataTypeFactory));
+        return RelOptCluster.create(createVolcanoPlanner(), new 
RexBuilder(relDataTypeFactory));
     }
 }
diff --git 
a/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
 
b/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
index f7bcbd4aaa8..04b3ba3978e 100644
--- 
a/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
+++ 
b/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
@@ -68,7 +68,7 @@ class SQLStatementCompilerIT {
     
     @BeforeEach
     void init() {
-        Map<String, ShardingSphereTable> tables = new HashMap<>();
+        Map<String, ShardingSphereTable> tables = new HashMap<>(12, 1F);
         tables.put("t_order_federate", createOrderFederationTableMetaData());
         tables.put("t_user_info", createUserInfoTableMetaData());
         tables.put("t_order", createTOrderTableMetaData());
diff --git 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
index 55c1b1fc99f..d8897224dbc 100644
--- 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
+++ 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
@@ -34,7 +34,7 @@ class SQLParserRuleTest {
     
     @BeforeEach
     void setup() {
-        sqlParserRule = new SQLParserRule(new SQLParserRuleConfiguration(new 
CacheOption(2, 4), new CacheOption(3, 7)));
+        sqlParserRule = new SQLParserRule(new SQLParserRuleConfiguration(new 
CacheOption(2, 4L), new CacheOption(3, 7L)));
     }
     
     @Test
diff --git 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
index 29746123292..3477814873f 100644
--- 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
+++ 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
@@ -24,7 +24,7 @@ import org.apache.shardingsphere.parser.rule.SQLParserRule;
 import org.apache.shardingsphere.sql.parser.api.CacheOption;
 import org.junit.jupiter.api.Test;
 
-import java.util.HashMap;
+import java.util.Collections;
 import java.util.Properties;
 
 import static org.hamcrest.MatcherAssert.assertThat;
@@ -34,8 +34,8 @@ class SQLParserRuleBuilderTest {
     
     @Test
     void assertBuild() {
-        SQLParserRuleConfiguration ruleConfig = new 
SQLParserRuleConfiguration(new CacheOption(2, 5), new CacheOption(4, 7));
-        SQLParserRule actualResult = new 
SQLParserRuleBuilder().build(ruleConfig, new HashMap<>(), new 
ConfigurationProperties(new Properties()));
+        SQLParserRuleConfiguration ruleConfig = new 
SQLParserRuleConfiguration(new CacheOption(2, 5L), new CacheOption(4, 7L));
+        SQLParserRule actualResult = new 
SQLParserRuleBuilder().build(ruleConfig, Collections.emptyMap(), new 
ConfigurationProperties(new Properties()));
         assertThat(actualResult.getConfiguration(), is(ruleConfig));
         assertThat(actualResult.getSqlStatementCache().getInitialCapacity(), 
is(4));
         assertThat(actualResult.getSqlStatementCache().getMaximumSize(), 
is(7L));
diff --git 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
index ec18c970232..2199069f743 100644
--- 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
+++ 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
@@ -28,7 +28,7 @@ class YamlSQLParserCacheOptionConfigurationSwapperTest {
     
     @Test
     void assertSwapToYamlConfiguration() {
-        YamlSQLParserCacheOptionRuleConfiguration actual = new 
YamlSQLParserCacheOptionConfigurationSwapper().swapToYamlConfiguration(new 
CacheOption(2, 5));
+        YamlSQLParserCacheOptionRuleConfiguration actual = new 
YamlSQLParserCacheOptionConfigurationSwapper().swapToYamlConfiguration(new 
CacheOption(2, 5L));
         assertThat(actual.getInitialCapacity(), is(2));
         assertThat(actual.getMaximumSize(), is(5L));
     }
diff --git 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
index 7f395ea8f27..0d4d6078f6b 100644
--- 
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
+++ 
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
@@ -31,7 +31,7 @@ class YamlSQLParserRuleConfigurationSwapperTest {
     @Test
     void assertSwapToYamlConfiguration() {
         YamlSQLParserRuleConfiguration actual =
-                new 
YamlSQLParserRuleConfigurationSwapper().swapToYamlConfiguration(new 
SQLParserRuleConfiguration(new CacheOption(2, 5), new CacheOption(4, 7)));
+                new 
YamlSQLParserRuleConfigurationSwapper().swapToYamlConfiguration(new 
SQLParserRuleConfiguration(new CacheOption(2, 5L), new CacheOption(4, 7L)));
         assertThat(actual.getParseTreeCache().getInitialCapacity(), is(2));
         assertThat(actual.getParseTreeCache().getMaximumSize(), is(5L));
         assertThat(actual.getSqlStatementCache().getInitialCapacity(), is(4));
diff --git 
a/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
 
b/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
index 2764e4158ec..7d737c4e046 100644
--- 
a/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
+++ 
b/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
@@ -51,7 +51,7 @@ class ShowSQLParserRuleExecutorTest {
     private ContextManager mockContextManager() {
         ContextManager result = mock(ContextManager.class, RETURNS_DEEP_STUBS);
         SQLParserRule rule = mock(SQLParserRule.class);
-        when(rule.getConfiguration()).thenReturn(new 
SQLParserRuleConfiguration(new CacheOption(128, 1024), new CacheOption(2000, 
65535)));
+        when(rule.getConfiguration()).thenReturn(new 
SQLParserRuleConfiguration(new CacheOption(128, 1024L), new CacheOption(2000, 
65535L)));
         
when(result.getMetaDataContexts().getMetaData().getGlobalRuleMetaData().findSingleRule(SQLParserRule.class)).thenReturn(Optional.of(rule));
         return result;
     }
diff --git 
a/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
 
b/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
index 348a9592c5f..3d4cb0e03ca 100644
--- 
a/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
+++ 
b/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
@@ -42,7 +42,7 @@ public final class SQLParserDistSQLStatementVisitor extends 
SQLParserDistSQLStat
     
     @Override
     public ASTNode visitAlterSQLParserRule(final AlterSQLParserRuleContext 
ctx) {
-        return super.visit(ctx.sqlParserRuleDefinition());
+        return visit(ctx.sqlParserRuleDefinition());
     }
     
     @Override
diff --git 
a/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
 
b/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
index 5adf0cc7b07..55425f3153d 100644
--- 
a/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
+++ 
b/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
@@ -40,9 +40,9 @@ public final class ConnectionTransaction {
     private final TransactionConnectionContext transactionContext;
     
     public ConnectionTransaction(final TransactionRule rule, final 
TransactionConnectionContext transactionContext) {
-        this.transactionType = 
transactionContext.getTransactionType().isPresent() ? 
TransactionType.valueOf(transactionContext.getTransactionType().get()) : 
rule.getDefaultType();
+        transactionType = transactionContext.getTransactionType().isPresent() 
? TransactionType.valueOf(transactionContext.getTransactionType().get()) : 
rule.getDefaultType();
         this.transactionContext = transactionContext;
-        transactionManager = TransactionType.LOCAL == this.transactionType ? 
null : rule.getResource().getTransactionManager(rule.getDefaultType());
+        transactionManager = TransactionType.LOCAL == transactionType ? null : 
rule.getResource().getTransactionManager(rule.getDefaultType());
     }
     
     /**
diff --git 
a/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
 
b/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
index 4986f7dd3db..ae60d42255c 100644
--- 
a/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
+++ 
b/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
@@ -61,8 +61,8 @@ public final class DataSourceUtils {
         result.setPassword("root");
         result.setMaximumPoolSize(10);
         result.setMinimumIdle(2);
-        result.setConnectionTimeout(15 * 1000L);
-        result.setIdleTimeout(40 * 1000L);
+        result.setConnectionTimeout(15L * 1000L);
+        result.setIdleTimeout(40L * 1000L);
         return result;
     }
     
diff --git 
a/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
 
b/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
index ae29706d9a1..2431cd90d56 100644
--- 
a/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
+++ 
b/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
@@ -117,7 +117,7 @@ public final class ResourceSwitchManager {
     }
     
     private Map<StorageNode, DataSource> getToBeRemovedStaleDataSource(final 
ResourceMetaData resourceMetaData, final Collection<String> storageUnitNames) {
-        Map<StorageNode, DataSource> result = new LinkedHashMap<>();
+        Map<StorageNode, DataSource> result = new 
LinkedHashMap<>(storageUnitNames.size(), 1F);
         for (String each : storageUnitNames) {
             if (!resourceMetaData.getStorageUnits().containsKey(each)) {
                 return Collections.emptyMap();
diff --git 
a/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
 
b/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
index dc0e794aa50..c6cdd7a8c59 100644
--- 
a/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
+++ 
b/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
@@ -53,7 +53,7 @@ public final class MetaDataContexts implements AutoCloseable {
     public MetaDataContexts(final MetaDataBasedPersistService persistService, 
final ShardingSphereMetaData metaData) {
         this.persistService = persistService;
         this.metaData = metaData;
-        this.statistics = initStatistics(metaData);
+        statistics = initStatistics(metaData);
     }
     
     private ShardingSphereStatistics initStatistics(final 
ShardingSphereMetaData metaData) {
diff --git 
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
 
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
index 80cfe7ebf0e..2efacde544d 100644
--- 
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
+++ 
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
@@ -109,7 +109,7 @@ public final class ClusterModeContextManager implements 
ModeContextManager, Cont
     }
     
     private Map<String, DataSourcePoolProperties> 
getToBeDroppedDataSourcePoolProperties(final Map<String, 
DataSourcePoolProperties> propsMap, final Collection<String> 
toBeDroppedResourceNames) {
-        Map<String, DataSourcePoolProperties> result = new LinkedHashMap<>();
+        Map<String, DataSourcePoolProperties> result = new 
LinkedHashMap<>(toBeDroppedResourceNames.size(), 1F);
         for (String each : toBeDroppedResourceNames) {
             if (propsMap.containsKey(each)) {
                 result.put(each, propsMap.get(each));
diff --git 
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
 
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
index 199ad54bb7c..96488f881f1 100644
--- 
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
+++ 
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
@@ -172,8 +172,8 @@ public final class ComputeNodeStatusService {
      * @return assigned worker ids
      */
     public Collection<Integer> getAssignedWorkerIds() {
-        Collection<Integer> result = new LinkedHashSet<>();
         Collection<String> childrenKeys = 
repository.getChildrenKeys(ComputeNode.getInstanceWorkerIdRootNodePath());
+        Collection<Integer> result = new LinkedHashSet<>(childrenKeys.size(), 
1F);
         for (String each : childrenKeys) {
             String workerId = 
repository.getDirectly(ComputeNode.getInstanceWorkerIdNodePath(each));
             if (null != workerId) {
diff --git 
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
 
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
index 6ca87662d91..4ec220a0359 100644
--- 
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
+++ 
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
@@ -127,7 +127,7 @@ class ResourceMetaDataChangedSubscriberTest {
     @Test
     void assertRenewForSchemaAdded() {
         subscriber.renew(new SchemaAddedEvent("db", "foo_schema"));
-        
verify(contextManager.getMetaDataContexts().getMetaData().getDatabase("db")).addSchema(argThat(argument
 -> argument.equals("foo_schema")), any(ShardingSphereSchema.class));
+        
verify(contextManager.getMetaDataContexts().getMetaData().getDatabase("db")).addSchema(argThat("foo_schema"::equals),
 any(ShardingSphereSchema.class));
     }
     
     @Test
diff --git 
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
 
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
index 6526dd265ef..71ee8ac0641 100644
--- 
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
+++ 
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
@@ -117,7 +117,7 @@ class StateChangedSubscriberTest {
         subscriber.renew(event);
         verify(ruleAttribute).updateStatus(
                 argThat(qualifiedDataSource -> 
Objects.equals(event.getQualifiedDataSource(), qualifiedDataSource)),
-                argThat(dataSourceState -> 
Objects.equals(event.getStatus().getStatus(), dataSourceState)));
+                argThat(dataSourceState -> event.getStatus().getStatus() == 
dataSourceState));
     }
     
     @Test
diff --git 
a/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
 
b/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
index 00c169180f6..4c1f99faada 100644
--- 
a/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
+++ 
b/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
@@ -36,14 +36,13 @@ public final class DistributedLockHolder {
     
     private final TypedProperties<?> props;
     
-    private final Map<String, DistributedLock> locks;
+    private final Map<String, DistributedLock> locks = new 
ConcurrentHashMap<>();
     
     @SuppressWarnings("unchecked")
     public DistributedLockHolder(final String type, final Object client, final 
TypedProperties<?> props) {
         creator = TypedSPILoader.getService(DistributedLockCreator.class, 
type);
         this.client = client;
         this.props = props;
-        locks = new ConcurrentHashMap<>();
     }
     
     /**
diff --git 
a/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
 
b/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
index a4288ae55cf..a03783f2be8 100644
--- 
a/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
+++ 
b/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
@@ -39,7 +39,7 @@ import java.util.Properties;
 @Slf4j
 public final class SessionConnectionListener implements 
ConnectionStateListener {
     
-    private static final int RECONNECT_INTERVAL_SECONDS = 5;
+    private static final long RECONNECT_INTERVAL_SECONDS = 5L;
     
     private final InstanceContext instanceContext;
     
diff --git 
a/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
 
b/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
index 41042cf9ee3..4e9d42fc673 100644
--- 
a/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
+++ 
b/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
@@ -177,7 +177,7 @@ public final class StandaloneModeContextManager implements 
ModeContextManager, C
     @Override
     public void dropSchema(final String databaseName, final Collection<String> 
schemaNames) {
         Collection<String> tobeRemovedTables = new LinkedHashSet<>();
-        Collection<String> tobeRemovedSchemas = new LinkedHashSet<>();
+        Collection<String> tobeRemovedSchemas = new 
LinkedHashSet<>(schemaNames.size(), 1F);
         ShardingSphereMetaData metaData = 
contextManager.getMetaDataContexts().getMetaData();
         ShardingSphereDatabase database = metaData.getDatabase(databaseName);
         for (String each : schemaNames) {


Reply via email to