lvyanquan commented on code in PR #3656:
URL: https://github.com/apache/flink-cdc/pull/3656#discussion_r1808363691


##########
flink-cdc-connect/flink-cdc-source-connectors/flink-connector-mysql-cdc/src/main/java/org/apache/flink/cdc/connectors/mysql/source/split/MySqlBinlogSplit.java:
##########
@@ -108,6 +108,16 @@ public boolean isCompletedSplit() {
         return totalFinishedSplitSize == finishedSnapshotSplitInfos.size();
     }
 
+    public String getTables() {
+        String tables;
+        if (tableSchemas != null) {
+            tables = tableSchemas.keySet().toString();

Review Comment:
   If we iterate every time, it will have an impact on performance. Can we 
assume that tableschemas will not change and we only need to build the tables 
variable once.



##########
flink-cdc-connect/flink-cdc-source-connectors/flink-connector-mysql-cdc/src/main/java/org/apache/flink/cdc/connectors/mysql/source/reader/MySqlSourceReader.java:
##########
@@ -530,7 +530,11 @@ private void logCurrentBinlogOffsets(List<MySqlSplit> 
splits, long checkpointId)
                 return;
             }
             BinlogOffset offset = split.asBinlogSplit().getStartingOffset();
-            LOG.info("Binlog offset on checkpoint {}: {}", checkpointId, 
offset);
+            LOG.info(
+                    "Binlog offset for tables {} on checkpoint {}: {}",
+                    split.asBinlogSplit().getTables(),

Review Comment:
   If there are a large number of tables, it will result in logs being quite 
long. Would you consider truncating them?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to