[ 
https://issues.apache.org/jira/browse/HIVE-26265?focusedWorklogId=785196&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-785196
 ]

ASF GitHub Bot logged work on HIVE-26265:
-----------------------------------------

                Author: ASF GitHub Bot
            Created on: 27/Jun/22 17:51
            Start Date: 27/Jun/22 17:51
    Worklog Time Spent: 10m 
      Work Description: cmunkey commented on code in PR #3365:
URL: https://github.com/apache/hive/pull/3365#discussion_r907648669


##########
itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/ReplicationTestUtils.java:
##########
@@ -17,28 +17,34 @@
  */
 package org.apache.hadoop.hive.ql.parse;
 
+import org.apache.hadoop.fs.FSDataInputStream;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.LocatedFileStatus;
 import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.fs.RemoteIterator;
 import org.apache.hadoop.hdfs.DistributedFileSystem;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.ql.exec.repl.DirCopyWork;
-import org.apache.hadoop.hive.ql.parse.repl.PathBuilder;
 import org.apache.hadoop.hive.ql.exec.repl.util.ReplUtils;
 import org.junit.Assert;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 
 import java.io.BufferedReader;
+import java.io.File;
 import java.io.IOException;
 import java.io.InputStream;
 import java.io.InputStreamReader;
 import java.util.ArrayList;
 import java.util.List;
-import java.util.Arrays;
 import java.util.Set;
 import java.util.HashSet;
 
 /**
  * ReplicationTestUtils - static helper functions for replication test
  */
 public class ReplicationTestUtils {
+  static final private Logger LOG = 
LoggerFactory.getLogger(ReplicationTestUtils.class);

Review Comment:
   I will fixed.



##########
common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:
##########
@@ -5615,7 +5615,12 @@ public static enum ConfVars {
 
     HIVE_ADDITIONAL_CONFIG_FILES("hive.additional.config.files", "",
             "The names of additional config files, such as ldap-site.xml," +
-                    "tez-site.xml, etc in comma separated list.");
+                    "tez-site.xml, etc in comma separated list."),
+
+    REPL_FILTER_TRANSACTIONS("hive.repl.filter.transactions", false,
+            "Enable transaction event filtering to save dump space.\n" +
+                    "When true, transactions are implicitly opened during REPL 
DUMP.\n" +

Review Comment:
   I will move.





Issue Time Tracking
-------------------

    Worklog Id:     (was: 785196)
    Time Spent: 3h 50m  (was: 3h 40m)

> REPL DUMP should filter out OpenXacts and unneeded CommitXact/Abort.
> --------------------------------------------------------------------
>
>                 Key: HIVE-26265
>                 URL: https://issues.apache.org/jira/browse/HIVE-26265
>             Project: Hive
>          Issue Type: Improvement
>          Components: HiveServer2
>            Reporter: francis pang
>            Assignee: francis pang
>            Priority: Major
>              Labels: pull-request-available
>          Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> REPL DUMP is replication all OpenXacts, even when they are from other non 
> replicated databases. This wastes space in the dump, and ends up opening 
> unneeded transactions during REPL LOAD.
>  
> Add a config property for replication that filters out OpenXact events during 
> REPL DUMP. During REPL LOAD, the txns can be implicitly opened when the 
> ALLOC_WRITE_ID is processed. For CommitTxn and AbortTxn, dump only if WRITE 
> ID was allocated.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

Reply via email to