This is an automated email from the ASF dual-hosted git repository.

yiguolei pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris.git


The following commit(s) were added to refs/heads/master by this push:
     new 5eb5444476 [fix](memtracker) Remove useless memory exceed check #11939
5eb5444476 is described below

commit 5eb5444476a32cf172aa462e4ab7755510868656
Author: Xinyi Zou <zouxiny...@gmail.com>
AuthorDate: Mon Aug 22 08:40:19 2022 +0800

    [fix](memtracker) Remove useless memory exceed check #11939
---
 be/src/exec/broker_scan_node.cpp      | 6 +-----
 be/src/exec/tablet_sink.cpp           | 3 +--
 be/src/vec/exec/file_scan_node.cpp    | 6 +-----
 be/src/vec/exec/vbroker_scan_node.cpp | 6 +-----
 be/src/vec/sink/vtablet_sink.cpp      | 8 ++------
 5 files changed, 6 insertions(+), 23 deletions(-)

diff --git a/be/src/exec/broker_scan_node.cpp b/be/src/exec/broker_scan_node.cpp
index 32baad6fcd..9cff2da4ff 100644
--- a/be/src/exec/broker_scan_node.cpp
+++ b/be/src/exec/broker_scan_node.cpp
@@ -348,11 +348,7 @@ Status BrokerScanNode::scanner_scan(const 
TBrokerScanRange& scan_range,
                    // stop pushing more batch if
                    // 1. too many batches in queue, or
                    // 2. at least one batch in queue and memory exceed limit.
-                   (_batch_queue.size() >= _max_buffered_batches ||
-                    (thread_context()
-                             ->_thread_mem_tracker_mgr->limiter_mem_tracker()
-                             ->any_limit_exceeded() &&
-                     !_batch_queue.empty()))) {
+                   (_batch_queue.size() >= _max_buffered_batches || 
!_batch_queue.empty())) {
                 _queue_writer_cond.wait_for(l, std::chrono::seconds(1));
             }
             // Process already set failed, so we just return OK
diff --git a/be/src/exec/tablet_sink.cpp b/be/src/exec/tablet_sink.cpp
index 281f94aee2..cdb503551b 100644
--- a/be/src/exec/tablet_sink.cpp
+++ b/be/src/exec/tablet_sink.cpp
@@ -303,8 +303,7 @@ Status NodeChannel::add_row(Tuple* input_tuple, int64_t 
tablet_id) {
     // _cancelled may be set by rpc callback, and it's possible that 
_cancelled might be set in any of the steps below.
     // It's fine to do a fake add_row() and return OK, because we will check 
_cancelled in next add_row() or mark_close().
     while (!_cancelled && _pending_batches_num > 0 &&
-           (_pending_batches_bytes > _max_pending_batches_bytes ||
-            
_parent->_mem_tracker->limit_exceeded(_max_pending_batches_bytes))) {
+           _pending_batches_bytes > _max_pending_batches_bytes) {
         SCOPED_ATOMIC_TIMER(&_mem_exceeded_block_ns);
         std::this_thread::sleep_for(std::chrono::milliseconds(10));
     }
diff --git a/be/src/vec/exec/file_scan_node.cpp 
b/be/src/vec/exec/file_scan_node.cpp
index ec8a5165ce..e84c0e9371 100644
--- a/be/src/vec/exec/file_scan_node.cpp
+++ b/be/src/vec/exec/file_scan_node.cpp
@@ -403,11 +403,7 @@ Status FileScanNode::scanner_scan(const TFileScanRange& 
scan_range, ScannerCount
                // stop pushing more batch if
                // 1. too many batches in queue, or
                // 2. at least one batch in queue and memory exceed limit.
-               (_block_queue.size() >= _max_buffered_batches ||
-                (thread_context()
-                         ->_thread_mem_tracker_mgr->limiter_mem_tracker()
-                         ->any_limit_exceeded() &&
-                 !_block_queue.empty()))) {
+               (_block_queue.size() >= _max_buffered_batches || 
!_block_queue.empty())) {
             _queue_writer_cond.wait_for(l, std::chrono::seconds(1));
         }
         // Process already set failed, so we just return OK
diff --git a/be/src/vec/exec/vbroker_scan_node.cpp 
b/be/src/vec/exec/vbroker_scan_node.cpp
index 98c16112c1..2f0593b3f6 100644
--- a/be/src/vec/exec/vbroker_scan_node.cpp
+++ b/be/src/vec/exec/vbroker_scan_node.cpp
@@ -246,11 +246,7 @@ Status VBrokerScanNode::scanner_scan(const 
TBrokerScanRange& scan_range, Scanner
                // stop pushing more batch if
                // 1. too many batches in queue, or
                // 2. at least one batch in queue and memory exceed limit.
-               (_block_queue.size() >= _max_buffered_batches ||
-                (thread_context()
-                         ->_thread_mem_tracker_mgr->limiter_mem_tracker()
-                         ->any_limit_exceeded() &&
-                 !_block_queue.empty()))) {
+               (_block_queue.size() >= _max_buffered_batches || 
!_block_queue.empty())) {
             _queue_writer_cond.wait_for(l, std::chrono::seconds(1));
         }
         // Process already set failed, so we just return OK
diff --git a/be/src/vec/sink/vtablet_sink.cpp b/be/src/vec/sink/vtablet_sink.cpp
index 4965562c3c..7a0faee0d6 100644
--- a/be/src/vec/sink/vtablet_sink.cpp
+++ b/be/src/vec/sink/vtablet_sink.cpp
@@ -178,12 +178,8 @@ Status VNodeChannel::add_row(const BlockRow& block_row, 
int64_t tablet_id) {
     // But there is still some unfinished things, we do mem limit here 
temporarily.
     // _cancelled may be set by rpc callback, and it's possible that 
_cancelled might be set in any of the steps below.
     // It's fine to do a fake add_row() and return OK, because we will check 
_cancelled in next add_row() or mark_close().
-    while (!_cancelled &&
-           (_pending_batches_bytes > _max_pending_batches_bytes ||
-            thread_context()
-                    ->_thread_mem_tracker_mgr->limiter_mem_tracker()
-                    ->any_limit_exceeded()) &&
-           _pending_batches_num > 0) {
+    while (!_cancelled && _pending_batches_num > 0 &&
+           _pending_batches_bytes > _max_pending_batches_bytes) {
         SCOPED_ATOMIC_TIMER(&_mem_exceeded_block_ns);
         std::this_thread::sleep_for(std::chrono::milliseconds(10));
     }


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org
For additional commands, e-mail: commits-h...@doris.apache.org

Reply via email to