On 10/9/24 16:05, Pierrick Bouvier wrote:
@@ -720,13 +728,10 @@ static void tlb_flush_range_locked(CPUState *cpu, int 
midx,
          return;
      }
+    tlbfast_flush_range_locked(d, f, addr, len, mask);
+
      for (vaddr i = 0; i < len; i += TARGET_PAGE_SIZE) {
          vaddr page = addr + i;
-        CPUTLBEntry *entry = tlb_entry(cpu, midx, page);
-
-        if (tlb_flush_entry_mask_locked(entry, page, mask)) {
-            tlb_n_used_entries_dec(cpu, midx);
-        }
          tlb_flush_vtlb_page_mask_locked(cpu, midx, page, mask);
      }
  }

Why don't we have the same kind of change for tlb_flush_vtlb_page_mask_locked?

We know have two loops (for entry mask, and for page mask).

It goes away in patch 15.

r~

Reply via email to