Using post-index st2g is a faster way of memory tagging/untagging.
Because a post-index 'st2g tag, [addr], #32' is equivalent to:
   stg tag, addr, #0
   stg tag, addr, #16
   add addr, addr, #32

TBD:
  - Currently generated by in the aarch64 backend.  Not sure if this is
    the right way to do it.
  - Also not clear how to weave in the generation of stz2g.

ChangeLog:
        * gcc/config/aarch64/aarch64.md

---
[New in RFC V2]
---
 gcc/config/aarch64/aarch64.md | 20 ++++++++++++++++++++
 1 file changed, 20 insertions(+)

diff --git a/gcc/config/aarch64/aarch64.md b/gcc/config/aarch64/aarch64.md
index d3223e275c51..175aed3146ac 100644
--- a/gcc/config/aarch64/aarch64.md
+++ b/gcc/config/aarch64/aarch64.md
@@ -8495,6 +8495,26 @@
   [(set_attr "type" "memtag")]
 )
 
+;; ST2G with post-index writeback.
+(define_insn "*st2g_post"
+  [(set (mem:QI (unspec:DI
+        [(plus:DI (match_operand:DI 1 "register_operand" "=&rk")
+                  (const_int 0))]
+        UNSPEC_TAG_SPACE))
+       (and:QI (lshiftrt:DI (match_operand:DI 0 "register_operand" "rk")
+                            (const_int 56)) (const_int 15)))
+   (set (mem:QI (unspec:DI
+        [(plus:DI (match_dup 1) (const_int -16))]
+        UNSPEC_TAG_SPACE))
+       (and:QI (lshiftrt:DI (match_dup 0)
+                            (const_int 56)) (const_int 15)))
+    (set (match_dup 1)
+         (plus:DI (match_dup 1) (match_operand:DI 2 "aarch64_granule16_simm9" 
"i")))]
+  "TARGET_MEMTAG"
+  "st2g\\t%0, [%1], #%2"
+  [(set_attr "type" "memtag")]
+)
+
 ;; Load/Store 64-bit (LS64) instructions.
 (define_insn "ld64b"
   [(set (match_operand:V8DI 0 "register_operand" "=r")
-- 
2.43.0

Reply via email to