Ping
> -----Original Message----- > From: Gcc-patches <gcc-patches- > bounces+tamar.christina=arm....@gcc.gnu.org> On Behalf Of Tamar > Christina via Gcc-patches > Sent: Friday, November 11, 2022 2:40 PM > To: Richard Sandiford <richard.sandif...@arm.com> > Cc: gcc-patches@gcc.gnu.org; nd <n...@arm.com>; Richard Earnshaw > <richard.earns...@arm.com>; Marcus Shawcroft > <marcus.shawcr...@arm.com>; Kyrylo Tkachov <kyrylo.tkac...@arm.com> > Subject: RE: [PATCH 5/8]AArch64 aarch64: Make existing V2HF be usable. > > Hi, > > > > This name might cause confusion with the SVE iterators, where FULL > > means "every bit of the register is used". How about something like > > VMOVE instead? > > > > With this change, I guess VALL_F16 represents "The set of all modes > > for which the vld1 intrinsics are provided" and VMOVE or whatever is > > "All Advanced SIMD modes suitable for moving, loading, and storing". > > That is, VMOVE extends VALL_F16 with modes that are not manifested via > > intrinsics. > > > > Done. > > > Where is the 2h used, and is it valid syntax in that context? > > > > Same for later instances of 2h. > > They are, but they weren't meant to be in this patch. They belong in a > separate FP16 series that I won't get to finish for GCC 13 due not being able > to finish writing all the tests. I have moved them to that patch series > though. > > While the addp patch series has been killed, this patch is still good > standalone > and improves codegen as shown in the updated testcase. > > Bootstrapped Regtested on aarch64-none-linux-gnu and no issues. > > Ok for master? > > Thanks, > Tamar > > gcc/ChangeLog: > > * config/aarch64/aarch64-simd.md (*aarch64_simd_movv2hf): New. > (mov<mode>, movmisalign<mode>, aarch64_dup_lane<mode>, > aarch64_store_lane0<mode>, aarch64_simd_vec_set<mode>, > @aarch64_simd_vec_copy_lane<mode>, vec_set<mode>, > reduc_<optab>_scal_<mode>, reduc_<fmaxmin>_scal_<mode>, > aarch64_reduc_<optab>_internal<mode>, > aarch64_get_lane<mode>, > vec_init<mode><Vel>, vec_extract<mode><Vel>): Support V2HF. > (aarch64_simd_dupv2hf): New. > * config/aarch64/aarch64.cc (aarch64_classify_vector_mode): > Add E_V2HFmode. > * config/aarch64/iterators.md (VHSDF_P): New. > (V2F, VMOVE, nunits, Vtype, Vmtype, Vetype, stype, VEL, > Vel, q, vp): Add V2HF. > * config/arm/types.md (neon_fp_reduc_add_h): New. > > gcc/testsuite/ChangeLog: > > * gcc.target/aarch64/sve/slp_1.c: Update testcase. > > --- inline copy of patch --- > > diff --git a/gcc/config/aarch64/aarch64-simd.md > b/gcc/config/aarch64/aarch64-simd.md > index > f4152160084d6b6f34bd69f0ba6386c1ab50f77e..487a31010245accec28e779661 > e6c2d578fca4b7 100644 > --- a/gcc/config/aarch64/aarch64-simd.md > +++ b/gcc/config/aarch64/aarch64-simd.md > @@ -19,10 +19,10 @@ > ;; <http://www.gnu.org/licenses/>. > > (define_expand "mov<mode>" > - [(set (match_operand:VALL_F16 0 "nonimmediate_operand") > - (match_operand:VALL_F16 1 "general_operand"))] > + [(set (match_operand:VMOVE 0 "nonimmediate_operand") > + (match_operand:VMOVE 1 "general_operand"))] > "TARGET_SIMD" > - " > +{ > /* Force the operand into a register if it is not an > immediate whose use can be replaced with xzr. > If the mode is 16 bytes wide, then we will be doing @@ -46,12 +46,11 @@ > (define_expand "mov<mode>" > aarch64_expand_vector_init (operands[0], operands[1]); > DONE; > } > - " > -) > +}) > > (define_expand "movmisalign<mode>" > - [(set (match_operand:VALL_F16 0 "nonimmediate_operand") > - (match_operand:VALL_F16 1 "general_operand"))] > + [(set (match_operand:VMOVE 0 "nonimmediate_operand") > + (match_operand:VMOVE 1 "general_operand"))] > "TARGET_SIMD && !STRICT_ALIGNMENT" > { > /* This pattern is not permitted to fail during expansion: if both > arguments > @@ -73,6 +72,16 @@ (define_insn "aarch64_simd_dup<mode>" > [(set_attr "type" "neon_dup<q>, neon_from_gp<q>")] > ) > > +(define_insn "aarch64_simd_dupv2hf" > + [(set (match_operand:V2HF 0 "register_operand" "=w") > + (vec_duplicate:V2HF > + (match_operand:HF 1 "register_operand" "0")))] > + "TARGET_SIMD" > + "@ > + sli\\t%d0, %d1, 16" > + [(set_attr "type" "neon_shift_imm")] > +) > + > (define_insn "aarch64_simd_dup<mode>" > [(set (match_operand:VDQF_F16 0 "register_operand" "=w,w") > (vec_duplicate:VDQF_F16 > @@ -85,10 +94,10 @@ (define_insn "aarch64_simd_dup<mode>" > ) > > (define_insn "aarch64_dup_lane<mode>" > - [(set (match_operand:VALL_F16 0 "register_operand" "=w") > - (vec_duplicate:VALL_F16 > + [(set (match_operand:VMOVE 0 "register_operand" "=w") > + (vec_duplicate:VMOVE > (vec_select:<VEL> > - (match_operand:VALL_F16 1 "register_operand" "w") > + (match_operand:VMOVE 1 "register_operand" "w") > (parallel [(match_operand:SI 2 "immediate_operand" "i")]) > )))] > "TARGET_SIMD" > @@ -142,6 +151,29 @@ (define_insn > "*aarch64_simd_mov<VDMOV:mode>" > mov_reg, neon_move<q>")] > ) > > +(define_insn "*aarch64_simd_movv2hf" > + [(set (match_operand:V2HF 0 "nonimmediate_operand" > + "=w, m, m, w, ?r, ?w, ?r, w, w") > + (match_operand:V2HF 1 "general_operand" > + "m, Dz, w, w, w, r, r, Dz, Dn"))] > + "TARGET_SIMD_F16INST > + && (register_operand (operands[0], V2HFmode) > + || aarch64_simd_reg_or_zero (operands[1], V2HFmode))" > + "@ > + ldr\\t%s0, %1 > + str\\twzr, %0 > + str\\t%s1, %0 > + mov\\t%0.2s[0], %1.2s[0] > + umov\\t%w0, %1.s[0] > + fmov\\t%s0, %1 > + mov\\t%0, %1 > + movi\\t%d0, 0 > + * return aarch64_output_simd_mov_immediate (operands[1], 32);" > + [(set_attr "type" "neon_load1_1reg, store_8, neon_store1_1reg,\ > + neon_logic, neon_to_gp, f_mcr,\ > + mov_reg, neon_move, neon_move")] > +) > + > (define_insn "*aarch64_simd_mov<VQMOV:mode>" > [(set (match_operand:VQMOV 0 "nonimmediate_operand" > "=w, Umn, m, w, ?r, ?w, ?r, w") > @@ -182,7 +214,7 @@ (define_insn "*aarch64_simd_mov<VQMOV:mode>" > > (define_insn "aarch64_store_lane0<mode>" > [(set (match_operand:<VEL> 0 "memory_operand" "=m") > - (vec_select:<VEL> (match_operand:VALL_F16 1 "register_operand" > "w") > + (vec_select:<VEL> (match_operand:VMOVE 1 "register_operand" > "w") > (parallel [(match_operand 2 "const_int_operand" > "n")])))] > "TARGET_SIMD > && ENDIAN_LANE_N (<nunits>, INTVAL (operands[2])) == 0" > @@ -1035,11 +1067,11 @@ (define_insn "one_cmpl<mode>2" > ) > > (define_insn "aarch64_simd_vec_set<mode>" > - [(set (match_operand:VALL_F16 0 "register_operand" "=w,w,w") > - (vec_merge:VALL_F16 > - (vec_duplicate:VALL_F16 > + [(set (match_operand:VMOVE 0 "register_operand" "=w,w,w") > + (vec_merge:VMOVE > + (vec_duplicate:VMOVE > (match_operand:<VEL> 1 > "aarch64_simd_nonimmediate_operand" "w,?r,Utv")) > - (match_operand:VALL_F16 3 "register_operand" "0,0,0") > + (match_operand:VMOVE 3 "register_operand" "0,0,0") > (match_operand:SI 2 "immediate_operand" "i,i,i")))] > "TARGET_SIMD" > { > @@ -1061,14 +1093,14 @@ (define_insn "aarch64_simd_vec_set<mode>" > ) > > (define_insn "@aarch64_simd_vec_copy_lane<mode>" > - [(set (match_operand:VALL_F16 0 "register_operand" "=w") > - (vec_merge:VALL_F16 > - (vec_duplicate:VALL_F16 > + [(set (match_operand:VMOVE 0 "register_operand" "=w") > + (vec_merge:VMOVE > + (vec_duplicate:VMOVE > (vec_select:<VEL> > - (match_operand:VALL_F16 3 "register_operand" "w") > + (match_operand:VMOVE 3 "register_operand" "w") > (parallel > [(match_operand:SI 4 "immediate_operand" "i")]))) > - (match_operand:VALL_F16 1 "register_operand" "0") > + (match_operand:VMOVE 1 "register_operand" "0") > (match_operand:SI 2 "immediate_operand" "i")))] > "TARGET_SIMD" > { > @@ -1376,7 +1408,7 @@ (define_insn "vec_shr_<mode>" > ) > > (define_expand "vec_set<mode>" > - [(match_operand:VALL_F16 0 "register_operand") > + [(match_operand:VMOVE 0 "register_operand") > (match_operand:<VEL> 1 "aarch64_simd_nonimmediate_operand") > (match_operand:SI 2 "immediate_operand")] > "TARGET_SIMD" > @@ -3495,7 +3527,7 @@ (define_insn "popcount<mode>2" > ;; gimple_fold'd to the IFN_REDUC_(MAX|MIN) function. (This is FP > smax/smin). > (define_expand "reduc_<optab>_scal_<mode>" > [(match_operand:<VEL> 0 "register_operand") > - (unspec:<VEL> [(match_operand:VHSDF 1 "register_operand")] > + (unspec:<VEL> [(match_operand:VHSDF_P 1 "register_operand")] > FMAXMINV)] > "TARGET_SIMD" > { > @@ -3510,7 +3542,7 @@ (define_expand "reduc_<optab>_scal_<mode>" > > (define_expand "reduc_<fmaxmin>_scal_<mode>" > [(match_operand:<VEL> 0 "register_operand") > - (unspec:<VEL> [(match_operand:VHSDF 1 "register_operand")] > + (unspec:<VEL> [(match_operand:VHSDF_P 1 "register_operand")] > FMAXMINNMV)] > "TARGET_SIMD" > { > @@ -3554,8 +3586,8 @@ (define_insn > "aarch64_reduc_<optab>_internalv2si" > ) > > (define_insn "aarch64_reduc_<optab>_internal<mode>" > - [(set (match_operand:VHSDF 0 "register_operand" "=w") > - (unspec:VHSDF [(match_operand:VHSDF 1 "register_operand" "w")] > + [(set (match_operand:VHSDF_P 0 "register_operand" "=w") > + (unspec:VHSDF_P [(match_operand:VHSDF_P 1 "register_operand" > + "w")] > FMAXMINV))] > "TARGET_SIMD" > "<maxmin_uns_op><vp>\\t%<Vetype>0, %1.<Vtype>" > @@ -4200,7 +4232,7 @@ (define_insn > "*aarch64_get_lane_zero_extend<GPI:mode><VDQQH:mode>" > (define_insn_and_split "aarch64_get_lane<mode>" > [(set (match_operand:<VEL> 0 "aarch64_simd_nonimmediate_operand" > "=?r, w, Utv") > (vec_select:<VEL> > - (match_operand:VALL_F16 1 "register_operand" "w, w, w") > + (match_operand:VMOVE 1 "register_operand" "w, w, w") > (parallel [(match_operand:SI 2 "immediate_operand" "i, i, i")])))] > "TARGET_SIMD" > { > @@ -7981,7 +8013,7 @@ (define_expand "aarch64_st1<VALL_F16:mode>" > ;; Standard pattern name vec_init<mode><Vel>. > > (define_expand "vec_init<mode><Vel>" > - [(match_operand:VALL_F16 0 "register_operand") > + [(match_operand:VMOVE 0 "register_operand") > (match_operand 1 "" "")] > "TARGET_SIMD" > { > @@ -8060,7 +8092,7 @@ (define_insn "aarch64_urecpe<mode>" > > (define_expand "vec_extract<mode><Vel>" > [(match_operand:<VEL> 0 "aarch64_simd_nonimmediate_operand") > - (match_operand:VALL_F16 1 "register_operand") > + (match_operand:VMOVE 1 "register_operand") > (match_operand:SI 2 "immediate_operand")] > "TARGET_SIMD" > { > diff --git a/gcc/config/aarch64/aarch64.cc b/gcc/config/aarch64/aarch64.cc > index > 84dbe2f4ea7d03b424602ed98a34e7824217dc91..35671cb86e374f9ded21d0e4 > 944c63bc2cbc0901 100644 > --- a/gcc/config/aarch64/aarch64.cc > +++ b/gcc/config/aarch64/aarch64.cc > @@ -3566,6 +3566,7 @@ aarch64_classify_vector_mode (machine_mode > mode) > case E_V8BFmode: > case E_V4SFmode: > case E_V2DFmode: > + case E_V2HFmode: > return TARGET_SIMD ? VEC_ADVSIMD : 0; > > default: > diff --git a/gcc/config/aarch64/iterators.md > b/gcc/config/aarch64/iterators.md index > 37d8161a33b1c399d80be82afa67613a087389d4..dfcf86a440e316c2abdbcc6463 > 63d39e458d1a91 100644 > --- a/gcc/config/aarch64/iterators.md > +++ b/gcc/config/aarch64/iterators.md > @@ -160,6 +160,10 @@ (define_mode_iterator VDQF [V2SF V4SF V2DF]) > (define_mode_iterator VHSDF [(V4HF "TARGET_SIMD_F16INST") > (V8HF "TARGET_SIMD_F16INST") > V2SF V4SF V2DF]) > +;; Advanced SIMD Float modes suitable for pairwise operations. > +(define_mode_iterator VHSDF_P [(V4HF "TARGET_SIMD_F16INST") > + (V8HF "TARGET_SIMD_F16INST") > + V2SF V4SF V2DF (V2HF > "TARGET_SIMD_F16INST")]) > > ;; Advanced SIMD Float modes, and DF. > (define_mode_iterator VDQF_DF [V2SF V4SF V2DF DF]) @@ -188,15 +192,23 > @@ (define_mode_iterator VDQF_COND [V2SF V2SI V4SF V4SI V2DF V2DI]) > (define_mode_iterator VALLF [V2SF V4SF V2DF SF DF]) > > ;; Advanced SIMD Float modes with 2 elements. > -(define_mode_iterator V2F [V2SF V2DF]) > +(define_mode_iterator V2F [V2SF V2DF V2HF]) > > ;; All Advanced SIMD modes on which we support any arithmetic operations. > (define_mode_iterator VALL [V8QI V16QI V4HI V8HI V2SI V4SI V2DI V2SF > V4SF V2DF]) > > -;; All Advanced SIMD modes suitable for moving, loading, and storing. > +;; All Advanced SIMD modes suitable for moving, loading, and storing ;; > +except V2HF. > (define_mode_iterator VALL_F16 [V8QI V16QI V4HI V8HI V2SI V4SI V2DI > V4HF V8HF V4BF V8BF V2SF V4SF V2DF]) > > +;; All Advanced SIMD modes suitable for moving, loading, and storing ;; > +including V2HF (define_mode_iterator VMOVE [V8QI V16QI V4HI V8HI V2SI > +V4SI V2DI > + V4HF V8HF V4BF V8BF V2SF V4SF V2DF > + (V2HF "TARGET_SIMD_F16INST")]) > + > + > ;; The VALL_F16 modes except the 128-bit 2-element ones. > (define_mode_iterator VALL_F16_NO_V2Q [V8QI V16QI V4HI V8HI V2SI > V4SI > V4HF V8HF V2SF V4SF]) > @@ -1076,7 +1088,7 @@ (define_mode_attr nunits [(V8QI "8") (V16QI "16") > (V2SF "2") (V4SF "4") > (V1DF "1") (V2DF "2") > (DI "1") (DF "1") > - (V8DI "8")]) > + (V8DI "8") (V2HF "2")]) > > ;; Map a mode to the number of bits in it, if the size of the mode ;; is > constant. > @@ -1090,6 +1102,7 @@ (define_mode_attr s [(HF "h") (SF "s") (DF "d") (SI > "s") (DI "d")]) > > ;; Give the length suffix letter for a sign- or zero-extension. > (define_mode_attr size [(QI "b") (HI "h") (SI "w")]) > +(define_mode_attr sizel [(QI "b") (HI "h") (SI "")]) > > ;; Give the number of bits in the mode > (define_mode_attr sizen [(QI "8") (HI "16") (SI "32") (DI "64")]) @@ -1193,7 > +1206,7 @@ (define_mode_attr Vmntype [(V8HI ".8b") (V4SI ".4h") > (define_mode_attr Vetype [(V8QI "b") (V16QI "b") > (V4HI "h") (V8HI "h") > (V2SI "s") (V4SI "s") > - (V2DI "d") > + (V2DI "d") (V2HF "h") > (V4HF "h") (V8HF "h") > (V2SF "s") (V4SF "s") > (V2DF "d") > @@ -1285,7 +1298,7 @@ (define_mode_attr Vcwtype [(VNx16QI "b") > (VNx8QI "h") (VNx4QI "w") (VNx2QI "d") ;; more accurately. > (define_mode_attr stype [(V8QI "b") (V16QI "b") (V4HI "s") (V8HI "s") > (V2SI "s") (V4SI "s") (V2DI "d") (V4HF "s") > - (V8HF "s") (V2SF "s") (V4SF "s") (V2DF "d") > + (V8HF "s") (V2SF "s") (V4SF "s") (V2DF "d") (V2HF > "s") > (HF "s") (SF "s") (DF "d") (QI "b") (HI "s") > (SI "s") (DI "d")]) > > @@ -1360,8 +1373,8 @@ (define_mode_attr VEL [(V8QI "QI") (V16QI "QI") > (V4HF "HF") (V8HF "HF") > (V2SF "SF") (V4SF "SF") > (DF "DF") (V2DF "DF") > - (SI "SI") (HI "HI") > - (QI "QI") > + (SI "SI") (V2HF "HF") > + (QI "QI") (HI "HI") > (V4BF "BF") (V8BF "BF") > (VNx16QI "QI") (VNx8QI "QI") (VNx4QI "QI") (VNx2QI > "QI") > (VNx8HI "HI") (VNx4HI "HI") (VNx2HI "HI") @@ -1381,7 > +1394,7 @@ (define_mode_attr Vel [(V8QI "qi") (V16QI "qi") > (V2SF "sf") (V4SF "sf") > (V2DF "df") (DF "df") > (SI "si") (HI "hi") > - (QI "qi") > + (QI "qi") (V2HF "hf") > (V4BF "bf") (V8BF "bf") > (VNx16QI "qi") (VNx8QI "qi") (VNx4QI "qi") (VNx2QI "qi") > (VNx8HI "hi") (VNx4HI "hi") (VNx2HI "hi") @@ -1866,7 > +1879,7 @@ (define_mode_attr q [(V8QI "") (V16QI "_q") > (V4HF "") (V8HF "_q") > (V4BF "") (V8BF "_q") > (V2SF "") (V4SF "_q") > - (V2DF "_q") > + (V2HF "") (V2DF "_q") > (QI "") (HI "") (SI "") (DI "") (HF "") (SF "") (DF "") > (V2x8QI "") (V2x16QI "_q") > (V2x4HI "") (V2x8HI "_q") > @@ -1905,6 +1918,7 @@ (define_mode_attr vp [(V8QI "v") (V16QI "v") > (V2SI "p") (V4SI "v") > (V2DI "p") (V2DF "p") > (V2SF "p") (V4SF "v") > + (V2HF "p") > (V4HF "v") (V8HF "v")]) > > (define_mode_attr vsi2qi [(V2SI "v8qi") (V4SI "v16qi") diff --git > a/gcc/config/arm/types.md b/gcc/config/arm/types.md index > 7d0504bdd944e9c0d1b545b0b66a9a1adc808714..3cfbc7a93cca1bea4925853e5 > 1d0a147c5722247 100644 > --- a/gcc/config/arm/types.md > +++ b/gcc/config/arm/types.md > @@ -483,6 +483,7 @@ (define_attr "autodetect_type" > ; neon_fp_minmax_s_q > ; neon_fp_minmax_d > ; neon_fp_minmax_d_q > +; neon_fp_reduc_add_h > ; neon_fp_reduc_add_s > ; neon_fp_reduc_add_s_q > ; neon_fp_reduc_add_d > @@ -1033,6 +1034,7 @@ (define_attr "type" > neon_fp_minmax_d,\ > neon_fp_minmax_d_q,\ > \ > + neon_fp_reduc_add_h,\ > neon_fp_reduc_add_s,\ > neon_fp_reduc_add_s_q,\ > neon_fp_reduc_add_d,\ > @@ -1257,8 +1259,8 @@ (define_attr "is_neon_type" "yes,no" > neon_fp_compare_d, neon_fp_compare_d_q, neon_fp_minmax_s,\ > neon_fp_minmax_s_q, neon_fp_minmax_d, neon_fp_minmax_d_q,\ > neon_fp_neg_s, neon_fp_neg_s_q, neon_fp_neg_d, > neon_fp_neg_d_q,\ > - neon_fp_reduc_add_s, neon_fp_reduc_add_s_q, > neon_fp_reduc_add_d,\ > - neon_fp_reduc_add_d_q, neon_fp_reduc_minmax_s, > + neon_fp_reduc_add_h, neon_fp_reduc_add_s, > neon_fp_reduc_add_s_q,\ > + neon_fp_reduc_add_d, neon_fp_reduc_add_d_q, > + neon_fp_reduc_minmax_s,\ > neon_fp_reduc_minmax_s_q, neon_fp_reduc_minmax_d,\ > neon_fp_reduc_minmax_d_q,\ > neon_fp_cvt_narrow_s_q, neon_fp_cvt_narrow_d_q,\ diff --git > a/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c > b/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c > index > 07d71a63414b1066ea431e287286ad048515711a..e6021c5a42748701e5326a5c3 > 87a39a0bbadc9e5 100644 > --- a/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c > +++ b/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c > @@ -30,11 +30,9 @@ vec_slp_##TYPE (TYPE *restrict a, TYPE b, TYPE c, int n) > \ > TEST_ALL (VEC_PERM) > > /* We should use one DUP for each of the 8-, 16- and 32-bit types, > - although we currently use LD1RW for _Float16. We should use two > - DUPs for each of the three 64-bit types. */ > + We should use two DUPs for each of the three 64-bit types. */ > /* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.h, [hw]} 2 } } */ > -/* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.s, [sw]} 2 } } */ > -/* { dg-final { scan-assembler-times {\tld1rw\tz[0-9]+\.s, } 1 } } */ > +/* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.s, [sw]} 3 } } */ > /* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.d, [dx]} 9 } } */ > /* { dg-final { scan-assembler-times {\tzip1\tz[0-9]+\.d, z[0-9]+\.d, z[0- > 9]+\.d\n} 3 } } */ > /* { dg-final { scan-assembler-not {\tzip2\t} } } */ @@ -53,7 +51,7 @@ > TEST_ALL (VEC_PERM) > /* { dg-final { scan-assembler-times {\twhilelo\tp[0-7]\.s} 6 } } */ > /* { dg-final { scan-assembler-times {\twhilelo\tp[0-7]\.d} 6 } } */ > /* { dg-final { scan-assembler-not {\tldr} } } */ > -/* { dg-final { scan-assembler-times {\tstr} 2 } } */ > -/* { dg-final { scan-assembler-times {\tstr\th[0-9]+} 2 } } */ > +/* { dg-final { scan-assembler-not {\tstr} } } */ > +/* { dg-final { scan-assembler-not {\tstr\th[0-9]+} } } */ > > /* { dg-final { scan-assembler-not {\tuqdec} } } */