Hi All, Attached is a new version that fixes the previous SVE fallouts in a new way.
Ok for master? Thanks, Tamar --- inline copy of patch --- diff --git a/gcc/config/aarch64/aarch64-cost-tables.h b/gcc/config/aarch64/aarch64-cost-tables.h index dd2e7e7cbb13d24f0b51092270cd7e2d75fabf29..bb499a1eae62a145f1665d521f57c98b49ac5389 100644 --- a/gcc/config/aarch64/aarch64-cost-tables.h +++ b/gcc/config/aarch64/aarch64-cost-tables.h @@ -124,7 +124,10 @@ const struct cpu_cost_table qdf24xx_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -229,7 +232,10 @@ const struct cpu_cost_table thunderx_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* Alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -333,7 +339,10 @@ const struct cpu_cost_table thunderx2t99_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* Alu. */ - COSTS_N_INSNS (4) /* Mult. */ + COSTS_N_INSNS (4), /* Mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -437,7 +446,10 @@ const struct cpu_cost_table thunderx3t110_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* Alu. */ - COSTS_N_INSNS (4) /* Mult. */ + COSTS_N_INSNS (4), /* Mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -542,7 +554,10 @@ const struct cpu_cost_table tsv110_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -646,7 +661,10 @@ const struct cpu_cost_table a64fx_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; diff --git a/gcc/config/aarch64/aarch64-simd.md b/gcc/config/aarch64/aarch64-simd.md index 29f381728a3b3d28bcd6a1002ba398c8b87713d2..61c3d7e195c510da88aa513f99af5f76f4d696e7 100644 --- a/gcc/config/aarch64/aarch64-simd.md +++ b/gcc/config/aarch64/aarch64-simd.md @@ -74,12 +74,14 @@ (define_insn "aarch64_simd_dup<mode>" ) (define_insn "aarch64_simd_dup<mode>" - [(set (match_operand:VDQF_F16 0 "register_operand" "=w") + [(set (match_operand:VDQF_F16 0 "register_operand" "=w,w") (vec_duplicate:VDQF_F16 - (match_operand:<VEL> 1 "register_operand" "w")))] + (match_operand:<VEL> 1 "register_operand" "w,r")))] "TARGET_SIMD" - "dup\\t%0.<Vtype>, %1.<Vetype>[0]" - [(set_attr "type" "neon_dup<q>")] + "@ + dup\\t%0.<Vtype>, %1.<Vetype>[0] + dup\\t%0.<Vtype>, %<vw>1" + [(set_attr "type" "neon_dup<q>, neon_from_gp<q>")] ) (define_insn "aarch64_dup_lane<mode>" diff --git a/gcc/config/aarch64/aarch64.c b/gcc/config/aarch64/aarch64.c index 699c105a42a613c06c462e2de686795279d85bc9..542fc874a4e224fb2cbe94e64eab590458fe935b 100644 --- a/gcc/config/aarch64/aarch64.c +++ b/gcc/config/aarch64/aarch64.c @@ -12705,7 +12705,7 @@ aarch64_rtx_costs (rtx x, machine_mode mode, int outer ATTRIBUTE_UNUSED, rtx op0, op1, op2; const struct cpu_cost_table *extra_cost = aarch64_tune_params.insn_extra_cost; - int code = GET_CODE (x); + rtx_code code = GET_CODE (x); scalar_int_mode int_mode; /* By default, assume that everything has equivalent cost to the @@ -13466,8 +13466,7 @@ cost_plus: we must cost the explicit register move. */ if (mode == DImode - && GET_MODE (op0) == SImode - && outer == SET) + && GET_MODE (op0) == SImode) { int op_cost = rtx_cost (op0, VOIDmode, ZERO_EXTEND, 0, speed); @@ -14006,8 +14005,39 @@ cost_plus: mode, MULT, 1, speed); return true; } + break; + case CONST_VECTOR: + { + /* Load using MOVI/MVNI. */ + if (aarch64_simd_valid_immediate (x, NULL)) + *cost = extra_cost->vect.movi; + else /* Load using constant pool. */ + *cost = extra_cost->ldst.load; + break; + } + case VEC_CONCAT: + /* depending on the operation, either DUP or INS. + For now, keep default costing. */ + break; + /* Load using a DUP. */ + case VEC_DUPLICATE: + *cost = extra_cost->vect.dup; + return false; + case VEC_SELECT: + { + rtx op0 = XEXP (x, 0); + *cost = rtx_cost (op0, GET_MODE (op0), VEC_SELECT, 0, speed); - /* Fall through. */ + /* cost subreg of 0 as free, otherwise as DUP */ + rtx op1 = XEXP (x, 1); + if (vec_series_lowpart_p (mode, GET_MODE (op1), op1)) + ; + else if (vec_series_highpart_p (mode, GET_MODE (op1), op1)) + *cost = extra_cost->vect.dup; + else + *cost = extra_cost->vect.extract; + return true; + } default: break; } diff --git a/gcc/config/arm/aarch-common-protos.h b/gcc/config/arm/aarch-common-protos.h index 6be5fb1e083d7ff130386dfa181b9a0c8fd5437c..55a470d8e1410bdbcfbea084ec11b468485c1400 100644 --- a/gcc/config/arm/aarch-common-protos.h +++ b/gcc/config/arm/aarch-common-protos.h @@ -133,6 +133,9 @@ struct vector_cost_table { const int alu; const int mult; + const int movi; + const int dup; + const int extract; }; struct cpu_cost_table diff --git a/gcc/config/arm/aarch-cost-tables.h b/gcc/config/arm/aarch-cost-tables.h index 25ff702f01fab50d749b9a7b7b072c2be2504562..0e6a62665c7e18debc382a294a37945188fb90ef 100644 --- a/gcc/config/arm/aarch-cost-tables.h +++ b/gcc/config/arm/aarch-cost-tables.h @@ -122,7 +122,10 @@ const struct cpu_cost_table generic_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -226,7 +229,10 @@ const struct cpu_cost_table cortexa53_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -330,7 +336,10 @@ const struct cpu_cost_table cortexa57_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -434,7 +443,10 @@ const struct cpu_cost_table cortexa76_extra_costs = /* Vector */ { COSTS_N_INSNS (1), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -538,7 +550,10 @@ const struct cpu_cost_table exynosm1_extra_costs = /* Vector */ { COSTS_N_INSNS (0), /* alu. */ - COSTS_N_INSNS (4) /* mult. */ + COSTS_N_INSNS (4), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; @@ -642,7 +657,10 @@ const struct cpu_cost_table xgene1_extra_costs = /* Vector */ { COSTS_N_INSNS (2), /* alu. */ - COSTS_N_INSNS (8) /* mult. */ + COSTS_N_INSNS (8), /* mult. */ + COSTS_N_INSNS (1), /* movi. */ + COSTS_N_INSNS (2), /* dup. */ + COSTS_N_INSNS (2) /* extract. */ } }; diff --git a/gcc/testsuite/gcc.target/aarch64/vect-cse-codegen.c b/gcc/testsuite/gcc.target/aarch64/vect-cse-codegen.c new file mode 100644 index 0000000000000000000000000000000000000000..d025e989a1e67f00f4f4ce94897a961d38abfab7 --- /dev/null +++ b/gcc/testsuite/gcc.target/aarch64/vect-cse-codegen.c @@ -0,0 +1,97 @@ +/* { dg-do compile { target { lp64 } } } */ +/* { dg-additional-options "-O3 -march=armv8.2-a+crypto -fno-schedule-insns -fno-schedule-insns2 -mcmodel=small" } */ +/* { dg-final { check-function-bodies "**" "" "" { target { le } } } } */ + +#include <arm_neon.h> + +/* +**test1: +** adrp x[0-9]+, .LC[0-9]+ +** ldr q[0-9]+, \[x[0-9]+, #:lo12:.LC[0-9]+\] +** add v[0-9]+.2d, v[0-9]+.2d, v[0-9]+.2d +** str q[0-9]+, \[x[0-9]+\] +** fmov x[0-9]+, d[0-9]+ +** orr x[0-9]+, x[0-9]+, x[0-9]+ +** ret +*/ + +uint64_t +test1 (uint64_t a, uint64x2_t b, uint64x2_t* rt) +{ + uint64_t arr[2] = { 0x0942430810234076UL, 0x0942430810234076UL}; + uint64_t res = a | arr[0]; + uint64x2_t val = vld1q_u64 (arr); + *rt = vaddq_u64 (val, b); + return res; +} + +/* +**test2: +** adrp x[0-9]+, .LC[0-1]+ +** ldr q[0-9]+, \[x[0-9]+, #:lo12:.LC[0-9]+\] +** add v[0-9]+.2d, v[0-9]+.2d, v[0-9]+.2d +** str q[0-9]+, \[x[0-9]+\] +** fmov x[0-9]+, d[0-9]+ +** orr x[0-9]+, x[0-9]+, x[0-9]+ +** ret +*/ + +uint64_t +test2 (uint64_t a, uint64x2_t b, uint64x2_t* rt) +{ + uint64x2_t val = vdupq_n_u64 (0x0424303242234076UL); + uint64_t arr = vgetq_lane_u64 (val, 0); + uint64_t res = a | arr; + *rt = vaddq_u64 (val, b); + return res; +} + +/* +**test3: +** adrp x[0-9]+, .LC[0-9]+ +** ldr q[0-9]+, \[x[0-9]+, #:lo12:.LC[0-9]+\] +** add v[0-9]+.4s, v[0-9]+.4s, v[0-9]+.4s +** str q[0-9]+, \[x1\] +** fmov w[0-9]+, s[0-9]+ +** orr w[0-9]+, w[0-9]+, w[0-9]+ +** ret +*/ + +uint32_t +test3 (uint32_t a, uint32x4_t b, uint32x4_t* rt) +{ + uint32_t arr[4] = { 0x094243, 0x094243, 0x094243, 0x094243 }; + uint32_t res = a | arr[0]; + uint32x4_t val = vld1q_u32 (arr); + *rt = vaddq_u32 (val, b); + return res; +} + +/* +**test4: +** ushr v[0-9]+.16b, v[0-9]+.16b, 7 +** mov x[0-9]+, 16512 +** movk x[0-9]+, 0x1020, lsl 16 +** movk x[0-9]+, 0x408, lsl 32 +** movk x[0-9]+, 0x102, lsl 48 +** fmov d[0-9]+, x[0-9]+ +** pmull v[0-9]+.1q, v[0-9]+.1d, v[0-9]+.1d +** dup v[0-9]+.2d, v[0-9]+.d\[0\] +** pmull2 v[0-9]+.1q, v[0-9]+.2d, v[0-9]+.2d +** trn2 v[0-9]+.8b, v[0-9]+.8b, v[0-9]+.8b +** umov w[0-9]+, v[0-9]+.h\[3\] +** ret +*/ + +uint64_t +test4 (uint8x16_t input) +{ + uint8x16_t bool_input = vshrq_n_u8(input, 7); + poly64x2_t mask = vdupq_n_p64(0x0102040810204080UL); + poly64_t prodL = vmull_p64((poly64_t)vgetq_lane_p64((poly64x2_t)bool_input, 0), + vgetq_lane_p64(mask, 0)); + poly64_t prodH = vmull_high_p64((poly64x2_t)bool_input, mask); + uint8x8_t res = vtrn2_u8((uint8x8_t)prodL, (uint8x8_t)prodH); + return vget_lane_u16((uint16x4_t)res, 3); +} +
rb14774.patch
Description: rb14774.patch