https://gcc.gnu.org/bugzilla/show_bug.cgi?id=105617
--- Comment #11 from Michael_S <already5chosen at yahoo dot com> --- (In reply to Richard Biener from comment #10) > (In reply to Hongtao.liu from comment #9) > > (In reply to Hongtao.liu from comment #8) > > > (In reply to Hongtao.liu from comment #7) > > > > Hmm, we have specific code to add scalar->vector(vmovq) cost to vector > > > > construct, but it seems not to work here, guess it's because &r0,and > > > > thought > > > > it was load not scalar? > > > Yes, true for as gimple_assign_load_p > > > > > > > > > (gdb) p debug_gimple_stmt (def) > > > 72# VUSE <.MEM_46> > > > 73r0.0_20 = r0; > > It's a load from stack, and finally eliminated in rtl dse1, but here the > > vectorizer doesn't know. > > Yes, it's difficult for the SLP vectorizer to guess whether rN will come > from memory or not. Some friendlier middle-end representation for > add-with-carry might be nice - the x86 backend could for example fold > __builtin_ia32_addcarryx_u64 to use a _Complex unsinged long long for the > return, ferrying the carry in __imag. Alternatively we could devise > some special GIMPLE_ASM kind ferrying RTL and not assembly so the > backend could fold it directly to RTL on GIMPLE with asm constraints > doing the plumbing ... (we'd need some match-scratch and RTL expansion > would still need to allocate the actual pseudos). > > <bb 2> [local count: 1073741824]: > _1 = *srcB_17(D); > _2 = *srcA_18(D); > _30 = __builtin_ia32_addcarryx_u64 (0, _2, _1, &r0); > _3 = MEM[(const uint64_t *)srcB_17(D) + 8B]; > _4 = MEM[(const uint64_t *)srcA_18(D) + 8B]; > _5 = (int) _30; > _29 = __builtin_ia32_addcarryx_u64 (_5, _4, _3, &r1); > _6 = MEM[(const uint64_t *)srcB_17(D) + 16B]; > _7 = MEM[(const uint64_t *)srcA_18(D) + 16B]; > _8 = (int) _29; > _28 = __builtin_ia32_addcarryx_u64 (_8, _7, _6, &r2); > _9 = MEM[(const uint64_t *)srcB_17(D) + 24B]; > _10 = MEM[(const uint64_t *)srcA_18(D) + 24B]; > _11 = (int) _28; > __builtin_ia32_addcarryx_u64 (_11, _10, _9, &r3); > r0.0_12 = r0; > r1.1_13 = r1; > _36 = {r0.0_12, r1.1_13}; > r2.2_14 = r2; > r3.3_15 = r3; > _37 = {r2.2_14, r3.3_15}; > vectp.9_35 = dst_19(D); > MEM <vector(2) long unsigned int> [(uint64_t *)vectp.9_35] = _36; > vectp.9_39 = vectp.9_35 + 16; > MEM <vector(2) long unsigned int> [(uint64_t *)vectp.9_39] = _37; > > so for the situation at hand I don't see any reasonable way out that > doesn't have the chance of regressing things in other places (like > treat loads from non-indexed auto variables specially or so). The > only real solution is to find a GIMPLE representation for > __builtin_ia32_addcarryx_u64 that doesn't force the alternate output > to memory. How about simple heuristic: "Never auto-vectorize integer code unless in inner loop"? It would be optimal >80% of time and even in cases where it's sub-optimal the impact is likely single-digit per cents. On the other hand, the impact of mistake in the opposit direction (i.e. over-vectorization) is often quite large.