On Mon, Apr 14, 2025 at 3:11 PM Kyrylo Tkachov <ktkac...@nvidia.com> wrote: > > Hi Honza, > > > On 13 Apr 2025, at 23:19, Jan Hubicka <hubi...@ucw.cz> wrote: > > > >> +@opindex fipa-reorder-for-locality > >> +@item -fipa-reorder-for-locality > >> +Group call chains close together in the binary layout to improve code code > >> +locality. This option is incompatible with an explicit > >> +@option{-flto-partition=} option since it enforces a custom partitioning > >> +scheme. > > > > Please also cross-link this with -fprofile-reorder-functions and > > -freorder-functions, which does similar thing. > > If you see how to clean-up the description of the other two so user is > > not confused. > > > > Perhaps say that -freorder-functions only partitions functions into > > never-executed/cold/normal/hot and -fprofile-reroder-functions is aiming > > for program startup optimization (it reorders by measured first time the > > function is executed. By accident it seems to kind of work for > > locality. > > Yeah, the option names are quite similar aren't they? > I’ve attempted to disambiguate them a bit in their description. > I’m attaching a diff from the previous version (as the full updated patch) to > make it easier to see what’s adjusted. > > > > > >> + > >> +/* Helper function of to accumulate call counts. */ > >> +static bool > >> +accumulate_profile_counts_after_cloning (cgraph_node *node, void *data) > >> +{ > >> + struct profile_stats *stats = (struct profile_stats *) data; > >> + for (cgraph_edge *e = node->callers; e; e = e->next_caller) > >> + { > >> + if (e->caller == stats->target) > >> + { > >> + if (stats->rec_count.compatible_p (e->count.ipa ())) > >> + stats->rec_count += e->count.ipa (); > >> + } > >> + else > >> + { > >> + if (stats->nonrec_count.compatible_p (e->count.ipa ())) > >> + stats->nonrec_count += e->count.ipa (); > >> + } > > In case part of profile is missing (which may happen if one unit has -O0 > > or so) , we may have counts to be uninitialized. Uninitialized counts are > > compatible with everything, but any arithmetics with it will produce > > uninitialized result which will likely confuse code later. So I would > > skip edges with uninitialized counts. > > > > On the other hand ipa counts are always compatible, so compatible_p > > should be redundat. Main reaosn for existence of compatible_p is that we > > can have local profiles that are 0 or unknown at IPA level. The ipa () > > conversion turns all counts into IPA counts and those are compatible > > with each other. > > > > I suppose compatibe_p test is there since the code ICEd in past,but I > > think it was because of missing ipa() conversion. > > > > > >> + } > >> + return false; > >> +} > >> + > >> +/* NEW_NODE is a previously created clone of ORIG_NODE already present in > >> + current partition. EDGES contains newly redirected edges to NEW_NODE. > >> + Adjust profile information for both nodes and the edge. */ > >> + > >> +static void > >> +adjust_profile_info_for_non_self_rec_edges (auto_vec<cgraph_edge *> > >> &edges, > >> + cgraph_node *new_node, > >> + cgraph_node *orig_node) > >> +{ > >> + profile_count orig_node_count = orig_node->count.ipa (); > >> + profile_count edge_count = profile_count::zero (); > >> + profile_count final_new_count = profile_count::zero (); > >> + profile_count final_orig_count = profile_count::zero (); > >> + > >> + for (unsigned i = 0; i < edges.length (); ++i) > >> + edge_count += edges[i]->count.ipa (); > > Here I would again skip uninitialized. It is probably legal for -O0 > > function to end up in partition. > >> + > >> + final_orig_count = orig_node_count - edge_count; > >> + > >> + /* NEW_NODE->count was adjusted for other callers when the clone was > >> + first created. Just add the new edge count. */ > >> + if (new_node->count.compatible_p (edge_count)) > >> + final_new_count = new_node->count + edge_count; > > And here compatible_p should be unnecesary. > >> +/* Accumulate frequency of all edges from EDGE->caller to EDGE->callee. > >> */ > >> + > >> +static sreal > >> +accumulate_incoming_edge_frequency (cgraph_edge *edge) > >> +{ > >> + sreal count = 0; > >> + struct cgraph_edge *e; > >> + for (e = edge->callee->callers; e; e = e->next_caller) > >> + { > >> + /* Make a local decision about all edges for EDGE->caller but not > >> the > >> + other nodes already in the partition. Their edges will be visited > >> + later or may have been visited before and not fit the > >> + cut-off criteria. */ > >> + if (e->caller == edge->caller) > >> + { > >> + profile_count caller_count = e->caller->inlined_to > >> + ? e->caller->inlined_to->count > >> + : e->caller->count; > >> + if (e->count.compatible_p (caller_count)) > > Here again compatiblity check should not be necessary, since the counts > > belong to one function body (after inlining) and should be compatible. > > inliner calls e->sreal_frequency all the time withotu further checks. > > > > Yeah, I’ve adjusted these uses and used checks for initialized_p where you > highlighted. > Indeed it seems to work with a profiled bootstrap and the ICEs we were seeing > earlier in development aren’t appearing. > > > Patch is OK with these changes. I apologize for letting the review slip > > for so long. I was setting up Firefox testing and LLVM builds to gather > > some data that took longer than I hoped for. On Firefox and LLVM on zen > > I can measure some improvements via instruction cache perofmrance > > counters, but the overall benefit seems to be close to noise, but this > > is likely very CPU specfic. Overall code locality is one of main missing > > parts of the LTO framework. As discussed on Cauldron, I think next > > stage1 we can untie this from partitining algorithm, but that needs more > > work on linker side as well as on gas fragments, so I think it is a good > > idea to move with this patch as it is and improve from it. > > Thank you very much for the evaluation and the feedback! I agree the effect > can be very CPU-specific. > > > > > I think the patch is modifying almost no code that is run w/o the > > -fipa-reorder-for-locality so I hope it is safe for 15.1, but I would > > like to let Richi and Jakub to comment on this. > > Thanks a lot for your reviews yet again. They were very helpful. > I’ve updated the patch as per your suggestion and did a profiled lto > bootstrap with the new bootstrap-lto-locality.mk that exercises this. > All looks good on aarch64-none-linux-gnu. > Richi, Jakub, is it okay for trunk now given Honza’s comments?
OK, prepare to back out if there are any issues though. Quickly skimming over the patch shows +/* Locality partitions, assigns nodes to partitions. These are used later in + WPA partitioning. */ +vec<locality_partition> locality_partitions; + +/* Map from original node to its latest clone. Gets overwritten whenever a new + clone is created from the same node. */ +hash_map<cgraph_node *, cgraph_node *> node_to_clone; +/* Map from clone to its original node. */ +hash_map<cgraph_node *, cgraph_node *> clone_to_node; those global CTORs are frowned upon (why are those not static?), we prefer those to be pointers. They are also not marked for garbage collection but cgraph_node generally are so I assume they are only live through the IPA pass itself. So they should be ideally created in a more local scope. Richard. > > > > > > Honza >