On Wed, Feb 26, 2025 at 04:00:25PM -0500, Zi Yan wrote:
> +static int __split_unmapped_folio(struct folio *folio, int new_order,
> +             struct page *split_at, struct page *lock_at,
> +             struct list_head *list, pgoff_t end,
> +             struct xa_state *xas, struct address_space *mapping,
> +             bool uniform_split)
> +{
[...]
> +             /* complete memcg works before add pages to LRU */
> +             split_page_memcg(&folio->page, old_order, split_order);
> +             split_page_owner(&folio->page, old_order, split_order);
> +             pgalloc_tag_split(folio, old_order, split_order);

At least split_page_memcg() needs to become aware of 'uniform_split'.

        if (folio_memcg_kmem(folio))
                obj_cgroup_get_many(__folio_objcg(folio), old_nr / new_nr - 1);

If we're doing uniform_split, that calculation should be
        old_order - new_order - 1


Reply via email to