on 21/09/2010 19:16 Alan Cox said the following:
> Actually, I think that there is a middle ground between "per-cpu caches" and
> "directly from the VM" that we are missing.  When I've looked at the default
> configuration of ZFS (without the extra UMA zones enabled), there is an
> incredible amount of churn on the kmem map caused by the implementation of
> uma_large_malloc() and uma_large_free() going directly to the kmem map.  Not
> only are the obvious things happening, like allocating and freeing kernel
> virtual addresses and underlying physical pages on every call, but also
> system-wide TLB shootdowns and sometimes superpage demotions are occurring.
> 
> I have some trouble believing that the large allocations being performed by 
> ZFS
> really need per-CPU caching, but I can certainly believe that they could 
> benefit
> from not going directly to the kmem map on every uma_large_malloc() and
> uma_large_free().  In other words, I think it would make a lot of sense to 
> have
> a thin layer between UMA and the kmem map that caches allocated but unused
> ranges of pages.

Alan,

thank you very much for the testing and analysis.
These are very good points.

So, for the reference, here are two patches that I came up with:
1. original patch that attempts to implement Solaris-like behavior but doesn't
go all the way to disabling per-CPU caches:
http://people.freebsd.org/~avg/uma-1.diff

2. patch that attempts to implement Jeff's three suggestions; I've tested
per-CPU cache size adaptive behavior, works well, but haven't tested per-CPU
cache draining yet:
http://people.freebsd.org/~avg/uma-2.diff

-- 
Andriy Gapon
_______________________________________________
freebsd-hackers@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-hackers
To unsubscribe, send any mail to "freebsd-hackers-unsubscr...@freebsd.org"

Reply via email to