linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [slub p4 0/7] slub: per cpu partial lists V4
@ 2011-08-09 21:12 Christoph Lameter
  2011-08-09 21:12 ` [slub p4 1/7] slub: free slabs without holding locks (V2) Christoph Lameter
                   ` (8 more replies)
  0 siblings, 9 replies; 20+ messages in thread
From: Christoph Lameter @ 2011-08-09 21:12 UTC (permalink / raw)
  To: Pekka Enberg
  Cc: David Rientjes, Andi Kleen, tj, Metathronius Galabant,
	Matt Mackall, Eric Dumazet, Adrian Drzewiecki, linux-kernel

V3->V4 : Use a single linked per cpu list instead of a per cpu array.
	This results in improvements even for the single threaded
	case. I think this is ready for more widespread testing (-next?)
	The number of partial pages per cpu is configurable via
	/sys/kernel/slab/<name>/cpu_partial

V2->V3 : Work on the todo list. Still some work to be done to reduce
         code impact and make this all cleaner. (Pekka: patch 1-3
         are cleanup patches of general usefulness. You got #1 already
         2+3 could be picked up w/o any issue).

The following patchset introduces per cpu partial lists which allow
a performance increase of around ~10-20% with hackbench on my Sandybridge
processor.

These lists help to avoid per node locking overhead. Allocator latency
could be further reduced by making these operations work without
disabling interrupts (like the fastpath and the free slowpath) but that
is another project.

It is interesting to note that BSD has gone to a scheme with partial
pages only per cpu (source: Adrian). Transfer of cpu ownerships is
done using IPIs. Probably too much overhead for our taste. The approach
here keeps the per node partial lists essentially meaning the "pages"
in there have no cpu owner.


^ permalink raw reply	[flat|nested] 20+ messages in thread

end of thread, other threads:[~2011-08-24 13:58 UTC | newest]

Thread overview: 20+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2011-08-09 21:12 [slub p4 0/7] slub: per cpu partial lists V4 Christoph Lameter
2011-08-09 21:12 ` [slub p4 1/7] slub: free slabs without holding locks (V2) Christoph Lameter
2011-08-20 10:32   ` Pekka Enberg
2011-08-20 15:58     ` Christoph Lameter
2011-08-09 21:12 ` [slub p4 2/7] slub: Remove useless statements in __slab_alloc Christoph Lameter
2011-08-20 10:44   ` Pekka Enberg
2011-08-20 16:01     ` Christoph Lameter
2011-08-09 21:12 ` [slub p4 3/7] slub: Prepare inuse field in new_slab() Christoph Lameter
2011-08-09 21:12 ` [slub p4 4/7] slub: pass kmem_cache_cpu pointer to get_partial() Christoph Lameter
2011-08-09 21:12 ` [slub p4 5/7] slub: return object pointer from get_partial() / new_slab() Christoph Lameter
2011-08-09 21:12 ` [slub p4 6/7] slub: per cpu cache for partial pages Christoph Lameter
2011-08-20 10:40   ` Pekka Enberg
2011-08-20 16:00     ` Christoph Lameter
     [not found]   ` <CAF1ivSaH9fh6_QvuBkLc5t=zC4mPEAD5ZzsxOuPruDwG9MiZzw@mail.gmail.com>
2011-08-24  7:26     ` Lin Ming
2011-08-24 13:57       ` Christoph Lameter
2011-08-09 21:12 ` [slub p4 7/7] slub: update slabinfo tools to report per cpu partial list statistics Christoph Lameter
2011-08-13 18:28 ` [slub p4 0/7] slub: per cpu partial lists V4 David Rientjes
2011-08-15  8:44   ` Pekka Enberg
2011-08-15 14:29   ` Christoph Lameter
2011-08-20 10:48 ` Pekka Enberg

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).