From 3b123520428533196629e1c6e1d38990249faef6 Mon Sep 17 00:00:00 2001 From: Christoph Lameter Date: Mon, 7 Jan 2008 23:20:30 -0800 Subject: [PATCH] SLUB: Alternate fast paths using cmpxchg_local Provide an alternate implementation of the SLUB fast paths for alloc and free using cmpxchg_local. The cmpxchg_local fast path is selected for arches that have CONFIG_FAST_CMPXCHG_LOCAL set. An arch should only set CONFIG_FAST_CMPXCHG_LOCAL if the cmpxchg_local is faster than an interrupt enable/disable sequence. This is known to be true for both x86 platforms so set FAST_CMPXCHG_LOCAL for both arches. Not all arches can support fast cmpxchg operations. Typically the architecture must have an optimized cmpxchg instruction. The cmpxchg fast path makes no sense on platforms whose cmpxchg is slower than interrupt enable/disable (like f.e. IA64). The advantages of a cmpxchg_local based fast path are: 1. Lower cycle count (30%-60% faster) 2. There is no need to disable and enable interrupts on the fast path. Currently interrupts have to be disabled and enabled on every slab operation. This is likely saving a significant percentage of interrupt off / on sequences in the kernel. 3. The disposal of freed slabs can occur with interrupts enabled. The alternate path is realized using #ifdef's. Several attempts to do the same with macros and in line functions resulted in a mess (in particular due to the strange way that local_interrupt_save() handles its argument and due to the need to define macros/functions that sometimes disable interrupts and sometimes do something else. The macro based approaches made it also #ifdef seems to be the way to go here to have a readable source. Signed-off-by: Christoph Lameter Reviewed-by: Pekka Enberg Cc: Signed-off-by: Andrew Morton --- arch/x86/Kconfig | 4 +++ mm/slub.c | 71 ++++++++++++++++++++++++++++++++++++++++++++++++++++- 2 files changed, 73 insertions(+), 2 deletions(-) diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig index 80b7ba4..ada1f70 100644 --- a/arch/x86/Kconfig +++ b/arch/x86/Kconfig @@ -53,6 +53,10 @@ config SEMAPHORE_SLEEPERS bool default y +config FAST_CMPXCHG_LOCAL + bool + default y + config MMU bool default y diff --git a/mm/slub.c b/mm/slub.c index 08be5a1..42d8a56 100644 --- a/mm/slub.c +++ b/mm/slub.c @@ -1499,7 +1499,12 @@ static void *__slab_alloc(struct kmem_cache *s, { void **object; struct page *new; +#ifdef CONFIG_FAST_CMPXCHG_LOCAL + unsigned long flags; + local_irq_save(flags); + preempt_enable_no_resched(); +#endif if (!c->page) goto new_slab; @@ -1521,6 +1526,10 @@ load_freelist: unlock_out: slab_unlock(c->page); out: +#ifdef CONFIG_FAST_CMPXCHG_LOCAL + preempt_disable(); + local_irq_restore(flags); +#endif return object; another_slab: @@ -1577,9 +1586,26 @@ static __always_inline void *slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, void *addr) { void **object; - unsigned long flags; struct kmem_cache_cpu *c; +#ifdef CONFIG_FAST_CMPXCHG_LOCAL + c = get_cpu_slab(s, get_cpu()); + do { + object = c->freelist; + if (unlikely(is_end(object) || !node_match(c, node))) { + object = __slab_alloc(s, gfpflags, node, addr, c); + if (unlikely(!object)) { + put_cpu(); + goto out; + } + break; + } + } while (cmpxchg_local(&c->freelist, object, object[c->offset]) + != object); + put_cpu(); +#else + unsigned long flags; + local_irq_save(flags); c = get_cpu_slab(s, smp_processor_id()); if (unlikely((is_end(c->freelist)) || !node_match(c, node))) { @@ -1594,6 +1620,7 @@ static __always_inline void *slab_alloc(struct kmem_cache *s, c->freelist = object[c->offset]; } local_irq_restore(flags); +#endif if (unlikely((gfpflags & __GFP_ZERO))) memset(object, 0, c->objsize); @@ -1629,6 +1656,11 @@ static void __slab_free(struct kmem_cache *s, struct page *page, void *prior; void **object = (void *)x; +#ifdef CONFIG_FAST_CMPXCHG_LOCAL + unsigned long flags; + + local_irq_save(flags); +#endif slab_lock(page); if (unlikely(SlabDebug(page))) @@ -1654,6 +1686,9 @@ checks_ok: out_unlock: slab_unlock(page); +#ifdef CONFIG_FAST_CMPXCHG_LOCAL + local_irq_restore(flags); +#endif return; slab_empty: @@ -1664,6 +1699,9 @@ slab_empty: remove_partial(s, page); slab_unlock(page); +#ifdef CONFIG_FAST_CMPXCHG_LOCAL + local_irq_restore(flags); +#endif discard_slab(s, page); return; @@ -1688,9 +1726,37 @@ static __always_inline void slab_free(struct kmem_cache *s, struct page *page, void *x, void *addr) { void **object = (void *)x; - unsigned long flags; struct kmem_cache_cpu *c; +#ifdef CONFIG_FAST_CMPXCHG_LOCAL + void **freelist; + + c = get_cpu_slab(s, get_cpu()); + debug_check_no_locks_freed(object, s->objsize); + do { + freelist = c->freelist; + barrier(); + /* + * If the compiler would reorder the retrieval of c->page to + * come before c->freelist then an interrupt could + * change the cpu slab before we retrieve c->freelist. We + * could be matching on a page no longer active and put the + * object onto the freelist of the wrong slab. + * + * On the other hand: If we already have the freelist pointer + * then any change of cpu_slab will cause the cmpxchg to fail + * since the freelist pointers are unique per slab. + */ + if (unlikely(page != c->page || c->node < 0)) { + __slab_free(s, page, x, addr, c->offset); + break; + } + object[c->offset] = freelist; + } while (cmpxchg_local(&c->freelist, freelist, object) != freelist); + put_cpu(); +#else + unsigned long flags; + local_irq_save(flags); debug_check_no_locks_freed(object, s->objsize); c = get_cpu_slab(s, smp_processor_id()); @@ -1701,6 +1767,7 @@ static __always_inline void slab_free(struct kmem_cache *s, __slab_free(s, page, x, addr, c->offset); local_irq_restore(flags); +#endif } void kmem_cache_free(struct kmem_cache *s, void *x) -- debian.1.5.3.7.1-dirty