Slabifier: A slab allocator with minimal meta information Lately I have started tinkering around with the slab in particular after Matt Mackal mentioned that the slab should be more modular at the KS. One particular design issue with the current slab is that it is build on the basic notion of shifting object references from list to list. Without NUMA this is wild enough with the per cpu caches and the shared cache but with NUMA we now have per node shared arrays, per node list and per node per node alien caches. Somehow this all works but one wonders does it have to be that way? On very large systems the number of these entities grows to unbelievable numbers. So I thought it may be best to try to develop another basic slab layer that does not have all the object queues and that does not have to carry so much state information. I also have had concerns about the way locking is handled for awhile. We could increase parallelism by finer grained locking. This in turn may avoid the need for object queues. After toying around for awhile I came to the realization that the page struct contains all the information necessary to manage a slab block. One can put all the management information there and that is also advantageous for performance since we constantly have to use the page struct anyways for reverse object lookups and during slab creation. So this also reduces the cache footprint of the slab. The alignment is naturally the best since the first object starts right at the page boundary. This reduces the complexity of alignment calculations. We use two locks: 1. The per slab list_lock to protect the partial list. This lock does not protect the slab and is only taken during partial list operations. List is reduced to necessary moves if the state of a page changes. An allocation of an object or the freeing of an object in a slab does not require that the list_lock is taken. 2. The page lock in struct page is used to protect the slab during allocation and freeing. This lock is conveniently placed in a cacheline that is already available. Other key information is also placed there. struct page overloading: - _mapcout => Used to count the objects in use in a slab - mapping => Reference to the slab structure - index => Pointer to the first free element in a slab - lru => Used for list management. Flag overloading: PageReferenced => Used to control per cpu slab freeing. PageActive => slab is under active allocation. PageLocked => slab locking The freelists of objects per page are managed as a chained list. The struct page contains a pointer to the first element. The first 4 bytes of the free element contains a pointer to the next free element etc until the chain ends with NULL. There is no freelist for slabs. slabs are immediately returned to the page allocator. The page allocator has its own per cpu page queues that should provide enough caching (only works for order 0 pages though). There is also no list of full slabs. Full slabs will come back only when an object is freed on the slab. Otherwise full slabs are of no concern (apart from counting them). Per cpu caches exist in the sense that each processor has a per processor "cpuslab". Objects in this slab will only be allocated from this processor. The page state is likely going to stay in the cache. Allocation will be very fast since we only need the page struct reference for all our needs which is likely not contended at all. Fetching the next free pointer from the location of the object nicely prefetches the object. Signed-off-by: Christoph Lameter Index: linux-2.6.18-rc4-mm2/mm/slabifier.c =================================================================== --- /dev/null 1970-01-01 00:00:00.000000000 +0000 +++ linux-2.6.18-rc4-mm2/mm/slabifier.c 2006-08-25 23:48:19.209830001 -0700 @@ -0,0 +1,1024 @@ +/* + * Generic Slabifier for the allocato abstraction framework. + * The allocator synchronizes using slab based locks and only + * uses a centralized list lock to manage the pool of partial slabs. + * + * (C) 2006 Silicon Graphics Inc., Christoph Lameter + */ + +#include +#include +#include +#include +#include + +#ifdef SLABIFIER_DEBUG +#define DBUG_ON(_x) BUG_ON(_x) +#else +#define DBUG_ON(_x) +#endif + +struct slab { + struct slab_cache sc; +#ifdef CONFIG_SMP + int flusher_active; + struct work_struct flush; +#endif + atomic_t refcount; /* Refcount for destroy */ + atomic_long_t nr_slabs; /* Total slabs used */ + /* Performance critical items follow */ + int size; /* Total size of an object */ + int offset; /* Free pointer offset. */ + int objects; /* Number of objects in slab */ + spinlock_t list_lock; + struct list_head partial; + unsigned long nr_partial; + struct page *active[NR_CPUS]; +}; + +/* + * The page struct is used to keep necessary information about a slab. + * For a compound page the first page keeps the slab state. + * + * Overloaded fields in struct page: + * + * lru -> used to a slab on the lists + * mapping -> pointer to struct slab + * index -> pointer to next free object + * _mapcount -> count number of elements in use + * + * Lock order: + * 1. slab_lock(page) + * 2. slab->list_lock + * + * The slabifier assigns one slab for allocation to each processor. + * Allocators only occur from these active slabs. + * If a cpu slab is active thena workqueue thread checks every 10 + * seconds if the cpu slab is still in use. The cpu slab is pushed back + * to the list if inactive. + * + * Leftover slabs with free elements are kept on the partial list. + * There is no list for full slabs. If an object in a full slab is + * freed then the slab will show up again on the partial lists. + * Otherwise we have no way of tracking used slabs. + * + * Slabs are freed when they become empty. Teardown and setup is + * minimal so we rely on the page allocators per cpu caches for + * fast frees and allocations. + */ + +#define lru_to_last_page(_head) (list_entry((_head)->next, struct page, lru)) +#define lru_to_first_page(_head) (list_entry((_head)->next, struct page, lru)) + +/* + * Some definitions to overload fields in struct page + */ +static __always_inline void *get_object_pointer(struct page *page) +{ + return (void *)page->index; +} + +static __always_inline void set_object_pointer(struct page *page, + void *object) +{ + page->index = (unsigned long)object; +} + +static __always_inline struct slab *get_slab(struct page *page) +{ + return (struct slab *)page->mapping; +} + +static __always_inline void set_slab(struct page *page, struct slab *s) +{ + page->mapping = (void *)s; +} + +static __always_inline int *object_counter(struct page *page) +{ + return (int *)&page->_mapcount; +} + +static __always_inline void inc_object_counter(struct page *page) +{ + (*object_counter(page))++; +} + +static __always_inline void dec_object_counter(struct page *page) +{ + (*object_counter(page))--; +} + +static __always_inline void set_object_counter(struct page *page, + int counter) +{ + (*object_counter(page))= counter; +} + +static __always_inline int get_object_counter(struct page *page) +{ + return (*object_counter(page)); +} + +/* + * Locking for each individual slab using the pagelock + */ +static __always_inline void slab_lock(struct page *page) +{ + bit_spin_lock(PG_locked, &page->flags); +} + +static __always_inline void slab_unlock(struct page *page) +{ + bit_spin_unlock(PG_locked, &page->flags); +} + +/* + * Management of partially allocated slabs + */ +static void __always_inline add_partial(struct slab *s, struct page *page) +{ + spin_lock(&s->list_lock); + s->nr_partial++; + list_add_tail(&page->lru, &s->partial); + spin_unlock(&s->list_lock); +} + +static void __always_inline remove_partial(struct slab *s, + struct page *page) +{ + spin_lock(&s->list_lock); + list_del(&page->lru); + s->nr_partial--; + spin_unlock(&s->list_lock); +} + +/* + * Get a page and remove it from the partial list + * Must hold list_lock + */ +static __always_inline int lock_and_del_slab(struct slab *s, + struct page *page) +{ + if (bit_spin_trylock(PG_locked, &page->flags)) { + list_del(&page->lru); + s->nr_partial--; + return 1; + } + return 0; +} + +struct page *numa_search(struct slab *s, int node) +{ +#ifdef CONFIG_NUMA + struct list_head *h; + struct page *page; + + /* + * Search for slab on the right node + */ + + if (node == -1) + node = numa_node_id(); + + list_for_each(h, &s->partial) { + page = container_of(h, struct page, lru); + + if (likely(page_to_nid(page) == node) && + lock_and_del_slab(s, page)) + return page; + } +#endif + return NULL; +} + +/* + * Get a partial page, lock it and return it. + */ +static struct page *get_partial(struct slab *s, int node) +{ + struct page *page; + struct list_head *h; + + spin_lock(&s->list_lock); + + page = numa_search(s, node); + if (page) + goto out; +#ifdef CONFIG_NUMA + if (node >= 0) + goto fail; +#endif + + list_for_each(h, &s->partial) { + page = container_of(h, struct page, lru); + + if (likely(lock_and_del_slab(s, page))) + goto out; + } +fail: + page = NULL; +out: + spin_unlock(&s->list_lock); + return page; +} + +/* + * Debugging checks + */ +static void check_slab(struct page *page) +{ +#ifdef SLABIFIER_DEBUG + if (!PageSlab(page)) { + printk(KERN_CRIT "Not a valid slab page @%p flags=%lx" + " mapping=%p count=%d \n", + page, page->flags, page->mapping, page_count(page)); + BUG(); + } +#endif +} + +static void check_active_slab(struct page *page) +{ +#ifdef SLABIFIER_DEBUG + if (!PageActive(page)) { + printk(KERN_CRIT "Not an active slab page @%p flags=%lx" + " mapping=%p count=%d \n", + page, page->flags, page->mapping, page_count(page)); + BUG(); + } +#endif +} + +static int check_valid_pointer(struct slab *s, struct page *page, + void *object, void *origin) +{ +#ifdef SLABIFIER_DEBUG + void *base = page_address(page); + + if (object < base || object >= base + s->objects * s->size) { + printk(KERN_CRIT "slab %s size %d: pointer %p->%p\nnot in" + " range (%p-%p) in page %p\n", s->sc.name, s->size, + origin, object, base, base + s->objects * s->size, + page); + return 0; + } + + if ((object - base) % s->size) { + printk(KERN_CRIT "slab %s size %d: pointer %p->%p\n" + "does not properly point" + "to an object in page %p\n", + s->sc.name, s->size, origin, object, page); + return 0; + } +#endif + return 1; +} + +/* + * Determine if a certain object on a page is on the freelist and + * therefore free. Must hold the slab lock for active slabs to + * guarantee that the chains are consistent. + */ +static int on_freelist(struct slab *s, struct page *page, void *search) +{ + int nr = 0; + void **object = get_object_pointer(page); + void *origin = &page->lru; + + if (s->objects == 1) + return 0; + + check_slab(page); + + while (object && nr <= s->objects) { + if (object == search) + return 1; + if (!check_valid_pointer(s, page, object, origin)) + goto try_recover; + origin = object; + object = object[s->offset]; + nr++; + } + + if (get_object_counter(page) != s->objects - nr) { + printk(KERN_CRIT "slab %s: page %p wrong object count." + " counter is %d but counted were %d\n", + s->sc.name, page, get_object_counter(page), + s->objects - nr); +try_recover: + printk(KERN_CRIT "****** Trying to continue by marking " + "all objects used (memory leak!)\n"); + set_object_counter(page, s->objects); + set_object_pointer(page, NULL); + } + return 0; +} + +void check_free_chain(struct slab *s, struct page *page) +{ +#ifdef SLABIFIER_DEBUG + on_freelist(s, page, NULL); +#endif +} + +/* + * Operations on slabs + */ +static void discard_slab(struct slab *s, struct page *page) +{ + DBUG_ON(PageActive(page)); + DBUG_ON(PageLocked(page)); + atomic_long_dec(&s->nr_slabs); + + /* Restore page state */ + page->mapping = NULL; + reset_page_mapcount(page); + __ClearPageSlab(page); + + s->sc.page_alloc->free(s->sc.page_alloc, page, s->sc.order); + sub_zone_page_state(page_zone(page), NR_SLAB, 1 << s->sc.order); +} + +/* + * Allocate a new slab and prepare an empty freelist and the basic struct + * page settings. + */ +static struct page *new_slab(struct slab *s, gfp_t flags, int node) +{ + struct page *page; + + page = s->sc.page_alloc->allocate(s->sc.page_alloc, s->sc.order, + flags, node < 0 ? s->sc.node : node); + if (!page) + return NULL; + + set_slab(page, s); + __SetPageSlab(page); + add_zone_page_state(page_zone(page), NR_SLAB, 1 << s->sc.order); + atomic_long_inc(&s->nr_slabs); + return page; +} + +/* + * Move a page back to the lists. + * + * Must be called with the slab lock held. + * On exit the slab lock will have been dropped. + */ +static void __always_inline putback_slab(struct slab *s, struct page *page) +{ + int inuse; + + inuse = get_object_counter(page); + + if (inuse) { + if (inuse < s->objects) + add_partial(s, page); + slab_unlock(page); + } else { + slab_unlock(page); + discard_slab(s, page); + } +} + +static void deactivate_slab(struct slab *s, struct page *page, int cpu) +{ + s->active[cpu] = NULL; + smp_wmb(); + ClearPageActive(page); + ClearPageReferenced(page); + + putback_slab(s, page); +} + +/* + * Acquire the slab lock from the active array. If there is no active + * slab for this processor then return NULL; + */ +static __always_inline struct page *get_and_lock_active(struct slab *s, + int cpu) +{ + struct page *page; + +redo: + page = s->active[cpu]; + if (unlikely(!page)) + return NULL; + slab_lock(page); + if (unlikely(s->active[cpu] != page)) { + slab_unlock(page); + goto redo; + } + check_active_slab(page); + check_free_chain(s, page); + return page; +} + +/* + * Flush an active slab back to the lists. + */ +static void flush_active(struct slab *s, int cpu) +{ + struct page *page; + unsigned long flags; + + local_irq_save(flags); + page = get_and_lock_active(s, cpu); + if (likely(page)) + deactivate_slab(s, page, cpu); + local_irq_restore(flags); +} + +#ifdef CONFIG_SMP +/* + * Flush per cpu slabs if they are not in use. + */ +void flusher(void *d) +{ + struct slab *s = d; + int cpu = smp_processor_id(); + struct page *page; + int nr_active = 0; + + for_each_online_cpu(cpu) { + + page = s->active[cpu]; + if (!page) + continue; + + if (PageReferenced(page)) { + ClearPageReferenced(page); + nr_active++; + } else + flush_active(s, cpu); + } + if (nr_active) + schedule_delayed_work(&s->flush, 10 * HZ); + else + s->flusher_active = 0; +} + +static void drain_all(struct slab *s) +{ + int cpu; + + if (s->flusher_active) { + cancel_delayed_work(&s->flush); + for_each_possible_cpu(cpu) + flush_active(s, cpu); + s->flusher_active = 0; + } +} +#else +static void drain_all(struct slab *s) +{ + flush_active(s, 0); +} +#endif + +/* + * slab_create produces objects aligned at size and the first object + * is placed at offset 0 in the slab (We have no metainformation on the + * slab, all slabs are in essence off slab). + * + * In order to get the desired alignment one just needs to align the + * size. + * + * Notice that the allocation order determines the sizes of the per cpu + * caches. Each processor has always one slab available for allocations. + * Increasing the allocation order reduces the number of times that slabs + * must be moved on and off the partial lists and therefore may influence + * locking overhead. + * + * The offset is used to relocate the free list link in each object. It is + * therefore possible to move the free list link behind the object. This + * is necessary for RCU to work properly and also useful for debugging. + * However no freelists are necessary if there is only one element per + * slab. + */ +static struct slab_cache *slab_create(struct slab_control *x, + const struct slab_cache *sc) +{ + struct slab *s = (void *)x; + int cpu; + + /* Verify that the generic structure is big enough for our data */ + BUG_ON(sizeof(struct slab_control) < sizeof(struct slab)); + + memcpy(&x->sc, sc, sizeof(struct slab_cache)); + + s->size = ALIGN(sc->size, sizeof(void *)); + + if (sc->offset > s->size - sizeof(void *) || + (sc->offset % sizeof(void*))) + return NULL; + + s->offset = sc->offset / sizeof(void *); + s->objects = (PAGE_SIZE << sc->order) / s->size; + atomic_long_set(&s->nr_slabs, 0); + s->nr_partial = 0; +#ifdef CONFIG_SMP + s->flusher_active = 0; + INIT_WORK(&s->flush, &flusher, s); +#endif + if (!s->objects) + return NULL; + + INIT_LIST_HEAD(&s->partial); + + atomic_set(&s->refcount, 1); + spin_lock_init(&s->list_lock); + + for_each_possible_cpu(cpu) + s->active[cpu] = NULL; + return &s->sc; +} + +/* + * Reload the per cpu slab + * + * If we have reloaded successfully then we exit with holding the slab lock + * and return the pointer to the new page. + * + * Return NULL if we cannot reload. + */ +static struct page *reload(struct slab *s, unsigned long cpu, gfp_t flags, + int node) +{ + void *p, *start, *end; + void **last; + struct page *page; + +redo: + /* Racy check. If we mistakenly see no partial slabs then we just + * expand the partial list. If we mistakenly try to get a partial + * slab then get_partials will return NULL. + */ + if (s->nr_partial) { + page = get_partial(s, node); + if (page) + goto gotpage; + } + + if ((flags & __GFP_WAIT)) { + local_irq_enable(); + page = new_slab(s, flags, node); + local_irq_disable(); + } else + page = new_slab(s, flags, node); + + if (!page) + return NULL; + + start = page_address(page); + set_object_pointer(page, start); + + end = start + s->objects * s->size; + last = start; + for (p = start + s->size; p < end; p += s->size) { + last[s->offset] = p; + last = p; + } + last[s->offset] = NULL; + set_object_counter(page, 0); + slab_lock(page); + check_free_chain(s, page); + +gotpage: + /* + * Now we have a page that is isolated from the lists and locked, + */ + SetPageActive(page); + ClearPageReferenced(page); + + if (cmpxchg(&s->active[cpu], NULL, page) != NULL) { + + ClearPageActive(page); + add_partial(s, page); + slab_unlock(page); + + page = get_and_lock_active(s, cpu); + if (page) + return page; + goto redo; + } + + check_free_chain(s, page); + +#ifdef CONFIG_SMP + if (keventd_up() && !s->flusher_active) { + s->flusher_active = 1; + schedule_delayed_work(&s->flush, 10 * HZ); + } +#endif + + return page; +} + +static __always_inline void *__slab_alloc(struct slab_cache *sc, + gfp_t gfpflags, int node) +{ + struct slab *s = (void *)sc; + struct page *page; + void **object; + void *next_object; + unsigned long flags; + int cpu; + + if (unlikely(s->objects == 1)) { + struct page *page = new_slab(s, gfpflags, node); + + if (page) + return page_address(page); + else + return NULL; + } + + local_irq_save(flags); + cpu = smp_processor_id(); + page = get_and_lock_active(s, cpu); + if (unlikely(!page)) + goto load; + + while (unlikely(!get_object_pointer(page) || + (node > 0 && page_to_nid(page) != node))) { + + deactivate_slab(s, page, cpu); +load: + page = reload(s, cpu, gfpflags, node); + if (unlikely(!page)) { + local_irq_restore(flags); + return NULL; + } + } + + inc_object_counter(page); + object = get_object_pointer(page); + next_object = object[s->offset]; + set_object_pointer(page, next_object); + check_free_chain(s, page); + SetPageReferenced(page); + slab_unlock(page); + local_irq_restore(flags); + return object; +} + +static void *slab_alloc(struct slab_cache *sc, gfp_t gfpflags) +{ + return __slab_alloc(sc, gfpflags, -1); +} + +static void *slab_alloc_node(struct slab_cache *sc, gfp_t gfpflags, + int node) +{ + return __slab_alloc(sc, gfpflags, node); +} + +/* Figure out on which slab object the object resides */ +static __always_inline struct page *get_object_page(const void *x) +{ + struct page * page = virt_to_page(x); + + if (unlikely(PageCompound(page))) + page = (struct page *)page_private(page); + + if (!PageSlab(page)) + return NULL; + + return page; +} + +static void slab_free(struct slab_cache *sc, const void *x) +{ + struct slab *s = (void *)sc; + struct page * page; + void *prior; + void **object = (void *)x; + unsigned long flags; + + if (!object) + return; + + page = get_object_page(object); + if (unlikely(!page)) { + printk(KERN_CRIT "slab_free %s size %d: attempt to free object" + "(%p) outside of slab.\n", s->sc.name, s->size, object); + goto dumpret; + } + + if (!s) { + s = get_slab(page); + + if (unlikely(!s)) { + printk(KERN_CRIT + "slab_free : no slab(NULL) for object %p.\n", + object); + goto dumpret; + } + } else + if (unlikely(s != get_slab(page))) { + printk(KERN_CRIT "slab_free %s: object at %p" + " belongs to slab %p\n", + s->sc.name, object, get_slab(page)); + dump_stack(); + s = get_slab(page); + } + + if (unlikely(!check_valid_pointer(s, page, object, NULL))) { +dumpret: + dump_stack(); + printk(KERN_CRIT "***** Trying to continue by not" + "freeing object.\n"); + return; + } + + if (unlikely(s->objects == 1)) { + discard_slab(s, page); + return; + } + + local_irq_save(flags); + slab_lock(page); + +#ifdef SLABIFIER_DEBUG + if (on_freelist(s, page, object)) { + printk(KERN_CRIT "slab_free %s: object %p already free.\n", + s->sc.name, object); + dump_stack(); + goto out_unlock; + } +#endif + + prior = get_object_pointer(page); + object[s->offset] = prior; + + set_object_pointer(page, object); + dec_object_counter(page); + + if (unlikely(PageActive(page))) + goto out_unlock; + + if (unlikely(get_object_counter(page) == 0)) { + if (s->objects > 1) + remove_partial(s, page); + check_free_chain(s, page); + slab_unlock(page); + discard_slab(s, page); + goto out; + } + + if (unlikely(!prior)) + /* + * Page was fully used before. It will only have one free + * object now. So move to the partial list. + */ + add_partial(s, page); + +out_unlock: + slab_unlock(page); +out: + local_irq_restore(flags); +} + +/* + * Check if a given pointer is valid + */ +static int slab_pointer_valid(struct slab_cache *sc, const void *object) +{ + struct slab *s = (void *)sc; + struct page * page; + void *addr; + + page = get_object_page(object); + + if (!page || s != get_slab(page)) + return 0; + + addr = page_address(page); + if (object < addr || object >= addr + s->objects * s->size) + return 0; + + if ((object - addr) & s->size) + return 0; + + return 1; +} + +/* + * Determine the size of a slab object + */ +static unsigned long slab_object_size(struct slab_cache *sc, + const void *object) +{ + struct page *page; + struct slab *s; + + page = get_object_page(object); + if (page) { + s = get_slab(page); + BUG_ON(sc && s != (void *)sc); + if (s) + return s->size; + } + BUG(); + return 0; /* Satisfy compiler */ +} + +/* + * Move slab objects in a given slab by calling the move_objects function. + * + * Must be called with the slab lock held but will drop and reacquire the + * slab lock. + */ +static int move_slab_objects(struct slab *s, struct page *page, + int (*move_objects)(struct slab_cache *, void *)) +{ + int unfreeable = 0; + void *addr = page_address(page); + + while (get_object_counter(page) - unfreeable > 0) { + void *p; + + for (p = addr; p < addr + s->objects; p+= s->size) { + if (!on_freelist(s, page, p)) { + /* + * Drop the lock here to allow the + * move_object function to do things + * with the slab_cache and maybe this + * page. + */ + slab_unlock(page); + local_irq_enable(); + if (move_objects((struct slab_cache *)s, p)) + slab_free(&s->sc, p); + else + unfreeable++; + local_irq_disable(); + slab_lock(page); + } + } + } + return unfreeable; +} + +/* + * Shrinking drops all the active per cpu slabs and also reaps all empty + * slabs off the partial list. Returns the number of slabs freed. + * + * If a move_object function is specified then the partial list is going + * to be compacted by calling the function on all slabs. + * The move_object function will be called for each objects in partially + * allocated slabs. move_object() needs to perform a new allocation for + * the object and move the contents of the object to the new location. + * If move_object() returns 1 for success then the object is going to be + * removed. If 0 then the object cannot be freed at all. As a result the + * slab containing the object will also not be freeable. + * + * Returns the number of slabs freed. + */ +static int slab_shrink(struct slab_cache *sc, + int (*move_object)(struct slab_cache *, void *)) +{ + struct slab *s = (void *)sc; + unsigned long flags; + int slabs_freed = 0; + int i; + + drain_all(s); + + local_irq_save(flags); + for(i = 0; s->nr_partial > 1 && i < s->nr_partial - 1; i++ ) { + struct page * page; + + page = get_partial(s, -1); + if (!page) + break; + + /* + * Pin page so that slab_free will not free even if we + * drop the slab lock. + */ + SetPageActive(page); + + if (get_object_counter(page) < s->objects && move_object) + if (move_slab_objects(s, + page, move_object) == 0) + slabs_freed++; + + /* + * This will put the slab on the front of the partial + * list, the used list or free it. + */ + putback_slab(s, page); + } + local_irq_restore(flags); + + return slabs_freed; + +} + +static struct slab_cache *slab_dup(struct slab_cache *sc) +{ + struct slab *s = (void *)sc; + + atomic_inc(&s->refcount); + return &s->sc; +} + +static void free_list(struct slab *s, struct list_head *list) +{ + unsigned long flags; + + spin_lock_irqsave(&s->list_lock, flags); + while (!list_empty(list)) + discard_slab(s, lru_to_last_page(list)); + + spin_unlock_irqrestore(&s->list_lock, flags); +} + +static int slab_destroy(struct slab_cache *sc) +{ + struct slab * s = (void *)sc; + + if (!atomic_dec_and_test(&s->refcount)) + return 0; + + drain_all(s); + free_list(s, &s->partial); + + if (atomic_long_read(&s->nr_slabs)) + return 1; + + /* Just to make sure that no one uses this again */ + s->size = 0; + return 0; + +} + +static unsigned long count_objects(struct slab *s, struct list_head *list) +{ + int count = 0; + struct list_head *h; + unsigned long flags; + + spin_lock_irqsave(&s->list_lock, flags); + list_for_each(h, list) { + struct page *page = lru_to_first_page(h); + + count += get_object_counter(page); + } + spin_unlock_irqrestore(&s->list_lock, flags); + return count; +} + +static unsigned long slab_objects(struct slab_cache *sc, + unsigned long *p_total, unsigned long *p_active, + unsigned long *p_partial) +{ + struct slab *s = (void *)sc; + int partial; + int active = 0; /* Active slabs */ + int nr_active = 0; /* Objects in active slabs */ + int cpu; + int nr_slabs = atomic_read(&s->nr_slabs); + + for_each_possible_cpu(cpu) { + struct page *page = s->active[cpu]; + + if (s->active[cpu]) { + nr_active++; + active += get_object_counter(page); + } + } + + partial = count_objects(s, &s->partial); + + if (p_partial) + *p_partial = s->nr_partial; + + if (p_active) + *p_active = nr_active; + + if (p_total) + *p_total = nr_slabs; + + return partial + active + + (nr_slabs - s->nr_partial - nr_active) * s->objects; +} + +const struct slab_allocator slabifier_allocator = { + .name = "Slabifier", + .create = slab_create, + .alloc = slab_alloc, + .alloc_node = slab_alloc_node, + .free = slab_free, + .valid_pointer = slab_pointer_valid, + .object_size = slab_object_size, + .get_objects = slab_objects, + .shrink = slab_shrink, + .dup = slab_dup, + .destroy = slab_destroy, + .destructor = null_slab_allocator_destructor, +}; +EXPORT_SYMBOL(slabifier_allocator); Index: linux-2.6.18-rc4-mm2/mm/Makefile =================================================================== --- linux-2.6.18-rc4-mm2.orig/mm/Makefile 2006-08-25 18:38:28.653471440 -0700 +++ linux-2.6.18-rc4-mm2/mm/Makefile 2006-08-25 19:08:07.053500816 -0700 @@ -25,4 +25,4 @@ obj-$(CONFIG_MEMORY_HOTPLUG) += memory_h obj-$(CONFIG_FS_XIP) += filemap_xip.o obj-$(CONFIG_MIGRATION) += migrate.o obj-$(CONFIG_SMP) += allocpercpu.o -obj-$(CONFIG_MODULAR_SLAB) += allocator.o +obj-$(CONFIG_MODULAR_SLAB) += allocator.o slabifier.o Index: linux-2.6.18-rc4-mm2/mm/memory.c =================================================================== --- linux-2.6.18-rc4-mm2.orig/mm/memory.c 2006-08-23 12:37:01.832149961 -0700 +++ linux-2.6.18-rc4-mm2/mm/memory.c 2006-08-25 19:08:07.055453820 -0700 @@ -2485,7 +2485,7 @@ int make_pages_present(unsigned long add /* * Map a vmalloc()-space virtual address to the physical page. */ -struct page * vmalloc_to_page(void * vmalloc_addr) +struct page * vmalloc_to_page(const void * vmalloc_addr) { unsigned long addr = (unsigned long) vmalloc_addr; struct page *page = NULL; Index: linux-2.6.18-rc4-mm2/include/linux/mm.h =================================================================== --- linux-2.6.18-rc4-mm2.orig/include/linux/mm.h 2006-08-23 12:37:01.087078845 -0700 +++ linux-2.6.18-rc4-mm2/include/linux/mm.h 2006-08-25 19:08:07.056430322 -0700 @@ -1152,7 +1152,7 @@ static inline unsigned long vma_pages(st pgprot_t vm_get_page_prot(unsigned long vm_flags); struct vm_area_struct *find_extend_vma(struct mm_struct *, unsigned long addr); -struct page *vmalloc_to_page(void *addr); +struct page *vmalloc_to_page(const void *addr); unsigned long vmalloc_to_pfn(void *addr); int remap_pfn_range(struct vm_area_struct *, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t);