From: Christoph Lameter Signed-off-by: Christoph Lameter Signed-off-by: Andrew Morton --- include/linux/mm.h | 6 ++++++ mm/slab.c | 9 ++++----- mm/slub.c | 10 ++++------ 3 files changed, 14 insertions(+), 11 deletions(-) diff -puN include/linux/mm.h~add-virt_to_head_page-and-consolidate-code-in-slab-and-slub include/linux/mm.h --- a/include/linux/mm.h~add-virt_to_head_page-and-consolidate-code-in-slab-and-slub +++ a/include/linux/mm.h @@ -313,6 +313,12 @@ static inline void get_page(struct page atomic_inc(&page->_count); } +static inline struct page *virt_to_head_page(const void *x) +{ + struct page *page = virt_to_page(x); + return compound_head(page); +} + /* * Setup the page count before being freed into the page allocator for * the first time (boot or memory hotplug) diff -puN mm/slab.c~add-virt_to_head_page-and-consolidate-code-in-slab-and-slub mm/slab.c --- a/mm/slab.c~add-virt_to_head_page-and-consolidate-code-in-slab-and-slub +++ a/mm/slab.c @@ -614,20 +614,19 @@ static inline void page_set_slab(struct static inline struct slab *page_get_slab(struct page *page) { - page = compound_head(page); BUG_ON(!PageSlab(page)); return (struct slab *)page->lru.prev; } static inline struct kmem_cache *virt_to_cache(const void *obj) { - struct page *page = virt_to_page(obj); + struct page *page = virt_to_head_page(obj); return page_get_cache(page); } static inline struct slab *virt_to_slab(const void *obj) { - struct page *page = virt_to_page(obj); + struct page *page = virt_to_head_page(obj); return page_get_slab(page); } @@ -2876,7 +2875,7 @@ static void *cache_free_debugcheck(struc objp -= obj_offset(cachep); kfree_debugcheck(objp); - page = virt_to_page(objp); + page = virt_to_head_page(objp); slabp = page_get_slab(page); @@ -3100,7 +3099,7 @@ static void *cache_alloc_debugcheck_afte struct slab *slabp; unsigned objnr; - slabp = page_get_slab(virt_to_page(objp)); + slabp = page_get_slab(virt_to_head_page(objp)); objnr = (unsigned)(objp - slabp->s_mem) / cachep->buffer_size; slab_bufctl(slabp)[objnr] = BUFCTL_ACTIVE; } diff -puN mm/slub.c~add-virt_to_head_page-and-consolidate-code-in-slab-and-slub mm/slub.c --- a/mm/slub.c~add-virt_to_head_page-and-consolidate-code-in-slab-and-slub +++ a/mm/slub.c @@ -1348,9 +1348,7 @@ void kmem_cache_free(struct kmem_cache * { struct page * page; - page = virt_to_page(x); - - page = compound_head(page); + page = virt_to_head_page(x); if (unlikely(PageError(page) && (s->flags & SLAB_STORE_USER))) set_tracking(s, x, TRACK_FREE); @@ -1361,7 +1359,7 @@ EXPORT_SYMBOL(kmem_cache_free); /* Figure out on which slab object the object resides */ static struct page *get_object_page(const void *x) { - struct page *page = compound_head(virt_to_page(x)); + struct page *page = virt_to_head_page(x); if (!PageSlab(page)) return NULL; @@ -2116,7 +2114,7 @@ void kfree(const void *x) if (!x) return; - page = compound_head(virt_to_page(x)); + page = virt_to_head_page(x); s = page->slab; @@ -2152,7 +2150,7 @@ void *krealloc(const void *p, size_t new return NULL; } - page = compound_head(virt_to_page(p)); + page = virt_to_head_page(p); new_cache = get_slab(new_size, flags); _