From: Andrew Morton Cc: Christoph Lameter Signed-off-by: Andrew Morton --- mm/page_alloc.c | 2 +- mm/vmscan.c | 7 ++++--- 2 files changed, 5 insertions(+), 4 deletions(-) diff -puN mm/page_alloc.c~zone_reclaim-dynamic-slab-reclaim-tidy mm/page_alloc.c --- a/mm/page_alloc.c~zone_reclaim-dynamic-slab-reclaim-tidy +++ a/mm/page_alloc.c @@ -2401,7 +2401,7 @@ static void __meminit free_area_init_cor #ifdef CONFIG_NUMA zone->min_unmapped_pages = (realsize*sysctl_min_unmapped_ratio) / 100; - zone->min_slab_pages = (realsize*sysctl_min_slab_ratio) / 100; + zone->min_slab_pages = (realsize * sysctl_min_slab_ratio) / 100; #endif zone->name = zone_names[j]; spin_lock_init(&zone->lock); diff -puN mm/vmscan.c~zone_reclaim-dynamic-slab-reclaim-tidy mm/vmscan.c --- a/mm/vmscan.c~zone_reclaim-dynamic-slab-reclaim-tidy +++ a/mm/vmscan.c @@ -1600,14 +1600,15 @@ static int __zone_reclaim(struct zone *z * by the same nr_pages that we used for reclaiming unmapped * pages. * - * Note that shrink_slab will free memory on all zones and may take - * a long time. + * Note that shrink_slab will free memory on all zones and may + * take a long time. */ unsigned long limit = zone_page_state(zone, NR_SLAB_RECLAIMABLE) - nr_pages; while (shrink_slab(sc.nr_scanned, gfp_mask, order) && - zone_page_state(zone, NR_SLAB_RECLAIMABLE) > limit) ; + zone_page_state(zone, NR_SLAB_RECLAIMABLE) > limit) + ; } p->reclaim_state = NULL; _