Add remove_from_swap remove_from_swap() allows the restoration of the pte entries that existed before page migration occurred for anonymous pages by walking the reverse maps. This reduces swap use and establishes regular pte's without the need for page faults. V3->V4: - Add new function remove_vma_swap in swapfile.c to encapsulate the functionality needed instead of exporting unuse_vma. - Add #ifdef CONFIG_MIGRATION Signed-off-by: Christoph Lameter Index: linux-2.6.15-rc1-mm2/include/linux/swap.h =================================================================== --- linux-2.6.15-rc1-mm2.orig/include/linux/swap.h 2005-11-18 11:52:08.000000000 -0800 +++ linux-2.6.15-rc1-mm2/include/linux/swap.h 2005-11-18 12:28:37.000000000 -0800 @@ -264,6 +264,9 @@ extern int remove_exclusive_swap_page(st struct backing_dev_info; extern spinlock_t swap_lock; +#ifdef CONFIG_MIGRATION +extern int remove_vma_swap(struct vm_area_struct *vma, struct page *page); +#endif /* linux/mm/thrash.c */ extern struct mm_struct * swap_token_mm; Index: linux-2.6.15-rc1-mm2/mm/swapfile.c =================================================================== --- linux-2.6.15-rc1-mm2.orig/mm/swapfile.c 2005-11-18 11:52:08.000000000 -0800 +++ linux-2.6.15-rc1-mm2/mm/swapfile.c 2005-11-18 12:28:37.000000000 -0800 @@ -532,6 +532,16 @@ static int unuse_mm(struct mm_struct *mm return 0; } +#ifdef CONFIG_MIGRATION +int remove_vma_swap(struct vm_area_struct *vma, struct page *page) +{ + swp_entry_t entry = { .val = page_private(page) }; + + return unuse_vma(vma, entry, page); +} +#endif + + /* * Scan swap_map from current position to next entry still in use. * Recycle to start on reaching the end, returning 0 when empty. Index: linux-2.6.15-rc1-mm2/mm/rmap.c =================================================================== --- linux-2.6.15-rc1-mm2.orig/mm/rmap.c 2005-11-18 09:47:15.000000000 -0800 +++ linux-2.6.15-rc1-mm2/mm/rmap.c 2005-11-18 12:28:37.000000000 -0800 @@ -205,6 +205,28 @@ out: return anon_vma; } +#ifdef CONFIG_MIGRATION +/* + * Remove an anonymous page from swap replacing the swap pte's + * through real pte's pointing to valid pages. + */ +void remove_from_swap(struct page *page) +{ + struct anon_vma *anon_vma; + struct vm_area_struct *vma; + + if (!PageAnon(page)) + return; + + anon_vma = page_lock_anon_vma(page); + if (!anon_vma) + return; + + list_for_each_entry(vma, &anon_vma->head, anon_vma_node) + remove_vma_swap(vma, page); +} +#endif + /* * At what user virtual address is page expected in vma? */ Index: linux-2.6.15-rc1-mm2/include/linux/rmap.h =================================================================== --- linux-2.6.15-rc1-mm2.orig/include/linux/rmap.h 2005-11-11 17:43:36.000000000 -0800 +++ linux-2.6.15-rc1-mm2/include/linux/rmap.h 2005-11-18 12:28:37.000000000 -0800 @@ -91,6 +91,9 @@ static inline void page_dup_rmap(struct */ int page_referenced(struct page *, int is_locked, int ignore_token); int try_to_unmap(struct page *); +#ifdef CONFIG_MIGRATION +void remove_from_swap(struct page *page); +#endif /* * Called from mm/filemap_xip.c to unmap empty zero page Index: linux-2.6.15-rc1-mm2/mm/vmscan.c =================================================================== --- linux-2.6.15-rc1-mm2.orig/mm/vmscan.c 2005-11-18 12:28:17.000000000 -0800 +++ linux-2.6.15-rc1-mm2/mm/vmscan.c 2005-11-18 12:28:37.000000000 -0800 @@ -981,13 +981,15 @@ next: else if (rc) { /* Permanent failure */ + remove_from_swap(page); list_move(&page->lru, failed); nr_failed++; } else { - if (newpage) + if (newpage) { /* Successful migration. Return new page to LRU */ + remove_from_swap(newpage); move_to_lru(newpage); - + } list_move(&page->lru, moved); } }