From: David Gibson Quite a long time back, prepare_hugepage_range() replaced is_aligned_hugepage_range() as the callback from mm/mmap.c to arch code to verify if an address range is suitable for a hugepage mapping. is_aligned_hugepage_range() stuck around, but only to implement prepare_hugepage_range() on archs which didn't implement their own. Most archs (everything except ia64 and powerpc) used the same implementation of is_aligned_hugepage_range(). On powerpc, which implements its own prepare_hugepage_range(), the custom version was never used. In addition, "is_aligned_hugepage_range()" was a bad name, because it suggests it returns true iff the given range is a good hugepage range, whereas in fact it returns 0-or-error (so the sense is reversed). This patch cleans up by abolishing is_aligned_hugepage_range(). Instead prepare_hugepage_range() is defined directly. Most archs use the default version, which simply checks the given region is aligned to the size of a hugepage. ia64 and powerpc define custom versions. The ia64 one simply checks that the range is in the correct address space region in addition to being suitably aligned. The powerpc version (just as previously) checks for suitable addresses, and if necessary performs low-level MMU frobbing to set up new areas for use by hugepages. No libhugetlbfs testsuite regressions on ppc64 (POWER5 LPAR). Signed-off-by: David Gibson Signed-off-by: Zhang Yanmin Cc: "David S. Miller" Cc: Benjamin Herrenschmidt Cc: Paul Mackerras Cc: William Lee Irwin III Signed-off-by: Andrew Morton --- arch/i386/mm/hugetlbpage.c | 12 ------------ arch/ia64/mm/hugetlbpage.c | 5 +++-- arch/powerpc/mm/hugetlbpage.c | 15 --------------- arch/sh/mm/hugetlbpage.c | 12 ------------ arch/sh64/mm/hugetlbpage.c | 12 ------------ arch/sparc64/mm/hugetlbpage.c | 12 ------------ include/asm-ia64/page.h | 1 + include/linux/hugetlb.h | 16 ++++++++++++---- 8 files changed, 16 insertions(+), 69 deletions(-) diff -puN arch/i386/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup arch/i386/mm/hugetlbpage.c --- devel/arch/i386/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/arch/i386/mm/hugetlbpage.c 2006-03-05 23:14:27.000000000 -0800 @@ -48,18 +48,6 @@ pte_t *huge_pte_offset(struct mm_struct return (pte_t *) pmd; } -/* - * This function checks for proper alignment of input addr and len parameters. - */ -int is_aligned_hugepage_range(unsigned long addr, unsigned long len) -{ - if (len & ~HPAGE_MASK) - return -EINVAL; - if (addr & ~HPAGE_MASK) - return -EINVAL; - return 0; -} - #if 0 /* This is just for testing */ struct page * follow_huge_addr(struct mm_struct *mm, unsigned long address, int write) diff -puN arch/ia64/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup arch/ia64/mm/hugetlbpage.c --- devel/arch/ia64/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/arch/ia64/mm/hugetlbpage.c 2006-03-05 23:14:27.000000000 -0800 @@ -68,9 +68,10 @@ huge_pte_offset (struct mm_struct *mm, u #define mk_pte_huge(entry) { pte_val(entry) |= _PAGE_P; } /* - * This function checks for proper alignment of input addr and len parameters. + * Don't actually need to do any preparation, but need to make sure + * the address is in the right region. */ -int is_aligned_hugepage_range(unsigned long addr, unsigned long len) +int prepare_hugepage_range(unsigned long addr, unsigned long len) { if (len & ~HPAGE_MASK) return -EINVAL; diff -puN arch/powerpc/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup arch/powerpc/mm/hugetlbpage.c --- devel/arch/powerpc/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/arch/powerpc/mm/hugetlbpage.c 2006-03-05 23:14:27.000000000 -0800 @@ -133,21 +133,6 @@ pte_t huge_ptep_get_and_clear(struct mm_ return __pte(old); } -/* - * This function checks for proper alignment of input addr and len parameters. - */ -int is_aligned_hugepage_range(unsigned long addr, unsigned long len) -{ - if (len & ~HPAGE_MASK) - return -EINVAL; - if (addr & ~HPAGE_MASK) - return -EINVAL; - if (! (within_hugepage_low_range(addr, len) - || within_hugepage_high_range(addr, len)) ) - return -EINVAL; - return 0; -} - struct slb_flush_info { struct mm_struct *mm; u16 newareas; diff -puN arch/sh64/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup arch/sh64/mm/hugetlbpage.c --- devel/arch/sh64/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/arch/sh64/mm/hugetlbpage.c 2006-03-05 23:14:27.000000000 -0800 @@ -84,18 +84,6 @@ pte_t huge_ptep_get_and_clear(struct mm_ return entry; } -/* - * This function checks for proper alignment of input addr and len parameters. - */ -int is_aligned_hugepage_range(unsigned long addr, unsigned long len) -{ - if (len & ~HPAGE_MASK) - return -EINVAL; - if (addr & ~HPAGE_MASK) - return -EINVAL; - return 0; -} - struct page *follow_huge_addr(struct mm_struct *mm, unsigned long address, int write) { diff -puN arch/sh/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup arch/sh/mm/hugetlbpage.c --- devel/arch/sh/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/arch/sh/mm/hugetlbpage.c 2006-03-05 23:14:27.000000000 -0800 @@ -84,18 +84,6 @@ pte_t huge_ptep_get_and_clear(struct mm_ return entry; } -/* - * This function checks for proper alignment of input addr and len parameters. - */ -int is_aligned_hugepage_range(unsigned long addr, unsigned long len) -{ - if (len & ~HPAGE_MASK) - return -EINVAL; - if (addr & ~HPAGE_MASK) - return -EINVAL; - return 0; -} - struct page *follow_huge_addr(struct mm_struct *mm, unsigned long address, int write) { diff -puN arch/sparc64/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup arch/sparc64/mm/hugetlbpage.c --- devel/arch/sparc64/mm/hugetlbpage.c~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/arch/sparc64/mm/hugetlbpage.c 2006-03-05 23:14:27.000000000 -0800 @@ -92,18 +92,6 @@ pte_t huge_ptep_get_and_clear(struct mm_ return entry; } -/* - * This function checks for proper alignment of input addr and len parameters. - */ -int is_aligned_hugepage_range(unsigned long addr, unsigned long len) -{ - if (len & ~HPAGE_MASK) - return -EINVAL; - if (addr & ~HPAGE_MASK) - return -EINVAL; - return 0; -} - struct page *follow_huge_addr(struct mm_struct *mm, unsigned long address, int write) { diff -puN include/asm-ia64/page.h~hugepage-is_aligned_hugepage_range-cleanup include/asm-ia64/page.h --- devel/include/asm-ia64/page.h~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/include/asm-ia64/page.h 2006-03-05 23:14:27.000000000 -0800 @@ -57,6 +57,7 @@ # define HAVE_ARCH_HUGETLB_UNMAPPED_AREA # define ARCH_HAS_HUGEPAGE_ONLY_RANGE +# define ARCH_HAS_PREPARE_HUGEPAGE_RANGE # define ARCH_HAS_HUGETLB_FREE_PGD_RANGE #endif /* CONFIG_HUGETLB_PAGE */ diff -puN include/linux/hugetlb.h~hugepage-is_aligned_hugepage_range-cleanup include/linux/hugetlb.h --- devel/include/linux/hugetlb.h~hugepage-is_aligned_hugepage_range-cleanup 2006-03-05 23:14:27.000000000 -0800 +++ devel-akpm/include/linux/hugetlb.h 2006-03-05 23:14:27.000000000 -0800 @@ -36,7 +36,6 @@ struct page *follow_huge_addr(struct mm_ int write); struct page *follow_huge_pmd(struct mm_struct *mm, unsigned long address, pmd_t *pmd, int write); -int is_aligned_hugepage_range(unsigned long addr, unsigned long len); int pmd_huge(pmd_t pmd); void hugetlb_change_protection(struct vm_area_struct *vma, unsigned long address, unsigned long end, pgprot_t newprot); @@ -54,8 +53,18 @@ void hugetlb_free_pgd_range(struct mmu_g #endif #ifndef ARCH_HAS_PREPARE_HUGEPAGE_RANGE -#define prepare_hugepage_range(addr, len) \ - is_aligned_hugepage_range(addr, len) +/* + * If the arch doesn't supply something else, assume that hugepage + * size aligned regions are ok without further preparation. + */ +static inline int prepare_hugepage_range(unsigned long addr, unsigned long len) +{ + if (len & ~HPAGE_MASK) + return -EINVAL; + if (addr & ~HPAGE_MASK) + return -EINVAL; + return 0; +} #else int prepare_hugepage_range(unsigned long addr, unsigned long len); #endif @@ -95,7 +104,6 @@ static inline unsigned long hugetlb_tota #define hugetlb_report_meminfo(buf) 0 #define hugetlb_report_node_meminfo(n, buf) 0 #define follow_huge_pmd(mm, addr, pmd, write) NULL -#define is_aligned_hugepage_range(addr, len) 0 #define prepare_hugepage_range(addr, len) (-EINVAL) #define pmd_huge(x) 0 #define is_hugepage_only_range(mm, addr, len) 0 _