is_aligned_hugepage_range() cleanup

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Andrew, please apply.

Quite a long time back, prepare_hugepage_range() replaced
is_aligned_hugepage_range() as the callback from mm/mmap.c to arch
code to verify is an address range is suitable for a hugepage mapping.
is_aligned_hugepage_range() stuck around, but only to implement
prepare_hugepage_range() on archs which didn't implement their own.

Most archs (everything except ia64 and powerpc) used the same
implementation of is_aligned_hugepage_range().  On powerpc, which
implements its own prepare_hugepage_range(), the custom version was
never used.

This patch cleans up the situation by abolishing
is_aligned_hugepage_range().  Instead prepare_hugepage_range() is
defined directly.  Most archs use the default version, which simple
checks the given region is aligned to the size of a hugepage.  ia64
and powerpc define custom versions.  The ia64 one simply checks that
the range is in the correct address space region in addition to being
suitably aligned.  The powerpc version (just as previously) checks for
suitable addresses, and if necessary performs low-level MMU frobbing
to set up new areas for use by hugepages.

Built and tested on POWER5 LPAR (powerpc), G5 (powerpc) and Pentium M
(i386).

Signed-off-by: David Gibson <[email protected]>

 arch/i386/mm/hugetlbpage.c    |   12 ------------
 arch/ia64/mm/hugetlbpage.c    |    5 +++--
 arch/powerpc/mm/hugetlbpage.c |   15 ---------------
 arch/sh/mm/hugetlbpage.c      |   12 ------------
 arch/sh64/mm/hugetlbpage.c    |   12 ------------
 arch/sparc64/mm/hugetlbpage.c |   12 ------------
 include/asm-ia64/page.h       |    1 +
 include/linux/hugetlb.h       |   16 ++++++++++++----
 8 files changed, 16 insertions(+), 69 deletions(-)

Index: working-2.6/arch/i386/mm/hugetlbpage.c
===================================================================
--- working-2.6.orig/arch/i386/mm/hugetlbpage.c	2005-10-25 11:59:52.000000000 +1000
+++ working-2.6/arch/i386/mm/hugetlbpage.c	2005-11-24 14:58:14.000000000 +1100
@@ -48,18 +48,6 @@ pte_t *huge_pte_offset(struct mm_struct 
 	return (pte_t *) pmd;
 }
 
-/*
- * This function checks for proper alignment of input addr and len parameters.
- */
-int is_aligned_hugepage_range(unsigned long addr, unsigned long len)
-{
-	if (len & ~HPAGE_MASK)
-		return -EINVAL;
-	if (addr & ~HPAGE_MASK)
-		return -EINVAL;
-	return 0;
-}
-
 #if 0	/* This is just for testing */
 struct page *
 follow_huge_addr(struct mm_struct *mm, unsigned long address, int write)
Index: working-2.6/arch/ia64/mm/hugetlbpage.c
===================================================================
--- working-2.6.orig/arch/ia64/mm/hugetlbpage.c	2005-10-25 11:59:52.000000000 +1000
+++ working-2.6/arch/ia64/mm/hugetlbpage.c	2005-11-24 14:58:14.000000000 +1100
@@ -68,9 +68,10 @@ huge_pte_offset (struct mm_struct *mm, u
 #define mk_pte_huge(entry) { pte_val(entry) |= _PAGE_P; }
 
 /*
- * This function checks for proper alignment of input addr and len parameters.
+ * Don't actually need to do any preparation, but need to make sure
+ * the address is in the right region.
  */
-int is_aligned_hugepage_range(unsigned long addr, unsigned long len)
+int prepare_hugepage_range(unsigned long addr, unsigned long len)
 {
 	if (len & ~HPAGE_MASK)
 		return -EINVAL;
Index: working-2.6/arch/powerpc/mm/hugetlbpage.c
===================================================================
--- working-2.6.orig/arch/powerpc/mm/hugetlbpage.c	2005-11-24 14:58:14.000000000 +1100
+++ working-2.6/arch/powerpc/mm/hugetlbpage.c	2005-11-24 14:58:14.000000000 +1100
@@ -133,21 +133,6 @@ pte_t huge_ptep_get_and_clear(struct mm_
 	return __pte(old);
 }
 
-/*
- * This function checks for proper alignment of input addr and len parameters.
- */
-int is_aligned_hugepage_range(unsigned long addr, unsigned long len)
-{
-	if (len & ~HPAGE_MASK)
-		return -EINVAL;
-	if (addr & ~HPAGE_MASK)
-		return -EINVAL;
-	if (! (within_hugepage_low_range(addr, len)
-	       || within_hugepage_high_range(addr, len)) )
-		return -EINVAL;
-	return 0;
-}
-
 static void flush_low_segments(void *parm)
 {
 	u16 areas = (unsigned long) parm;
Index: working-2.6/arch/sh/mm/hugetlbpage.c
===================================================================
--- working-2.6.orig/arch/sh/mm/hugetlbpage.c	2005-11-23 15:56:23.000000000 +1100
+++ working-2.6/arch/sh/mm/hugetlbpage.c	2005-11-24 14:58:14.000000000 +1100
@@ -84,18 +84,6 @@ pte_t huge_ptep_get_and_clear(struct mm_
 	return entry;
 }
 
-/*
- * This function checks for proper alignment of input addr and len parameters.
- */
-int is_aligned_hugepage_range(unsigned long addr, unsigned long len)
-{
-	if (len & ~HPAGE_MASK)
-		return -EINVAL;
-	if (addr & ~HPAGE_MASK)
-		return -EINVAL;
-	return 0;
-}
-
 struct page *follow_huge_addr(struct mm_struct *mm,
 			      unsigned long address, int write)
 {
Index: working-2.6/arch/sh64/mm/hugetlbpage.c
===================================================================
--- working-2.6.orig/arch/sh64/mm/hugetlbpage.c	2005-11-23 15:56:23.000000000 +1100
+++ working-2.6/arch/sh64/mm/hugetlbpage.c	2005-11-24 14:58:14.000000000 +1100
@@ -84,18 +84,6 @@ pte_t huge_ptep_get_and_clear(struct mm_
 	return entry;
 }
 
-/*
- * This function checks for proper alignment of input addr and len parameters.
- */
-int is_aligned_hugepage_range(unsigned long addr, unsigned long len)
-{
-	if (len & ~HPAGE_MASK)
-		return -EINVAL;
-	if (addr & ~HPAGE_MASK)
-		return -EINVAL;
-	return 0;
-}
-
 struct page *follow_huge_addr(struct mm_struct *mm,
 			      unsigned long address, int write)
 {
Index: working-2.6/arch/sparc64/mm/hugetlbpage.c
===================================================================
--- working-2.6.orig/arch/sparc64/mm/hugetlbpage.c	2005-10-25 11:59:53.000000000 +1000
+++ working-2.6/arch/sparc64/mm/hugetlbpage.c	2005-11-24 14:58:14.000000000 +1100
@@ -92,18 +92,6 @@ pte_t huge_ptep_get_and_clear(struct mm_
 	return entry;
 }
 
-/*
- * This function checks for proper alignment of input addr and len parameters.
- */
-int is_aligned_hugepage_range(unsigned long addr, unsigned long len)
-{
-	if (len & ~HPAGE_MASK)
-		return -EINVAL;
-	if (addr & ~HPAGE_MASK)
-		return -EINVAL;
-	return 0;
-}
-
 struct page *follow_huge_addr(struct mm_struct *mm,
 			      unsigned long address, int write)
 {
Index: working-2.6/include/asm-ia64/page.h
===================================================================
--- working-2.6.orig/include/asm-ia64/page.h	2005-11-23 15:56:35.000000000 +1100
+++ working-2.6/include/asm-ia64/page.h	2005-11-24 14:58:14.000000000 +1100
@@ -57,6 +57,7 @@
 
 # define HAVE_ARCH_HUGETLB_UNMAPPED_AREA
 # define ARCH_HAS_HUGEPAGE_ONLY_RANGE
+# define ARCH_HAS_PREPARE_HUGEPAGE_RANGE
 #endif /* CONFIG_HUGETLB_PAGE */
 
 #ifdef __ASSEMBLY__
Index: working-2.6/include/linux/hugetlb.h
===================================================================
--- working-2.6.orig/include/linux/hugetlb.h	2005-11-23 15:56:36.000000000 +1100
+++ working-2.6/include/linux/hugetlb.h	2005-11-24 14:58:14.000000000 +1100
@@ -39,7 +39,6 @@ struct page *follow_huge_addr(struct mm_
 			      int write);
 struct page *follow_huge_pmd(struct mm_struct *mm, unsigned long address,
 				pmd_t *pmd, int write);
-int is_aligned_hugepage_range(unsigned long addr, unsigned long len);
 int pmd_huge(pmd_t pmd);
 
 #ifndef ARCH_HAS_HUGEPAGE_ONLY_RANGE
@@ -49,8 +48,18 @@ int pmd_huge(pmd_t pmd);
 #endif
 
 #ifndef ARCH_HAS_PREPARE_HUGEPAGE_RANGE
-#define prepare_hugepage_range(addr, len)	\
-	is_aligned_hugepage_range(addr, len)
+/*
+ * If the arch doesn't supply something else, assume that hugepage
+ * size aligned regions are ok without further preparation.
+ */
+static inline int prepare_hugepage_range(unsigned long addr, unsigned long len)
+{
+	if (len & ~HPAGE_MASK)
+		return -EINVAL;
+	if (addr & ~HPAGE_MASK)
+		return -EINVAL;
+	return 0;
+}
 #else
 int prepare_hugepage_range(unsigned long addr, unsigned long len);
 #endif
@@ -91,7 +100,6 @@ static inline unsigned long hugetlb_tota
 #define hugetlb_report_meminfo(buf)		0
 #define hugetlb_report_node_meminfo(n, buf)	0
 #define follow_huge_pmd(mm, addr, pmd, write)	NULL
-#define is_aligned_hugepage_range(addr, len)	0
 #define prepare_hugepage_range(addr, len)	(-EINVAL)
 #define pmd_huge(x)	0
 #define is_hugepage_only_range(mm, addr, len)	0

-- 
David Gibson			| I'll have my music baroque, and my code
david AT gibson.dropbear.id.au	| minimalist, thank you.  NOT _the_ _other_
				| _way_ _around_!
http://www.ozlabs.org/~dgibson
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux