|
@@ -392,15 +392,13 @@ static inline pgoff_t basepage_index(struct page *page)
|
|
|
|
|
|
extern void dissolve_free_huge_pages(unsigned long start_pfn,
|
|
|
unsigned long end_pfn);
|
|
|
-int pmd_huge_support(void);
|
|
|
-/*
|
|
|
- * Currently hugepage migration is enabled only for pmd-based hugepage.
|
|
|
- * This function will be updated when hugepage migration is more widely
|
|
|
- * supported.
|
|
|
- */
|
|
|
static inline int hugepage_migration_support(struct hstate *h)
|
|
|
{
|
|
|
- return pmd_huge_support() && (huge_page_shift(h) == PMD_SHIFT);
|
|
|
+#ifdef CONFIG_ARCH_ENABLE_HUGEPAGE_MIGRATION
|
|
|
+ return huge_page_shift(h) == PMD_SHIFT;
|
|
|
+#else
|
|
|
+ return 0;
|
|
|
+#endif
|
|
|
}
|
|
|
|
|
|
static inline spinlock_t *huge_pte_lockptr(struct hstate *h,
|
|
@@ -450,7 +448,6 @@ static inline pgoff_t basepage_index(struct page *page)
|
|
|
return page->index;
|
|
|
}
|
|
|
#define dissolve_free_huge_pages(s, e) do {} while (0)
|
|
|
-#define pmd_huge_support() 0
|
|
|
#define hugepage_migration_support(h) 0
|
|
|
|
|
|
static inline spinlock_t *huge_pte_lockptr(struct hstate *h,
|