|
@@ -363,7 +363,7 @@ static bool shmem_confirm_swap(struct address_space *mapping,
|
|
|
#define SHMEM_HUGE_DENY (-1)
|
|
|
#define SHMEM_HUGE_FORCE (-2)
|
|
|
|
|
|
-#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
|
|
+#ifdef CONFIG_TRANSPARENT_HUGE_PAGECACHE
|
|
|
/* ifdef here to avoid bloating shmem.o when not necessary */
|
|
|
|
|
|
int shmem_huge __read_mostly;
|
|
@@ -406,11 +406,11 @@ static const char *shmem_format_huge(int huge)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-#else /* !CONFIG_TRANSPARENT_HUGEPAGE */
|
|
|
+#else /* !CONFIG_TRANSPARENT_HUGE_PAGECACHE */
|
|
|
|
|
|
#define shmem_huge SHMEM_HUGE_DENY
|
|
|
|
|
|
-#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
|
|
|
+#endif /* CONFIG_TRANSPARENT_HUGE_PAGECACHE */
|
|
|
|
|
|
/*
|
|
|
* Like add_to_page_cache_locked, but error if expected item has gone.
|
|
@@ -1229,7 +1229,7 @@ static struct page *shmem_alloc_hugepage(gfp_t gfp,
|
|
|
void __rcu **results;
|
|
|
struct page *page;
|
|
|
|
|
|
- if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE))
|
|
|
+ if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGE_PAGECACHE))
|
|
|
return NULL;
|
|
|
|
|
|
rcu_read_lock();
|
|
@@ -1270,7 +1270,7 @@ static struct page *shmem_alloc_and_acct_page(gfp_t gfp,
|
|
|
int nr;
|
|
|
int err = -ENOSPC;
|
|
|
|
|
|
- if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE))
|
|
|
+ if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGE_PAGECACHE))
|
|
|
huge = false;
|
|
|
nr = huge ? HPAGE_PMD_NR : 1;
|
|
|
|
|
@@ -1773,7 +1773,7 @@ unsigned long shmem_get_unmapped_area(struct file *file,
|
|
|
get_area = current->mm->get_unmapped_area;
|
|
|
addr = get_area(file, uaddr, len, pgoff, flags);
|
|
|
|
|
|
- if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE))
|
|
|
+ if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGE_PAGECACHE))
|
|
|
return addr;
|
|
|
if (IS_ERR_VALUE(addr))
|
|
|
return addr;
|
|
@@ -1890,7 +1890,7 @@ static int shmem_mmap(struct file *file, struct vm_area_struct *vma)
|
|
|
{
|
|
|
file_accessed(file);
|
|
|
vma->vm_ops = &shmem_vm_ops;
|
|
|
- if (IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) &&
|
|
|
+ if (IS_ENABLED(CONFIG_TRANSPARENT_HUGE_PAGECACHE) &&
|
|
|
((vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK) <
|
|
|
(vma->vm_end & HPAGE_PMD_MASK)) {
|
|
|
khugepaged_enter(vma, vma->vm_flags);
|
|
@@ -3287,7 +3287,7 @@ static int shmem_parse_options(char *options, struct shmem_sb_info *sbinfo,
|
|
|
sbinfo->gid = make_kgid(current_user_ns(), gid);
|
|
|
if (!gid_valid(sbinfo->gid))
|
|
|
goto bad_val;
|
|
|
-#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
|
|
+#ifdef CONFIG_TRANSPARENT_HUGE_PAGECACHE
|
|
|
} else if (!strcmp(this_char, "huge")) {
|
|
|
int huge;
|
|
|
huge = shmem_parse_huge(value);
|
|
@@ -3384,7 +3384,7 @@ static int shmem_show_options(struct seq_file *seq, struct dentry *root)
|
|
|
if (!gid_eq(sbinfo->gid, GLOBAL_ROOT_GID))
|
|
|
seq_printf(seq, ",gid=%u",
|
|
|
from_kgid_munged(&init_user_ns, sbinfo->gid));
|
|
|
-#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
|
|
+#ifdef CONFIG_TRANSPARENT_HUGE_PAGECACHE
|
|
|
/* Rightly or wrongly, show huge mount option unmasked by shmem_huge */
|
|
|
if (sbinfo->huge)
|
|
|
seq_printf(seq, ",huge=%s", shmem_format_huge(sbinfo->huge));
|
|
@@ -3730,7 +3730,7 @@ int __init shmem_init(void)
|
|
|
goto out1;
|
|
|
}
|
|
|
|
|
|
-#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
|
|
+#ifdef CONFIG_TRANSPARENT_HUGE_PAGECACHE
|
|
|
if (has_transparent_hugepage() && shmem_huge < SHMEM_HUGE_DENY)
|
|
|
SHMEM_SB(shm_mnt->mnt_sb)->huge = shmem_huge;
|
|
|
else
|
|
@@ -3747,7 +3747,7 @@ out3:
|
|
|
return error;
|
|
|
}
|
|
|
|
|
|
-#if defined(CONFIG_TRANSPARENT_HUGEPAGE) && defined(CONFIG_SYSFS)
|
|
|
+#if defined(CONFIG_TRANSPARENT_HUGE_PAGECACHE) && defined(CONFIG_SYSFS)
|
|
|
static ssize_t shmem_enabled_show(struct kobject *kobj,
|
|
|
struct kobj_attribute *attr, char *buf)
|
|
|
{
|
|
@@ -3830,7 +3830,7 @@ bool shmem_huge_enabled(struct vm_area_struct *vma)
|
|
|
return false;
|
|
|
}
|
|
|
}
|
|
|
-#endif /* CONFIG_TRANSPARENT_HUGEPAGE && CONFIG_SYSFS */
|
|
|
+#endif /* CONFIG_TRANSPARENT_HUGE_PAGECACHE && CONFIG_SYSFS */
|
|
|
|
|
|
#else /* !CONFIG_SHMEM */
|
|
|
|
|
@@ -4010,7 +4010,7 @@ int shmem_zero_setup(struct vm_area_struct *vma)
|
|
|
vma->vm_file = file;
|
|
|
vma->vm_ops = &shmem_vm_ops;
|
|
|
|
|
|
- if (IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) &&
|
|
|
+ if (IS_ENABLED(CONFIG_TRANSPARENT_HUGE_PAGECACHE) &&
|
|
|
((vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK) <
|
|
|
(vma->vm_end & HPAGE_PMD_MASK)) {
|
|
|
khugepaged_enter(vma, vma->vm_flags);
|