|
@@ -92,7 +92,7 @@ retry:
|
|
|
*/
|
|
|
mark_page_accessed(page);
|
|
|
}
|
|
|
- if ((flags & FOLL_MLOCK) && (vma->vm_flags & VM_LOCKED)) {
|
|
|
+ if ((flags & FOLL_POPULATE) && (vma->vm_flags & VM_LOCKED)) {
|
|
|
/*
|
|
|
* The preliminary mapping check is mainly to avoid the
|
|
|
* pointless overhead of lock_page on the ZERO_PAGE
|
|
@@ -265,8 +265,8 @@ static int faultin_page(struct task_struct *tsk, struct vm_area_struct *vma,
|
|
|
unsigned int fault_flags = 0;
|
|
|
int ret;
|
|
|
|
|
|
- /* For mlock, just skip the stack guard page. */
|
|
|
- if ((*flags & FOLL_MLOCK) &&
|
|
|
+ /* For mm_populate(), just skip the stack guard page. */
|
|
|
+ if ((*flags & FOLL_POPULATE) &&
|
|
|
(stack_guard_page_start(vma, address) ||
|
|
|
stack_guard_page_end(vma, address + PAGE_SIZE)))
|
|
|
return -ENOENT;
|