|
@@ -60,6 +60,16 @@ static int follow_pfn_pte(struct vm_area_struct *vma, unsigned long address,
|
|
|
return -EEXIST;
|
|
|
}
|
|
|
|
|
|
+/*
|
|
|
+ * FOLL_FORCE can write to even unwritable pte's, but only
|
|
|
+ * after we've gone through a COW cycle and they are dirty.
|
|
|
+ */
|
|
|
+static inline bool can_follow_write_pte(pte_t pte, unsigned int flags)
|
|
|
+{
|
|
|
+ return pte_write(pte) ||
|
|
|
+ ((flags & FOLL_FORCE) && (flags & FOLL_COW) && pte_dirty(pte));
|
|
|
+}
|
|
|
+
|
|
|
static struct page *follow_page_pte(struct vm_area_struct *vma,
|
|
|
unsigned long address, pmd_t *pmd, unsigned int flags)
|
|
|
{
|
|
@@ -95,7 +105,7 @@ retry:
|
|
|
}
|
|
|
if ((flags & FOLL_NUMA) && pte_protnone(pte))
|
|
|
goto no_page;
|
|
|
- if ((flags & FOLL_WRITE) && !pte_write(pte)) {
|
|
|
+ if ((flags & FOLL_WRITE) && !can_follow_write_pte(pte, flags)) {
|
|
|
pte_unmap_unlock(ptep, ptl);
|
|
|
return NULL;
|
|
|
}
|
|
@@ -412,7 +422,7 @@ static int faultin_page(struct task_struct *tsk, struct vm_area_struct *vma,
|
|
|
* reCOWed by userspace write).
|
|
|
*/
|
|
|
if ((ret & VM_FAULT_WRITE) && !(vma->vm_flags & VM_WRITE))
|
|
|
- *flags &= ~FOLL_WRITE;
|
|
|
+ *flags |= FOLL_COW;
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
@@ -729,7 +739,6 @@ static __always_inline long __get_user_pages_locked(struct task_struct *tsk,
|
|
|
struct mm_struct *mm,
|
|
|
unsigned long start,
|
|
|
unsigned long nr_pages,
|
|
|
- int write, int force,
|
|
|
struct page **pages,
|
|
|
struct vm_area_struct **vmas,
|
|
|
int *locked, bool notify_drop,
|
|
@@ -747,10 +756,6 @@ static __always_inline long __get_user_pages_locked(struct task_struct *tsk,
|
|
|
|
|
|
if (pages)
|
|
|
flags |= FOLL_GET;
|
|
|
- if (write)
|
|
|
- flags |= FOLL_WRITE;
|
|
|
- if (force)
|
|
|
- flags |= FOLL_FORCE;
|
|
|
|
|
|
pages_done = 0;
|
|
|
lock_dropped = false;
|
|
@@ -843,12 +848,12 @@ static __always_inline long __get_user_pages_locked(struct task_struct *tsk,
|
|
|
* up_read(&mm->mmap_sem);
|
|
|
*/
|
|
|
long get_user_pages_locked(unsigned long start, unsigned long nr_pages,
|
|
|
- int write, int force, struct page **pages,
|
|
|
+ unsigned int gup_flags, struct page **pages,
|
|
|
int *locked)
|
|
|
{
|
|
|
return __get_user_pages_locked(current, current->mm, start, nr_pages,
|
|
|
- write, force, pages, NULL, locked, true,
|
|
|
- FOLL_TOUCH);
|
|
|
+ pages, NULL, locked, true,
|
|
|
+ gup_flags | FOLL_TOUCH);
|
|
|
}
|
|
|
EXPORT_SYMBOL(get_user_pages_locked);
|
|
|
|
|
@@ -864,14 +869,14 @@ EXPORT_SYMBOL(get_user_pages_locked);
|
|
|
*/
|
|
|
__always_inline long __get_user_pages_unlocked(struct task_struct *tsk, struct mm_struct *mm,
|
|
|
unsigned long start, unsigned long nr_pages,
|
|
|
- int write, int force, struct page **pages,
|
|
|
- unsigned int gup_flags)
|
|
|
+ struct page **pages, unsigned int gup_flags)
|
|
|
{
|
|
|
long ret;
|
|
|
int locked = 1;
|
|
|
+
|
|
|
down_read(&mm->mmap_sem);
|
|
|
- ret = __get_user_pages_locked(tsk, mm, start, nr_pages, write, force,
|
|
|
- pages, NULL, &locked, false, gup_flags);
|
|
|
+ ret = __get_user_pages_locked(tsk, mm, start, nr_pages, pages, NULL,
|
|
|
+ &locked, false, gup_flags);
|
|
|
if (locked)
|
|
|
up_read(&mm->mmap_sem);
|
|
|
return ret;
|
|
@@ -896,10 +901,10 @@ EXPORT_SYMBOL(__get_user_pages_unlocked);
|
|
|
* "force" parameter).
|
|
|
*/
|
|
|
long get_user_pages_unlocked(unsigned long start, unsigned long nr_pages,
|
|
|
- int write, int force, struct page **pages)
|
|
|
+ struct page **pages, unsigned int gup_flags)
|
|
|
{
|
|
|
return __get_user_pages_unlocked(current, current->mm, start, nr_pages,
|
|
|
- write, force, pages, FOLL_TOUCH);
|
|
|
+ pages, gup_flags | FOLL_TOUCH);
|
|
|
}
|
|
|
EXPORT_SYMBOL(get_user_pages_unlocked);
|
|
|
|
|
@@ -910,9 +915,7 @@ EXPORT_SYMBOL(get_user_pages_unlocked);
|
|
|
* @mm: mm_struct of target mm
|
|
|
* @start: starting user address
|
|
|
* @nr_pages: number of pages from start to pin
|
|
|
- * @write: whether pages will be written to by the caller
|
|
|
- * @force: whether to force access even when user mapping is currently
|
|
|
- * protected (but never forces write access to shared mapping).
|
|
|
+ * @gup_flags: flags modifying lookup behaviour
|
|
|
* @pages: array that receives pointers to the pages pinned.
|
|
|
* Should be at least nr_pages long. Or NULL, if caller
|
|
|
* only intends to ensure the pages are faulted in.
|
|
@@ -941,9 +944,9 @@ EXPORT_SYMBOL(get_user_pages_unlocked);
|
|
|
* or similar operation cannot guarantee anything stronger anyway because
|
|
|
* locks can't be held over the syscall boundary.
|
|
|
*
|
|
|
- * If write=0, the page must not be written to. If the page is written to,
|
|
|
- * set_page_dirty (or set_page_dirty_lock, as appropriate) must be called
|
|
|
- * after the page is finished with, and before put_page is called.
|
|
|
+ * If gup_flags & FOLL_WRITE == 0, the page must not be written to. If the page
|
|
|
+ * is written to, set_page_dirty (or set_page_dirty_lock, as appropriate) must
|
|
|
+ * be called after the page is finished with, and before put_page is called.
|
|
|
*
|
|
|
* get_user_pages is typically used for fewer-copy IO operations, to get a
|
|
|
* handle on the memory by some means other than accesses via the user virtual
|
|
@@ -960,12 +963,12 @@ EXPORT_SYMBOL(get_user_pages_unlocked);
|
|
|
*/
|
|
|
long get_user_pages_remote(struct task_struct *tsk, struct mm_struct *mm,
|
|
|
unsigned long start, unsigned long nr_pages,
|
|
|
- int write, int force, struct page **pages,
|
|
|
+ unsigned int gup_flags, struct page **pages,
|
|
|
struct vm_area_struct **vmas)
|
|
|
{
|
|
|
- return __get_user_pages_locked(tsk, mm, start, nr_pages, write, force,
|
|
|
- pages, vmas, NULL, false,
|
|
|
- FOLL_TOUCH | FOLL_REMOTE);
|
|
|
+ return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas,
|
|
|
+ NULL, false,
|
|
|
+ gup_flags | FOLL_TOUCH | FOLL_REMOTE);
|
|
|
}
|
|
|
EXPORT_SYMBOL(get_user_pages_remote);
|
|
|
|
|
@@ -976,12 +979,12 @@ EXPORT_SYMBOL(get_user_pages_remote);
|
|
|
* obviously don't pass FOLL_REMOTE in here.
|
|
|
*/
|
|
|
long get_user_pages(unsigned long start, unsigned long nr_pages,
|
|
|
- int write, int force, struct page **pages,
|
|
|
+ unsigned int gup_flags, struct page **pages,
|
|
|
struct vm_area_struct **vmas)
|
|
|
{
|
|
|
return __get_user_pages_locked(current, current->mm, start, nr_pages,
|
|
|
- write, force, pages, vmas, NULL, false,
|
|
|
- FOLL_TOUCH);
|
|
|
+ pages, vmas, NULL, false,
|
|
|
+ gup_flags | FOLL_TOUCH);
|
|
|
}
|
|
|
EXPORT_SYMBOL(get_user_pages);
|
|
|
|
|
@@ -1505,7 +1508,8 @@ int get_user_pages_fast(unsigned long start, int nr_pages, int write,
|
|
|
start += nr << PAGE_SHIFT;
|
|
|
pages += nr;
|
|
|
|
|
|
- ret = get_user_pages_unlocked(start, nr_pages - nr, write, 0, pages);
|
|
|
+ ret = get_user_pages_unlocked(start, nr_pages - nr, pages,
|
|
|
+ write ? FOLL_WRITE : 0);
|
|
|
|
|
|
/* Have to be a bit careful with return values */
|
|
|
if (nr > 0) {
|