|
|
@@ -140,10 +140,10 @@ unsigned int kobjsize(const void *objp)
|
|
|
return PAGE_SIZE << compound_order(page);
|
|
|
}
|
|
|
|
|
|
-int __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
|
|
|
- unsigned long start, int nr_pages, unsigned int foll_flags,
|
|
|
- struct page **pages, struct vm_area_struct **vmas,
|
|
|
- int *retry)
|
|
|
+long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
|
|
|
+ unsigned long start, unsigned long nr_pages,
|
|
|
+ unsigned int foll_flags, struct page **pages,
|
|
|
+ struct vm_area_struct **vmas, int *nonblocking)
|
|
|
{
|
|
|
struct vm_area_struct *vma;
|
|
|
unsigned long vm_flags;
|
|
|
@@ -190,9 +190,10 @@ int __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
|
|
|
* slab page or a secondary page from a compound page
|
|
|
* - don't permit access to VMAs that don't support it, such as I/O mappings
|
|
|
*/
|
|
|
-int get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
|
|
|
- unsigned long start, int nr_pages, int write, int force,
|
|
|
- struct page **pages, struct vm_area_struct **vmas)
|
|
|
+long get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
|
|
|
+ unsigned long start, unsigned long nr_pages,
|
|
|
+ int write, int force, struct page **pages,
|
|
|
+ struct vm_area_struct **vmas)
|
|
|
{
|
|
|
int flags = 0;
|
|
|
|