|
@@ -1476,10 +1476,9 @@ static void __vunmap(const void *addr, int deallocate_pages)
|
|
|
if (!addr)
|
|
|
return;
|
|
|
|
|
|
- if ((PAGE_SIZE-1) & (unsigned long)addr) {
|
|
|
- WARN(1, KERN_ERR "Trying to vfree() bad address (%p)\n", addr);
|
|
|
+ if (WARN(!PAGE_ALIGNED(addr), "Trying to vfree() bad address (%p)\n",
|
|
|
+ addr));
|
|
|
return;
|
|
|
- }
|
|
|
|
|
|
area = remove_vm_area(addr);
|
|
|
if (unlikely(!area)) {
|
|
@@ -2148,42 +2147,43 @@ finished:
|
|
|
}
|
|
|
|
|
|
/**
|
|
|
- * remap_vmalloc_range - map vmalloc pages to userspace
|
|
|
- * @vma: vma to cover (map full range of vma)
|
|
|
- * @addr: vmalloc memory
|
|
|
- * @pgoff: number of pages into addr before first page to map
|
|
|
+ * remap_vmalloc_range_partial - map vmalloc pages to userspace
|
|
|
+ * @vma: vma to cover
|
|
|
+ * @uaddr: target user address to start at
|
|
|
+ * @kaddr: virtual address of vmalloc kernel memory
|
|
|
+ * @size: size of map area
|
|
|
*
|
|
|
* Returns: 0 for success, -Exxx on failure
|
|
|
*
|
|
|
- * This function checks that addr is a valid vmalloc'ed area, and
|
|
|
- * that it is big enough to cover the vma. Will return failure if
|
|
|
- * that criteria isn't met.
|
|
|
+ * This function checks that @kaddr is a valid vmalloc'ed area,
|
|
|
+ * and that it is big enough to cover the range starting at
|
|
|
+ * @uaddr in @vma. Will return failure if that criteria isn't
|
|
|
+ * met.
|
|
|
*
|
|
|
* Similar to remap_pfn_range() (see mm/memory.c)
|
|
|
*/
|
|
|
-int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
|
|
|
- unsigned long pgoff)
|
|
|
+int remap_vmalloc_range_partial(struct vm_area_struct *vma, unsigned long uaddr,
|
|
|
+ void *kaddr, unsigned long size)
|
|
|
{
|
|
|
struct vm_struct *area;
|
|
|
- unsigned long uaddr = vma->vm_start;
|
|
|
- unsigned long usize = vma->vm_end - vma->vm_start;
|
|
|
|
|
|
- if ((PAGE_SIZE-1) & (unsigned long)addr)
|
|
|
+ size = PAGE_ALIGN(size);
|
|
|
+
|
|
|
+ if (!PAGE_ALIGNED(uaddr) || !PAGE_ALIGNED(kaddr))
|
|
|
return -EINVAL;
|
|
|
|
|
|
- area = find_vm_area(addr);
|
|
|
+ area = find_vm_area(kaddr);
|
|
|
if (!area)
|
|
|
return -EINVAL;
|
|
|
|
|
|
if (!(area->flags & VM_USERMAP))
|
|
|
return -EINVAL;
|
|
|
|
|
|
- if (usize + (pgoff << PAGE_SHIFT) > area->size - PAGE_SIZE)
|
|
|
+ if (kaddr + size > area->addr + area->size)
|
|
|
return -EINVAL;
|
|
|
|
|
|
- addr += pgoff << PAGE_SHIFT;
|
|
|
do {
|
|
|
- struct page *page = vmalloc_to_page(addr);
|
|
|
+ struct page *page = vmalloc_to_page(kaddr);
|
|
|
int ret;
|
|
|
|
|
|
ret = vm_insert_page(vma, uaddr, page);
|
|
@@ -2191,14 +2191,37 @@ int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
|
|
|
return ret;
|
|
|
|
|
|
uaddr += PAGE_SIZE;
|
|
|
- addr += PAGE_SIZE;
|
|
|
- usize -= PAGE_SIZE;
|
|
|
- } while (usize > 0);
|
|
|
+ kaddr += PAGE_SIZE;
|
|
|
+ size -= PAGE_SIZE;
|
|
|
+ } while (size > 0);
|
|
|
|
|
|
vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP;
|
|
|
|
|
|
return 0;
|
|
|
}
|
|
|
+EXPORT_SYMBOL(remap_vmalloc_range_partial);
|
|
|
+
|
|
|
+/**
|
|
|
+ * remap_vmalloc_range - map vmalloc pages to userspace
|
|
|
+ * @vma: vma to cover (map full range of vma)
|
|
|
+ * @addr: vmalloc memory
|
|
|
+ * @pgoff: number of pages into addr before first page to map
|
|
|
+ *
|
|
|
+ * Returns: 0 for success, -Exxx on failure
|
|
|
+ *
|
|
|
+ * This function checks that addr is a valid vmalloc'ed area, and
|
|
|
+ * that it is big enough to cover the vma. Will return failure if
|
|
|
+ * that criteria isn't met.
|
|
|
+ *
|
|
|
+ * Similar to remap_pfn_range() (see mm/memory.c)
|
|
|
+ */
|
|
|
+int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
|
|
|
+ unsigned long pgoff)
|
|
|
+{
|
|
|
+ return remap_vmalloc_range_partial(vma, vma->vm_start,
|
|
|
+ addr + (pgoff << PAGE_SHIFT),
|
|
|
+ vma->vm_end - vma->vm_start);
|
|
|
+}
|
|
|
EXPORT_SYMBOL(remap_vmalloc_range);
|
|
|
|
|
|
/*
|