|
@@ -460,43 +460,90 @@ static void region_abort(struct resv_map *resv, long f, long t)
|
|
|
}
|
|
|
|
|
|
/*
|
|
|
- * Truncate the reserve map at index 'end'. Modify/truncate any
|
|
|
- * region which contains end. Delete any regions past end.
|
|
|
- * Return the number of huge pages removed from the map.
|
|
|
+ * Delete the specified range [f, t) from the reserve map. If the
|
|
|
+ * t parameter is LONG_MAX, this indicates that ALL regions after f
|
|
|
+ * should be deleted. Locate the regions which intersect [f, t)
|
|
|
+ * and either trim, delete or split the existing regions.
|
|
|
+ *
|
|
|
+ * Returns the number of huge pages deleted from the reserve map.
|
|
|
+ * In the normal case, the return value is zero or more. In the
|
|
|
+ * case where a region must be split, a new region descriptor must
|
|
|
+ * be allocated. If the allocation fails, -ENOMEM will be returned.
|
|
|
+ * NOTE: If the parameter t == LONG_MAX, then we will never split
|
|
|
+ * a region and possibly return -ENOMEM. Callers specifying
|
|
|
+ * t == LONG_MAX do not need to check for -ENOMEM error.
|
|
|
*/
|
|
|
-static long region_truncate(struct resv_map *resv, long end)
|
|
|
+static long region_del(struct resv_map *resv, long f, long t)
|
|
|
{
|
|
|
struct list_head *head = &resv->regions;
|
|
|
struct file_region *rg, *trg;
|
|
|
- long chg = 0;
|
|
|
+ struct file_region *nrg = NULL;
|
|
|
+ long del = 0;
|
|
|
|
|
|
+retry:
|
|
|
spin_lock(&resv->lock);
|
|
|
- /* Locate the region we are either in or before. */
|
|
|
- list_for_each_entry(rg, head, link)
|
|
|
- if (end <= rg->to)
|
|
|
+ list_for_each_entry_safe(rg, trg, head, link) {
|
|
|
+ if (rg->to <= f)
|
|
|
+ continue;
|
|
|
+ if (rg->from >= t)
|
|
|
break;
|
|
|
- if (&rg->link == head)
|
|
|
- goto out;
|
|
|
|
|
|
- /* If we are in the middle of a region then adjust it. */
|
|
|
- if (end > rg->from) {
|
|
|
- chg = rg->to - end;
|
|
|
- rg->to = end;
|
|
|
- rg = list_entry(rg->link.next, typeof(*rg), link);
|
|
|
- }
|
|
|
+ if (f > rg->from && t < rg->to) { /* Must split region */
|
|
|
+ /*
|
|
|
+ * Check for an entry in the cache before dropping
|
|
|
+ * lock and attempting allocation.
|
|
|
+ */
|
|
|
+ if (!nrg &&
|
|
|
+ resv->region_cache_count > resv->adds_in_progress) {
|
|
|
+ nrg = list_first_entry(&resv->region_cache,
|
|
|
+ struct file_region,
|
|
|
+ link);
|
|
|
+ list_del(&nrg->link);
|
|
|
+ resv->region_cache_count--;
|
|
|
+ }
|
|
|
|
|
|
- /* Drop any remaining regions. */
|
|
|
- list_for_each_entry_safe(rg, trg, rg->link.prev, link) {
|
|
|
- if (&rg->link == head)
|
|
|
+ if (!nrg) {
|
|
|
+ spin_unlock(&resv->lock);
|
|
|
+ nrg = kmalloc(sizeof(*nrg), GFP_KERNEL);
|
|
|
+ if (!nrg)
|
|
|
+ return -ENOMEM;
|
|
|
+ goto retry;
|
|
|
+ }
|
|
|
+
|
|
|
+ del += t - f;
|
|
|
+
|
|
|
+ /* New entry for end of split region */
|
|
|
+ nrg->from = t;
|
|
|
+ nrg->to = rg->to;
|
|
|
+ INIT_LIST_HEAD(&nrg->link);
|
|
|
+
|
|
|
+ /* Original entry is trimmed */
|
|
|
+ rg->to = f;
|
|
|
+
|
|
|
+ list_add(&nrg->link, &rg->link);
|
|
|
+ nrg = NULL;
|
|
|
break;
|
|
|
- chg += rg->to - rg->from;
|
|
|
- list_del(&rg->link);
|
|
|
- kfree(rg);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (f <= rg->from && t >= rg->to) { /* Remove entire region */
|
|
|
+ del += rg->to - rg->from;
|
|
|
+ list_del(&rg->link);
|
|
|
+ kfree(rg);
|
|
|
+ continue;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (f <= rg->from) { /* Trim beginning of region */
|
|
|
+ del += t - rg->from;
|
|
|
+ rg->from = t;
|
|
|
+ } else { /* Trim end of region */
|
|
|
+ del += rg->to - f;
|
|
|
+ rg->to = f;
|
|
|
+ }
|
|
|
}
|
|
|
|
|
|
-out:
|
|
|
spin_unlock(&resv->lock);
|
|
|
- return chg;
|
|
|
+ kfree(nrg);
|
|
|
+ return del;
|
|
|
}
|
|
|
|
|
|
/*
|
|
@@ -647,7 +694,7 @@ void resv_map_release(struct kref *ref)
|
|
|
struct file_region *rg, *trg;
|
|
|
|
|
|
/* Clear out any active regions before we release the map. */
|
|
|
- region_truncate(resv_map, 0);
|
|
|
+ region_del(resv_map, 0, LONG_MAX);
|
|
|
|
|
|
/* ... and any entries left in the cache */
|
|
|
list_for_each_entry_safe(rg, trg, head, link) {
|
|
@@ -1572,7 +1619,7 @@ static void return_unused_surplus_pages(struct hstate *h,
|
|
|
|
|
|
|
|
|
/*
|
|
|
- * vma_needs_reservation, vma_commit_reservation and vma_abort_reservation
|
|
|
+ * vma_needs_reservation, vma_commit_reservation and vma_end_reservation
|
|
|
* are used by the huge page allocation routines to manage reservations.
|
|
|
*
|
|
|
* vma_needs_reservation is called to determine if the huge page at addr
|
|
@@ -1580,8 +1627,9 @@ static void return_unused_surplus_pages(struct hstate *h,
|
|
|
* needed, the value 1 is returned. The caller is then responsible for
|
|
|
* managing the global reservation and subpool usage counts. After
|
|
|
* the huge page has been allocated, vma_commit_reservation is called
|
|
|
- * to add the page to the reservation map. If the reservation must be
|
|
|
- * aborted instead of committed, vma_abort_reservation is called.
|
|
|
+ * to add the page to the reservation map. If the page allocation fails,
|
|
|
+ * the reservation must be ended instead of committed. vma_end_reservation
|
|
|
+ * is called in such cases.
|
|
|
*
|
|
|
* In the normal case, vma_commit_reservation returns the same value
|
|
|
* as the preceding vma_needs_reservation call. The only time this
|
|
@@ -1592,7 +1640,7 @@ static void return_unused_surplus_pages(struct hstate *h,
|
|
|
enum vma_resv_mode {
|
|
|
VMA_NEEDS_RESV,
|
|
|
VMA_COMMIT_RESV,
|
|
|
- VMA_ABORT_RESV,
|
|
|
+ VMA_END_RESV,
|
|
|
};
|
|
|
static long __vma_reservation_common(struct hstate *h,
|
|
|
struct vm_area_struct *vma, unsigned long addr,
|
|
@@ -1614,7 +1662,7 @@ static long __vma_reservation_common(struct hstate *h,
|
|
|
case VMA_COMMIT_RESV:
|
|
|
ret = region_add(resv, idx, idx + 1);
|
|
|
break;
|
|
|
- case VMA_ABORT_RESV:
|
|
|
+ case VMA_END_RESV:
|
|
|
region_abort(resv, idx, idx + 1);
|
|
|
ret = 0;
|
|
|
break;
|
|
@@ -1640,10 +1688,10 @@ static long vma_commit_reservation(struct hstate *h,
|
|
|
return __vma_reservation_common(h, vma, addr, VMA_COMMIT_RESV);
|
|
|
}
|
|
|
|
|
|
-static void vma_abort_reservation(struct hstate *h,
|
|
|
+static void vma_end_reservation(struct hstate *h,
|
|
|
struct vm_area_struct *vma, unsigned long addr)
|
|
|
{
|
|
|
- (void)__vma_reservation_common(h, vma, addr, VMA_ABORT_RESV);
|
|
|
+ (void)__vma_reservation_common(h, vma, addr, VMA_END_RESV);
|
|
|
}
|
|
|
|
|
|
static struct page *alloc_huge_page(struct vm_area_struct *vma,
|
|
@@ -1670,7 +1718,7 @@ static struct page *alloc_huge_page(struct vm_area_struct *vma,
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
if (chg || avoid_reserve)
|
|
|
if (hugepage_subpool_get_pages(spool, 1) < 0) {
|
|
|
- vma_abort_reservation(h, vma, addr);
|
|
|
+ vma_end_reservation(h, vma, addr);
|
|
|
return ERR_PTR(-ENOSPC);
|
|
|
}
|
|
|
|
|
@@ -1718,7 +1766,7 @@ out_uncharge_cgroup:
|
|
|
out_subpool_put:
|
|
|
if (chg || avoid_reserve)
|
|
|
hugepage_subpool_put_pages(spool, 1);
|
|
|
- vma_abort_reservation(h, vma, addr);
|
|
|
+ vma_end_reservation(h, vma, addr);
|
|
|
return ERR_PTR(-ENOSPC);
|
|
|
}
|
|
|
|
|
@@ -3365,7 +3413,7 @@ retry:
|
|
|
goto backout_unlocked;
|
|
|
}
|
|
|
/* Just decrements count, does not deallocate */
|
|
|
- vma_abort_reservation(h, vma, address);
|
|
|
+ vma_end_reservation(h, vma, address);
|
|
|
}
|
|
|
|
|
|
ptl = huge_pte_lockptr(h, mm, ptep);
|
|
@@ -3514,7 +3562,7 @@ int hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
goto out_mutex;
|
|
|
}
|
|
|
/* Just decrements count, does not deallocate */
|
|
|
- vma_abort_reservation(h, vma, address);
|
|
|
+ vma_end_reservation(h, vma, address);
|
|
|
|
|
|
if (!(vma->vm_flags & VM_MAYSHARE))
|
|
|
pagecache_page = hugetlbfs_pagecache_page(h,
|
|
@@ -3870,7 +3918,7 @@ void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed)
|
|
|
long gbl_reserve;
|
|
|
|
|
|
if (resv_map)
|
|
|
- chg = region_truncate(resv_map, offset);
|
|
|
+ chg = region_del(resv_map, offset, LONG_MAX);
|
|
|
spin_lock(&inode->i_lock);
|
|
|
inode->i_blocks -= (blocks_per_huge_page(h) * freed);
|
|
|
spin_unlock(&inode->i_lock);
|