|
@@ -2138,7 +2138,7 @@ static unsigned long set_max_huge_pages(struct hstate *h, unsigned long count,
|
|
|
* First take pages out of surplus state. Then make up the
|
|
|
* remaining difference by allocating fresh huge pages.
|
|
|
*
|
|
|
- * We might race with alloc_buddy_huge_page() here and be unable
|
|
|
+ * We might race with __alloc_buddy_huge_page() here and be unable
|
|
|
* to convert a surplus huge page to a normal huge page. That is
|
|
|
* not critical, though, it just means the overall size of the
|
|
|
* pool might be one hugepage larger than it needs to be, but
|
|
@@ -2180,7 +2180,7 @@ static unsigned long set_max_huge_pages(struct hstate *h, unsigned long count,
|
|
|
* By placing pages into the surplus state independent of the
|
|
|
* overcommit value, we are allowing the surplus pool size to
|
|
|
* exceed overcommit. There are few sane options here. Since
|
|
|
- * alloc_buddy_huge_page() is checking the global counter,
|
|
|
+ * __alloc_buddy_huge_page() is checking the global counter,
|
|
|
* though, we'll note that we're not allowed to exceed surplus
|
|
|
* and won't grow the pool anywhere else. Not until one of the
|
|
|
* sysctls are changed, or the surplus pages go out of use.
|