aboutsummaryrefslogtreecommitdiff
path: root/sys/vm
diff options
context:
space:
mode:
authorAlan Cox <alc@FreeBSD.org>2011-12-05 18:29:25 +0000
committerAlan Cox <alc@FreeBSD.org>2011-12-05 18:29:25 +0000
commitc68c35372e0a727fda00b556d87512666a9b4204 (patch)
tree169ae815456d0b52e73adc8ac35eff5e095e269b /sys/vm
parente9d7727e6b8bba6485fb9295ecef65541741615b (diff)
downloadsrc-c68c35372e0a727fda00b556d87512666a9b4204.tar.gz
src-c68c35372e0a727fda00b556d87512666a9b4204.zip
Introduce vm_reserv_alloc_contig() and teach vm_page_alloc_contig() how to
use superpage reservations. So, for the first time, kernel virtual memory that is allocated by contigmalloc(), kmem_alloc_attr(), and kmem_alloc_contig() can be promoted to superpages. In fact, even a series of small contigmalloc() allocations may collectively result in a promoted superpage. Eliminate some duplication of code in vm_reserv_alloc_page(). Change the type of vm_reserv_reclaim_contig()'s first parameter in order that it be consistent with other vm_*_contig() functions. Tested by: marius (sparc64)
Notes
Notes: svn path=/head/; revision=228287
Diffstat (limited to 'sys/vm')
-rw-r--r--sys/vm/vm_page.c11
-rw-r--r--sys/vm/vm_reserv.c321
-rw-r--r--sys/vm/vm_reserv.h8
3 files changed, 269 insertions, 71 deletions
diff --git a/sys/vm/vm_page.c b/sys/vm/vm_page.c
index 405bf25f422b..ee4f2a44e792 100644
--- a/sys/vm/vm_page.c
+++ b/sys/vm/vm_page.c
@@ -1554,9 +1554,12 @@ vm_page_alloc_contig(vm_object_t object, vm_pindex_t pindex, int req,
cnt.v_free_count + cnt.v_cache_count >= npages)) {
#if VM_NRESERVLEVEL > 0
retry:
+ if (object == NULL || (object->flags & OBJ_COLORED) == 0 ||
+ (m_ret = vm_reserv_alloc_contig(object, pindex, npages,
+ low, high, alignment, boundary)) == NULL)
#endif
- m_ret = vm_phys_alloc_contig(npages, low, high, alignment,
- boundary);
+ m_ret = vm_phys_alloc_contig(npages, low, high,
+ alignment, boundary);
} else {
mtx_unlock(&vm_page_queue_free_mtx);
atomic_add_int(&vm_pageout_deficit, npages);
@@ -1581,8 +1584,8 @@ retry:
}
else {
#if VM_NRESERVLEVEL > 0
- if (vm_reserv_reclaim_contig(npages << PAGE_SHIFT, low, high,
- alignment, boundary))
+ if (vm_reserv_reclaim_contig(npages, low, high, alignment,
+ boundary))
goto retry;
#endif
}
diff --git a/sys/vm/vm_reserv.c b/sys/vm/vm_reserv.c
index 12340ff6c79d..1450f8d098be 100644
--- a/sys/vm/vm_reserv.c
+++ b/sys/vm/vm_reserv.c
@@ -31,6 +31,9 @@
/*
* Superpage reservation management module
+ *
+ * Any external functions defined by this module are only to be used by the
+ * virtual memory system.
*/
#include <sys/cdefs.h>
@@ -285,25 +288,54 @@ vm_reserv_populate(vm_reserv_t rv)
}
/*
- * Allocates a page from an existing or newly-created reservation.
+ * Allocates a contiguous set of physical pages of the given size "npages"
+ * from an existing or newly-created reservation. All of the physical pages
+ * must be at or above the given physical address "low" and below the given
+ * physical address "high". The given value "alignment" determines the
+ * alignment of the first physical page in the set. If the given value
+ * "boundary" is non-zero, then the set of physical pages cannot cross any
+ * physical address boundary that is a multiple of that value. Both
+ * "alignment" and "boundary" must be a power of two.
*
* The object and free page queue must be locked.
*/
vm_page_t
-vm_reserv_alloc_page(vm_object_t object, vm_pindex_t pindex)
+vm_reserv_alloc_contig(vm_object_t object, vm_pindex_t pindex, u_long npages,
+ vm_paddr_t low, vm_paddr_t high, u_long alignment, vm_paddr_t boundary)
{
- vm_page_t m, mpred, msucc;
+ vm_paddr_t pa, size;
+ vm_page_t m, m_ret, mpred, msucc;
vm_pindex_t first, leftcap, rightcap;
vm_reserv_t rv;
+ u_long allocpages, maxpages, minpages;
+ int i, index, n;
mtx_assert(&vm_page_queue_free_mtx, MA_OWNED);
+ VM_OBJECT_LOCK_ASSERT(object, MA_OWNED);
+ KASSERT(npages != 0, ("vm_reserv_alloc_contig: npages is 0"));
/*
- * Is a reservation fundamentally not possible?
+ * Is a reservation fundamentally impossible?
*/
- VM_OBJECT_LOCK_ASSERT(object, MA_OWNED);
if (pindex < VM_RESERV_INDEX(object, pindex) ||
- pindex >= object->size)
+ pindex + npages > object->size)
+ return (NULL);
+
+ /*
+ * All reservations of a particular size have the same alignment.
+ * Assuming that the first page is allocated from a reservation, the
+ * least significant bits of its physical address can be determined
+ * from its offset from the beginning of the reservation and the size
+ * of the reservation.
+ *
+ * Could the specified index within a reservation of the smallest
+ * possible size satisfy the alignment and boundary requirements?
+ */
+ pa = VM_RESERV_INDEX(object, pindex) << PAGE_SHIFT;
+ if ((pa & (alignment - 1)) != 0)
+ return (NULL);
+ size = npages << PAGE_SHIFT;
+ if (((pa ^ (pa + size - 1)) & ~(boundary - 1)) != 0)
return (NULL);
/*
@@ -313,31 +345,21 @@ vm_reserv_alloc_page(vm_object_t object, vm_pindex_t pindex)
mpred = object->root;
while (mpred != NULL) {
KASSERT(mpred->pindex != pindex,
- ("vm_reserv_alloc_page: pindex already allocated"));
+ ("vm_reserv_alloc_contig: pindex already allocated"));
rv = vm_reserv_from_page(mpred);
- if (rv->object == object && vm_reserv_has_pindex(rv, pindex)) {
- m = &rv->pages[VM_RESERV_INDEX(object, pindex)];
- /* Handle vm_page_rename(m, new_object, ...). */
- if ((m->flags & (PG_CACHED | PG_FREE)) == 0)
- return (NULL);
- vm_reserv_populate(rv);
- return (m);
- } else if (mpred->pindex < pindex) {
+ if (rv->object == object && vm_reserv_has_pindex(rv, pindex))
+ goto found;
+ else if (mpred->pindex < pindex) {
if (msucc != NULL ||
(msucc = TAILQ_NEXT(mpred, listq)) == NULL)
break;
KASSERT(msucc->pindex != pindex,
- ("vm_reserv_alloc_page: pindex already allocated"));
+ ("vm_reserv_alloc_contig: pindex already allocated"));
rv = vm_reserv_from_page(msucc);
if (rv->object == object &&
- vm_reserv_has_pindex(rv, pindex)) {
- m = &rv->pages[VM_RESERV_INDEX(object, pindex)];
- /* Handle vm_page_rename(m, new_object, ...). */
- if ((m->flags & (PG_CACHED | PG_FREE)) == 0)
- return (NULL);
- vm_reserv_populate(rv);
- return (m);
- } else if (pindex < msucc->pindex)
+ vm_reserv_has_pindex(rv, pindex))
+ goto found;
+ else if (pindex < msucc->pindex)
break;
} else if (msucc == NULL) {
msucc = mpred;
@@ -349,72 +371,238 @@ vm_reserv_alloc_page(vm_object_t object, vm_pindex_t pindex)
}
/*
- * Determine the first index to the left that can be used.
- */
- if (mpred == NULL)
- leftcap = 0;
- else if ((rv = vm_reserv_from_page(mpred))->object != object)
- leftcap = mpred->pindex + 1;
- else
- leftcap = rv->pindex + VM_LEVEL_0_NPAGES;
-
- /*
- * Determine the first index to the right that cannot be used.
- */
- if (msucc == NULL)
- rightcap = pindex + VM_LEVEL_0_NPAGES;
- else if ((rv = vm_reserv_from_page(msucc))->object != object)
- rightcap = msucc->pindex;
- else
- rightcap = rv->pindex;
-
- /*
- * Determine if a reservation fits between the first index to
- * the left that can be used and the first index to the right
- * that cannot be used.
+ * Could at least one reservation fit between the first index to the
+ * left that can be used and the first index to the right that cannot
+ * be used?
*/
first = pindex - VM_RESERV_INDEX(object, pindex);
- if (first < leftcap || first + VM_LEVEL_0_NPAGES > rightcap)
- return (NULL);
+ if (mpred != NULL) {
+ if ((rv = vm_reserv_from_page(mpred))->object != object)
+ leftcap = mpred->pindex + 1;
+ else
+ leftcap = rv->pindex + VM_LEVEL_0_NPAGES;
+ if (leftcap > first)
+ return (NULL);
+ }
+ minpages = VM_RESERV_INDEX(object, pindex) + npages;
+ maxpages = roundup2(minpages, VM_LEVEL_0_NPAGES);
+ allocpages = maxpages;
+ if (msucc != NULL) {
+ if ((rv = vm_reserv_from_page(msucc))->object != object)
+ rightcap = msucc->pindex;
+ else
+ rightcap = rv->pindex;
+ if (first + maxpages > rightcap) {
+ if (maxpages == VM_LEVEL_0_NPAGES)
+ return (NULL);
+ allocpages = minpages;
+ }
+ }
/*
- * Would a new reservation extend past the end of the given object?
+ * Would the last new reservation extend past the end of the object?
*/
- if (object->size < first + VM_LEVEL_0_NPAGES) {
+ if (first + maxpages > object->size) {
/*
- * Don't allocate a new reservation if the object is a vnode or
- * backed by another object that is a vnode.
+ * Don't allocate the last new reservation if the object is a
+ * vnode or backed by another object that is a vnode.
*/
if (object->type == OBJT_VNODE ||
(object->backing_object != NULL &&
- object->backing_object->type == OBJT_VNODE))
- return (NULL);
+ object->backing_object->type == OBJT_VNODE)) {
+ if (maxpages == VM_LEVEL_0_NPAGES)
+ return (NULL);
+ allocpages = minpages;
+ }
/* Speculate that the object may grow. */
}
/*
- * Allocate a new reservation.
+ * Allocate and populate the new reservations. The alignment and
+ * boundary specified for this allocation may be different from the
+ * alignment and boundary specified for the requested pages. For
+ * instance, the specified index may not be the first page within the
+ * first new reservation.
*/
- m = vm_phys_alloc_pages(VM_FREEPOOL_DEFAULT, VM_LEVEL_0_ORDER);
- if (m != NULL) {
+ m = vm_phys_alloc_contig(allocpages, low, high, ulmax(alignment,
+ VM_LEVEL_0_SIZE), boundary > VM_LEVEL_0_SIZE ? boundary : 0);
+ if (m == NULL)
+ return (NULL);
+ m_ret = NULL;
+ index = VM_RESERV_INDEX(object, pindex);
+ do {
rv = vm_reserv_from_page(m);
KASSERT(rv->pages == m,
- ("vm_reserv_alloc_page: reserv %p's pages is corrupted",
+ ("vm_reserv_alloc_contig: reserv %p's pages is corrupted",
rv));
KASSERT(rv->object == NULL,
- ("vm_reserv_alloc_page: reserv %p isn't free", rv));
+ ("vm_reserv_alloc_contig: reserv %p isn't free", rv));
LIST_INSERT_HEAD(&object->rvq, rv, objq);
rv->object = object;
rv->pindex = first;
KASSERT(rv->popcnt == 0,
- ("vm_reserv_alloc_page: reserv %p's popcnt is corrupted",
+ ("vm_reserv_alloc_contig: reserv %p's popcnt is corrupted",
rv));
KASSERT(!rv->inpartpopq,
- ("vm_reserv_alloc_page: reserv %p's inpartpopq is TRUE",
+ ("vm_reserv_alloc_contig: reserv %p's inpartpopq is TRUE",
rv));
+ n = ulmin(VM_LEVEL_0_NPAGES - index, npages);
+ for (i = 0; i < n; i++)
+ vm_reserv_populate(rv);
+ npages -= n;
+ if (m_ret == NULL) {
+ m_ret = &rv->pages[index];
+ index = 0;
+ }
+ m += VM_LEVEL_0_NPAGES;
+ first += VM_LEVEL_0_NPAGES;
+ allocpages -= VM_LEVEL_0_NPAGES;
+ } while (allocpages > VM_LEVEL_0_NPAGES);
+ return (m_ret);
+
+ /*
+ * Found a matching reservation.
+ */
+found:
+ index = VM_RESERV_INDEX(object, pindex);
+ /* Does the allocation fit within the reservation? */
+ if (index + npages > VM_LEVEL_0_NPAGES)
+ return (NULL);
+ m = &rv->pages[index];
+ pa = VM_PAGE_TO_PHYS(m);
+ if (pa < low || pa + size > high || (pa & (alignment - 1)) != 0 ||
+ ((pa ^ (pa + size - 1)) & ~(boundary - 1)) != 0)
+ return (NULL);
+ /* Handle vm_page_rename(m, new_object, ...). */
+ for (i = 0; i < npages; i++)
+ if ((rv->pages[index + i].flags & (PG_CACHED | PG_FREE)) == 0)
+ return (NULL);
+ for (i = 0; i < npages; i++)
vm_reserv_populate(rv);
- m = &rv->pages[VM_RESERV_INDEX(object, pindex)];
+ return (m);
+}
+
+/*
+ * Allocates a page from an existing or newly-created reservation.
+ *
+ * The object and free page queue must be locked.
+ */
+vm_page_t
+vm_reserv_alloc_page(vm_object_t object, vm_pindex_t pindex)
+{
+ vm_page_t m, mpred, msucc;
+ vm_pindex_t first, leftcap, rightcap;
+ vm_reserv_t rv;
+
+ mtx_assert(&vm_page_queue_free_mtx, MA_OWNED);
+ VM_OBJECT_LOCK_ASSERT(object, MA_OWNED);
+
+ /*
+ * Is a reservation fundamentally impossible?
+ */
+ if (pindex < VM_RESERV_INDEX(object, pindex) ||
+ pindex >= object->size)
+ return (NULL);
+
+ /*
+ * Look for an existing reservation.
+ */
+ msucc = NULL;
+ mpred = object->root;
+ while (mpred != NULL) {
+ KASSERT(mpred->pindex != pindex,
+ ("vm_reserv_alloc_page: pindex already allocated"));
+ rv = vm_reserv_from_page(mpred);
+ if (rv->object == object && vm_reserv_has_pindex(rv, pindex))
+ goto found;
+ else if (mpred->pindex < pindex) {
+ if (msucc != NULL ||
+ (msucc = TAILQ_NEXT(mpred, listq)) == NULL)
+ break;
+ KASSERT(msucc->pindex != pindex,
+ ("vm_reserv_alloc_page: pindex already allocated"));
+ rv = vm_reserv_from_page(msucc);
+ if (rv->object == object &&
+ vm_reserv_has_pindex(rv, pindex))
+ goto found;
+ else if (pindex < msucc->pindex)
+ break;
+ } else if (msucc == NULL) {
+ msucc = mpred;
+ mpred = TAILQ_PREV(msucc, pglist, listq);
+ continue;
+ }
+ msucc = NULL;
+ mpred = object->root = vm_page_splay(pindex, object->root);
+ }
+
+ /*
+ * Could a reservation fit between the first index to the left that
+ * can be used and the first index to the right that cannot be used?
+ */
+ first = pindex - VM_RESERV_INDEX(object, pindex);
+ if (mpred != NULL) {
+ if ((rv = vm_reserv_from_page(mpred))->object != object)
+ leftcap = mpred->pindex + 1;
+ else
+ leftcap = rv->pindex + VM_LEVEL_0_NPAGES;
+ if (leftcap > first)
+ return (NULL);
+ }
+ if (msucc != NULL) {
+ if ((rv = vm_reserv_from_page(msucc))->object != object)
+ rightcap = msucc->pindex;
+ else
+ rightcap = rv->pindex;
+ if (first + VM_LEVEL_0_NPAGES > rightcap)
+ return (NULL);
+ }
+
+ /*
+ * Would a new reservation extend past the end of the object?
+ */
+ if (first + VM_LEVEL_0_NPAGES > object->size) {
+ /*
+ * Don't allocate a new reservation if the object is a vnode or
+ * backed by another object that is a vnode.
+ */
+ if (object->type == OBJT_VNODE ||
+ (object->backing_object != NULL &&
+ object->backing_object->type == OBJT_VNODE))
+ return (NULL);
+ /* Speculate that the object may grow. */
}
+
+ /*
+ * Allocate and populate the new reservation.
+ */
+ m = vm_phys_alloc_pages(VM_FREEPOOL_DEFAULT, VM_LEVEL_0_ORDER);
+ if (m == NULL)
+ return (NULL);
+ rv = vm_reserv_from_page(m);
+ KASSERT(rv->pages == m,
+ ("vm_reserv_alloc_page: reserv %p's pages is corrupted", rv));
+ KASSERT(rv->object == NULL,
+ ("vm_reserv_alloc_page: reserv %p isn't free", rv));
+ LIST_INSERT_HEAD(&object->rvq, rv, objq);
+ rv->object = object;
+ rv->pindex = first;
+ KASSERT(rv->popcnt == 0,
+ ("vm_reserv_alloc_page: reserv %p's popcnt is corrupted", rv));
+ KASSERT(!rv->inpartpopq,
+ ("vm_reserv_alloc_page: reserv %p's inpartpopq is TRUE", rv));
+ vm_reserv_populate(rv);
+ return (&rv->pages[VM_RESERV_INDEX(object, pindex)]);
+
+ /*
+ * Found a matching reservation.
+ */
+found:
+ m = &rv->pages[VM_RESERV_INDEX(object, pindex)];
+ /* Handle vm_page_rename(m, new_object, ...). */
+ if ((m->flags & (PG_CACHED | PG_FREE)) == 0)
+ return (NULL);
+ vm_reserv_populate(rv);
return (m);
}
@@ -627,16 +815,17 @@ vm_reserv_reclaim_inactive(void)
* The free page queue lock must be held.
*/
boolean_t
-vm_reserv_reclaim_contig(vm_paddr_t size, vm_paddr_t low, vm_paddr_t high,
+vm_reserv_reclaim_contig(u_long npages, vm_paddr_t low, vm_paddr_t high,
u_long alignment, vm_paddr_t boundary)
{
- vm_paddr_t pa, pa_length;
+ vm_paddr_t pa, pa_length, size;
vm_reserv_t rv;
int i;
mtx_assert(&vm_page_queue_free_mtx, MA_OWNED);
- if (size > VM_LEVEL_0_SIZE - PAGE_SIZE)
+ if (npages > VM_LEVEL_0_NPAGES - 1)
return (FALSE);
+ size = npages << PAGE_SHIFT;
TAILQ_FOREACH(rv, &vm_rvq_partpop, partpopq) {
pa = VM_PAGE_TO_PHYS(&rv->pages[VM_LEVEL_0_NPAGES - 1]);
if (pa + PAGE_SIZE - size < low) {
diff --git a/sys/vm/vm_reserv.h b/sys/vm/vm_reserv.h
index 9ab95682558d..a2b664778b23 100644
--- a/sys/vm/vm_reserv.h
+++ b/sys/vm/vm_reserv.h
@@ -42,13 +42,19 @@
#if VM_NRESERVLEVEL > 0
+/*
+ * The following functions are only to be used by the virtual memory system.
+ */
+vm_page_t vm_reserv_alloc_contig(vm_object_t object, vm_pindex_t pindex,
+ u_long npages, vm_paddr_t low, vm_paddr_t high,
+ u_long alignment, vm_paddr_t boundary);
vm_page_t vm_reserv_alloc_page(vm_object_t object, vm_pindex_t pindex);
void vm_reserv_break_all(vm_object_t object);
boolean_t vm_reserv_free_page(vm_page_t m);
void vm_reserv_init(void);
int vm_reserv_level_iffullpop(vm_page_t m);
boolean_t vm_reserv_reactivate_page(vm_page_t m);
-boolean_t vm_reserv_reclaim_contig(vm_paddr_t size, vm_paddr_t low,
+boolean_t vm_reserv_reclaim_contig(u_long npages, vm_paddr_t low,
vm_paddr_t high, u_long alignment, vm_paddr_t boundary);
boolean_t vm_reserv_reclaim_inactive(void);
void vm_reserv_rename(vm_page_t m, vm_object_t new_object,