aboutsummaryrefslogtreecommitdiff
path: root/sys
diff options
context:
space:
mode:
authorJake Burkholder <jake@FreeBSD.org>2001-12-29 08:12:22 +0000
committerJake Burkholder <jake@FreeBSD.org>2001-12-29 08:12:22 +0000
commitb45b19a44eaf77d89ddd4f519f41932ddc422865 (patch)
treea8077a88e053bc734a29c5f3967b2b03ecb5f6b8 /sys
parent6e5f166d2c438a056bc01bb62b63ea4171cf3d40 (diff)
Implement pv entries as separate structures from the ttes, as other
architectures do. Many crucial bug fixes from: tmm
Notes
Notes: svn path=/head/; revision=88648
Diffstat (limited to 'sys')
-rw-r--r--sys/sparc64/sparc64/pv.c349
1 files changed, 177 insertions, 172 deletions
diff --git a/sys/sparc64/sparc64/pv.c b/sys/sparc64/sparc64/pv.c
index 881fd8634bb7..7785bf1a4d3d 100644
--- a/sys/sparc64/sparc64/pv.c
+++ b/sys/sparc64/sparc64/pv.c
@@ -1,5 +1,10 @@
/*-
- * Copyright (c) 1997 Berkeley Software Design, Inc. All rights reserved.
+ * Copyright (c) 1991 Regents of the University of California.
+ * All rights reserved.
+ *
+ * This code is derived from software contributed to Berkeley by
+ * the Systems Programming Group of the University of Utah Computer
+ * Science Department and William Jolitz of UUNET Technologies Inc.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
@@ -9,14 +14,18 @@
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
- * 3. Berkeley Software Design Inc's name may not be used to endorse or
- * promote products derived from this software without specific prior
- * written permission.
+ * 3. All advertising materials mentioning features or use of this software
+ * must display the following acknowledgement:
+ * This product includes software developed by the University of
+ * California, Berkeley and its contributors.
+ * 4. Neither the name of the University nor the names of its contributors
+ * may be used to endorse or promote products derived from this software
+ * without specific prior written permission.
*
- * THIS SOFTWARE IS PROVIDED BY BERKELEY SOFTWARE DESIGN INC ``AS IS'' AND
+ * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL BERKELEY SOFTWARE DESIGN INC BE LIABLE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
@@ -25,7 +34,6 @@
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
- * from BSDI: pmap.c,v 1.28.2.15 2000/04/27 03:10:31 cp Exp
* $FreeBSD$
*/
@@ -37,7 +45,10 @@
#include <vm/vm.h>
#include <vm/vm_param.h>
#include <vm/vm_kern.h>
+#include <vm/vm_object.h>
#include <vm/vm_page.h>
+#include <vm/vm_pageout.h>
+#include <vm/vm_zone.h>
#include <machine/asi.h>
#include <machine/frame.h>
@@ -47,219 +58,213 @@
#include <machine/tlb.h>
#include <machine/tsb.h>
-/*
- * Physical address of array of physical addresses of stte alias chain heads,
- * and generation count of alias chains.
- */
-vm_offset_t pv_table;
-u_long pv_generation;
+vm_zone_t pvzone;
+struct vm_zone pvzone_store;
+struct vm_object pvzone_obj;
+int pv_entry_count;
+int pv_entry_max;
+int pv_entry_high_water;
+struct pv_entry *pvinit;
+
+pv_entry_t
+pv_alloc(void)
+{
+
+ pv_entry_count++;
+ if (pv_entry_high_water && (pv_entry_count > pv_entry_high_water) &&
+ (pmap_pagedaemon_waken == 0)) {
+ pmap_pagedaemon_waken = 1;
+ wakeup(&vm_pages_needed);
+ }
+ return (zalloc(pvzone));
+}
-static void pv_local_remove_all(vm_offset_t pvh);
+void
+pv_free(pv_entry_t pv)
+{
+
+ pv_entry_count--;
+ zfree(pvzone, pv);
+}
/*
* Insert a mapped stte at the tail of an address alias chain.
*/
void
-pv_insert(pmap_t pm, vm_offset_t pa, vm_offset_t va, struct stte *stp)
+pv_insert(pmap_t pm, vm_page_t m, vm_offset_t va)
{
- vm_offset_t pstp;
- vm_offset_t pvh;
+ pv_entry_t pv;
- pstp = tsb_stte_vtophys(pm, stp);
- pvh = pv_lookup(pa);
- PV_LOCK();
- if ((stp->st_next = pvh_get_first(pvh)) != 0)
- pv_set_prev(stp->st_next, pstp + ST_NEXT);
- pvh_set_first(pvh, pstp);
- stp->st_prev = pvh;
- pv_generation++;
- PV_UNLOCK();
+ pv = pv_alloc();
+ pv->pv_va = va;
+ pv->pv_m = m;
+ pv->pv_pmap = pm;
+ TAILQ_INSERT_TAIL(&pm->pm_pvlist, pv, pv_plist);
+ TAILQ_INSERT_TAIL(&m->md.pv_list, pv, pv_list);
+ m->md.pv_list_count++;
+ pm->pm_stats.resident_count++;
+}
+
+pv_entry_t
+pv_lookup(pmap_t pm, vm_page_t m, vm_offset_t va)
+{
+ pv_entry_t pv;
+
+ if (m != NULL && m->md.pv_list_count < pm->pm_stats.resident_count) {
+ TAILQ_FOREACH(pv, &m->md.pv_list, pv_list)
+ if (pm == pv->pv_pmap && va == pv->pv_va)
+ break;
+ } else {
+ TAILQ_FOREACH(pv, &pm->pm_pvlist, pv_plist)
+ if (va == pv->pv_va)
+ break;
+ }
+ return (pv);
}
/*
* Remove a mapped tte from its address alias chain.
*/
void
-pv_remove_virt(struct stte *stp)
+pv_remove(pmap_t pm, vm_page_t m, vm_offset_t va)
{
- PV_LOCK();
- if (stp->st_next != 0)
- pv_set_prev(stp->st_next, stp->st_prev);
- stxp(stp->st_prev, stp->st_next);
- pv_generation++;
- PV_UNLOCK();
+ pv_entry_t pv;
+
+ if ((pv = pv_lookup(pm, m, va)) != NULL) {
+ m->md.pv_list_count--;
+ pm->pm_stats.resident_count--;
+ TAILQ_REMOVE(&m->md.pv_list, pv, pv_list);
+ TAILQ_REMOVE(&pm->pm_pvlist, pv, pv_plist);
+ if (TAILQ_EMPTY(&m->md.pv_list))
+ vm_page_flag_clear(m, PG_MAPPED | PG_WRITEABLE);
+ pv_free(pv);
+ }
}
void
pv_bit_clear(vm_page_t m, u_long bits)
{
- vm_offset_t pstp;
- vm_offset_t pvh;
- vm_offset_t pa;
- vm_offset_t va;
- struct tte tte;
+ struct tte *tp;
+ pv_entry_t pv;
- pa = VM_PAGE_TO_PHYS(m);
- pvh = pv_lookup(pa);
- PV_LOCK();
-#ifdef notyet
-restart:
-#endif
- for (pstp = pvh_get_first(pvh); pstp != 0; pstp = pv_get_next(pstp)) {
- tte = pv_get_tte(pstp);
- KASSERT(TD_PA(tte.tte_data) == pa,
- ("pv_bit_clear: corrupt alias chain"));
- if ((tte.tte_data & bits) == 0)
- continue;
- va = tte_get_va(tte);
- if (bits & (TD_W | TD_SW) && !pmap_track_modified(va))
- continue;
- if (bits & (TD_W | TD_SW) && tte.tte_data & TD_W)
- vm_page_dirty(m);
- pv_atomic_bit_clear(pstp, bits);
-#ifdef notyet
- generation = pv_generation;
- PV_UNLOCK();
- /* XXX pass function and parameter to ipi call */
- ipi_all(IPI_TLB_PAGE_DEMAP);
- PV_LOCK();
- if (generation != pv_generation)
- goto restart;
-#else
- tlb_page_demap(TLB_DTLB, TT_GET_CTX(tte.tte_tag), va);
-#endif
+ TAILQ_FOREACH(pv, &m->md.pv_list, pv_list) {
+ KASSERT(pv->pv_pmap != NULL, ("pv_bit_clear: null pmap"));
+ if ((tp = tsb_tte_lookup(pv->pv_pmap, pv->pv_va)) != NULL &&
+ (tp->tte_data & bits) != 0) {
+ if ((bits & TD_SW) != 0 &&
+ pmap_track_modified(pv->pv_pmap, pv->pv_va)) {
+ if (tp->tte_data & TD_W)
+ vm_page_dirty(m);
+ }
+ atomic_clear_long(&tp->tte_data, bits);
+ tlb_page_demap(TLB_DTLB | TLB_ITLB,
+ pv->pv_pmap->pm_context, pv->pv_va);
+ }
}
- PV_UNLOCK();
}
int
pv_bit_count(vm_page_t m, u_long bits)
{
- vm_offset_t pstp;
- vm_offset_t pvh;
- vm_offset_t pa;
- vm_offset_t va;
- struct tte tte;
+ struct tte *tp;
+ pv_entry_t pvf;
+ pv_entry_t pvn;
+ pv_entry_t pv;
int count;
count = 0;
- pa = VM_PAGE_TO_PHYS(m);
- pvh = pv_lookup(pa);
- PV_LOCK();
-#ifdef notyet
-restart:
-#endif
- for (pstp = pvh_get_first(pvh); pstp != 0; pstp = pv_get_next(pstp)) {
- tte = pv_get_tte(pstp);
- va = tte_get_va(tte);
- KASSERT(TD_PA(tte.tte_data) == pa,
- ("pv_bit_count: corrupt alias chain"));
- if (tte.tte_data & bits)
- count++;
- pv_atomic_bit_clear(pstp, bits);
-#ifdef notyet
- generation = pv_generation;
- PV_UNLOCK();
- ipi_all(IPI_TLB_PAGE_DEMAP);
- PV_LOCK();
- if (generation != pv_generation)
- goto restart;
-#else
- tlb_page_demap(TLB_DTLB, TT_GET_CTX(tte.tte_tag), va);
-#endif
+ if ((pv = TAILQ_FIRST(&m->md.pv_list)) != NULL) {
+ pvf = pv;
+ do {
+ pvn = TAILQ_NEXT(pv, pv_list);
+ TAILQ_REMOVE(&m->md.pv_list, pv, pv_list);
+ TAILQ_INSERT_TAIL(&m->md.pv_list, pv, pv_list);
+ if (!pmap_track_modified(pv->pv_pmap, pv->pv_va))
+ continue;
+ tp = tsb_tte_lookup(pv->pv_pmap, pv->pv_va);
+ if (tp != NULL) {
+ if ((tp->tte_data & bits) != 0) {
+ atomic_clear_long(&tp->tte_data, bits);
+ if (++count > 4)
+ break;
+ }
+ }
+ } while ((pv = pvn) != NULL && pv != pvf);
}
- PV_UNLOCK();
return (count);
}
-void
-pv_bit_set(vm_page_t m, u_long bits)
+int
+pv_bit_test(vm_page_t m, u_long bits)
{
- vm_offset_t pstp;
- vm_offset_t pvh;
- vm_offset_t pa;
- vm_offset_t va;
- struct tte tte;
+ struct tte *tp;
+ pv_entry_t pv;
- pa = VM_PAGE_TO_PHYS(m);
- pvh = pv_lookup(pa);
- PV_LOCK();
-#ifdef notyet
-restart:
-#endif
- for (pstp = pvh_get_first(pvh); pstp != 0; pstp = pv_get_next(pstp)) {
- tte = pv_get_tte(pstp);
- va = tte_get_va(tte);
- KASSERT(TD_PA(tte.tte_data) == pa,
- ("pv_bit_set: corrupt alias chain"));
- if (tte.tte_data & bits)
- continue;
- pv_atomic_bit_set(pstp, bits);
-#ifdef notyet
- generation = pv_generation;
- PV_UNLOCK();
- /* XXX pass function and parameter to ipi call */
- ipi_all(IPI_TLB_PAGE_DEMAP);
- PV_LOCK();
- if (generation != pv_generation)
- goto restart;
-#else
- tlb_page_demap(TLB_DTLB, TT_GET_CTX(tte.tte_tag), va);
-#endif
+ TAILQ_FOREACH(pv, &m->md.pv_list, pv_list) {
+ if (bits & (TD_REF | TD_W)) {
+ if (!pmap_track_modified(pv->pv_pmap, pv->pv_va))
+ continue;
+ }
+ KASSERT(pv->pv_pmap != NULL, ("pv_bit_test: null pmap"));
+ if ((tp = tsb_tte_lookup(pv->pv_pmap, pv->pv_va)) != NULL) {
+ if (atomic_load_long(&tp->tte_data) & bits) {
+ return (TRUE);
+ }
+ }
}
- PV_UNLOCK();
+ return (FALSE);
}
int
-pv_bit_test(vm_page_t m, u_long bits)
+pv_page_exists(pmap_t pm, vm_page_t m)
{
- vm_offset_t pstp;
- vm_offset_t pvh;
- vm_offset_t pa;
+ pv_entry_t pv;
- pa = VM_PAGE_TO_PHYS(m);
- pvh = pv_lookup(pa);
- PV_LOCK();
- for (pstp = pvh_get_first(pvh); pstp != 0; pstp = pv_get_next(pstp)) {
- if (pv_atomic_bit_test(pstp, bits)) {
- PV_UNLOCK();
- return (1);
+ TAILQ_FOREACH(pv, &m->md.pv_list, pv_list) {
+ if (pv->pv_pmap == pm) {
+ return (TRUE);
}
}
- PV_UNLOCK();
- return (0);
+ return (FALSE);
}
void
-pv_global_remove_all(vm_page_t m)
+pv_remove_all(vm_page_t m)
{
- vm_offset_t pstp;
- vm_offset_t pvh;
- vm_offset_t pa;
-
- pa = VM_PAGE_TO_PHYS(m);
- pvh = pv_lookup(pa);
- PV_LOCK();
- for (pstp = pvh_get_first(pvh); pstp != 0; pstp = pv_get_next(pstp))
- pv_atomic_bit_clear(pstp, TD_V);
- PV_UNLOCK();
- pv_local_remove_all(pvh);
- PV_LOCK();
- while ((pstp = pvh_get_first(pvh)) != 0)
- pv_remove_phys(pstp);
- PV_UNLOCK();
-}
+ struct tte *tp;
+ pv_entry_t pv;
+ u_long data;
-static void
-pv_local_remove_all(vm_offset_t pvh)
-{
- vm_offset_t pstp;
- struct tte tte;
-
- PV_LOCK();
- for (pstp = pvh_get_first(pvh); pstp != 0; pstp = pv_get_next(pstp)) {
- tte = pv_get_tte(pstp);
- tsb_tte_local_remove(&tte);
+ KASSERT((m->flags & (PG_FICTITIOUS|PG_UNMANAGED)) == 0,
+ ("pv_remove_all: illegal for unmanaged page %#lx",
+ VM_PAGE_TO_PHYS(m)));
+ while ((pv = TAILQ_FIRST(&m->md.pv_list)) != NULL) {
+ tp = tsb_tte_lookup(pv->pv_pmap, pv->pv_va);
+ KASSERT(tp != NULL, ("pv_remove_all: mapping lost"));
+ data = atomic_load_long(&tp->tte_data);
+ if ((data & TD_WIRED) != 0)
+ pv->pv_pmap->pm_stats.wired_count--;
+ if ((data & TD_REF) != 0)
+ vm_page_flag_set(m, PG_REFERENCED);
+ if ((data & TD_W) != 0) {
+ if (pmap_track_modified(pv->pv_pmap, pv->pv_va))
+ vm_page_dirty(m);
+ }
+ atomic_clear_long(&tp->tte_data, TD_V);
+ tp->tte_tag = 0;
+ tp->tte_data = 0;
+ tlb_page_demap(TLB_DTLB | TLB_ITLB,
+ pv->pv_pmap->pm_context, pv->pv_va);
+ pv->pv_pmap->pm_stats.resident_count--;
+ m->md.pv_list_count--;
+ TAILQ_REMOVE(&pv->pv_pmap->pm_pvlist, pv, pv_plist);
+ TAILQ_REMOVE(&m->md.pv_list, pv, pv_list);
+ pmap_cache_remove(pv->pv_m, pv->pv_va);
+ pv_free(pv);
}
- PV_UNLOCK();
+ KASSERT(m->md.pv_list_count == 0,
+ ("pv_remove_all: leaking pv entries 0 != %d", m->md.pv_list_count));
+ KASSERT(TAILQ_EMPTY(&m->md.pv_list),
+ ("pv_remove_all: leaking pv entries"));
+ vm_page_flag_clear(m, PG_MAPPED | PG_WRITEABLE);
}