aboutsummaryrefslogtreecommitdiff
path: root/sys/geom/mirror/g_mirror.c
diff options
context:
space:
mode:
authorPawel Jakub Dawidek <pjd@FreeBSD.org>2004-07-30 23:13:45 +0000
committerPawel Jakub Dawidek <pjd@FreeBSD.org>2004-07-30 23:13:45 +0000
commitfa4a1febf719c322c3c7db0e6b45e827e6e116b9 (patch)
treee04bc051596e48d58d986a0c7491de41c9fd0a31 /sys/geom/mirror/g_mirror.c
parentbd58d1d222719382577564cecd3e879d08c0f239 (diff)
downloadsrc-fa4a1febf719c322c3c7db0e6b45e827e6e116b9.tar.gz
src-fa4a1febf719c322c3c7db0e6b45e827e6e116b9.zip
Add GEOM_MIRROR class which provide RAID1 functionality and has many useful
features. The gmirror(8) utility should be used for control of this class. There is no manual page yet, but I'm working on it with keramida@. Many useful tests provided by: simon (thank you!) Some ideas from: scottl, simon, phk
Notes
Notes: svn path=/head/; revision=132904
Diffstat (limited to 'sys/geom/mirror/g_mirror.c')
-rw-r--r--sys/geom/mirror/g_mirror.c2591
1 files changed, 2591 insertions, 0 deletions
diff --git a/sys/geom/mirror/g_mirror.c b/sys/geom/mirror/g_mirror.c
new file mode 100644
index 000000000000..132527b5ef77
--- /dev/null
+++ b/sys/geom/mirror/g_mirror.c
@@ -0,0 +1,2591 @@
+/*-
+ * Copyright (c) 2004 Pawel Jakub Dawidek <pjd@FreeBSD.org>
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+#include <sys/cdefs.h>
+__FBSDID("$FreeBSD$");
+
+#include <sys/param.h>
+#include <sys/systm.h>
+#include <sys/kernel.h>
+#include <sys/module.h>
+#include <sys/limits.h>
+#include <sys/lock.h>
+#include <sys/mutex.h>
+#include <sys/bio.h>
+#include <sys/sysctl.h>
+#include <sys/malloc.h>
+#include <sys/bitstring.h>
+#include <vm/uma.h>
+#include <machine/atomic.h>
+#include <geom/geom.h>
+#include <sys/proc.h>
+#include <sys/kthread.h>
+#include <geom/mirror/g_mirror.h>
+
+
+static MALLOC_DEFINE(M_MIRROR, "mirror data", "GEOM_MIRROR Data");
+
+SYSCTL_DECL(_kern_geom);
+SYSCTL_NODE(_kern_geom, OID_AUTO, mirror, CTLFLAG_RW, 0, "GEOM_MIRROR stuff");
+u_int g_mirror_debug = 0;
+SYSCTL_UINT(_kern_geom_mirror, OID_AUTO, debug, CTLFLAG_RW, &g_mirror_debug, 0,
+ "Debug level");
+static u_int g_mirror_sync_block_size = 131072;
+SYSCTL_UINT(_kern_geom_mirror, OID_AUTO, sync_block_size, CTLFLAG_RW,
+ &g_mirror_sync_block_size, 0, "Synchronization block size");
+static u_int g_mirror_timeout = 8;
+SYSCTL_UINT(_kern_geom_mirror, OID_AUTO, timeout, CTLFLAG_RW, &g_mirror_timeout,
+ 0, "Time to wait on all mirror components");
+static u_int g_mirror_reqs_per_sync = 5;
+SYSCTL_UINT(_kern_geom_mirror, OID_AUTO, reqs_per_sync, CTLFLAG_RW,
+ &g_mirror_reqs_per_sync, 0,
+ "Number of regular I/O requests per synchronization request");
+static u_int g_mirror_syncs_per_sec = 100;
+SYSCTL_UINT(_kern_geom_mirror, OID_AUTO, syncs_per_sec, CTLFLAG_RW,
+ &g_mirror_syncs_per_sec, 0,
+ "Number of synchronizations requests per second");
+
+#define MSLEEP(ident, mtx, priority, wmesg, timeout) do { \
+ G_MIRROR_DEBUG(4, "%s: Sleeping %p.", __func__, (ident)); \
+ msleep((ident), (mtx), (priority), (wmesg), (timeout)); \
+ G_MIRROR_DEBUG(4, "%s: Woken up %p.", __func__, (ident)); \
+} while (0)
+
+
+static int g_mirror_destroy_geom(struct gctl_req *req, struct g_class *mp,
+ struct g_geom *gp);
+static g_taste_t g_mirror_taste;
+
+struct g_class g_mirror_class = {
+ .name = G_MIRROR_CLASS_NAME,
+ .ctlreq = g_mirror_config,
+ .taste = g_mirror_taste,
+ .destroy_geom = g_mirror_destroy_geom
+};
+
+
+static void g_mirror_destroy_provider(struct g_mirror_softc *sc);
+static int g_mirror_update_disk(struct g_mirror_disk *disk, u_int state);
+static void g_mirror_update_device(struct g_mirror_softc *sc, boolean_t force);
+static void g_mirror_dumpconf(struct sbuf *sb, const char *indent,
+ struct g_geom *gp, struct g_consumer *cp, struct g_provider *pp);
+static void g_mirror_sync_stop(struct g_mirror_disk *disk, int type);
+
+
+static const char *
+g_mirror_disk_state2str(int state)
+{
+
+ switch (state) {
+ case G_MIRROR_DISK_STATE_NONE:
+ return ("NONE");
+ case G_MIRROR_DISK_STATE_NEW:
+ return ("NEW");
+ case G_MIRROR_DISK_STATE_ACTIVE:
+ return ("ACTIVE");
+ case G_MIRROR_DISK_STATE_STALE:
+ return ("STALE");
+ case G_MIRROR_DISK_STATE_SYNCHRONIZING:
+ return ("SYNCHRONIZING");
+ case G_MIRROR_DISK_STATE_DISCONNECTED:
+ return ("DISCONNECTED");
+ case G_MIRROR_DISK_STATE_DESTROY:
+ return ("DESTROY");
+ default:
+ return ("INVALID");
+ }
+}
+
+static const char *
+g_mirror_device_state2str(int state)
+{
+
+ switch (state) {
+ case G_MIRROR_DEVICE_STATE_STARTING:
+ return ("STARTING");
+ case G_MIRROR_DEVICE_STATE_RUNNING:
+ return ("RUNNING");
+ default:
+ return ("INVALID");
+ }
+}
+
+static const char *
+g_mirror_get_diskname(struct g_mirror_disk *disk)
+{
+
+ if (disk->d_consumer == NULL || disk->d_consumer->provider == NULL)
+ return ("[unknown]");
+ return (disk->d_name);
+}
+
+/*
+ * --- Events handling functions ---
+ * Events in geom_mirror are used to maintain disks and device status
+ * from one thread to simplify locking.
+ */
+static void
+g_mirror_event_free(struct g_mirror_event *ep)
+{
+
+ free(ep, M_MIRROR);
+}
+
+int
+g_mirror_event_send(void *arg, int state, int flags)
+{
+ struct g_mirror_softc *sc;
+ struct g_mirror_disk *disk;
+ struct g_mirror_event *ep;
+ int error;
+
+ ep = malloc(sizeof(*ep), M_MIRROR, M_WAITOK);
+ G_MIRROR_DEBUG(4, "%s: Sending event %p.", __func__, ep);
+ if ((flags & G_MIRROR_EVENT_DEVICE) != 0) {
+ disk = NULL;
+ sc = arg;
+ } else {
+ disk = arg;
+ sc = disk->d_softc;
+ }
+ ep->e_disk = disk;
+ ep->e_state = state;
+ ep->e_flags = flags;
+ ep->e_error = 0;
+ mtx_lock(&sc->sc_events_mtx);
+ TAILQ_INSERT_TAIL(&sc->sc_events, ep, e_next);
+ mtx_unlock(&sc->sc_events_mtx);
+ G_MIRROR_DEBUG(4, "%s: Waking up %p.", __func__, sc);
+ mtx_lock(&sc->sc_queue_mtx);
+ wakeup(sc);
+ mtx_unlock(&sc->sc_queue_mtx);
+ if ((flags & G_MIRROR_EVENT_DONTWAIT) != 0)
+ return (0);
+ g_topology_assert();
+ G_MIRROR_DEBUG(4, "%s: Sleeping %p.", __func__, ep);
+ g_topology_unlock();
+ while ((ep->e_flags & G_MIRROR_EVENT_DONE) == 0) {
+ mtx_lock(&sc->sc_events_mtx);
+ MSLEEP(ep, &sc->sc_events_mtx, PRIBIO | PDROP, "m:event",
+ hz * 5);
+ }
+ /* Don't even try to use 'sc' here, because it could be already dead. */
+ g_topology_lock();
+ error = ep->e_error;
+ g_mirror_event_free(ep);
+ return (error);
+}
+
+static struct g_mirror_event *
+g_mirror_event_get(struct g_mirror_softc *sc)
+{
+ struct g_mirror_event *ep;
+
+ mtx_lock(&sc->sc_events_mtx);
+ ep = TAILQ_FIRST(&sc->sc_events);
+ if (ep != NULL)
+ TAILQ_REMOVE(&sc->sc_events, ep, e_next);
+ mtx_unlock(&sc->sc_events_mtx);
+ return (ep);
+}
+
+static void
+g_mirror_event_cancel(struct g_mirror_disk *disk)
+{
+ struct g_mirror_softc *sc;
+ struct g_mirror_event *ep, *tmpep;
+
+ g_topology_assert();
+
+ sc = disk->d_softc;
+ mtx_lock(&sc->sc_events_mtx);
+ TAILQ_FOREACH_SAFE(ep, &sc->sc_events, e_next, tmpep) {
+ if ((ep->e_flags & G_MIRROR_EVENT_DEVICE) != 0)
+ continue;
+ if (ep->e_disk != disk)
+ continue;
+ TAILQ_REMOVE(&sc->sc_events, ep, e_next);
+ if ((ep->e_flags & G_MIRROR_EVENT_DONTWAIT) != 0)
+ g_mirror_event_free(ep);
+ else {
+ ep->e_error = ECANCELED;
+ wakeup(ep);
+ }
+ }
+ mtx_unlock(&sc->sc_events_mtx);
+}
+
+/*
+ * Return the number of disks in given state.
+ * If state is equal to -1, count all connected disks.
+ */
+u_int
+g_mirror_ndisks(struct g_mirror_softc *sc, int state)
+{
+ struct g_mirror_disk *disk;
+ u_int n = 0;
+
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (state == -1 || disk->d_state == state)
+ n++;
+ }
+ return (n);
+}
+
+/*
+ * Find a disk in mirror by its disk ID.
+ */
+static struct g_mirror_disk *
+g_mirror_id2disk(struct g_mirror_softc *sc, uint32_t id)
+{
+ struct g_mirror_disk *disk;
+
+ g_topology_assert();
+
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_id == id)
+ return (disk);
+ }
+ return (NULL);
+}
+
+static u_int
+g_mirror_nrequests(struct g_mirror_softc *sc, struct g_consumer *cp)
+{
+ struct bio *bp;
+ u_int nreqs = 0;
+
+ mtx_lock(&sc->sc_queue_mtx);
+ TAILQ_FOREACH(bp, &sc->sc_queue.queue, bio_queue) {
+ if (bp->bio_from == cp)
+ nreqs++;
+ }
+ mtx_unlock(&sc->sc_queue_mtx);
+ return (nreqs);
+}
+
+static void
+g_mirror_kill_consumer(struct g_mirror_softc *sc, struct g_consumer *cp)
+{
+
+ g_topology_assert();
+
+ cp->private = NULL;
+ if (cp->nstart != cp->nend) {
+ G_MIRROR_DEBUG(2,
+ "I/O requests for %s exist, can't destroy it now.",
+ cp->provider->name);
+ return;
+ }
+ if (g_mirror_nrequests(sc, cp) > 0) {
+ G_MIRROR_DEBUG(2,
+ "I/O requests for %s in queue, can't destroy it now.",
+ cp->provider->name);
+ return;
+ }
+ G_MIRROR_DEBUG(2, "Consumer %s destroyed.", cp->provider->name);
+ g_detach(cp);
+ g_destroy_consumer(cp);
+}
+
+static int
+g_mirror_connect_disk(struct g_mirror_disk *disk, struct g_provider *pp)
+{
+ int error;
+
+ g_topology_assert();
+ KASSERT(disk->d_consumer == NULL,
+ ("Disk already connected (device %s).", disk->d_softc->sc_name));
+
+ disk->d_consumer = g_new_consumer(disk->d_softc->sc_geom);
+ disk->d_consumer->private = disk;
+ error = g_attach(disk->d_consumer, pp);
+ if (error != 0)
+ return (error);
+ G_MIRROR_DEBUG(2, "Disk %s connected.", g_mirror_get_diskname(disk));
+ return (0);
+}
+
+static void
+g_mirror_disconnect_disk(struct g_mirror_disk *disk)
+{
+ struct g_consumer *cp;
+
+ g_topology_assert();
+
+ cp = disk->d_consumer;
+ if (cp == NULL)
+ return;
+ if (cp->provider != NULL) {
+ G_MIRROR_DEBUG(2, "Disk %s disconnected.",
+ g_mirror_get_diskname(disk));
+ if (cp->acr > 0 || cp->acw > 0 || cp->ace > 0) {
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d",
+ cp->provider->name, -cp->acr, -cp->acw, -cp->ace,
+ 0);
+ g_access(cp, -cp->acr, -cp->acw, -cp->ace);
+ }
+ g_mirror_kill_consumer(disk->d_softc, cp);
+ } else {
+ g_destroy_consumer(cp);
+ }
+}
+
+/*
+ * Initialize disk. This means allocate memory, create consumer, attach it
+ * to the provider and open access (r1w1e1) to it.
+ */
+static struct g_mirror_disk *
+g_mirror_init_disk(struct g_mirror_softc *sc, struct g_provider *pp,
+ struct g_mirror_metadata *md, int *errorp)
+{
+ struct g_mirror_disk *disk;
+ int error;
+
+ disk = malloc(sizeof(*disk), M_MIRROR, M_NOWAIT | M_ZERO);
+ if (disk == NULL) {
+ error = ENOMEM;
+ goto fail;
+ }
+ disk->d_softc = sc;
+ error = g_mirror_connect_disk(disk, pp);
+ if (error != 0)
+ goto fail;
+ disk->d_id = md->md_did;
+ disk->d_state = G_MIRROR_DISK_STATE_NONE;
+ disk->d_priority = md->md_priority;
+ disk->d_delay.sec = 0;
+ disk->d_delay.frac = 0;
+ binuptime(&disk->d_last_used);
+ disk->d_flags = md->md_dflags;
+ disk->d_sync.ds_consumer = NULL;
+ disk->d_sync.ds_offset = md->md_sync_offset;
+ disk->d_sync.ds_offset_done = md->md_sync_offset;
+ disk->d_sync.ds_syncid = md->md_syncid;
+ if (errorp != NULL)
+ *errorp = 0;
+ return (disk);
+fail:
+ if (errorp != NULL)
+ *errorp = error;
+ if (disk != NULL) {
+ g_mirror_disconnect_disk(disk);
+ free(disk, M_MIRROR);
+ }
+ return (NULL);
+}
+
+/*
+ * Free the disk.
+ */
+static void
+g_mirror_free_disk(struct g_mirror_disk *disk)
+{
+
+ g_topology_assert();
+
+ g_mirror_disconnect_disk(disk);
+ free(disk, M_MIRROR);
+}
+
+static void
+g_mirror_destroy_disk(struct g_mirror_disk *disk)
+{
+ struct g_mirror_softc *sc;
+
+ g_topology_assert();
+
+ LIST_REMOVE(disk, d_next);
+ g_mirror_event_cancel(disk);
+ sc = disk->d_softc;
+ if (sc->sc_hint == disk)
+ sc->sc_hint = NULL;
+ switch (disk->d_state) {
+ case G_MIRROR_DISK_STATE_SYNCHRONIZING:
+ g_mirror_sync_stop(disk, 1);
+ /* FALLTHROUGH */
+ case G_MIRROR_DISK_STATE_NEW:
+ case G_MIRROR_DISK_STATE_STALE:
+ case G_MIRROR_DISK_STATE_ACTIVE:
+ g_mirror_free_disk(disk);
+ break;
+ default:
+ KASSERT(0 == 1, ("Wrong disk state (%s, %s).",
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ }
+}
+
+static void
+g_mirror_destroy_device(struct g_mirror_softc *sc)
+{
+ struct g_mirror_disk *disk;
+ struct g_mirror_event *ep;
+ struct g_geom *gp;
+
+ g_topology_assert();
+
+ gp = sc->sc_geom;
+ if (sc->sc_provider != NULL)
+ g_mirror_destroy_provider(sc);
+ for (disk = LIST_FIRST(&sc->sc_disks); disk != NULL;
+ disk = LIST_FIRST(&sc->sc_disks)) {
+ g_mirror_destroy_disk(disk);
+ }
+ while ((ep = g_mirror_event_get(sc)) != NULL) {
+ if ((ep->e_flags & G_MIRROR_EVENT_DONTWAIT) != 0)
+ g_mirror_event_free(ep);
+ else {
+ ep->e_error = ECANCELED;
+ ep->e_flags |= G_MIRROR_EVENT_DONE;
+ G_MIRROR_DEBUG(4, "%s: Waking up %p.", __func__, ep);
+ mtx_lock(&sc->sc_events_mtx);
+ wakeup(ep);
+ mtx_unlock(&sc->sc_events_mtx);
+ }
+ }
+ callout_drain(&sc->sc_callout);
+ gp->softc = NULL;
+ uma_zdestroy(sc->sc_sync.ds_zone);
+ g_wither_geom(sc->sc_sync.ds_geom, ENXIO);
+ mtx_destroy(&sc->sc_queue_mtx);
+ mtx_destroy(&sc->sc_events_mtx);
+ G_MIRROR_DEBUG(0, "Device %s destroyed.", gp->name);
+ g_wither_geom(gp, ENXIO);
+}
+
+static void
+g_mirror_orphan(struct g_consumer *cp)
+{
+ struct g_mirror_disk *disk;
+
+ g_topology_assert();
+
+ disk = cp->private;
+ if (disk == NULL)
+ return;
+ disk->d_softc->sc_bump_syncid = G_MIRROR_BUMP_ON_FIRST_WRITE;
+ g_mirror_event_send(disk, G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+}
+
+/*
+ * Function should return the next active disk on the list.
+ * It is possible that it will be the same disk as given.
+ * If there are no active disks on list, NULL is returned.
+ */
+static __inline struct g_mirror_disk *
+g_mirror_find_next(struct g_mirror_softc *sc, struct g_mirror_disk *disk)
+{
+ struct g_mirror_disk *dp;
+
+ for (dp = LIST_NEXT(disk, d_next); dp != disk;
+ dp = LIST_NEXT(dp, d_next)) {
+ if (dp == NULL)
+ dp = LIST_FIRST(&sc->sc_disks);
+ if (dp->d_state == G_MIRROR_DISK_STATE_ACTIVE)
+ break;
+ }
+ if (dp->d_state != G_MIRROR_DISK_STATE_ACTIVE)
+ return (NULL);
+ return (dp);
+}
+
+static struct g_mirror_disk *
+g_mirror_get_disk(struct g_mirror_softc *sc)
+{
+ struct g_mirror_disk *disk;
+
+ if (sc->sc_hint == NULL) {
+ sc->sc_hint = LIST_FIRST(&sc->sc_disks);
+ if (sc->sc_hint == NULL)
+ return (NULL);
+ }
+ disk = sc->sc_hint;
+ if (disk->d_state != G_MIRROR_DISK_STATE_ACTIVE) {
+ disk = g_mirror_find_next(sc, disk);
+ if (disk == NULL)
+ return (NULL);
+ }
+ sc->sc_hint = g_mirror_find_next(sc, disk);
+ return (disk);
+}
+
+static int
+g_mirror_clear_metadata(struct g_mirror_disk *disk)
+{
+ struct g_mirror_softc *sc;
+ struct g_consumer *cp;
+ off_t offset, length;
+ u_char *sector;
+ int close = 0, error = 0;
+
+ g_topology_assert();
+
+ sc = disk->d_softc;
+ cp = disk->d_consumer;
+ KASSERT(cp != NULL, ("NULL consumer (%s).", sc->sc_name));
+ KASSERT(cp->provider != NULL, ("NULL provider (%s).", sc->sc_name));
+ length = cp->provider->sectorsize;
+ offset = cp->provider->mediasize - length;
+ sector = malloc((size_t)length, M_MIRROR, M_WAITOK | M_ZERO);
+ /*
+ * Open consumer if it wasn't opened and remember to close it.
+ */
+ if ((disk->d_flags & G_MIRROR_DISK_FLAG_DIRTY) == 0) {
+ error = g_access(cp, 0, 1, 1);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d",
+ cp->provider->name, 0, 1, 1, error);
+ if (error == 0)
+ close = 1;
+#ifdef INVARIANTS
+ } else {
+ KASSERT(cp->acw > 0 && cp->ace > 0,
+ ("Consumer %s not opened (r%dw%de%d).", cp->provider->name,
+ cp->acr, cp->acw, cp->ace));
+#endif
+ }
+ if (error == 0) {
+ g_topology_unlock();
+ error = g_write_data(cp, offset, sector, length);
+ g_topology_lock();
+ }
+ free(sector, M_MIRROR);
+ if (close) {
+ g_access(cp, 0, -1, -1);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d",
+ cp->provider->name, 0, -1, -1, 0);
+ }
+ if (error != 0) {
+ G_MIRROR_DEBUG(0, "Cannot clear metadata on disk %s.",
+ g_mirror_get_diskname(disk));
+ disk->d_softc->sc_bump_syncid = G_MIRROR_BUMP_IMMEDIATELY;
+ g_mirror_event_send(disk, G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+ return (error);
+ }
+ G_MIRROR_DEBUG(2, "Metadata on %s cleared.",
+ g_mirror_get_diskname(disk));
+ return (0);
+}
+
+void
+g_mirror_fill_metadata(struct g_mirror_softc *sc, struct g_mirror_disk *disk,
+ struct g_mirror_metadata *md)
+{
+
+ strlcpy(md->md_magic, G_MIRROR_MAGIC, sizeof(md->md_magic));
+ md->md_version = G_MIRROR_VERSION;
+ strlcpy(md->md_name, sc->sc_name, sizeof(md->md_name));
+ md->md_mid = sc->sc_id;
+ md->md_all = sc->sc_ndisks;
+ md->md_slice = sc->sc_slice;
+ md->md_balance = sc->sc_balance;
+ md->md_mediasize = sc->sc_mediasize;
+ md->md_sectorsize = sc->sc_sectorsize;
+ md->md_mflags = (sc->sc_flags & G_MIRROR_DEVICE_FLAG_MASK);
+ if (disk == NULL) {
+ md->md_did = arc4random();
+ md->md_priority = 0;
+ md->md_syncid = 0;
+ md->md_dflags = 0;
+ md->md_sync_offset = 0;
+ } else {
+ md->md_did = disk->d_id;
+ md->md_priority = disk->d_priority;
+ md->md_syncid = disk->d_sync.ds_syncid;
+ md->md_dflags = (disk->d_flags & G_MIRROR_DISK_FLAG_MASK);
+ if (disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING)
+ md->md_sync_offset = disk->d_sync.ds_offset_done;
+ else
+ md->md_sync_offset = 0;
+ }
+}
+
+void
+g_mirror_update_metadata(struct g_mirror_disk *disk)
+{
+ struct g_mirror_softc *sc;
+ struct g_mirror_metadata md;
+ struct g_consumer *cp;
+ off_t offset, length;
+ u_char *sector;
+ int close = 0, error = 0;
+
+ g_topology_assert();
+
+ sc = disk->d_softc;
+ cp = disk->d_consumer;
+ KASSERT(cp != NULL, ("NULL consumer (%s).", sc->sc_name));
+ KASSERT(cp->provider != NULL, ("NULL provider (%s).", sc->sc_name));
+ length = cp->provider->sectorsize;
+ offset = cp->provider->mediasize - length;
+ sector = malloc((size_t)length, M_MIRROR, M_WAITOK);
+ /*
+ * Open consumer if it wasn't opened and remember to close it.
+ */
+ if ((disk->d_flags & G_MIRROR_DISK_FLAG_DIRTY) == 0) {
+ error = g_access(cp, 0, 1, 1);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d",
+ cp->provider->name, 0, 1, 1, error);
+ if (error == 0)
+ close = 1;
+#ifdef INVARIANTS
+ } else {
+ KASSERT(cp->acw > 0 && cp->ace > 0,
+ ("Consumer %s not opened (r%dw%de%d).", cp->provider->name,
+ cp->acr, cp->acw, cp->ace));
+#endif
+ }
+ if (error == 0) {
+ g_mirror_fill_metadata(sc, disk, &md);
+ mirror_metadata_encode(&md, sector);
+ g_topology_unlock();
+ error = g_write_data(cp, offset, sector, length);
+ g_topology_lock();
+ }
+ free(sector, M_MIRROR);
+ if (close) {
+ g_access(cp, 0, -1, -1);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d",
+ cp->provider->name, 0, -1, -1, 0);
+ }
+ if (error != 0) {
+ G_MIRROR_DEBUG(0,
+ "Cannot update metadata on disk %s (error=%d).",
+ g_mirror_get_diskname(disk), error);
+ disk->d_softc->sc_bump_syncid = G_MIRROR_BUMP_IMMEDIATELY;
+ g_mirror_event_send(disk, G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+ return;
+ }
+ G_MIRROR_DEBUG(2, "Metadata on %s updated.",
+ g_mirror_get_diskname(disk));
+}
+
+static void
+g_mirror_bump_syncid(struct g_mirror_softc *sc)
+{
+ struct g_mirror_disk *disk;
+
+ g_topology_assert();
+ KASSERT(g_mirror_ndisks(sc, G_MIRROR_DISK_STATE_ACTIVE) > 0,
+ ("%s called with no active disks (device=%s).", __func__,
+ sc->sc_name));
+
+ sc->sc_syncid++;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_state == G_MIRROR_DISK_STATE_ACTIVE ||
+ disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING) {
+ disk->d_sync.ds_syncid = sc->sc_syncid;
+ g_mirror_update_metadata(disk);
+ }
+ }
+}
+
+static __inline int
+bintime_cmp(struct bintime *bt1, struct bintime *bt2)
+{
+
+ if (bt1->sec < bt2->sec)
+ return (-1);
+ else if (bt1->sec > bt2->sec)
+ return (1);
+ if (bt1->frac < bt2->frac)
+ return (-1);
+ else if (bt1->frac > bt2->frac)
+ return (1);
+ return (0);
+}
+
+static void
+g_mirror_update_delay(struct g_mirror_disk *disk, struct bio *bp)
+{
+
+ if (disk->d_softc->sc_balance != G_MIRROR_BALANCE_LOAD)
+ return;
+ binuptime(&disk->d_delay);
+ bintime_sub(&disk->d_delay, &bp->bio_t0);
+}
+
+static void
+g_mirror_done(struct bio *bp)
+{
+ struct g_mirror_softc *sc;
+
+ sc = bp->bio_from->geom->softc;
+ bp->bio_flags = BIO_FLAG1;
+ mtx_lock(&sc->sc_queue_mtx);
+ bioq_disksort(&sc->sc_queue, bp);
+ wakeup(sc);
+ mtx_unlock(&sc->sc_queue_mtx);
+}
+
+static void
+g_mirror_regular_request(struct bio *bp)
+{
+ struct g_mirror_softc *sc;
+ struct g_mirror_disk *disk;
+ struct bio *pbp;
+
+ g_topology_assert_not();
+
+ pbp = bp->bio_parent;
+ sc = pbp->bio_to->geom->softc;
+ disk = bp->bio_from->private;
+ if (disk == NULL) {
+ g_topology_lock();
+ g_mirror_kill_consumer(sc, bp->bio_from);
+ g_topology_unlock();
+ } else {
+ g_mirror_update_delay(disk, bp);
+ }
+
+ pbp->bio_inbed++;
+ KASSERT(pbp->bio_inbed <= pbp->bio_children,
+ ("bio_inbed (%u) is bigger than bio_children (%u).", pbp->bio_inbed,
+ pbp->bio_children));
+ if (bp->bio_error == 0 && pbp->bio_error == 0) {
+ G_MIRROR_LOGREQ(3, bp, "Request delivered.");
+ g_destroy_bio(bp);
+ if (pbp->bio_children == pbp->bio_inbed) {
+ G_MIRROR_LOGREQ(3, pbp, "Request delivered.");
+ pbp->bio_completed = pbp->bio_length;
+ g_io_deliver(pbp, pbp->bio_error);
+ }
+ return;
+ } else if (bp->bio_error != 0) {
+ if (pbp->bio_error == 0)
+ pbp->bio_error = bp->bio_error;
+ G_MIRROR_LOGREQ(0, bp, "Request failed (error=%d).",
+ bp->bio_error);
+ if (disk != NULL) {
+ sc->sc_bump_syncid = G_MIRROR_BUMP_IMMEDIATELY;
+ g_mirror_event_send(disk,
+ G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+ }
+ switch (pbp->bio_cmd) {
+ case BIO_DELETE:
+ case BIO_WRITE:
+ pbp->bio_inbed--;
+ pbp->bio_children--;
+ break;
+ }
+ }
+ g_destroy_bio(bp);
+
+ switch (pbp->bio_cmd) {
+ case BIO_READ:
+ if (pbp->bio_children == pbp->bio_inbed) {
+ pbp->bio_error = 0;
+ mtx_lock(&sc->sc_queue_mtx);
+ bioq_disksort(&sc->sc_queue, pbp);
+ G_MIRROR_DEBUG(4, "%s: Waking up %p.", __func__, sc);
+ wakeup(sc);
+ mtx_unlock(&sc->sc_queue_mtx);
+ }
+ break;
+ case BIO_DELETE:
+ case BIO_WRITE:
+ if (pbp->bio_children == 0) {
+ /*
+ * All requests failed.
+ */
+ } else if (pbp->bio_inbed < pbp->bio_children) {
+ /* Do nothing. */
+ break;
+ } else if (pbp->bio_children == pbp->bio_inbed) {
+ /* Some requests succeeded. */
+ pbp->bio_error = 0;
+ pbp->bio_completed = pbp->bio_length;
+ }
+ g_io_deliver(pbp, pbp->bio_error);
+ break;
+ default:
+ KASSERT(1 == 0, ("Invalid request: %u.", pbp->bio_cmd));
+ break;
+ }
+}
+
+static void
+g_mirror_sync_done(struct bio *bp)
+{
+ struct g_mirror_softc *sc;
+
+ G_MIRROR_LOGREQ(3, bp, "Synchronization request delivered.");
+ sc = bp->bio_from->geom->softc;
+ bp->bio_flags = BIO_FLAG2;
+ mtx_lock(&sc->sc_queue_mtx);
+ bioq_disksort(&sc->sc_queue, bp);
+ wakeup(sc);
+ mtx_unlock(&sc->sc_queue_mtx);
+}
+
+static void
+g_mirror_start(struct bio *bp)
+{
+ struct g_mirror_softc *sc;
+
+ sc = bp->bio_to->geom->softc;
+ /*
+ * If sc == NULL or there are no valid disks, provider's error
+ * should be set and g_mirror_start() should not be called at all.
+ */
+ KASSERT(sc != NULL && sc->sc_state == G_MIRROR_DEVICE_STATE_RUNNING,
+ ("Provider's error should be set (error=%d)(mirror=%s).",
+ bp->bio_to->error, bp->bio_to->name));
+ G_MIRROR_LOGREQ(3, bp, "Request received.");
+
+ switch (bp->bio_cmd) {
+ case BIO_READ:
+ case BIO_WRITE:
+ case BIO_DELETE:
+ break;
+ case BIO_GETATTR:
+ default:
+ g_io_deliver(bp, EOPNOTSUPP);
+ return;
+ }
+ mtx_lock(&sc->sc_queue_mtx);
+ bioq_disksort(&sc->sc_queue, bp);
+ G_MIRROR_DEBUG(4, "%s: Waking up %p.", __func__, sc);
+ wakeup(sc);
+ mtx_unlock(&sc->sc_queue_mtx);
+}
+
+/*
+ * Send one synchronization request.
+ */
+static void
+g_mirror_sync_one(struct g_mirror_disk *disk)
+{
+ struct g_mirror_softc *sc;
+ struct bio *bp;
+
+ sc = disk->d_softc;
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING,
+ ("Disk %s is not marked for synchronization.",
+ g_mirror_get_diskname(disk)));
+
+ bp = g_new_bio();
+ if (bp == NULL)
+ return;
+ bp->bio_parent = NULL;
+ bp->bio_cmd = BIO_READ;
+ bp->bio_offset = disk->d_sync.ds_offset;
+ bp->bio_length = MIN(sc->sc_sync.ds_block,
+ sc->sc_mediasize - bp->bio_offset);
+ bp->bio_flags = 0;
+ bp->bio_done = g_mirror_sync_done;
+ bp->bio_data = uma_zalloc(sc->sc_sync.ds_zone, M_NOWAIT | M_ZERO);
+ if (bp->bio_data == NULL) {
+ g_destroy_bio(bp);
+ return;
+ }
+ disk->d_sync.ds_offset += bp->bio_length;
+ bp->bio_to = sc->sc_provider;
+ G_MIRROR_LOGREQ(3, bp, "Sending synchronization request.");
+ g_io_request(bp, disk->d_sync.ds_consumer);
+}
+
+static void
+g_mirror_sync_request(struct bio *bp)
+{
+ struct g_mirror_softc *sc;
+ struct g_mirror_disk *disk;
+
+ sc = bp->bio_from->geom->softc;
+ disk = bp->bio_from->private;
+ if (disk == NULL) {
+ g_topology_lock();
+ g_mirror_kill_consumer(sc, bp->bio_from);
+ g_topology_unlock();
+ uma_zfree(sc->sc_sync.ds_zone, bp->bio_data);
+ g_destroy_bio(bp);
+ return;
+ }
+
+ /*
+ * Synchronization request.
+ */
+ switch (bp->bio_cmd) {
+ case BIO_READ:
+ {
+ struct g_consumer *cp;
+
+ if (bp->bio_error != 0) {
+ G_MIRROR_LOGREQ(0, bp,
+ "Synchronization request failed (error=%d).",
+ bp->bio_error);
+ uma_zfree(sc->sc_sync.ds_zone, bp->bio_data);
+ g_destroy_bio(bp);
+ return;
+ }
+ bp->bio_cmd = BIO_WRITE;
+ bp->bio_flags = 0;
+ G_MIRROR_LOGREQ(3, bp, "Synchronization request finished.");
+ cp = disk->d_consumer;
+ KASSERT(cp->acr == 0 && cp->acw == 1 && cp->ace == 1,
+ ("Consumer %s not opened (r%dw%de%d).", cp->provider->name,
+ cp->acr, cp->acw, cp->ace));
+ g_io_request(bp, cp);
+ return;
+ }
+ case BIO_WRITE:
+ uma_zfree(sc->sc_sync.ds_zone, bp->bio_data);
+ if (bp->bio_error != 0) {
+ G_MIRROR_LOGREQ(0, bp,
+ "Synchronization request failed (error=%d).",
+ bp->bio_error);
+ g_destroy_bio(bp);
+ disk->d_softc->sc_bump_syncid = G_MIRROR_BUMP_IMMEDIATELY;
+ g_mirror_event_send(disk,
+ G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+ return;
+ }
+ G_MIRROR_LOGREQ(3, bp, "Synchronization request finished.");
+ g_destroy_bio(bp);
+ disk->d_sync.ds_offset_done = bp->bio_offset + bp->bio_length;
+ if (bp->bio_offset + bp->bio_length ==
+ sc->sc_provider->mediasize) {
+ /*
+ * Disk up-to-date, activate it.
+ */
+ g_mirror_event_send(disk, G_MIRROR_DISK_STATE_ACTIVE,
+ G_MIRROR_EVENT_DONTWAIT);
+ return;
+ } else if ((disk->d_sync.ds_offset_done %
+ (sc->sc_sync.ds_block * 100)) == 0) {
+ /*
+ * Update offset_done on every 100 blocks.
+ * XXX: This should be configurable.
+ */
+ g_topology_lock();
+ g_mirror_update_metadata(disk);
+ g_topology_unlock();
+ }
+ return;
+ default:
+ KASSERT(1 == 0, ("Invalid command here: %u (device=%s)",
+ bp->bio_cmd, sc->sc_name));
+ break;
+ }
+}
+
+static void
+g_mirror_request_round_robin(struct g_mirror_softc *sc, struct bio *bp)
+{
+ struct g_mirror_disk *disk;
+ struct g_consumer *cp;
+ struct bio *cbp;
+
+ disk = g_mirror_get_disk(sc);
+ if (disk == NULL) {
+ if (bp->bio_error == 0)
+ bp->bio_error = ENXIO;
+ g_io_deliver(bp, bp->bio_error);
+ return;
+ }
+ cbp = g_clone_bio(bp);
+ if (cbp == NULL) {
+ if (bp->bio_error == 0)
+ bp->bio_error = ENOMEM;
+ g_io_deliver(bp, bp->bio_error);
+ return;
+ }
+ /*
+ * Fill in the component buf structure.
+ */
+ cp = disk->d_consumer;
+ cbp->bio_done = g_mirror_done;
+ cbp->bio_to = cp->provider;
+ G_MIRROR_LOGREQ(3, cbp, "Sending request.");
+ KASSERT(cp->acr > 0 && cp->ace > 0,
+ ("Consumer %s not opened (r%dw%de%d).", cp->provider->name, cp->acr,
+ cp->acw, cp->ace));
+ g_io_request(cbp, cp);
+}
+
+static void
+g_mirror_request_load(struct g_mirror_softc *sc, struct bio *bp)
+{
+ struct g_mirror_disk *disk, *dp;
+ struct g_consumer *cp;
+ struct bio *cbp;
+ struct bintime curtime;
+
+ binuptime(&curtime);
+ /*
+ * Find a disk which the smallest load.
+ */
+ disk = NULL;
+ LIST_FOREACH(dp, &sc->sc_disks, d_next) {
+ if (dp->d_state != G_MIRROR_DISK_STATE_ACTIVE)
+ continue;
+ /* If disk wasn't used for more than 2 sec, use it. */
+ if (curtime.sec - dp->d_last_used.sec >= 2) {
+ disk = dp;
+ break;
+ }
+ if (disk == NULL ||
+ bintime_cmp(&dp->d_delay, &disk->d_delay) < 0) {
+ disk = dp;
+ }
+ }
+ cbp = g_clone_bio(bp);
+ if (cbp == NULL) {
+ if (bp->bio_error == 0)
+ bp->bio_error = ENOMEM;
+ g_io_deliver(bp, bp->bio_error);
+ return;
+ }
+ /*
+ * Fill in the component buf structure.
+ */
+ cp = disk->d_consumer;
+ cbp->bio_done = g_mirror_done;
+ cbp->bio_to = cp->provider;
+ binuptime(&disk->d_last_used);
+ G_MIRROR_LOGREQ(3, cbp, "Sending request.");
+ KASSERT(cp->acr > 0 && cp->ace > 0,
+ ("Consumer %s not opened (r%dw%de%d).", cp->provider->name, cp->acr,
+ cp->acw, cp->ace));
+ g_io_request(cbp, cp);
+}
+
+static void
+g_mirror_request_split(struct g_mirror_softc *sc, struct bio *bp)
+{
+ struct bio_queue_head queue;
+ struct g_mirror_disk *disk;
+ struct g_consumer *cp;
+ struct bio *cbp;
+ off_t left, mod, offset, slice;
+ u_char *data;
+ u_int ndisks;
+
+ if (bp->bio_length <= sc->sc_slice) {
+ g_mirror_request_round_robin(sc, bp);
+ return;
+ }
+ ndisks = g_mirror_ndisks(sc, G_MIRROR_DISK_STATE_ACTIVE);
+ slice = bp->bio_length / ndisks;
+ mod = slice % sc->sc_provider->sectorsize;
+ if (mod != 0)
+ slice += sc->sc_provider->sectorsize - mod;
+ /*
+ * Allocate all bios before sending any request, so we can
+ * return ENOMEM in nice and clean way.
+ */
+ left = bp->bio_length;
+ offset = bp->bio_offset;
+ data = bp->bio_data;
+ bioq_init(&queue);
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_state != G_MIRROR_DISK_STATE_ACTIVE)
+ continue;
+ cbp = g_clone_bio(bp);
+ if (cbp == NULL) {
+ for (cbp = bioq_first(&queue); cbp != NULL;
+ cbp = bioq_first(&queue)) {
+ bioq_remove(&queue, cbp);
+ g_destroy_bio(cbp);
+ }
+ if (bp->bio_error == 0)
+ bp->bio_error = ENOMEM;
+ g_io_deliver(bp, bp->bio_error);
+ return;
+ }
+ bioq_insert_tail(&queue, cbp);
+ cbp->bio_done = g_mirror_done;
+ cbp->bio_caller1 = disk;
+ cbp->bio_to = disk->d_consumer->provider;
+ cbp->bio_offset = offset;
+ cbp->bio_data = data;
+ cbp->bio_length = MIN(left, slice);
+ left -= cbp->bio_length;
+ if (left == 0)
+ break;
+ offset += cbp->bio_length;
+ data += cbp->bio_length;
+ }
+ for (cbp = bioq_first(&queue); cbp != NULL; cbp = bioq_first(&queue)) {
+ bioq_remove(&queue, cbp);
+ G_MIRROR_LOGREQ(3, cbp, "Sending request.");
+ disk = cbp->bio_caller1;
+ cbp->bio_caller1 = NULL;
+ cp = disk->d_consumer;
+ KASSERT(cp->acr > 0 && cp->ace > 0,
+ ("Consumer %s not opened (r%dw%de%d).", cp->provider->name,
+ cp->acr, cp->acw, cp->ace));
+ g_io_request(cbp, disk->d_consumer);
+ }
+}
+
+static void
+g_mirror_register_request(struct bio *bp)
+{
+ struct g_mirror_softc *sc;
+
+ sc = bp->bio_to->geom->softc;
+ switch (bp->bio_cmd) {
+ case BIO_READ:
+ switch (sc->sc_balance) {
+ case G_MIRROR_BALANCE_ROUND_ROBIN:
+ g_mirror_request_round_robin(sc, bp);
+ break;
+ case G_MIRROR_BALANCE_LOAD:
+ g_mirror_request_load(sc, bp);
+ break;
+ case G_MIRROR_BALANCE_SPLIT:
+ g_mirror_request_split(sc, bp);
+ break;
+ }
+ return;
+ case BIO_WRITE:
+ case BIO_DELETE:
+ {
+ struct g_mirror_disk *disk;
+ struct bio_queue_head queue;
+ struct g_consumer *cp;
+ struct bio *cbp;
+
+ /*
+ * Allocate all bios before sending any request, so we can
+ * return ENOMEM in nice and clean way.
+ */
+ bioq_init(&queue);
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ switch (disk->d_state) {
+ case G_MIRROR_DISK_STATE_ACTIVE:
+ break;
+ case G_MIRROR_DISK_STATE_SYNCHRONIZING:
+ if (bp->bio_offset >= disk->d_sync.ds_offset)
+ continue;
+ break;
+ default:
+ continue;
+ }
+ cbp = g_clone_bio(bp);
+ if (cbp == NULL) {
+ for (cbp = bioq_first(&queue); cbp != NULL;
+ cbp = bioq_first(&queue)) {
+ bioq_remove(&queue, cbp);
+ g_destroy_bio(cbp);
+ }
+ if (bp->bio_error == 0)
+ bp->bio_error = ENOMEM;
+ g_io_deliver(bp, bp->bio_error);
+ return;
+ }
+ bioq_insert_tail(&queue, cbp);
+ }
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ switch (disk->d_state) {
+ case G_MIRROR_DISK_STATE_ACTIVE:
+ break;
+ case G_MIRROR_DISK_STATE_SYNCHRONIZING:
+ if (bp->bio_offset >= disk->d_sync.ds_offset)
+ continue;
+ break;
+ default:
+ continue;
+ }
+ cbp = bioq_first(&queue);
+ KASSERT(cbp != NULL, ("NULL cbp! (device %s).",
+ sc->sc_name));
+ bioq_remove(&queue, cbp);
+ cp = disk->d_consumer;
+ cbp->bio_done = g_mirror_done;
+ cbp->bio_to = cp->provider;
+ G_MIRROR_LOGREQ(3, cbp, "Sending request.");
+ KASSERT(cp->acw > 0 && cp->ace > 0,
+ ("Consumer %s not opened (r%dw%de%d).",
+ cp->provider->name, cp->acr, cp->acw, cp->ace));
+ g_io_request(cbp, cp);
+ }
+ /*
+ * Bump syncid on first write.
+ */
+ if (sc->sc_bump_syncid == G_MIRROR_BUMP_ON_FIRST_WRITE) {
+ sc->sc_bump_syncid = 0;
+ g_topology_lock();
+ g_mirror_bump_syncid(sc);
+ g_topology_unlock();
+ }
+ return;
+ }
+ default:
+ KASSERT(1 == 0, ("Invalid command here: %u (device=%s)",
+ bp->bio_cmd, sc->sc_name));
+ break;
+ }
+}
+
+/*
+ * Worker thread.
+ */
+static void
+g_mirror_worker(void *arg)
+{
+ struct g_mirror_softc *sc;
+ struct g_mirror_disk *disk;
+ struct g_mirror_event *ep;
+ struct bio *bp;
+ u_int nreqs;
+
+ sc = arg;
+ curthread->td_base_pri = PRIBIO;
+
+ nreqs = 0;
+ for (;;) {
+ G_MIRROR_DEBUG(5, "%s: Let's see...", __func__);
+ /*
+ * First take a look at events.
+ * This is important to handle events before any I/O requests.
+ */
+ ep = g_mirror_event_get(sc);
+ if (ep != NULL) {
+ g_topology_lock();
+ if ((ep->e_flags & G_MIRROR_EVENT_DEVICE) != 0) {
+ /* Update only device status. */
+ G_MIRROR_DEBUG(3,
+ "Running event for device %s.",
+ sc->sc_name);
+ ep->e_error = 0;
+ g_mirror_update_device(sc, 1);
+ } else {
+ /* Update disk status. */
+ G_MIRROR_DEBUG(3, "Running event for disk %s.",
+ g_mirror_get_diskname(ep->e_disk));
+ ep->e_error = g_mirror_update_disk(ep->e_disk,
+ ep->e_state);
+ if (ep->e_error == 0)
+ g_mirror_update_device(sc, 0);
+ }
+ g_topology_unlock();
+ if ((ep->e_flags & G_MIRROR_EVENT_DONTWAIT) != 0) {
+ KASSERT(ep->e_error == 0,
+ ("Error cannot be handled."));
+ g_mirror_event_free(ep);
+ } else {
+ ep->e_flags |= G_MIRROR_EVENT_DONE;
+ G_MIRROR_DEBUG(4, "%s: Waking up %p.", __func__,
+ ep);
+ mtx_lock(&sc->sc_events_mtx);
+ wakeup(ep);
+ mtx_unlock(&sc->sc_events_mtx);
+ }
+ if ((sc->sc_flags &
+ G_MIRROR_DEVICE_FLAG_DESTROY) != 0) {
+end:
+ if ((sc->sc_flags &
+ G_MIRROR_DEVICE_FLAG_WAIT) != 0) {
+ G_MIRROR_DEBUG(4, "%s: Waking up %p.",
+ __func__, &sc->sc_worker);
+ wakeup(&sc->sc_worker);
+ sc->sc_worker = NULL;
+ } else {
+ g_topology_lock();
+ g_mirror_destroy_device(sc);
+ g_topology_unlock();
+ free(sc, M_MIRROR);
+ }
+ kthread_exit(0);
+ }
+ G_MIRROR_DEBUG(5, "%s: I'm here 1.", __func__);
+ continue;
+ }
+ /*
+ * Now I/O requests.
+ */
+ /* Get first request from the queue. */
+ mtx_lock(&sc->sc_queue_mtx);
+ bp = bioq_first(&sc->sc_queue);
+ if (bp == NULL) {
+ if ((sc->sc_flags &
+ G_MIRROR_DEVICE_FLAG_DESTROY) != 0) {
+ mtx_unlock(&sc->sc_queue_mtx);
+ goto end;
+ }
+ }
+ if (sc->sc_sync.ds_ndisks > 0 &&
+ (bp == NULL || nreqs > g_mirror_reqs_per_sync)) {
+ mtx_unlock(&sc->sc_queue_mtx);
+ /*
+ * It is time for synchronization...
+ */
+ nreqs = 0;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_state !=
+ G_MIRROR_DISK_STATE_SYNCHRONIZING) {
+ continue;
+ }
+ if (disk->d_sync.ds_offset >=
+ sc->sc_provider->mediasize) {
+ continue;
+ }
+ if (disk->d_sync.ds_offset >
+ disk->d_sync.ds_offset_done) {
+ continue;
+ }
+ g_mirror_sync_one(disk);
+ }
+ G_MIRROR_DEBUG(5, "%s: I'm here 2.", __func__);
+ goto sleep;
+ }
+ if (bp == NULL) {
+ MSLEEP(sc, &sc->sc_queue_mtx, PRIBIO | PDROP, "m:w1", 0);
+ G_MIRROR_DEBUG(5, "%s: I'm here 3.", __func__);
+ continue;
+ }
+ nreqs++;
+ bioq_remove(&sc->sc_queue, bp);
+ mtx_unlock(&sc->sc_queue_mtx);
+
+ if ((bp->bio_flags & BIO_FLAG1) != 0) {
+ g_mirror_regular_request(bp);
+ } else if ((bp->bio_flags & BIO_FLAG2) != 0) {
+ u_int timeout, sps;
+
+ g_mirror_sync_request(bp);
+sleep:
+ sps = atomic_load_acq_int(&g_mirror_syncs_per_sec);
+ if (sps == 0) {
+ G_MIRROR_DEBUG(5, "%s: I'm here 5.", __func__);
+ continue;
+ }
+ mtx_lock(&sc->sc_queue_mtx);
+ if (bioq_first(&sc->sc_queue) != NULL) {
+ mtx_unlock(&sc->sc_queue_mtx);
+ G_MIRROR_DEBUG(5, "%s: I'm here 4.", __func__);
+ continue;
+ }
+ timeout = hz / sps;
+ if (timeout == 0)
+ timeout = 1;
+ MSLEEP(sc, &sc->sc_queue_mtx, PRIBIO | PDROP, "m:w2",
+ timeout);
+ } else {
+ g_mirror_register_request(bp);
+ }
+ G_MIRROR_DEBUG(5, "%s: I'm here 6.", __func__);
+ }
+}
+
+/*
+ * Open disk's consumer if needed.
+ */
+static void
+g_mirror_update_access(struct g_mirror_disk *disk)
+{
+ struct g_provider *pp;
+ struct g_consumer *cp;
+ int acr, acw, ace, cpw, error;
+
+ g_topology_assert();
+
+ cp = disk->d_consumer;
+ pp = disk->d_softc->sc_provider;
+ if (pp == NULL) {
+ acr = -cp->acr;
+ acw = -cp->acw;
+ ace = -cp->ace;
+ } else {
+ acr = pp->acr - cp->acr;
+ acw = pp->acw - cp->acw;
+ ace = pp->ace - cp->ace;
+ /* Grab an extra "exclusive" bit. */
+ if (pp->acr > 0 || pp->acw > 0 || pp->ace > 0)
+ ace++;
+ }
+ if (acr == 0 && acw == 0 && ace == 0)
+ return;
+ cpw = cp->acw;
+ error = g_access(cp, acr, acw, ace);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d", cp->provider->name, acr,
+ acw, ace, error);
+ if (error != 0) {
+ disk->d_softc->sc_bump_syncid = G_MIRROR_BUMP_ON_FIRST_WRITE;
+ g_mirror_event_send(disk, G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+ return;
+ }
+ if (cpw == 0 && cp->acw > 0) {
+ G_MIRROR_DEBUG(1, "Disk %s (device %s) marked as dirty.",
+ g_mirror_get_diskname(disk), disk->d_softc->sc_name);
+ disk->d_flags |= G_MIRROR_DISK_FLAG_DIRTY;
+ } else if (cpw > 0 && cp->acw == 0) {
+ G_MIRROR_DEBUG(1, "Disk %s (device %s) marked as clean.",
+ g_mirror_get_diskname(disk), disk->d_softc->sc_name);
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_DIRTY;
+ }
+}
+
+static void
+g_mirror_sync_start(struct g_mirror_disk *disk)
+{
+ struct g_mirror_softc *sc;
+ struct g_consumer *cp;
+ int error;
+
+ g_topology_assert();
+
+ sc = disk->d_softc;
+ KASSERT(sc->sc_state == G_MIRROR_DEVICE_STATE_RUNNING,
+ ("Device not in RUNNING state (%s, %u).", sc->sc_name,
+ sc->sc_state));
+ cp = disk->d_consumer;
+ KASSERT(cp->acr == 0 && cp->acw == 0 && cp->ace == 0,
+ ("Consumer %s already opened.", cp->provider->name));
+
+ G_MIRROR_DEBUG(0, "Device %s: rebuilding provider %s.", sc->sc_name,
+ g_mirror_get_diskname(disk));
+ error = g_access(cp, 0, 1, 1);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d", cp->provider->name, 0, 1,
+ 1, error);
+ if (error != 0) {
+ g_mirror_event_send(disk, G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+ return;
+ }
+ disk->d_flags |= G_MIRROR_DISK_FLAG_DIRTY;
+ KASSERT(disk->d_sync.ds_consumer == NULL,
+ ("Sync consumer already exists (device=%s, disk=%s).",
+ sc->sc_name, g_mirror_get_diskname(disk)));
+ disk->d_sync.ds_consumer = g_new_consumer(sc->sc_sync.ds_geom);
+ disk->d_sync.ds_consumer->private = disk;
+ error = g_attach(disk->d_sync.ds_consumer, disk->d_softc->sc_provider);
+ KASSERT(error == 0, ("Cannot attach to %s (error=%d).",
+ disk->d_softc->sc_name, error));
+ error = g_access(disk->d_sync.ds_consumer, 1, 0, 0);
+ KASSERT(error == 0, ("Cannot open %s (error=%d).",
+ disk->d_softc->sc_name, error));
+ sc->sc_sync.ds_ndisks++;
+}
+
+/*
+ * Stop synchronization process.
+ * type: 0 - synchronization finished
+ * 1 - synchronization stopped
+ */
+static void
+g_mirror_sync_stop(struct g_mirror_disk *disk, int type)
+{
+ struct g_consumer *cp;
+
+ g_topology_assert();
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING,
+ ("Wrong disk state (%s, %s).", g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ if (disk->d_sync.ds_consumer == NULL)
+ return;
+
+ if (type == 0) {
+ G_MIRROR_DEBUG(0, "Device %s: rebuilding provider %s finished.",
+ disk->d_softc->sc_name, g_mirror_get_diskname(disk));
+ } else /* if (type == 1) */ {
+ G_MIRROR_DEBUG(0, "Device %s: rebuilding provider %s stopped.",
+ disk->d_softc->sc_name, g_mirror_get_diskname(disk));
+ }
+ cp = disk->d_sync.ds_consumer;
+ g_access(cp, -1, 0, 0);
+ g_mirror_kill_consumer(disk->d_softc, cp);
+ disk->d_sync.ds_consumer = NULL;
+ disk->d_softc->sc_sync.ds_ndisks--;
+ cp = disk->d_consumer;
+ KASSERT(cp->acr == 0 && cp->acw == 1 && cp->ace == 1,
+ ("Consumer %s not opened.", cp->provider->name));
+ g_access(cp, 0, -1, -1);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d", cp->provider->name, 0, -1,
+ -1, 0);
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_DIRTY;
+}
+
+static void
+g_mirror_launch_provider(struct g_mirror_softc *sc)
+{
+ struct g_mirror_disk *disk;
+ struct g_provider *pp;
+
+ g_topology_assert();
+
+ pp = g_new_providerf(sc->sc_geom, "mirror/%s", sc->sc_name);
+ pp->mediasize = sc->sc_mediasize;
+ pp->sectorsize = sc->sc_sectorsize;
+ sc->sc_provider = pp;
+ g_error_provider(pp, 0);
+ G_MIRROR_DEBUG(0, "Device %s: provider %s launched.", sc->sc_name,
+ pp->name);
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING)
+ g_mirror_sync_start(disk);
+ }
+}
+
+static void
+g_mirror_destroy_provider(struct g_mirror_softc *sc)
+{
+ struct g_mirror_disk *disk;
+ struct bio *bp;
+
+ g_topology_assert();
+ KASSERT(sc->sc_provider != NULL, ("NULL provider (device=%s).",
+ sc->sc_name));
+
+ g_error_provider(sc->sc_provider, ENXIO);
+ mtx_lock(&sc->sc_queue_mtx);
+ while ((bp = bioq_first(&sc->sc_queue)) != NULL) {
+ bioq_remove(&sc->sc_queue, bp);
+ g_io_deliver(bp, ENXIO);
+ }
+ mtx_unlock(&sc->sc_queue_mtx);
+ G_MIRROR_DEBUG(0, "Device %s: provider %s destroyed.", sc->sc_name,
+ sc->sc_provider->name);
+ sc->sc_provider->flags |= G_PF_WITHER;
+ g_orphan_provider(sc->sc_provider, ENXIO);
+ sc->sc_provider = NULL;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING)
+ g_mirror_sync_stop(disk, 1);
+ }
+}
+
+static void
+g_mirror_go(void *arg)
+{
+ struct g_mirror_softc *sc;
+
+ sc = arg;
+ G_MIRROR_DEBUG(0, "Force device %s start due to timeout.", sc->sc_name);
+ g_mirror_event_send(sc, 0,
+ G_MIRROR_EVENT_DONTWAIT | G_MIRROR_EVENT_DEVICE);
+}
+
+static u_int
+g_mirror_determine_state(struct g_mirror_disk *disk)
+{
+ struct g_mirror_softc *sc;
+ u_int state;
+
+ sc = disk->d_softc;
+ if (sc->sc_syncid == disk->d_sync.ds_syncid) {
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_SYNCHRONIZING) == 0) {
+ /* Disk does not need synchronization. */
+ state = G_MIRROR_DISK_STATE_ACTIVE;
+ } else {
+ if ((sc->sc_flags &
+ G_MIRROR_DEVICE_FLAG_NOAUTOSYNC) == 0 ||
+ (disk->d_flags &
+ G_MIRROR_DISK_FLAG_FORCE_SYNC) != 0) {
+ /*
+ * We can start synchronization from
+ * the stored offset.
+ */
+ state = G_MIRROR_DISK_STATE_SYNCHRONIZING;
+ } else {
+ state = G_MIRROR_DISK_STATE_STALE;
+ }
+ }
+ } else if (disk->d_sync.ds_syncid < sc->sc_syncid) {
+ /*
+ * Reset all synchronization data for this disk,
+ * because if it even was synchronized, it was
+ * synchronized to disks with different syncid.
+ */
+ disk->d_flags |= G_MIRROR_DISK_FLAG_SYNCHRONIZING;
+ disk->d_sync.ds_offset = 0;
+ disk->d_sync.ds_offset_done = 0;
+ disk->d_sync.ds_syncid = sc->sc_syncid;
+ if ((sc->sc_flags & G_MIRROR_DEVICE_FLAG_NOAUTOSYNC) == 0 ||
+ (disk->d_flags & G_MIRROR_DISK_FLAG_FORCE_SYNC) != 0) {
+ state = G_MIRROR_DISK_STATE_SYNCHRONIZING;
+ } else {
+ state = G_MIRROR_DISK_STATE_STALE;
+ }
+ } else /* if (sc->sc_syncid < disk->d_sync.ds_syncid) */ {
+ /*
+ * Not good, NOT GOOD!
+ * It means that mirror was started on stale disks
+ * and more fresh disk just arrive.
+ * If there were writes, mirror is fucked up, sorry.
+ * I think the best choice here is don't touch
+ * this disk and inform the user laudly.
+ */
+ G_MIRROR_DEBUG(0, "Device %s was started before the freshest "
+ "disk (%s) arrives!! It will not be connected to the "
+ "running device.", sc->sc_name,
+ g_mirror_get_diskname(disk));
+ g_mirror_destroy_disk(disk);
+ state = G_MIRROR_DISK_STATE_NONE;
+ /* Return immediately, because disk was destroyed. */
+ return (state);
+ }
+ G_MIRROR_DEBUG(3, "State for %s disk: %s.",
+ g_mirror_get_diskname(disk), g_mirror_disk_state2str(state));
+ return (state);
+}
+
+/*
+ * Update device state.
+ */
+static void
+g_mirror_update_device(struct g_mirror_softc *sc, boolean_t force)
+{
+ struct g_mirror_disk *disk;
+ u_int state;
+
+ g_topology_assert();
+
+ switch (sc->sc_state) {
+ case G_MIRROR_DEVICE_STATE_STARTING:
+ {
+ struct g_mirror_disk *pdisk;
+ u_int dirty, ndisks, syncid;
+
+ KASSERT(sc->sc_provider == NULL,
+ ("Non-NULL provider in STARTING state (%s).", sc->sc_name));
+ /*
+ * Are we ready? We are, if all disks are connected or
+ * if we have any disks and 'force' is true.
+ */
+ if ((force && g_mirror_ndisks(sc, -1) > 0) ||
+ sc->sc_ndisks == g_mirror_ndisks(sc, -1)) {
+ ;
+ } else if (g_mirror_ndisks(sc, -1) == 0) {
+ /*
+ * Disks went down in starting phase, so destroy
+ * device.
+ */
+ callout_drain(&sc->sc_callout);
+ sc->sc_flags |= G_MIRROR_DEVICE_FLAG_DESTROY;
+ return;
+ } else {
+ return;
+ }
+
+ /*
+ * Activate all disks with the biggest syncid.
+ */
+ if (force) {
+ /*
+ * If called with 'force' true, we're called from
+ * timeout * procedure, so don't bother canceling
+ * timeout.
+ */
+ ndisks = 0;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_SYNCHRONIZING) == 0) {
+ ndisks++;
+ }
+ }
+ if (ndisks == 0) {
+ int timeout;
+
+ /* No valid disks still, wait some more. */
+ timeout =
+ atomic_load_acq_int(&g_mirror_timeout);
+ callout_reset(&sc->sc_callout, timeout * hz,
+ g_mirror_go, sc);
+ return;
+ }
+ } else {
+ /* Cancel timeout. */
+ callout_drain(&sc->sc_callout);
+ }
+
+ /*
+ * Find disk with the biggest syncid.
+ */
+ syncid = 0;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_sync.ds_syncid > syncid)
+ syncid = disk->d_sync.ds_syncid;
+ }
+
+ /*
+ * Here we need to look for dirty disks and if all disks
+ * with the biggest syncid are dirty, we have to choose
+ * one with the biggest priority and rebuild the rest.
+ */
+ /*
+ * Find the number of dirty disks with the biggest syncid.
+ * Find the number of disks with the biggest syncid.
+ * While here, find a disk with the biggest priority.
+ */
+ dirty = ndisks = 0;
+ pdisk = NULL;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_sync.ds_syncid != syncid)
+ continue;
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_SYNCHRONIZING) != 0) {
+ continue;
+ }
+ ndisks++;
+ if ((disk->d_flags & G_MIRROR_DISK_FLAG_DIRTY) != 0) {
+ dirty++;
+ if (pdisk == NULL ||
+ pdisk->d_priority < disk->d_priority) {
+ pdisk = disk;
+ }
+ }
+ }
+ if (dirty == 0) {
+ /* No dirty disks at all, great. */
+ } else if (dirty == ndisks) {
+ /*
+ * Force synchronization for all dirty disks except one
+ * with the biggest priority.
+ */
+ KASSERT(pdisk != NULL, ("pdisk == NULL"));
+ G_MIRROR_DEBUG(1, "Using disk %s (device %s) as a "
+ "master disk for synchronization.",
+ g_mirror_get_diskname(pdisk), sc->sc_name);
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_sync.ds_syncid != syncid)
+ continue;
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_SYNCHRONIZING) != 0) {
+ continue;
+ }
+ KASSERT((disk->d_flags &
+ G_MIRROR_DISK_FLAG_DIRTY) != 0,
+ ("Disk %s isn't marked as dirty.",
+ g_mirror_get_diskname(disk)));
+ /* Skip the disk with the biggest priority. */
+ if (disk == pdisk)
+ continue;
+ disk->d_sync.ds_syncid = 0;
+ }
+ } else if (dirty < ndisks) {
+ /*
+ * Force synchronization for all dirty disks.
+ * We have some non-dirty disks.
+ */
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_sync.ds_syncid != syncid)
+ continue;
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_SYNCHRONIZING) != 0) {
+ continue;
+ }
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_DIRTY) == 0) {
+ continue;
+ }
+ disk->d_sync.ds_syncid = 0;
+ }
+ }
+
+ /* Reset hint. */
+ sc->sc_hint = NULL;
+ sc->sc_syncid = syncid;
+ if (force) {
+ /* Remember to bump syncid on first write. */
+ sc->sc_bump_syncid = G_MIRROR_BUMP_ON_FIRST_WRITE;
+ }
+ state = G_MIRROR_DEVICE_STATE_RUNNING;
+ G_MIRROR_DEBUG(1, "Device %s state changed from %s to %s.",
+ sc->sc_name, g_mirror_device_state2str(sc->sc_state),
+ g_mirror_device_state2str(state));
+ sc->sc_state = state;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ state = g_mirror_determine_state(disk);
+ g_mirror_event_send(disk, state,
+ G_MIRROR_EVENT_DONTWAIT);
+ if (state == G_MIRROR_DISK_STATE_STALE) {
+ sc->sc_bump_syncid =
+ G_MIRROR_BUMP_ON_FIRST_WRITE;
+ }
+ }
+ break;
+ }
+ case G_MIRROR_DEVICE_STATE_RUNNING:
+ /*
+ * Bump syncid here, if we need to do it immediately.
+ */
+ if (sc->sc_bump_syncid == G_MIRROR_BUMP_IMMEDIATELY) {
+ sc->sc_bump_syncid = 0;
+ g_mirror_bump_syncid(sc);
+ }
+ if (g_mirror_ndisks(sc, -1) == 0) {
+ /*
+ * No disks at all, we need to destroy device.
+ */
+ sc->sc_flags |= G_MIRROR_DEVICE_FLAG_DESTROY;
+ } else if (g_mirror_ndisks(sc,
+ G_MIRROR_DISK_STATE_ACTIVE) == 0 &&
+ g_mirror_ndisks(sc, G_MIRROR_DISK_STATE_NEW) == 0) {
+ /*
+ * No active disks, destroy provider.
+ */
+ if (sc->sc_provider != NULL)
+ g_mirror_destroy_provider(sc);
+ } else if (g_mirror_ndisks(sc,
+ G_MIRROR_DISK_STATE_ACTIVE) > 0) {
+ /*
+ * We have active disks, launch provider if it doesn't
+ * exist.
+ */
+ if (sc->sc_provider == NULL)
+ g_mirror_launch_provider(sc);
+ }
+ break;
+ default:
+ KASSERT(1 == 0, ("Wrong device state (%s, %s).",
+ sc->sc_name, g_mirror_device_state2str(sc->sc_state)));
+ break;
+ }
+}
+
+/*
+ * Update disk state and device state if needed.
+ */
+#define DISK_STATE_CHANGED() G_MIRROR_DEBUG(1, \
+ "Disk %s state changed from %s to %s (device %s).", \
+ g_mirror_get_diskname(disk), \
+ g_mirror_disk_state2str(disk->d_state), \
+ g_mirror_disk_state2str(state), sc->sc_name)
+static int
+g_mirror_update_disk(struct g_mirror_disk *disk, u_int state)
+{
+ struct g_mirror_softc *sc;
+
+ g_topology_assert();
+
+ sc = disk->d_softc;
+again:
+ G_MIRROR_DEBUG(3, "Changing disk %s state from %s to %s.",
+ g_mirror_get_diskname(disk), g_mirror_disk_state2str(disk->d_state),
+ g_mirror_disk_state2str(state));
+ switch (state) {
+ case G_MIRROR_DISK_STATE_NEW:
+ /*
+ * Possible scenarios:
+ * 1. New disk arrive.
+ */
+ /* Previous state should be NONE. */
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_NONE,
+ ("Wrong disk state (%s, %s).", g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ DISK_STATE_CHANGED();
+
+ disk->d_state = state;
+ LIST_INSERT_HEAD(&sc->sc_disks, disk, d_next);
+ G_MIRROR_DEBUG(0, "Device %s: provider %s detected.",
+ sc->sc_name, g_mirror_get_diskname(disk));
+ if (sc->sc_state == G_MIRROR_DEVICE_STATE_STARTING)
+ break;
+ KASSERT(sc->sc_state == G_MIRROR_DEVICE_STATE_RUNNING,
+ ("Wrong device state (%s, %s, %s, %s).", sc->sc_name,
+ g_mirror_device_state2str(sc->sc_state),
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ state = g_mirror_determine_state(disk);
+ if (state != G_MIRROR_DISK_STATE_NONE)
+ goto again;
+ break;
+ case G_MIRROR_DISK_STATE_ACTIVE:
+ /*
+ * Possible scenarios:
+ * 1. New disk does not need synchronization.
+ * 2. Synchronization process finished successfully.
+ */
+ KASSERT(sc->sc_state == G_MIRROR_DEVICE_STATE_RUNNING,
+ ("Wrong device state (%s, %s, %s, %s).", sc->sc_name,
+ g_mirror_device_state2str(sc->sc_state),
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ /* Previous state should be NEW or SYNCHRONIZING. */
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_NEW ||
+ disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING,
+ ("Wrong disk state (%s, %s).", g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ DISK_STATE_CHANGED();
+
+ if (disk->d_state == G_MIRROR_DISK_STATE_NEW)
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_DIRTY;
+ else if (disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING) {
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_SYNCHRONIZING;
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_FORCE_SYNC;
+ g_mirror_sync_stop(disk, 0);
+ }
+ disk->d_state = state;
+ disk->d_sync.ds_offset = 0;
+ disk->d_sync.ds_offset_done = 0;
+ g_mirror_update_access(disk);
+ g_mirror_update_metadata(disk);
+ G_MIRROR_DEBUG(0, "Device %s: provider %s activated.",
+ sc->sc_name, g_mirror_get_diskname(disk));
+ break;
+ case G_MIRROR_DISK_STATE_STALE:
+ /*
+ * Possible scenarios:
+ * 1. Stale disk was connected.
+ */
+ /* Previous state should be NEW. */
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_NEW,
+ ("Wrong disk state (%s, %s).", g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ KASSERT(sc->sc_state == G_MIRROR_DEVICE_STATE_RUNNING,
+ ("Wrong device state (%s, %s, %s, %s).", sc->sc_name,
+ g_mirror_device_state2str(sc->sc_state),
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ /*
+ * STALE state is only possible if device is marked
+ * NOAUTOSYNC.
+ */
+ KASSERT((sc->sc_flags & G_MIRROR_DEVICE_FLAG_NOAUTOSYNC) != 0,
+ ("Wrong device state (%s, %s, %s, %s).", sc->sc_name,
+ g_mirror_device_state2str(sc->sc_state),
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ DISK_STATE_CHANGED();
+
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_DIRTY;
+ disk->d_state = state;
+ g_mirror_update_metadata(disk);
+ G_MIRROR_DEBUG(0, "Device %s: provider %s is stale.",
+ sc->sc_name, g_mirror_get_diskname(disk));
+ break;
+ case G_MIRROR_DISK_STATE_SYNCHRONIZING:
+ /*
+ * Possible scenarios:
+ * 1. Disk which needs synchronization was connected.
+ */
+ /* Previous state should be NEW. */
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_NEW,
+ ("Wrong disk state (%s, %s).", g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ KASSERT(sc->sc_state == G_MIRROR_DEVICE_STATE_RUNNING,
+ ("Wrong device state (%s, %s, %s, %s).", sc->sc_name,
+ g_mirror_device_state2str(sc->sc_state),
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ DISK_STATE_CHANGED();
+
+ if (disk->d_state == G_MIRROR_DISK_STATE_NEW)
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_DIRTY;
+ disk->d_state = state;
+ if (sc->sc_provider != NULL) {
+ g_mirror_sync_start(disk);
+ g_mirror_update_metadata(disk);
+ }
+ break;
+ case G_MIRROR_DISK_STATE_DISCONNECTED:
+ /*
+ * Possible scenarios:
+ * 1. Device wasn't running yet, but disk disappear.
+ * 2. Disk was active and disapppear.
+ * 3. Disk disappear during synchronization process.
+ */
+ if (sc->sc_state == G_MIRROR_DEVICE_STATE_RUNNING) {
+ /*
+ * Previous state should be ACTIVE, STALE or
+ * SYNCHRONIZING.
+ */
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_ACTIVE ||
+ disk->d_state == G_MIRROR_DISK_STATE_STALE ||
+ disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING,
+ ("Wrong disk state (%s, %s).",
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ } else if (sc->sc_state == G_MIRROR_DEVICE_STATE_STARTING) {
+ /* Previous state should be NEW. */
+ KASSERT(disk->d_state == G_MIRROR_DISK_STATE_NEW,
+ ("Wrong disk state (%s, %s).",
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+ /*
+ * Reset bumping syncid if disk disappeared in STARTING
+ * state.
+ */
+ if (sc->sc_bump_syncid == G_MIRROR_BUMP_ON_FIRST_WRITE)
+ sc->sc_bump_syncid = 0;
+#ifdef INVARIANTS
+ } else {
+ KASSERT(1 == 0, ("Wrong device state (%s, %s, %s, %s).",
+ sc->sc_name,
+ g_mirror_device_state2str(sc->sc_state),
+ g_mirror_get_diskname(disk),
+ g_mirror_disk_state2str(disk->d_state)));
+#endif
+ }
+ DISK_STATE_CHANGED();
+ G_MIRROR_DEBUG(0, "Device %s: provider %s disconnected.",
+ sc->sc_name, g_mirror_get_diskname(disk));
+
+ g_mirror_destroy_disk(disk);
+ break;
+ case G_MIRROR_DISK_STATE_DESTROY:
+ {
+ int error;
+
+ error = g_mirror_clear_metadata(disk);
+ if (error != 0)
+ return (error);
+ DISK_STATE_CHANGED();
+ G_MIRROR_DEBUG(0, "Device %s: provider %s destroyed.",
+ sc->sc_name, g_mirror_get_diskname(disk));
+
+ g_mirror_destroy_disk(disk);
+ sc->sc_ndisks--;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ g_mirror_update_metadata(disk);
+ }
+ break;
+ }
+ default:
+ KASSERT(1 == 0, ("Unknown state (%u).", state));
+ break;
+ }
+ return (0);
+}
+#undef DISK_STATE_CHANGED
+
+static int
+g_mirror_read_metadata(struct g_consumer *cp, struct g_mirror_metadata *md)
+{
+ struct g_provider *pp;
+ u_char *buf;
+ int error;
+
+ g_topology_assert();
+
+ error = g_access(cp, 1, 0, 0);
+ if (error != 0)
+ return (error);
+ pp = cp->provider;
+ g_topology_unlock();
+ /* Metadata are stored on last sector. */
+ buf = g_read_data(cp, pp->mediasize - pp->sectorsize, pp->sectorsize,
+ &error);
+ g_topology_lock();
+ if (buf == NULL) {
+ g_access(cp, -1, 0, 0);
+ return (error);
+ }
+ if (error != 0) {
+ g_access(cp, -1, 0, 0);
+ g_free(buf);
+ return (error);
+ }
+ error = g_access(cp, -1, 0, 0);
+ KASSERT(error == 0, ("Cannot decrease access count for %s.", pp->name));
+
+ /* Decode metadata. */
+ error = mirror_metadata_decode(buf, md);
+ g_free(buf);
+ if (strcmp(md->md_magic, G_MIRROR_MAGIC) != 0)
+ return (EINVAL);
+ if (error != 0) {
+ G_MIRROR_DEBUG(1, "MD5 metadata hash mismatch for provider %s.",
+ cp->provider->name);
+ return (error);
+ }
+
+ return (0);
+}
+
+static int
+g_mirror_check_metadata(struct g_mirror_softc *sc, struct g_provider *pp,
+ struct g_mirror_metadata *md)
+{
+
+ if (g_mirror_id2disk(sc, md->md_did) != NULL) {
+ G_MIRROR_DEBUG(1, "Disk %s (id=%u) already exists, skipping.",
+ pp->name, md->md_did);
+ return (EEXIST);
+ }
+ if (md->md_all != sc->sc_ndisks) {
+ G_MIRROR_DEBUG(1,
+ "Invalid '%s' field on disk %s (device %s), skipping.",
+ "md_all", pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ if (md->md_slice != sc->sc_slice) {
+ G_MIRROR_DEBUG(1,
+ "Invalid '%s' field on disk %s (device %s), skipping.",
+ "md_slice", pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ if (md->md_balance != sc->sc_balance) {
+ G_MIRROR_DEBUG(1,
+ "Invalid '%s' field on disk %s (device %s), skipping.",
+ "md_balance", pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ if (md->md_mediasize != sc->sc_mediasize) {
+ G_MIRROR_DEBUG(1,
+ "Invalid '%s' field on disk %s (device %s), skipping.",
+ "md_mediasize", pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ if (sc->sc_mediasize > pp->mediasize) {
+ G_MIRROR_DEBUG(1,
+ "Invalid size of disk %s (device %s), skipping.", pp->name,
+ sc->sc_name);
+ return (EINVAL);
+ }
+ if (md->md_sectorsize != sc->sc_sectorsize) {
+ G_MIRROR_DEBUG(1,
+ "Invalid '%s' field on disk %s (device %s), skipping.",
+ "md_sectorsize", pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ if ((sc->sc_sectorsize % pp->sectorsize) != 0) {
+ G_MIRROR_DEBUG(1,
+ "Invalid sector size of disk %s (device %s), skipping.",
+ pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ if ((md->md_mflags & ~G_MIRROR_DEVICE_FLAG_MASK) != 0) {
+ G_MIRROR_DEBUG(1,
+ "Invalid device flags on disk %s (device %s), skipping.",
+ pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ if ((md->md_dflags & ~G_MIRROR_DISK_FLAG_MASK) != 0) {
+ G_MIRROR_DEBUG(1,
+ "Invalid disk flags on disk %s (device %s), skipping.",
+ pp->name, sc->sc_name);
+ return (EINVAL);
+ }
+ return (0);
+}
+
+static int
+g_mirror_add_disk(struct g_mirror_softc *sc, struct g_provider *pp,
+ struct g_mirror_metadata *md)
+{
+ struct g_mirror_disk *disk;
+ int error;
+
+ g_topology_assert();
+ G_MIRROR_DEBUG(2, "Adding disk %s.", pp->name);
+
+ error = g_mirror_check_metadata(sc, pp, md);
+ if (error != 0)
+ return (error);
+ disk = g_mirror_init_disk(sc, pp, md, &error);
+ if (disk == NULL)
+ return (error);
+ error = g_mirror_event_send(disk, G_MIRROR_DISK_STATE_NEW,
+ G_MIRROR_EVENT_WAIT);
+ return (error);
+}
+
+static int
+g_mirror_access(struct g_provider *pp, int acr, int acw, int ace)
+{
+ struct g_mirror_softc *sc;
+ struct g_mirror_disk *disk;
+ int dcr, dcw, dce, err, error;
+
+ g_topology_assert();
+ G_MIRROR_DEBUG(2, "Access request for %s: r%dw%de%d.", pp->name, acr,
+ acw, ace);
+
+ dcr = pp->acr + acr;
+ dcw = pp->acw + acw;
+ dce = pp->ace + ace;
+
+ /* On first open, grab an extra "exclusive" bit */
+ if (pp->acr == 0 && pp->acw == 0 && pp->ace == 0)
+ ace++;
+ /* ... and let go of it on last close */
+ if (dcr == 0 && dcw == 0 && dce == 0)
+ ace--;
+
+ sc = pp->geom->softc;
+ if (sc == NULL || LIST_EMPTY(&sc->sc_disks)) {
+ if (acr <= 0 && acw <= 0 && ace <= 0)
+ return (0);
+ else
+ return (ENXIO);
+ }
+ error = ENXIO;
+ LIST_FOREACH(disk, &sc->sc_disks, d_next) {
+ if (disk->d_state != G_MIRROR_DISK_STATE_ACTIVE)
+ continue;
+ err = g_access(disk->d_consumer, acr, acw, ace);
+ G_MIRROR_DEBUG(2, "Access %s r%dw%de%d = %d",
+ g_mirror_get_diskname(disk), acr, acw, ace, err);
+ if (err == 0) {
+ /*
+ * Mark disk as dirty on open and unmark on close.
+ */
+ if (pp->acw == 0 && dcw > 0) {
+ G_MIRROR_DEBUG(1,
+ "Disk %s (device %s) marked as dirty.",
+ g_mirror_get_diskname(disk), sc->sc_name);
+ disk->d_flags |= G_MIRROR_DISK_FLAG_DIRTY;
+ g_mirror_update_metadata(disk);
+ } else if (pp->acw > 0 && dcw == 0) {
+ G_MIRROR_DEBUG(1,
+ "Disk %s (device %s) marked as clean.",
+ g_mirror_get_diskname(disk), sc->sc_name);
+ disk->d_flags &= ~G_MIRROR_DISK_FLAG_DIRTY;
+ g_mirror_update_metadata(disk);
+ }
+ error = 0;
+ } else {
+ sc->sc_bump_syncid = G_MIRROR_BUMP_ON_FIRST_WRITE;
+ g_mirror_event_send(disk,
+ G_MIRROR_DISK_STATE_DISCONNECTED,
+ G_MIRROR_EVENT_DONTWAIT);
+ }
+ }
+ return (error);
+}
+
+static struct g_geom *
+g_mirror_create(struct g_class *mp, const struct g_mirror_metadata *md)
+{
+ struct g_mirror_softc *sc;
+ struct g_geom *gp;
+ int error, timeout;
+
+ g_topology_assert();
+ G_MIRROR_DEBUG(1, "Creating device %s (id=%u).", md->md_name,
+ md->md_mid);
+
+ /* One disk is minimum. */
+ if (md->md_all < 1)
+ return (NULL);
+ /*
+ * Action geom.
+ */
+ gp = g_new_geomf(mp, "%s", md->md_name);
+ sc = malloc(sizeof(*sc), M_MIRROR, M_WAITOK | M_ZERO);
+ gp->start = g_mirror_start;
+ gp->spoiled = g_mirror_orphan;
+ gp->orphan = g_mirror_orphan;
+ gp->access = g_mirror_access;
+ gp->dumpconf = g_mirror_dumpconf;
+
+ sc->sc_id = md->md_mid;
+ sc->sc_slice = md->md_slice;
+ sc->sc_balance = md->md_balance;
+ sc->sc_mediasize = md->md_mediasize;
+ sc->sc_sectorsize = md->md_sectorsize;
+ sc->sc_ndisks = md->md_all;
+ sc->sc_flags = md->md_mflags;
+ sc->sc_bump_syncid = 0;
+ bioq_init(&sc->sc_queue);
+ mtx_init(&sc->sc_queue_mtx, "gmirror:queue", NULL, MTX_DEF);
+ LIST_INIT(&sc->sc_disks);
+ TAILQ_INIT(&sc->sc_events);
+ mtx_init(&sc->sc_events_mtx, "gmirror:events", NULL, MTX_DEF);
+ callout_init(&sc->sc_callout, CALLOUT_MPSAFE);
+ sc->sc_state = G_MIRROR_DEVICE_STATE_STARTING;
+ gp->softc = sc;
+ sc->sc_geom = gp;
+ sc->sc_provider = NULL;
+ /*
+ * Synchronization geom.
+ */
+ gp = g_new_geomf(mp, "%s.sync", md->md_name);
+ gp->softc = sc;
+ gp->spoiled = g_mirror_orphan;
+ gp->orphan = g_mirror_orphan;
+ sc->sc_sync.ds_geom = gp;
+ sc->sc_sync.ds_block = atomic_load_acq_int(&g_mirror_sync_block_size);
+ sc->sc_sync.ds_ndisks = 0;
+ sc->sc_sync.ds_zone = uma_zcreate("gmirror:sync", sc->sc_sync.ds_block,
+ NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, 0);
+ error = kthread_create(g_mirror_worker, sc, &sc->sc_worker, 0, 0,
+ "g_mirror %s", md->md_name);
+ if (error != 0) {
+ G_MIRROR_DEBUG(1, "Cannot create kernel thread for %s.",
+ sc->sc_name);
+ uma_zdestroy(sc->sc_sync.ds_zone);
+ g_destroy_geom(sc->sc_sync.ds_geom);
+ mtx_destroy(&sc->sc_events_mtx);
+ mtx_destroy(&sc->sc_queue_mtx);
+ g_destroy_geom(sc->sc_geom);
+ free(sc, M_MIRROR);
+ return (NULL);
+ }
+
+ G_MIRROR_DEBUG(0, "Device %s created (id=%u).", sc->sc_name, sc->sc_id);
+
+ /*
+ * Run timeout.
+ */
+ timeout = atomic_load_acq_int(&g_mirror_timeout);
+ callout_reset(&sc->sc_callout, timeout * hz, g_mirror_go, sc);
+ return (sc->sc_geom);
+}
+
+int
+g_mirror_destroy(struct g_mirror_softc *sc, boolean_t force)
+{
+ struct g_provider *pp;
+
+ g_topology_assert();
+
+ if (sc == NULL)
+ return (ENXIO);
+ pp = sc->sc_provider;
+ if (pp != NULL && (pp->acr != 0 || pp->acw != 0 || pp->ace != 0)) {
+ if (force) {
+ G_MIRROR_DEBUG(0, "Device %s is still open, so it "
+ "can't be definitely removed.", pp->name);
+ } else {
+ G_MIRROR_DEBUG(1,
+ "Device %s is still open (r%dw%de%d).", pp->name,
+ pp->acr, pp->acw, pp->ace);
+ return (EBUSY);
+ }
+ }
+
+ sc->sc_flags |= G_MIRROR_DEVICE_FLAG_DESTROY;
+ sc->sc_flags |= G_MIRROR_DEVICE_FLAG_WAIT;
+ G_MIRROR_DEBUG(4, "%s: Waking up %p.", __func__, sc);
+ mtx_lock(&sc->sc_queue_mtx);
+ wakeup(sc);
+ mtx_unlock(&sc->sc_queue_mtx);
+ G_MIRROR_DEBUG(4, "%s: Sleeping %p.", __func__, &sc->sc_worker);
+ while (sc->sc_worker != NULL)
+ tsleep(&sc->sc_worker, PRIBIO, "m:destroy", hz / 5);
+ G_MIRROR_DEBUG(4, "%s: Woken up %p.", __func__, &sc->sc_worker);
+ g_mirror_destroy_device(sc);
+ free(sc, M_MIRROR);
+ return (0);
+}
+
+static void
+g_mirror_taste_orphan(struct g_consumer *cp)
+{
+
+ KASSERT(1 == 0, ("%s called while tasting %s.", __func__,
+ cp->provider->name));
+}
+
+static struct g_geom *
+g_mirror_taste(struct g_class *mp, struct g_provider *pp, int flags __unused)
+{
+ struct g_mirror_metadata md;
+ struct g_mirror_softc *sc;
+ struct g_consumer *cp;
+ struct g_geom *gp;
+ int error;
+
+ g_topology_assert();
+ g_trace(G_T_TOPOLOGY, "%s(%s, %s)", __func__, mp->name, pp->name);
+ G_MIRROR_DEBUG(2, "Tasting %s.", pp->name);
+
+ gp = g_new_geomf(mp, "mirror:taste");
+ /*
+ * This orphan function should be never called.
+ */
+ gp->orphan = g_mirror_taste_orphan;
+ cp = g_new_consumer(gp);
+ g_attach(cp, pp);
+ error = g_mirror_read_metadata(cp, &md);
+ g_detach(cp);
+ g_destroy_consumer(cp);
+ g_destroy_geom(gp);
+ if (error != 0)
+ return (NULL);
+ gp = NULL;
+
+ if (md.md_version > G_MIRROR_VERSION) {
+ printf("geom_mirror.ko module is too old to handle %s.\n",
+ pp->name);
+ return (NULL);
+ }
+ if ((md.md_dflags & G_MIRROR_DISK_FLAG_INACTIVE) != 0) {
+ G_MIRROR_DEBUG(0,
+ "Device %s: provider %s marked as inactive, skipping.",
+ md.md_name, pp->name);
+ return (NULL);
+ }
+ if (g_mirror_debug >= 2)
+ mirror_metadata_dump(&md);
+
+ /*
+ * Let's check if device already exists.
+ */
+ LIST_FOREACH(gp, &mp->geom, geom) {
+ sc = gp->softc;
+ if (sc == NULL)
+ continue;
+ if (sc->sc_sync.ds_geom == gp)
+ continue;
+ if (strcmp(md.md_name, sc->sc_name) != 0)
+ continue;
+ if (md.md_mid != sc->sc_id) {
+ G_MIRROR_DEBUG(0, "Device %s already configured.",
+ sc->sc_name);
+ return (NULL);
+ }
+ break;
+ }
+ if (gp == NULL) {
+ gp = g_mirror_create(mp, &md);
+ if (gp == NULL) {
+ G_MIRROR_DEBUG(0, "Cannot create device %s.mirror",
+ md.md_name);
+ return (NULL);
+ }
+ sc = gp->softc;
+ }
+ G_MIRROR_DEBUG(1, "Adding disk %s to %s.", pp->name, gp->name);
+ error = g_mirror_add_disk(sc, pp, &md);
+ if (error != 0) {
+ G_MIRROR_DEBUG(0, "Cannot add disk %s to %s (error=%d).",
+ pp->name, gp->name, error);
+ if (LIST_EMPTY(&sc->sc_disks))
+ g_mirror_destroy(sc, 1);
+ return (NULL);
+ }
+ return (gp);
+}
+
+static int
+g_mirror_destroy_geom(struct gctl_req *req __unused,
+ struct g_class *mp __unused, struct g_geom *gp)
+{
+
+ return (g_mirror_destroy(gp->softc, 0));
+}
+
+static void
+g_mirror_dumpconf(struct sbuf *sb, const char *indent, struct g_geom *gp,
+ struct g_consumer *cp, struct g_provider *pp)
+{
+ struct g_mirror_softc *sc;
+
+ g_topology_assert();
+
+ sc = gp->softc;
+ if (sc == NULL)
+ return;
+ /* Skip synchronization geom. */
+ if (gp == sc->sc_sync.ds_geom)
+ return;
+ if (pp != NULL) {
+ /* Nothing here. */
+ } else if (cp != NULL) {
+ struct g_mirror_disk *disk;
+
+ disk = cp->private;
+ if (disk == NULL)
+ return;
+ sbuf_printf(sb, "%s<ID>%u</ID>\n", indent, (u_int)disk->d_id);
+ if (disk->d_state == G_MIRROR_DISK_STATE_SYNCHRONIZING) {
+ sbuf_printf(sb, "%s<Synchronized>", indent);
+ if (disk->d_sync.ds_offset_done == 0)
+ sbuf_printf(sb, "0%%");
+ else {
+ sbuf_printf(sb, "%u%%",
+ (u_int)((disk->d_sync.ds_offset_done * 100) /
+ sc->sc_provider->mediasize));
+ }
+ sbuf_printf(sb, "</Synchronized>\n");
+ }
+ sbuf_printf(sb, "%s<SyncID>%u</SyncID>\n", indent,
+ disk->d_sync.ds_syncid);
+ sbuf_printf(sb, "%s<Flags>", indent);
+ if (disk->d_flags == 0)
+ sbuf_printf(sb, "NONE");
+ else {
+ int first = 1;
+
+ if ((disk->d_flags & G_MIRROR_DISK_FLAG_DIRTY) != 0) {
+ if (!first)
+ sbuf_printf(sb, ", ");
+ else
+ first = 0;
+ sbuf_printf(sb, "DIRTY");
+ }
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_INACTIVE) != 0) {
+ if (!first)
+ sbuf_printf(sb, ", ");
+ else
+ first = 0;
+ sbuf_printf(sb, "INACTIVE");
+ }
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_SYNCHRONIZING) != 0) {
+ if (!first)
+ sbuf_printf(sb, ", ");
+ else
+ first = 0;
+ sbuf_printf(sb, "SYNCHRONIZING");
+ }
+ if ((disk->d_flags &
+ G_MIRROR_DISK_FLAG_FORCE_SYNC) != 0) {
+ if (!first)
+ sbuf_printf(sb, ", ");
+ else
+ first = 0;
+ sbuf_printf(sb, "FORCE_SYNC");
+ }
+ }
+ sbuf_printf(sb, "</Flags>\n");
+ sbuf_printf(sb, "%s<Priority>%u</Priority>\n", indent,
+ disk->d_sync.ds_syncid);
+ sbuf_printf(sb, "%s<State>%s</State>\n", indent,
+ g_mirror_disk_state2str(disk->d_state));
+ } else {
+ sbuf_printf(sb, "%s<ID>%u</ID>\n", indent, (u_int)sc->sc_id);
+ sbuf_printf(sb, "%s<SyncID>%u</SyncID>\n", indent, sc->sc_syncid);
+ sbuf_printf(sb, "%s<Flags>", indent);
+ if (sc->sc_flags == 0)
+ sbuf_printf(sb, "NONE");
+ else {
+ int first = 1;
+
+ if ((sc->sc_flags &
+ G_MIRROR_DEVICE_FLAG_NOAUTOSYNC) != 0) {
+ if (!first)
+ sbuf_printf(sb, ", ");
+ else
+ first = 0;
+ sbuf_printf(sb, "NOAUTOSYNC");
+ }
+ }
+ sbuf_printf(sb, "</Flags>\n");
+ sbuf_printf(sb, "%s<Slice>%u</Slice>\n", indent,
+ (u_int)sc->sc_slice);
+ sbuf_printf(sb, "%s<Balance>%s</Balance>\n", indent,
+ balance_name(sc->sc_balance));
+ sbuf_printf(sb, "%s<Components>%u</Components>\n", indent,
+ sc->sc_ndisks);
+ }
+}
+
+DECLARE_GEOM_CLASS(g_mirror_class, g_mirror);