aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorSean Bruno <sbruno@FreeBSD.org>2017-01-24 15:55:52 +0000
committerSean Bruno <sbruno@FreeBSD.org>2017-01-24 15:55:52 +0000
commit36fa5d5b645727132c69620eba752aaf975e2982 (patch)
treea00983a672d1be4033d50cb1e45dcd09192a1c7c
parent562a3182f6e4aad86dddad30d10a9ac902db923c (diff)
downloadsrc-36fa5d5b645727132c69620eba752aaf975e2982.tar.gz
src-36fa5d5b645727132c69620eba752aaf975e2982.zip
Revert 312696 due to build tests.
Notes
Notes: svn path=/head/; revision=312697
-rw-r--r--sys/dev/e1000/em_txrx.c11
-rw-r--r--sys/kern/subr_gtaskqueue.c36
-rw-r--r--sys/net/iflib.c43
-rw-r--r--sys/sys/gtaskqueue.h20
4 files changed, 35 insertions, 75 deletions
diff --git a/sys/dev/e1000/em_txrx.c b/sys/dev/e1000/em_txrx.c
index 96bc7d2dc4bc..33a9b9d115f4 100644
--- a/sys/dev/e1000/em_txrx.c
+++ b/sys/dev/e1000/em_txrx.c
@@ -408,13 +408,10 @@ em_isc_txd_credits_update(void *arg, uint16_t txqid, uint32_t cidx_init, bool cl
cidx = cidx_init;
buf = &txr->tx_buffers[cidx];
tx_desc = &txr->tx_base[cidx];
- last = buf->eop;
- if (last == -1)
- return (processed);
+ last = buf->eop;
eop_desc = &txr->tx_base[last];
- DPRINTF(iflib_get_dev(adapter->ctx),
- "credits_update: cidx_init=%d clear=%d last=%d\n",
+ DPRINTF(iflib_get_dev(adapter->ctx), "credits_update: cidx_init=%d clear=%d last=%d\n",
cidx_init, clear, last);
/*
* What this does is get the index of the
@@ -423,7 +420,7 @@ em_isc_txd_credits_update(void *arg, uint16_t txqid, uint32_t cidx_init, bool cl
* simple comparison on the inner while loop.
*/
if (++last == scctx->isc_ntxd[0])
- last = 0;
+ last = 0;
done = last;
@@ -439,7 +436,7 @@ em_isc_txd_credits_update(void *arg, uint16_t txqid, uint32_t cidx_init, bool cl
tx_desc++;
buf++;
processed++;
-
+
/* wrap the ring ? */
if (++cidx == scctx->isc_ntxd[0]) {
cidx = 0;
diff --git a/sys/kern/subr_gtaskqueue.c b/sys/kern/subr_gtaskqueue.c
index 2fa39ac8c46a..a5633bce7ea7 100644
--- a/sys/kern/subr_gtaskqueue.c
+++ b/sys/kern/subr_gtaskqueue.c
@@ -630,29 +630,6 @@ taskqgroup_find(struct taskqgroup *qgroup, void *uniq)
return (idx);
}
-/*
- * smp_started is unusable since it is not set for UP kernels or even for
- * SMP kernels when there is 1 CPU. This is usually handled by adding a
- * (mp_ncpus == 1) test, but that would be broken here since we need to
- * to synchronize with the SI_SUB_SMP ordering. Even in the pure SMP case
- * smp_started only gives a fuzzy ordering relative to SI_SUB_SMP.
- *
- * So maintain our own flag. It must be set after all CPUs are started
- * and before SI_SUB_SMP:SI_ORDER_ANY so that the SYSINIT for delayed
- * adjustment is properly delayed. SI_ORDER_FOURTH is clearly before
- * SI_ORDER_ANY and unclearly after the CPUs are started. It would be
- * simpler for adjustment to pass a flag indicating if it is delayed.
- */
-static int tqg_smp_started;
-
-static void
-tqg_record_smp_started(void *arg)
-{
- tqg_smp_started = 1;
-}
-
-SYSINIT(tqg_record_smp_started, SI_SUB_SMP, SI_ORDER_FOURTH,
- tqg_record_smp_started, NULL);
void
taskqgroup_attach(struct taskqgroup *qgroup, struct grouptask *gtask,
@@ -670,7 +647,7 @@ taskqgroup_attach(struct taskqgroup *qgroup, struct grouptask *gtask,
qgroup->tqg_queue[qid].tgc_cnt++;
LIST_INSERT_HEAD(&qgroup->tqg_queue[qid].tgc_tasks, gtask, gt_list);
gtask->gt_taskqueue = qgroup->tqg_queue[qid].tgc_taskq;
- if (irq != -1 && tqg_smp_started ) {
+ if (irq != -1 && (smp_started || mp_ncpus == 1)) {
gtask->gt_cpu = qgroup->tqg_queue[qid].tgc_cpu;
CPU_ZERO(&mask);
CPU_SET(qgroup->tqg_queue[qid].tgc_cpu, &mask);
@@ -720,7 +697,7 @@ taskqgroup_attach_cpu(struct taskqgroup *qgroup, struct grouptask *gtask,
gtask->gt_irq = irq;
gtask->gt_cpu = cpu;
mtx_lock(&qgroup->tqg_lock);
- if (tqg_smp_started)
+ if (smp_started || mp_ncpus == 1) {
for (i = 0; i < qgroup->tqg_cnt; i++)
if (qgroup->tqg_queue[i].tgc_cpu == cpu) {
qid = i;
@@ -754,7 +731,7 @@ taskqgroup_attach_cpu_deferred(struct taskqgroup *qgroup, struct grouptask *gtas
qid = -1;
irq = gtask->gt_irq;
cpu = gtask->gt_cpu;
- MPASS(tqg_smp_started);
+ MPASS(smp_started || mp_ncpus == 1);
mtx_lock(&qgroup->tqg_lock);
for (i = 0; i < qgroup->tqg_cnt; i++)
if (qgroup->tqg_queue[i].tgc_cpu == cpu) {
@@ -847,10 +824,9 @@ _taskqgroup_adjust(struct taskqgroup *qgroup, int cnt, int stride)
mtx_assert(&qgroup->tqg_lock, MA_OWNED);
- if (cnt < 1 || cnt * stride > mp_ncpus || !tqg_smp_started) {
- printf("%s: failed cnt: %d stride: %d "
- "mp_ncpus: %d smp_started: %d\n",
- __func__, cnt, stride, mp_ncpus, smp_started);
+ if (cnt < 1 || cnt * stride > mp_ncpus || (!smp_started && (mp_ncpus != 1))) {
+ printf("taskqgroup_adjust failed cnt: %d stride: %d mp_ncpus: %d smp_started: %d\n",
+ cnt, stride, mp_ncpus, smp_started);
return (EINVAL);
}
if (qgroup->tqg_adjusting) {
diff --git a/sys/net/iflib.c b/sys/net/iflib.c
index 550470fd6078..a1bf0ab9eb31 100644
--- a/sys/net/iflib.c
+++ b/sys/net/iflib.c
@@ -1193,36 +1193,13 @@ iflib_dma_free_multi(iflib_dma_info_t *dmalist, int count)
iflib_dma_free(*dmaiter);
}
-#ifdef EARLY_AP_STARTUP
-static const int iflib_started = 1;
-#else
-/*
- * We used to abuse the smp_started flag to decide if the queues have been
- * fully initialized (by late taskqgroup_adjust() calls in a SYSINIT()).
- * That gave bad races, since the SYSINIT() runs strictly after smp_started
- * is set. Run a SYSINIT() strictly after that to just set a usable
- * completion flag.
- */
-
-static int iflib_started;
-
-static void
-iflib_record_started(void *arg)
-{
- iflib_started = 1;
-}
-
-SYSINIT(iflib_record_started, SI_SUB_SMP + 1, SI_ORDER_FIRST,
- iflib_record_started, NULL);
-#endif
-
static int
iflib_fast_intr(void *arg)
{
iflib_filter_info_t info = arg;
struct grouptask *gtask = info->ifi_task;
- if (!iflib_started)
+ if (!smp_started && mp_ncpus > 1)
return (FILTER_HANDLED);
DBG_COUNTER_INC(fast_intrs);
@@ -3751,16 +3728,7 @@ iflib_device_register(device_t dev, void *sc, if_shared_ctx_t sctx, if_ctx_t *ct
device_printf(dev, "qset structure setup failed %d\n", err);
goto fail_queues;
}
- /*
- * Group taskqueues aren't properly set up until SMP is started,
- * so we disable interrupts until we can handle them post
- * SI_SUB_SMP.
- *
- * XXX: disabling interrupts doesn't actually work, at least for
- * the non-MSI case. When they occur before SI_SUB_SMP completes,
- * we do null handling and depend on this not causing too large an
- * interrupt storm.
- */
+
IFDI_INTR_DISABLE(ctx);
if (msix > 1 && (err = IFDI_MSIX_INTR_ASSIGN(ctx, msix)) != 0) {
device_printf(dev, "IFDI_MSIX_INTR_ASSIGN failed %d\n", err);
@@ -4588,6 +4556,13 @@ iflib_legacy_setup(if_ctx_t ctx, driver_filter_t filter, void *filter_arg, int *
void *q;
int err;
+ /*
+ * group taskqueues aren't properly set up until SMP is started
+ * so we disable interrupts until we can handle them post
+ * SI_SUB_SMP
+ */
+ IFDI_INTR_DISABLE(ctx);
+
q = &ctx->ifc_rxqs[0];
info = &rxq[0].ifr_filter_info;
gtask = &rxq[0].ifr_task;
diff --git a/sys/sys/gtaskqueue.h b/sys/sys/gtaskqueue.h
index fed9718b8c2e..3a9aeadc0a40 100644
--- a/sys/sys/gtaskqueue.h
+++ b/sys/sys/gtaskqueue.h
@@ -81,7 +81,7 @@ int taskqgroup_adjust(struct taskqgroup *qgroup, int cnt, int stride);
extern struct taskqgroup *qgroup_##name
-#ifdef EARLY_AP_STARTUP
+#if (!defined(SMP) || defined(EARLY_AP_STARTUP))
#define TASKQGROUP_DEFINE(name, cnt, stride) \
\
struct taskqgroup *qgroup_##name; \
@@ -95,8 +95,7 @@ taskqgroup_define_##name(void *arg) \
\
SYSINIT(taskqgroup_##name, SI_SUB_INIT_IF, SI_ORDER_FIRST, \
taskqgroup_define_##name, NULL)
-
-#else /* !EARLY_AP_STARTUP */
+#else /* SMP && !EARLY_AP_STARTUP */
#define TASKQGROUP_DEFINE(name, cnt, stride) \
\
struct taskqgroup *qgroup_##name; \
@@ -105,6 +104,15 @@ static void \
taskqgroup_define_##name(void *arg) \
{ \
qgroup_##name = taskqgroup_create(#name); \
+ /* Adjustment will be null unless smp_cpus == 1. */ \
+ /* \
+ * XXX this was intended to fix the smp_cpus == 1 case, but \
+ * doesn't actually work for that. It gives thes same strange \
+ * panic as adjustment at SI_SUB_INIT_IF:SI_ORDER_ANY for a \
+ * device that works with a pure UP kernel. \
+ */ \
+ /* XXX this code is common now, so should not be ifdefed. */ \
+ taskqgroup_adjust(qgroup_##name, (cnt), (stride)); \
} \
\
SYSINIT(taskqgroup_##name, SI_SUB_INIT_IF, SI_ORDER_FIRST, \
@@ -113,13 +121,17 @@ SYSINIT(taskqgroup_##name, SI_SUB_INIT_IF, SI_ORDER_FIRST, \
static void \
taskqgroup_adjust_##name(void *arg) \
{ \
+ /* \
+ * Adjustment when smp_cpus > 1 only works accidentally \
+ * (when there is no device interrupt before adjustment). \
+ */ \
taskqgroup_adjust(qgroup_##name, (cnt), (stride)); \
} \
\
SYSINIT(taskqgroup_adj_##name, SI_SUB_SMP, SI_ORDER_ANY, \
taskqgroup_adjust_##name, NULL); \
-#endif /* EARLY_AP_STARTUP */
+#endif /* !SMP || EARLY_AP_STARTUP */
TASKQGROUP_DECLARE(net);