aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--sys/dev/mpt/mpt.c640
-rw-r--r--sys/dev/mpt/mpt.h196
-rw-r--r--sys/dev/mpt/mpt_cam.c2595
-rw-r--r--sys/dev/mpt/mpt_debug.c64
-rw-r--r--sys/dev/mpt/mpt_pci.c78
-rw-r--r--sys/dev/mpt/mpt_raid.c57
-rw-r--r--sys/dev/mpt/mpt_reg.h2
7 files changed, 3102 insertions, 530 deletions
diff --git a/sys/dev/mpt/mpt.c b/sys/dev/mpt/mpt.c
index 12fb146694cc..f780e8825c08 100644
--- a/sys/dev/mpt/mpt.c
+++ b/sys/dev/mpt/mpt.c
@@ -100,6 +100,8 @@ __FBSDID("$FreeBSD$");
#include <dev/mpt/mpilib/mpi.h>
#include <dev/mpt/mpilib/mpi_ioc.h>
+#include <dev/mpt/mpilib/mpi_fc.h>
+#include <dev/mpt/mpilib/mpi_targ.h>
#include <sys/sysctl.h>
@@ -123,7 +125,7 @@ static int mpt_send_event_request(struct mpt_softc *mpt, int onoff);
static int mpt_soft_reset(struct mpt_softc *mpt);
static void mpt_hard_reset(struct mpt_softc *mpt);
static int mpt_configure_ioc(struct mpt_softc *mpt);
-static int mpt_enable_ioc(struct mpt_softc *mpt);
+static int mpt_enable_ioc(struct mpt_softc *mpt, int);
/************************* Personality Module Support *************************/
/*
@@ -150,7 +152,7 @@ mpt_pers_find(struct mpt_softc *mpt, u_int start_at)
}
/*
- * Used infrequenstly, so no need to optimize like a forward
+ * Used infrequently, so no need to optimize like a forward
* traversal where we use the MAX+1 is guaranteed to be NULL
* trick.
*/
@@ -179,6 +181,7 @@ mpt_pers_find_reverse(struct mpt_softc *mpt, u_int start_at)
static mpt_load_handler_t mpt_stdload;
static mpt_probe_handler_t mpt_stdprobe;
static mpt_attach_handler_t mpt_stdattach;
+static mpt_enable_handler_t mpt_stdenable;
static mpt_event_handler_t mpt_stdevent;
static mpt_reset_handler_t mpt_stdreset;
static mpt_shutdown_handler_t mpt_stdshutdown;
@@ -189,6 +192,7 @@ static struct mpt_personality mpt_default_personality =
.load = mpt_stdload,
.probe = mpt_stdprobe,
.attach = mpt_stdattach,
+ .enable = mpt_stdenable,
.event = mpt_stdevent,
.reset = mpt_stdreset,
.shutdown = mpt_stdshutdown,
@@ -198,6 +202,7 @@ static struct mpt_personality mpt_default_personality =
static mpt_load_handler_t mpt_core_load;
static mpt_attach_handler_t mpt_core_attach;
+static mpt_enable_handler_t mpt_core_enable;
static mpt_reset_handler_t mpt_core_ioc_reset;
static mpt_event_handler_t mpt_core_event;
static mpt_shutdown_handler_t mpt_core_shutdown;
@@ -208,6 +213,7 @@ static struct mpt_personality mpt_core_personality =
.name = "mpt_core",
.load = mpt_core_load,
.attach = mpt_core_attach,
+ .enable = mpt_core_enable,
.event = mpt_core_event,
.reset = mpt_core_ioc_reset,
.shutdown = mpt_core_shutdown,
@@ -226,8 +232,7 @@ static moduledata_t mpt_core_mod = {
DECLARE_MODULE(mpt_core, mpt_core_mod, SI_SUB_DRIVERS, SI_ORDER_FIRST);
MODULE_VERSION(mpt_core, 1);
-#define MPT_PERS_ATACHED(pers, mpt) \
- ((mpt)->pers_mask & (0x1 << pers->id))
+#define MPT_PERS_ATTACHED(pers, mpt) ((mpt)->mpt_pers_mask & (0x1 << pers->id))
int
@@ -274,8 +279,10 @@ mpt_modevent(module_t mod, int type, void *data)
}
case MOD_SHUTDOWN:
break;
+#if __FreeBSD_version >= 500000
case MOD_QUIESCE:
break;
+#endif
case MOD_UNLOAD:
error = pers->unload(pers);
mpt_personalities[pers->id] = NULL;
@@ -309,6 +316,13 @@ mpt_stdattach(struct mpt_softc *mpt)
}
int
+mpt_stdenable(struct mpt_softc *mpt)
+{
+ /* Enable is always successfull. */
+ return (0);
+}
+
+int
mpt_stdevent(struct mpt_softc *mpt, request_t *req, MSG_EVENT_NOTIFY_REPLY *msg)
{
mpt_lprt(mpt, MPT_PRT_DEBUG, "mpt_stdevent: 0x%x\n", msg->Event & 0xFF);
@@ -385,13 +399,14 @@ mpt_register_handler(struct mpt_softc *mpt, mpt_handler_type type,
* that the full table is checked to see if
* this handler was previously registered.
*/
- if (free_cbi == MPT_HANDLER_ID_NONE
- && (mpt_reply_handlers[cbi]
+ if (free_cbi == MPT_HANDLER_ID_NONE &&
+ (mpt_reply_handlers[cbi]
== mpt_default_reply_handler))
free_cbi = cbi;
}
- if (free_cbi == MPT_HANDLER_ID_NONE)
+ if (free_cbi == MPT_HANDLER_ID_NONE) {
return (ENOMEM);
+ }
mpt_reply_handlers[free_cbi] = handler.reply_handler;
*phandler_id = MPT_CBI_TO_HID(free_cbi);
break;
@@ -429,22 +444,23 @@ mpt_deregister_handler(struct mpt_softc *mpt, mpt_handler_type type,
static int
mpt_default_reply_handler(struct mpt_softc *mpt, request_t *req,
- MSG_DEFAULT_REPLY *reply_frame)
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
{
- mpt_prt(mpt, "XXXX Default Handler Called. Req %p, Frame %p\n",
- req, reply_frame);
+ mpt_prt(mpt,
+ "Default Handler Called: req=%p:%u reply_descriptor=%x frame=%p\n",
+ req, req->serno, reply_desc, reply_frame);
if (reply_frame != NULL)
mpt_dump_reply_frame(mpt, reply_frame);
- mpt_prt(mpt, "XXXX Reply Frame Ignored\n");
+ mpt_prt(mpt, "Reply Frame Ignored\n");
return (/*free_reply*/TRUE);
}
static int
mpt_config_reply_handler(struct mpt_softc *mpt, request_t *req,
- MSG_DEFAULT_REPLY *reply_frame)
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
{
if (req != NULL) {
@@ -471,7 +487,7 @@ mpt_config_reply_handler(struct mpt_softc *mpt, request_t *req,
static int
mpt_handshake_reply_handler(struct mpt_softc *mpt, request_t *req,
- MSG_DEFAULT_REPLY *reply_frame)
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
{
/* Nothing to be done. */
return (/*free_reply*/TRUE);
@@ -479,12 +495,13 @@ mpt_handshake_reply_handler(struct mpt_softc *mpt, request_t *req,
static int
mpt_event_reply_handler(struct mpt_softc *mpt, request_t *req,
- MSG_DEFAULT_REPLY *reply_frame)
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
{
int free_reply;
if (reply_frame == NULL) {
- mpt_prt(mpt, "Event Handler: req %p - Unexpected NULL reply\n");
+ mpt_prt(mpt, "Event Handler: req %p:%u - Unexpected NULL reply\n",
+ req, req->serno);
return (/*free_reply*/TRUE);
}
@@ -619,13 +636,15 @@ void
mpt_intr(void *arg)
{
struct mpt_softc *mpt;
- uint32_t reply_desc;
+ uint32_t reply_desc;
+ int ntrips = 0;
mpt = (struct mpt_softc *)arg;
while ((reply_desc = mpt_pop_reply_queue(mpt)) != MPT_REPLY_EMPTY) {
request_t *req;
MSG_DEFAULT_REPLY *reply_frame;
uint32_t reply_baddr;
+ uint32_t ctxt_idx;
u_int cb_index;
u_int req_index;
int free_rf;
@@ -635,27 +654,102 @@ mpt_intr(void *arg)
reply_baddr = 0;
if ((reply_desc & MPI_ADDRESS_REPLY_A_BIT) != 0) {
u_int offset;
-
/*
* Insure that the reply frame is coherent.
*/
reply_baddr = (reply_desc << 1);
offset = reply_baddr - (mpt->reply_phys & 0xFFFFFFFF);
- bus_dmamap_sync_range(mpt->reply_dmat, mpt->reply_dmap,
- offset, MPT_REPLY_SIZE,
- BUS_DMASYNC_POSTREAD);
+ bus_dmamap_sync_range(mpt->reply_dmat,
+ mpt->reply_dmap, offset, MPT_REPLY_SIZE,
+ BUS_DMASYNC_POSTREAD);
reply_frame = MPT_REPLY_OTOV(mpt, offset);
- reply_desc = le32toh(reply_frame->MsgContext);
+ ctxt_idx = le32toh(reply_frame->MsgContext);
+ } else {
+ uint32_t type;
+
+ type = MPI_GET_CONTEXT_REPLY_TYPE(reply_desc);
+ ctxt_idx = reply_desc;
+ mpt_lprt(mpt, MPT_PRT_DEBUG1, "Context Reply: 0x%08x\n",
+ reply_desc);
+
+ switch (type) {
+ case MPI_CONTEXT_REPLY_TYPE_SCSI_INIT:
+ ctxt_idx &= MPI_CONTEXT_REPLY_CONTEXT_MASK;
+ break;
+ case MPI_CONTEXT_REPLY_TYPE_SCSI_TARGET:
+ ctxt_idx = GET_IO_INDEX(reply_desc);
+ if (mpt->tgt_cmd_ptrs == NULL) {
+ mpt_prt(mpt,
+ "mpt_intr: no target cmd ptrs\n");
+ reply_desc = MPT_REPLY_EMPTY;
+ break;
+ }
+ if (ctxt_idx >= mpt->tgt_cmds_allocated) {
+ mpt_prt(mpt,
+ "mpt_intr: bad tgt cmd ctxt %u\n",
+ ctxt_idx);
+ reply_desc = MPT_REPLY_EMPTY;
+ ntrips = 1000;
+ break;
+ }
+ req = mpt->tgt_cmd_ptrs[ctxt_idx];
+ if (req == NULL) {
+ mpt_prt(mpt, "no request backpointer "
+ "at index %u", ctxt_idx);
+ reply_desc = MPT_REPLY_EMPTY;
+ ntrips = 1000;
+ break;
+ }
+ /*
+ * Reformulate ctxt_idx to be just as if
+ * it were another type of context reply
+ * so the code below will find the request
+ * via indexing into the pool.
+ */
+ ctxt_idx =
+ req->index | mpt->scsi_tgt_handler_id;
+ req = NULL;
+ break;
+ case MPI_CONTEXT_REPLY_TYPE_LAN:
+ mpt_prt(mpt, "LAN CONTEXT REPLY: 0x%08x\n",
+ reply_desc);
+ reply_desc = MPT_REPLY_EMPTY;
+ break;
+ default:
+ mpt_prt(mpt, "Context Reply 0x%08x?\n", type);
+ reply_desc = MPT_REPLY_EMPTY;
+ break;
+ }
+ if (reply_desc == MPT_REPLY_EMPTY) {
+ if (ntrips++ > 1000) {
+ break;
+ }
+ continue;
+ }
}
- cb_index = MPT_CONTEXT_TO_CBI(reply_desc);
- req_index = MPT_CONTEXT_TO_REQI(reply_desc);
- if (req_index < MPT_MAX_REQUESTS(mpt))
+
+ cb_index = MPT_CONTEXT_TO_CBI(ctxt_idx);
+ req_index = MPT_CONTEXT_TO_REQI(ctxt_idx);
+ if (req_index < MPT_MAX_REQUESTS(mpt)) {
req = &mpt->request_pool[req_index];
+ }
- free_rf = mpt_reply_handlers[cb_index](mpt, req, reply_frame);
+ free_rf = mpt_reply_handlers[cb_index](mpt, req,
+ reply_desc, reply_frame);
if (reply_frame != NULL && free_rf)
mpt_free_reply(mpt, reply_baddr);
+
+ /*
+ * If we got ourselves disabled, don't get stuck in a loop
+ */
+ if (mpt->disabled) {
+ mpt_disable_ints(mpt);
+ break;
+ }
+ if (ntrips++ > 1000) {
+ break;
+ }
}
}
@@ -678,7 +772,8 @@ mpt_complete_request_chain(struct mpt_softc *mpt, struct req_queue *chain,
ioc_status_frame.Function = msg_hdr->Function;
ioc_status_frame.MsgContext = msg_hdr->MsgContext;
cb_index = MPT_CONTEXT_TO_CBI(le32toh(msg_hdr->MsgContext));
- mpt_reply_handlers[cb_index](mpt, req, &ioc_status_frame);
+ mpt_reply_handlers[cb_index](mpt, req, msg_hdr->MsgContext,
+ &ioc_status_frame);
}
}
@@ -718,12 +813,11 @@ mpt_wait_db_ack(struct mpt_softc *mpt)
for (i=0; i < MPT_MAX_WAIT; i++) {
if (!MPT_DB_IS_BUSY(mpt_rd_intr(mpt))) {
maxwait_ack = i > maxwait_ack ? i : maxwait_ack;
- return MPT_OK;
+ return (MPT_OK);
}
-
- DELAY(1000);
+ DELAY(200);
}
- return MPT_FAIL;
+ return (MPT_FAIL);
}
/* Busy wait for a door bell interrupt */
@@ -738,7 +832,7 @@ mpt_wait_db_int(struct mpt_softc *mpt)
}
DELAY(100);
}
- return MPT_FAIL;
+ return (MPT_FAIL);
}
/* Wait for IOC to transition to a give state */
@@ -782,7 +876,7 @@ mpt_soft_reset(struct mpt_softc *mpt)
/* Have to use hard reset if we are not in Running state */
if (MPT_STATE(mpt_rd_db(mpt)) != MPT_DB_STATE_RUNNING) {
mpt_prt(mpt, "soft reset failed: device not running\n");
- return MPT_FAIL;
+ return (MPT_FAIL);
}
/* If door bell is in use we don't have a chance of getting
@@ -791,7 +885,7 @@ mpt_soft_reset(struct mpt_softc *mpt)
*/
if (MPT_DB_IS_IN_USE(mpt_rd_db(mpt))) {
mpt_prt(mpt, "soft reset failed: doorbell wedged\n");
- return MPT_FAIL;
+ return (MPT_FAIL);
}
/* Send the reset request to the IOC */
@@ -799,13 +893,13 @@ mpt_soft_reset(struct mpt_softc *mpt)
MPI_FUNCTION_IOC_MESSAGE_UNIT_RESET << MPI_DOORBELL_FUNCTION_SHIFT);
if (mpt_wait_db_ack(mpt) != MPT_OK) {
mpt_prt(mpt, "soft reset failed: ack timeout\n");
- return MPT_FAIL;
+ return (MPT_FAIL);
}
/* Wait for the IOC to reload and come out of reset state */
if (mpt_wait_state(mpt, MPT_DB_STATE_READY) != MPT_OK) {
mpt_prt(mpt, "soft reset failed: device did not restart\n");
- return MPT_FAIL;
+ return (MPT_FAIL);
}
return MPT_OK;
@@ -979,7 +1073,7 @@ mpt_reset(struct mpt_softc *mpt, int reinit)
}
if (reinit != 0) {
- ret = mpt_enable_ioc(mpt);
+ ret = mpt_enable_ioc(mpt, 1);
if (ret == MPT_OK) {
mpt_enable_ints(mpt);
}
@@ -1010,7 +1104,10 @@ mpt_free_request(struct mpt_softc *mpt, request_t *req)
req->ccb = NULL;
req->state = REQ_STATE_FREE;
if (LIST_EMPTY(&mpt->ack_frames)) {
- TAILQ_INSERT_HEAD(&mpt->request_free_list, req, links);
+ /*
+ * Insert free ones at the tail
+ */
+ TAILQ_INSERT_TAIL(&mpt->request_free_list, req, links);
if (mpt->getreqwaiter != 0) {
mpt->getreqwaiter = 0;
wakeup(&mpt->request_free_list);
@@ -1043,9 +1140,7 @@ retry:
TAILQ_REMOVE(&mpt->request_free_list, req, links);
req->state = REQ_STATE_ALLOCATED;
req->chain = NULL;
- if ((req->serno = ++(mpt->cmd_serno)) == 0) {
- req->serno = ++(mpt->cmd_serno);
- }
+ req->serno = mpt->sequence++;
} else if (sleep_ok != 0) {
mpt->getreqwaiter = 1;
mpt_sleep(mpt, &mpt->request_free_list, PUSER, "mptgreq", 0);
@@ -1063,8 +1158,13 @@ mpt_send_cmd(struct mpt_softc *mpt, request_t *req)
pReq = req->req_vbuf;
if (mpt->verbose > MPT_PRT_TRACE) {
int offset;
- mpt_prt(mpt, "Send Request %d (0x%x):",
- req->index, req->req_pbuf);
+#if __FreeBSD_version >= 500000
+ mpt_prt(mpt, "Send Request %d (%jx):",
+ req->index, (uintmax_t) req->req_pbuf);
+#else
+ mpt_prt(mpt, "Send Request %d (%llx):",
+ req->index, (unsigned long long) req->req_pbuf);
+#endif
for (offset = 0; offset < mpt->request_frame_size; offset++) {
if ((offset & 0x7) == 0) {
mpt_prtc(mpt, "\n");
@@ -1108,16 +1208,15 @@ mpt_wait_req(struct mpt_softc *mpt, request_t *req,
* Convert to ticks or 500us units depending on
* our sleep mode.
*/
- if (sleep_ok != 0)
+ if (sleep_ok != 0) {
timeout = (time_ms * hz) / 1000;
- else
+ } else {
timeout = time_ms * 2;
+ }
req->state |= REQ_STATE_NEED_WAKEUP;
mask &= ~REQ_STATE_NEED_WAKEUP;
saved_cnt = mpt->reset_cnt;
- while ((req->state & mask) != state
- && mpt->reset_cnt == saved_cnt) {
-
+ while ((req->state & mask) != state && mpt->reset_cnt == saved_cnt) {
if (sleep_ok != 0) {
error = mpt_sleep(mpt, req, PUSER, "mptreq", timeout);
if (error == EWOULDBLOCK) {
@@ -1126,7 +1225,6 @@ mpt_wait_req(struct mpt_softc *mpt, request_t *req,
}
} else {
if (time_ms != 0 && --timeout == 0) {
- mpt_prt(mpt, "mpt_wait_req timed out\n");
break;
}
DELAY(500);
@@ -1134,10 +1232,14 @@ mpt_wait_req(struct mpt_softc *mpt, request_t *req,
}
}
req->state &= ~REQ_STATE_NEED_WAKEUP;
- if (mpt->reset_cnt != saved_cnt)
+ if (mpt->reset_cnt != saved_cnt) {
return (EIO);
- if (time_ms && timeout <= 0)
+ }
+ if (time_ms && timeout <= 0) {
+ MSG_REQUEST_HEADER *msg_hdr = req->req_vbuf;
+ mpt_prt(mpt, "mpt_wait_req(%x) timed out\n", msg_hdr->Function);
return (ETIMEDOUT);
+ }
return (0);
}
@@ -1436,8 +1538,6 @@ mpt_read_cfg_header(struct mpt_softc *mpt, int PageType, int PageNumber,
return (error);
}
-#define CFG_DATA_OFF 128
-
int
mpt_read_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress,
CONFIG_PAGE_HEADER *hdr, size_t len, int sleep_ok,
@@ -1455,7 +1555,7 @@ mpt_read_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress,
error = mpt_issue_cfg_req(mpt, req, Action, hdr->PageVersion,
hdr->PageLength, hdr->PageNumber,
hdr->PageType & MPI_CONFIG_PAGETYPE_MASK,
- PageAddress, req->req_pbuf + CFG_DATA_OFF,
+ PageAddress, req->req_pbuf + MPT_RQSL(mpt),
len, sleep_ok, timeout_ms);
if (error != 0) {
mpt_prt(mpt, "read_cfg_page(%d) timed out\n", Action);
@@ -1470,7 +1570,7 @@ mpt_read_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress,
}
bus_dmamap_sync(mpt->request_dmat, mpt->request_dmap,
BUS_DMASYNC_POSTREAD);
- memcpy(hdr, ((uint8_t *)req->req_vbuf)+CFG_DATA_OFF, len);
+ memcpy(hdr, ((uint8_t *)req->req_vbuf)+MPT_RQSL(mpt), len);
mpt_free_request(mpt, req);
return (0);
}
@@ -1497,14 +1597,14 @@ mpt_write_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress,
if (req == NULL)
return (-1);
- memcpy(((caddr_t)req->req_vbuf)+CFG_DATA_OFF, hdr, len);
+ memcpy(((caddr_t)req->req_vbuf)+MPT_RQSL(mpt), hdr, len);
/* Restore stripped out attributes */
hdr->PageType |= hdr_attr;
error = mpt_issue_cfg_req(mpt, req, Action, hdr->PageVersion,
hdr->PageLength, hdr->PageNumber,
hdr->PageType & MPI_CONFIG_PAGETYPE_MASK,
- PageAddress, req->req_pbuf + CFG_DATA_OFF,
+ PageAddress, req->req_pbuf + MPT_RQSL(mpt),
len, sleep_ok, timeout_ms);
if (error != 0) {
mpt_prt(mpt, "mpt_write_cfg_page timed out\n");
@@ -1674,236 +1774,6 @@ mpt_read_config_info_ioc(struct mpt_softc *mpt)
}
/*
- * Read SCSI configuration information
- */
-static int
-mpt_read_config_info_spi(struct mpt_softc *mpt)
-{
- int rv, i;
-
- rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 0,
- 0, &mpt->mpt_port_page0.Header,
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv)
- return (-1);
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Port Page 0 Header: %x %x %x %x\n",
- mpt->mpt_port_page0.Header.PageVersion,
- mpt->mpt_port_page0.Header.PageLength,
- mpt->mpt_port_page0.Header.PageNumber,
- mpt->mpt_port_page0.Header.PageType);
-
- rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 1,
- 0, &mpt->mpt_port_page1.Header,
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv)
- return (-1);
-
- mpt_lprt(mpt, MPT_PRT_DEBUG, "SPI Port Page 1 Header: %x %x %x %x\n",
- mpt->mpt_port_page1.Header.PageVersion,
- mpt->mpt_port_page1.Header.PageLength,
- mpt->mpt_port_page1.Header.PageNumber,
- mpt->mpt_port_page1.Header.PageType);
-
- rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 2,
- /*PageAddress*/0, &mpt->mpt_port_page2.Header,
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv)
- return (-1);
-
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Port Page 2 Header: %x %x %x %x\n",
- mpt->mpt_port_page1.Header.PageVersion,
- mpt->mpt_port_page1.Header.PageLength,
- mpt->mpt_port_page1.Header.PageNumber,
- mpt->mpt_port_page1.Header.PageType);
-
- for (i = 0; i < 16; i++) {
- rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE,
- 0, i, &mpt->mpt_dev_page0[i].Header,
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv)
- return (-1);
-
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Target %d Device Page 0 Header: %x %x %x %x\n",
- i, mpt->mpt_dev_page0[i].Header.PageVersion,
- mpt->mpt_dev_page0[i].Header.PageLength,
- mpt->mpt_dev_page0[i].Header.PageNumber,
- mpt->mpt_dev_page0[i].Header.PageType);
-
- rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE,
- 1, i, &mpt->mpt_dev_page1[i].Header,
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv)
- return (-1);
-
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Target %d Device Page 1 Header: %x %x %x %x\n",
- i, mpt->mpt_dev_page1[i].Header.PageVersion,
- mpt->mpt_dev_page1[i].Header.PageLength,
- mpt->mpt_dev_page1[i].Header.PageNumber,
- mpt->mpt_dev_page1[i].Header.PageType);
- }
-
- /*
- * At this point, we don't *have* to fail. As long as we have
- * valid config header information, we can (barely) lurch
- * along.
- */
-
- rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
- &mpt->mpt_port_page0.Header,
- sizeof(mpt->mpt_port_page0),
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv) {
- mpt_prt(mpt, "failed to read SPI Port Page 0\n");
- } else {
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Port Page 0: Capabilities %x PhysicalInterface %x\n",
- mpt->mpt_port_page0.Capabilities,
- mpt->mpt_port_page0.PhysicalInterface);
- }
-
- rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
- &mpt->mpt_port_page1.Header,
- sizeof(mpt->mpt_port_page1),
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv) {
- mpt_prt(mpt, "failed to read SPI Port Page 1\n");
- } else {
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Port Page 1: Configuration %x OnBusTimerValue %x\n",
- mpt->mpt_port_page1.Configuration,
- mpt->mpt_port_page1.OnBusTimerValue);
- }
-
- rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
- &mpt->mpt_port_page2.Header,
- sizeof(mpt->mpt_port_page2),
- /*sleep_ok*/FALSE, /*timeout_ms*/5000);
- if (rv) {
- mpt_prt(mpt, "failed to read SPI Port Page 2\n");
- } else {
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Port Page 2: Flags %x Settings %x\n",
- mpt->mpt_port_page2.PortFlags,
- mpt->mpt_port_page2.PortSettings);
- for (i = 0; i < 16; i++) {
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Port Page 2 Tgt %d: timo %x SF %x Flags %x\n",
- i, mpt->mpt_port_page2.DeviceSettings[i].Timeout,
- mpt->mpt_port_page2.DeviceSettings[i].SyncFactor,
- mpt->mpt_port_page2.DeviceSettings[i].DeviceFlags);
- }
- }
-
- for (i = 0; i < 16; i++) {
- rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i,
- &mpt->mpt_dev_page0[i].Header,
- sizeof(*mpt->mpt_dev_page0),
- /*sleep_ok*/FALSE,
- /*timeout_ms*/5000);
- if (rv) {
- mpt_prt(mpt,
- "cannot read SPI Tgt %d Device Page 0\n", i);
- continue;
- }
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Tgt %d Page 0: NParms %x Information %x",
- i, mpt->mpt_dev_page0[i].NegotiatedParameters,
- mpt->mpt_dev_page0[i].Information);
-
- rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i,
- &mpt->mpt_dev_page1[i].Header,
- sizeof(*mpt->mpt_dev_page1),
- /*sleep_ok*/FALSE,
- /*timeout_ms*/5000);
- if (rv) {
- mpt_prt(mpt,
- "cannot read SPI Tgt %d Device Page 1\n", i);
- continue;
- }
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Tgt %d Page 1: RParms %x Configuration %x\n",
- i, mpt->mpt_dev_page1[i].RequestedParameters,
- mpt->mpt_dev_page1[i].Configuration);
- }
- return (0);
-}
-
-/*
- * Validate SPI configuration information.
- *
- * In particular, validate SPI Port Page 1.
- */
-static int
-mpt_set_initial_config_spi(struct mpt_softc *mpt)
-{
- int i, pp1val = ((1 << mpt->mpt_ini_id) << 16) | mpt->mpt_ini_id;
- int error;
-
- mpt->mpt_disc_enable = 0xff;
- mpt->mpt_tag_enable = 0;
-
- if (mpt->mpt_port_page1.Configuration != pp1val) {
- CONFIG_PAGE_SCSI_PORT_1 tmp;
-
- mpt_prt(mpt,
- "SPI Port Page 1 Config value bad (%x)- should be %x\n",
- mpt->mpt_port_page1.Configuration, pp1val);
- tmp = mpt->mpt_port_page1;
- tmp.Configuration = pp1val;
- error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/0,
- &tmp.Header, sizeof(tmp),
- /*sleep_ok*/FALSE,
- /*timeout_ms*/5000);
- if (error)
- return (-1);
- error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
- &tmp.Header, sizeof(tmp),
- /*sleep_ok*/FALSE,
- /*timeout_ms*/5000);
- if (error)
- return (-1);
- if (tmp.Configuration != pp1val) {
- mpt_prt(mpt,
- "failed to reset SPI Port Page 1 Config value\n");
- return (-1);
- }
- mpt->mpt_port_page1 = tmp;
- }
-
- for (i = 0; i < 16; i++) {
- CONFIG_PAGE_SCSI_DEVICE_1 tmp;
- tmp = mpt->mpt_dev_page1[i];
- tmp.RequestedParameters = 0;
- tmp.Configuration = 0;
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "Set Tgt %d SPI DevicePage 1 values to %x 0 %x\n",
- i, tmp.RequestedParameters, tmp.Configuration);
- error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/i,
- &tmp.Header, sizeof(tmp),
- /*sleep_ok*/FALSE,
- /*timeout_ms*/5000);
- if (error)
- return (-1);
- error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i,
- &tmp.Header, sizeof(tmp),
- /*sleep_ok*/FALSE,
- /*timeout_ms*/5000);
- if (error)
- return (-1);
- mpt->mpt_dev_page1[i] = tmp;
- mpt_lprt(mpt, MPT_PRT_DEBUG,
- "SPI Tgt %d Page 1: RParm %x Configuration %x\n", i,
- mpt->mpt_dev_page1[i].RequestedParameters,
- mpt->mpt_dev_page1[i].Configuration);
- }
- return (0);
-}
-
-/*
* Enable IOC port
*/
static int
@@ -1918,7 +1788,7 @@ mpt_send_port_enable(struct mpt_softc *mpt, int port)
return (-1);
enable_req = req->req_vbuf;
- bzero(enable_req, sizeof *enable_req);
+ bzero(enable_req, MPT_RQSL(mpt));
enable_req->Function = MPI_FUNCTION_PORT_ENABLE;
enable_req->MsgContext = htole32(req->index | MPT_REPLY_HANDLER_CONFIG);
@@ -1932,10 +1802,11 @@ mpt_send_port_enable(struct mpt_softc *mpt, int port)
/*sleep_ok*/FALSE,
/*time_ms*/(mpt->is_sas || mpt->is_fc)? 30000 : 3000);
if (error != 0) {
- mpt_prt(mpt, "port enable timed out\n");
+ mpt_prt(mpt, "port %d enable timed out\n", port);
return (-1);
}
mpt_free_request(mpt, req);
+ mpt_lprt(mpt, MPT_PRT_DEBUG, "enabled port %d\n", port);
return (0);
}
@@ -1992,27 +1863,28 @@ mpt_disable_ints(struct mpt_softc *mpt)
static void
mpt_sysctl_attach(struct mpt_softc *mpt)
{
+#if __FreeBSD_version >= 500000
struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(mpt->dev);
struct sysctl_oid *tree = device_get_sysctl_tree(mpt->dev);
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(tree), OID_AUTO,
"debug", CTLFLAG_RW, &mpt->verbose, 0,
"Debugging/Verbose level");
+#endif
}
int
mpt_attach(struct mpt_softc *mpt)
{
+ struct mpt_personality *pers;
int i;
+ int error;
for (i = 0; i < MPT_MAX_PERSONALITIES; i++) {
- struct mpt_personality *pers;
- int error;
-
pers = mpt_personalities[i];
- if (pers == NULL)
+ if (pers == NULL) {
continue;
-
+ }
if (pers->probe(mpt) == 0) {
error = pers->attach(mpt);
if (error != 0) {
@@ -2024,6 +1896,24 @@ mpt_attach(struct mpt_softc *mpt)
}
}
+ /*
+ * Now that we've attached everything, do the enable function
+ * for all of the personalities. This allows the personalities
+ * to do setups that are appropriate for them prior to enabling
+ * any ports.
+ */
+ for (i = 0; i < MPT_MAX_PERSONALITIES; i++) {
+ pers = mpt_personalities[i];
+ if (pers != NULL && MPT_PERS_ATTACHED(pers, mpt) != 0) {
+ error = pers->enable(mpt);
+ if (error != 0) {
+ mpt_prt(mpt, "personality %s attached but would"
+ " not enable (%d)\n", pers->name, error);
+ mpt_detach(mpt);
+ return (error);
+ }
+ }
+ }
return (0);
}
@@ -2032,10 +1922,9 @@ mpt_shutdown(struct mpt_softc *mpt)
{
struct mpt_personality *pers;
- MPT_PERS_FOREACH_REVERSE(mpt, pers)
+ MPT_PERS_FOREACH_REVERSE(mpt, pers) {
pers->shutdown(mpt);
-
- mpt_reset(mpt, /*reinit*/FALSE);
+ }
return (0);
}
@@ -2062,8 +1951,9 @@ mpt_core_load(struct mpt_personality *pers)
* Setup core handlers and insert the default handler
* into all "empty slots".
*/
- for (i = 0; i < MPT_NUM_REPLY_HANDLERS; i++)
+ for (i = 0; i < MPT_NUM_REPLY_HANDLERS; i++) {
mpt_reply_handlers[i] = mpt_default_reply_handler;
+ }
mpt_reply_handlers[MPT_CBI(MPT_REPLY_HANDLER_EVENTS)] =
mpt_event_reply_handler;
@@ -2071,7 +1961,6 @@ mpt_core_load(struct mpt_personality *pers)
mpt_config_reply_handler;
mpt_reply_handlers[MPT_CBI(MPT_REPLY_HANDLER_HANDSHAKE)] =
mpt_handshake_reply_handler;
-
return (0);
}
@@ -2085,32 +1974,92 @@ mpt_core_attach(struct mpt_softc *mpt)
int val;
int error;
+
LIST_INIT(&mpt->ack_frames);
/* Put all request buffers on the free list */
TAILQ_INIT(&mpt->request_pending_list);
TAILQ_INIT(&mpt->request_free_list);
- for (val = 0; val < MPT_MAX_REQUESTS(mpt); val++)
+ for (val = 0; val < MPT_MAX_REQUESTS(mpt); val++) {
mpt_free_request(mpt, &mpt->request_pool[val]);
+ }
+
+ for (val = 0; val < MPT_MAX_LUNS; val++) {
+ STAILQ_INIT(&mpt->trt[val].atios);
+ STAILQ_INIT(&mpt->trt[val].inots);
+ }
+ STAILQ_INIT(&mpt->trt_wildcard.atios);
+ STAILQ_INIT(&mpt->trt_wildcard.inots);
+
+ mpt->scsi_tgt_handler_id = MPT_HANDLER_ID_NONE;
mpt_sysctl_attach(mpt);
mpt_lprt(mpt, MPT_PRT_DEBUG, "doorbell req = %s\n",
- mpt_ioc_diag(mpt_read(mpt, MPT_OFFSET_DOORBELL)));
+ mpt_ioc_diag(mpt_read(mpt, MPT_OFFSET_DOORBELL)));
error = mpt_configure_ioc(mpt);
return (error);
}
+int
+mpt_core_enable(struct mpt_softc *mpt)
+{
+ /*
+ * We enter with the IOC enabled, but async events
+ * not enabled, ports not enabled and interrupts
+ * not enabled.
+ */
+
+ /*
+ * Enable asynchronous event reporting- all personalities
+ * have attached so that they should be able to now field
+ * async events.
+ */
+ mpt_send_event_request(mpt, 1);
+
+ /*
+ * Catch any pending interrupts
+ *
+ * This seems to be crucial- otherwise
+ * the portenable below times out.
+ */
+ mpt_intr(mpt);
+
+ /*
+ * Enable Interrupts
+ */
+ mpt_enable_ints(mpt);
+
+ /*
+ * Catch any pending interrupts
+ *
+ * This seems to be crucial- otherwise
+ * the portenable below times out.
+ */
+ mpt_intr(mpt);
+
+ /*
+ * Enable the port- but only if we are not MPT_ROLE_NONE.
+ */
+ if (mpt_send_port_enable(mpt, 0) != MPT_OK) {
+ mpt_prt(mpt, "failed to enable port 0\n");
+ return (ENXIO);
+ }
+ return (0);
+}
+
void
mpt_core_shutdown(struct mpt_softc *mpt)
{
+ mpt_disable_ints(mpt);
}
void
mpt_core_detach(struct mpt_softc *mpt)
{
+ mpt_disable_ints(mpt);
}
int
@@ -2262,10 +2211,11 @@ mpt_configure_ioc(struct mpt_softc *mpt)
* first channel is ok, the second will not require a hard
* reset.
*/
- if (needreset || (mpt_rd_db(mpt) & MPT_DB_STATE_MASK) !=
+ if (needreset || MPT_STATE(mpt_rd_db(mpt)) !=
MPT_DB_STATE_READY) {
- if (mpt_reset(mpt, /*reinit*/FALSE) != MPT_OK)
+ if (mpt_reset(mpt, FALSE) != MPT_OK) {
continue;
+ }
}
needreset = 0;
@@ -2408,10 +2358,8 @@ mpt_configure_ioc(struct mpt_softc *mpt)
pfp.PortType);
return (ENXIO);
}
- if (!(pfp.ProtocolFlags & MPI_PORTFACTS_PROTOCOL_INITIATOR)) {
- mpt_prt(mpt, "initiator role unsupported\n");
- return (ENXIO);
- }
+ mpt->mpt_max_tgtcmds = le16toh(pfp.MaxPostedCmdBuffers);
+
if (pfp.PortType == MPI_PORTFACTS_PORTTYPE_FC) {
mpt->is_fc = 1;
mpt->is_sas = 0;
@@ -2425,30 +2373,49 @@ mpt_configure_ioc(struct mpt_softc *mpt)
mpt->mpt_ini_id = pfp.PortSCSIID;
mpt->mpt_max_devices = pfp.MaxDevices;
- if (mpt_enable_ioc(mpt) != 0) {
- mpt_prt(mpt, "Unable to initialize IOC\n");
+ /*
+ * Match our expected role with what this port supports.
+ *
+ * We only do this to meet expectations. That is, if the
+ * user has specified they want initiator role, and we
+ * don't support it, that's an error we return back upstream.
+ */
+
+ mpt->cap = MPT_ROLE_NONE;
+ if (pfp.ProtocolFlags & MPI_PORTFACTS_PROTOCOL_INITIATOR) {
+ mpt->cap |= MPT_ROLE_INITIATOR;
+ }
+ if (pfp.ProtocolFlags & MPI_PORTFACTS_PROTOCOL_TARGET) {
+ mpt->cap |= MPT_ROLE_TARGET;
+ }
+ if (mpt->cap == MPT_ROLE_NONE) {
+ mpt_prt(mpt, "port does not support either target or "
+ "initiator role\n");
+ return (ENXIO);
+ }
+
+ if ((mpt->role & MPT_ROLE_INITIATOR) &&
+ (mpt->cap & MPT_ROLE_INITIATOR) == 0) {
+ mpt_prt(mpt, "port does not support initiator role\n");
+ return (ENXIO);
+ }
+
+ if ((mpt->role & MPT_ROLE_TARGET) &&
+ (mpt->cap & MPT_ROLE_TARGET) == 0) {
+ mpt_prt(mpt, "port does not support target role\n");
+ return (ENXIO);
+ }
+
+ if (mpt_enable_ioc(mpt, 0) != MPT_OK) {
+ mpt_prt(mpt, "unable to initialize IOC\n");
return (ENXIO);
}
/*
- * Read and set up initial configuration information
- * (IOC and SPI only for now)
- *
- * XXX Should figure out what "personalities" are
- * available and defer all initialization junk to
- * them.
+ * Read IOC configuration information.
*/
mpt_read_config_info_ioc(mpt);
- if (mpt->is_fc == 0 && mpt->is_sas == 0) {
- if (mpt_read_config_info_spi(mpt)) {
- return (EIO);
- }
- if (mpt_set_initial_config_spi(mpt)) {
- return (EIO);
- }
- }
-
/* Everything worked */
break;
}
@@ -2458,14 +2425,11 @@ mpt_configure_ioc(struct mpt_softc *mpt)
return (EIO);
}
- mpt_lprt(mpt, MPT_PRT_DEBUG, "enabling interrupts\n");
-
- mpt_enable_ints(mpt);
return (0);
}
static int
-mpt_enable_ioc(struct mpt_softc *mpt)
+mpt_enable_ioc(struct mpt_softc *mpt, int portenable)
{
uint32_t pptr;
int val;
@@ -2496,20 +2460,20 @@ mpt_enable_ioc(struct mpt_softc *mpt)
break;
}
- /*
- * Enable asynchronous event reporting
- */
- mpt_send_event_request(mpt, 1);
/*
- * Enable the port
+ * Enable the port if asked
*/
- if (mpt_send_port_enable(mpt, 0) != MPT_OK) {
- mpt_prt(mpt, "failed to enable port 0\n");
- return (ENXIO);
- }
- mpt_lprt(mpt, MPT_PRT_DEBUG, "enabled port 0\n");
-
+ if (portenable) {
+ /*
+ * Enable asynchronous event reporting
+ */
+ mpt_send_event_request(mpt, 1);
+ if (mpt_send_port_enable(mpt, 0) != MPT_OK) {
+ mpt_prt(mpt, "failed to enable port 0\n");
+ return (ENXIO);
+ }
+ }
return (MPT_OK);
}
diff --git a/sys/dev/mpt/mpt.h b/sys/dev/mpt/mpt.h
index b88d94b11729..a38731c957e4 100644
--- a/sys/dev/mpt/mpt.h
+++ b/sys/dev/mpt/mpt.h
@@ -121,6 +121,10 @@
#include <machine/cpu.h>
#include <machine/resource.h>
+#if __FreeBSD_version < 500000
+#include <machine/bus.h>
+#endif
+
#include <sys/rman.h>
#if __FreeBSD_version < 500000
@@ -153,6 +157,12 @@
#define NUM_ELEMENTS(array) (sizeof(array) / sizeof(*array))
+#define MPT_ROLE_NONE 0
+#define MPT_ROLE_INITIATOR 1
+#define MPT_ROLE_TARGET 2
+#define MPT_ROLE_BOTH 3
+#define MPT_ROLE_DEFAULT MPT_ROLE_INITIATOR
+
/**************************** Forward Declarations ****************************/
struct mpt_softc;
struct mpt_personality;
@@ -162,6 +172,7 @@ typedef struct req_entry request_t;
typedef int mpt_load_handler_t(struct mpt_personality *);
typedef int mpt_probe_handler_t(struct mpt_softc *);
typedef int mpt_attach_handler_t(struct mpt_softc *);
+typedef int mpt_enable_handler_t(struct mpt_softc *);
typedef int mpt_event_handler_t(struct mpt_softc *, request_t *,
MSG_EVENT_NOTIFY_REPLY *);
typedef void mpt_reset_handler_t(struct mpt_softc *, int /*type*/);
@@ -179,6 +190,7 @@ struct mpt_personality
#define MPT_PERS_FIRST_HANDLER(pers) (&(pers)->load)
mpt_probe_handler_t *probe; /* configure personailty */
mpt_attach_handler_t *attach; /* initialize device instance */
+ mpt_enable_handler_t *enable; /* enable device */
mpt_event_handler_t *event; /* Handle MPI event. */
mpt_reset_handler_t *reset; /* Re-init after reset. */
mpt_shutdown_handler_t *shutdown; /* Shutdown instance. */
@@ -280,7 +292,7 @@ struct req_entry {
mpt_req_state_t state; /* Request State Information */
uint16_t index; /* Index of this entry */
uint16_t IOCStatus; /* Completion status */
- uint32_t serno; /* serial number */
+ uint16_t serno; /* serial number */
union ccb *ccb; /* CAM request */
void *req_vbuf; /* Virtual Address of Entry */
void *sense_vbuf; /* Virtual Address of sense data */
@@ -290,6 +302,59 @@ struct req_entry {
struct req_entry *chain; /* for SGE overallocations */
};
+/**************************** MPI Target State Info ***************************/
+
+typedef struct {
+ uint32_t reply_desc; /* current reply descriptor */
+ uint32_t resid; /* current data residual */
+ uint32_t bytes_xfered; /* current relative offset */
+ union ccb *ccb; /* pointer to currently active ccb */
+ request_t *req; /* pointer to currently active assist request */
+ int flags;
+#define BOGUS_JO 0x01
+ int nxfers;
+ enum {
+ TGT_STATE_NIL,
+ TGT_STATE_LOADED,
+ TGT_STATE_IN_CAM,
+ TGT_STATE_SETTING_UP_FOR_DATA,
+ TGT_STATE_MOVING_DATA,
+ TGT_STATE_MOVING_DATA_AND_STATUS,
+ TGT_STATE_SENDING_STATUS
+ } state;
+} mpt_tgt_state_t;
+
+/*
+ * When we get an incoming command it has its own tag which is called the
+ * IoIndex. This is the value we gave that particular command buffer when
+ * we originally assigned it. It's just a number, really. The FC card uses
+ * it as an RX_ID. We can use it to index into mpt->tgt_cmd_ptrs, which
+ * contains pointers the request_t structures related to that IoIndex.
+ *
+ * What *we* do is construct a tag out of the index for the target command
+ * which owns the incoming ATIO plus a rolling sequence number.
+ */
+#define MPT_MAKE_TAGID(mpt, req, ioindex) \
+ ((ioindex << 16) | (mpt->sequence++))
+
+#ifdef INVARIANTS
+#define MPT_TAG_2_REQ(a, b) mpt_tag_2_req(a, (uint32_t) b)
+#else
+#define MPT_TAG_2_REQ(mpt, tag) mpt->tgt_cmd_ptrs[tag >> 16]
+#endif
+
+#define MPT_TGT_STATE(mpt, req) ((mpt_tgt_state_t *) \
+ (&((uint8_t *)req->req_vbuf)[MPT_RQSL(mpt) - sizeof (mpt_tgt_state_t)]))
+
+STAILQ_HEAD(mpt_hdr_stailq, ccb_hdr);
+#define MPT_MAX_LUNS 256
+typedef struct {
+ struct mpt_hdr_stailq atios;
+ struct mpt_hdr_stailq inots;
+ int enabled;
+} tgt_resource_t;
+#define MPT_MAX_ELS 8
+
/**************************** Handler Registration ****************************/
/*
* Global table of registered reply handlers. The
@@ -311,12 +376,12 @@ struct req_entry {
* all commonly executed handlers fit in a single cache
* line.
*/
-#define MPT_NUM_REPLY_HANDLERS (16)
+#define MPT_NUM_REPLY_HANDLERS (32)
#define MPT_REPLY_HANDLER_EVENTS MPT_CBI_TO_HID(0)
#define MPT_REPLY_HANDLER_CONFIG MPT_CBI_TO_HID(MPT_NUM_REPLY_HANDLERS-1)
#define MPT_REPLY_HANDLER_HANDSHAKE MPT_CBI_TO_HID(MPT_NUM_REPLY_HANDLERS-2)
typedef int mpt_reply_handler_t(struct mpt_softc *mpt, request_t *request,
- MSG_DEFAULT_REPLY *reply_frame);
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame);
typedef union {
mpt_reply_handler_t *reply_handler;
} mpt_handler_t;
@@ -417,21 +482,24 @@ struct mpt_softc {
struct mtx mpt_lock;
#endif
uint32_t mpt_pers_mask;
- uint32_t : 14,
- is_sas : 1,
+ uint32_t : 8,
+ unit : 8,
+ : 1,
+ twildcard : 1,
+ tenabled : 1,
+ cap : 2, /* none, ini, target, both */
+ role : 2, /* none, ini, target, both */
raid_mwce_set : 1,
getreqwaiter : 1,
shutdwn_raid : 1,
shutdwn_recovery: 1,
- unit : 8,
outofbeer : 1,
mpt_locksetup : 1,
disabled : 1,
- is_fc : 1,
- bus : 1; /* FC929/1030 have two busses */
+ is_sas : 1,
+ is_fc : 1;
u_int verbose;
- uint32_t cmd_serno;
/*
* IOC Facts
@@ -449,6 +517,7 @@ struct mpt_softc {
uint16_t mpt_ini_id;
uint16_t mpt_port_type;
uint16_t mpt_proto_flags;
+ uint16_t mpt_max_tgtcmds;
/*
* Device Configuration Information
@@ -475,7 +544,8 @@ struct mpt_softc {
#define mpt_update_params0 cfg.spi._update_params0
#define mpt_update_params1 cfg.spi._update_params1
struct mpi_fc_cfg {
- uint8_t nada;
+ CONFIG_PAGE_FC_PORT_0 _port_page0;
+#define mpt_fcport_page0 cfg.fc._port_page0
} fc;
} cfg;
@@ -528,7 +598,7 @@ struct mpt_softc {
bus_dma_tag_t request_dmat; /* DMA tag for request memroy */
bus_dmamap_t request_dmap; /* DMA map for request memroy */
uint8_t *request; /* KVA of Request memory */
- bus_addr_t request_phys; /* BusADdr of request memory */
+ bus_addr_t request_phys; /* BusAddr of request memory */
uint32_t max_seg_cnt; /* calculated after IOC facts */
@@ -560,9 +630,28 @@ struct mpt_softc {
struct proc *recovery_thread;
request_t *tmf_req;
- uint32_t sequence; /* Sequence Number */
- uint32_t timeouts; /* timeout count */
- uint32_t success; /* successes afer timeout */
+ /*
+ * Target Mode Support
+ */
+ uint32_t scsi_tgt_handler_id;
+ request_t ** tgt_cmd_ptrs;
+
+ /*
+ * *snork*- this is chosen to be here *just in case* somebody
+ * forgets to point to it exactly and we index off of trt with
+ * CAM_LUN_WILDCARD.
+ */
+ tgt_resource_t trt_wildcard; /* wildcard luns */
+ tgt_resource_t trt[MPT_MAX_LUNS];
+ uint16_t tgt_cmds_allocated;
+
+ /*
+ * Stuff..
+ */
+ uint16_t sequence; /* Sequence Number */
+ uint16_t timeouts; /* timeout count */
+ uint16_t success; /* successes afer timeout */
+
/* Opposing port in a 929 or 1030, or NULL */
struct mpt_softc * mpt2;
@@ -580,7 +669,7 @@ struct mpt_softc {
TAILQ_ENTRY(mpt_softc) links;
};
-/***************************** Locking Primatives *****************************/
+/***************************** Locking Primitives *****************************/
#if __FreeBSD_version < 500000
#define MPT_IFLAGS INTR_TYPE_CAM
#define MPT_LOCK(mpt) mpt_lockspl(mpt)
@@ -631,7 +720,7 @@ mpt_sleep(struct mpt_softc *mpt, void *ident, int priority,
saved_spl = mpt->mpt_splsaved;
mpt->mpt_islocked = 0;
error = tsleep(ident, priority, wmesg, timo);
- KASSERT(mpt->mpt_islocked = 0, ("Invalid lock count on wakeup"));
+ KASSERT(mpt->mpt_islocked == 0, ("Invalid lock count on wakeup"));
mpt->mpt_islocked = saved_cnt;
mpt->mpt_splsaved = saved_spl;
return (error);
@@ -709,19 +798,27 @@ mpt_pio_read(struct mpt_softc *mpt, int offset)
/* Max MPT Reply we are willing to accept (must be power of 2) */
#define MPT_REPLY_SIZE 256
+/*
+ * Must be less than 16384 in order for target mode to work
+ */
#define MPT_MAX_REQUESTS(mpt) 512
#define MPT_REQUEST_AREA 512
#define MPT_SENSE_SIZE 32 /* included in MPT_REQUEST_AREA */
#define MPT_REQ_MEM_SIZE(mpt) (MPT_MAX_REQUESTS(mpt) * MPT_REQUEST_AREA)
-#define MPT_CONTEXT_CB_SHIFT (16)
-#define MPT_CBI(handle) (handle >> MPT_CONTEXT_CB_SHIFT)
+/*
+ * Currently we try to pack both callbacks and request indices into 14 bits
+ * so that we don't have to get fancy when we get a target mode context
+ * reply (which only has 14 bits of IoIndex value) or a normal scsi
+ * initiator context reply (where we get bits 28..0 of context).
+ */
+#define MPT_CONTEXT_CB_SHIFT (14)
+#define MPT_CBI(handle) (handle >> MPT_CONTEXT_CB_SHIFT)
#define MPT_CBI_TO_HID(cbi) ((cbi) << MPT_CONTEXT_CB_SHIFT)
#define MPT_CONTEXT_TO_CBI(x) \
(((x) >> MPT_CONTEXT_CB_SHIFT) & (MPT_NUM_REPLY_HANDLERS - 1))
-#define MPT_CONTEXT_REQI_MASK 0xFFFF
-#define MPT_CONTEXT_TO_REQI(x) \
- ((x) & MPT_CONTEXT_REQI_MASK)
+#define MPT_CONTEXT_REQI_MASK 0x3FFF
+#define MPT_CONTEXT_TO_REQI(x) ((x) & MPT_CONTEXT_REQI_MASK)
/*
* Convert a 32bit physical address returned from IOC to an
@@ -786,7 +883,7 @@ void mpt_complete_request_chain(struct mpt_softc *mpt,
/***************************** IOC Initialization *****************************/
int mpt_reset(struct mpt_softc *, int /*reinit*/);
-/****************************** Debugging/Logging *****************************/
+/****************************** Debugging ************************************/
typedef struct mpt_decode_entry {
char *name;
u_int value;
@@ -804,10 +901,14 @@ enum {
MPT_PRT_WARN,
MPT_PRT_INFO,
MPT_PRT_DEBUG,
+ MPT_PRT_DEBUG1,
+ MPT_PRT_DEBUG2,
+ MPT_PRT_DEBUG3,
MPT_PRT_TRACE,
MPT_PRT_NONE=100
};
+#if __FreeBSD_version > 500000
#define mpt_lprt(mpt, level, ...) \
do { \
if (level <= (mpt)->verbose) \
@@ -819,9 +920,56 @@ do { \
if (level <= (mpt)->debug_level) \
mpt_prtc(mpt, __VA_ARGS__); \
} while (0)
+#else
+void mpt_lprt(struct mpt_softc *, int, const char *, ...)
+ __printflike(3, 4);
+void mpt_lprtc(struct mpt_softc *, int, const char *, ...)
+ __printflike(3, 4);
+#endif
+void mpt_prt(struct mpt_softc *, const char *, ...)
+ __printflike(2, 3);
+void mpt_prtc(struct mpt_softc *, const char *, ...)
+ __printflike(2, 3);
-void mpt_prt(struct mpt_softc *, const char *, ...);
-void mpt_prtc(struct mpt_softc *, const char *, ...);
+/**************************** Target Mode Related ***************************/
+static __inline int mpt_cdblen(uint8_t, int);
+static __inline int
+mpt_cdblen(uint8_t cdb0, int maxlen)
+{
+ int group = cdb0 >> 5;
+ switch (group) {
+ case 0:
+ return (6);
+ case 1:
+ return (10);
+ case 4:
+ case 5:
+ return (12);
+ default:
+ return (16);
+ }
+}
+#ifdef INVARIANTS
+static __inline request_t * mpt_tag_2_req(struct mpt_softc *, uint32_t);
+static __inline request_t *
+mpt_tag_2_req(struct mpt_softc *mpt, uint32_t tag)
+{
+ uint16_t rtg = (tag >> 16);
+ KASSERT(rtg < mpt->tgt_cmds_allocated, ("bad tag %d\n", tag));
+ KASSERT(mpt->tgt_cmd_ptrs, ("no cmd backpointer array"));
+ KASSERT(mpt->tgt_cmd_ptrs[rtg], ("no cmd backpointer"));
+ return (mpt->tgt_cmd_ptrs[rtg]);
+}
+#endif
+
+typedef enum {
+ MPT_ABORT_TASK_SET=1234,
+ MPT_CLEAR_TASK_SET,
+ MPT_TARGET_RESET,
+ MPT_CLEAR_ACA,
+ MPT_TERMINATE_TASK,
+ MPT_NIL_TMT_VALUE=5678
+} mpt_task_mgmt_t;
/**************************** Unclassified Routines ***************************/
void mpt_send_cmd(struct mpt_softc *mpt, request_t *req);
diff --git a/sys/dev/mpt/mpt_cam.c b/sys/dev/mpt/mpt_cam.c
index 7785f7d5fe27..95d8438431c0 100644
--- a/sys/dev/mpt/mpt_cam.c
+++ b/sys/dev/mpt/mpt_cam.c
@@ -100,6 +100,7 @@ __FBSDID("$FreeBSD$");
#include "dev/mpt/mpilib/mpi_ioc.h" /* XXX Fix Event Handling!!! */
#include "dev/mpt/mpilib/mpi_init.h"
#include "dev/mpt/mpilib/mpi_targ.h"
+#include "dev/mpt/mpilib/mpi_fc.h"
#include <sys/callout.h>
#include <sys/kthread.h>
@@ -112,9 +113,12 @@ static int mpt_setsync(struct mpt_softc *, int, int, int);
static void mpt_calc_geometry(struct ccb_calc_geometry *ccg, int extended);
static mpt_reply_handler_t mpt_scsi_reply_handler;
static mpt_reply_handler_t mpt_scsi_tmf_reply_handler;
+static mpt_reply_handler_t mpt_fc_els_reply_handler;
+static mpt_reply_handler_t mpt_scsi_tgt_reply_handler;
static int mpt_scsi_reply_frame_handler(struct mpt_softc *mpt, request_t *req,
MSG_DEFAULT_REPLY *reply_frame);
static int mpt_bus_reset(struct mpt_softc *, int /*sleep_ok*/);
+static int mpt_fc_reset_link(struct mpt_softc *, int);
static int mpt_spawn_recovery_thread(struct mpt_softc *mpt);
static void mpt_terminate_recovery_thread(struct mpt_softc *mpt);
@@ -123,13 +127,25 @@ static int mpt_scsi_send_tmf(struct mpt_softc *, u_int /*type*/,
u_int /*flags*/, u_int /*channel*/,
u_int /*target*/, u_int /*lun*/,
u_int /*abort_ctx*/, int /*sleep_ok*/);
+
+static void mpt_fc_add_els(struct mpt_softc *mpt, request_t *);
+static void mpt_post_target_command(struct mpt_softc *, request_t *, int);
+static void mpt_add_target_commands(struct mpt_softc *mpt);
+static int mpt_enable_lun(struct mpt_softc *, target_id_t, lun_id_t);
+static int mpt_disable_lun(struct mpt_softc *, target_id_t, lun_id_t);
+static void mpt_target_start_io(struct mpt_softc *, union ccb *);
+static cam_status mpt_abort_target_ccb(struct mpt_softc *, union ccb *);
+static cam_status mpt_abort_target_cmd(struct mpt_softc *, request_t *);
+
static void mpt_recover_commands(struct mpt_softc *mpt);
static uint32_t scsi_io_handler_id = MPT_HANDLER_ID_NONE;
static uint32_t scsi_tmf_handler_id = MPT_HANDLER_ID_NONE;
+static uint32_t fc_els_handler_id = MPT_HANDLER_ID_NONE;
static mpt_probe_handler_t mpt_cam_probe;
static mpt_attach_handler_t mpt_cam_attach;
+static mpt_enable_handler_t mpt_cam_enable;
static mpt_event_handler_t mpt_cam_event;
static mpt_reset_handler_t mpt_cam_ioc_reset;
static mpt_detach_handler_t mpt_cam_detach;
@@ -139,6 +155,7 @@ static struct mpt_personality mpt_cam_personality =
.name = "mpt_cam",
.probe = mpt_cam_probe,
.attach = mpt_cam_attach,
+ .enable = mpt_cam_enable,
.event = mpt_cam_event,
.reset = mpt_cam_ioc_reset,
.detach = mpt_cam_detach,
@@ -150,13 +167,14 @@ int
mpt_cam_probe(struct mpt_softc *mpt)
{
/*
- * Only attach to nodes that support the initiator
- * role or have RAID physical devices that need
- * CAM pass-thru support.
+ * Only attach to nodes that support the initiator or target
+ * role or have RAID physical devices that need CAM pass-thru support.
*/
if ((mpt->mpt_proto_flags & MPI_PORTFACTS_PROTOCOL_INITIATOR) != 0
- || (mpt->ioc_page2 != NULL && mpt->ioc_page2->MaxPhysDisks != 0))
+ || (mpt->mpt_proto_flags & MPI_PORTFACTS_PROTOCOL_TARGET) != 0
+ || (mpt->ioc_page2 != NULL && mpt->ioc_page2->MaxPhysDisks != 0)) {
return (0);
+ }
return (ENODEV);
}
@@ -168,22 +186,73 @@ mpt_cam_attach(struct mpt_softc *mpt)
int maxq;
int error;
- MPTLOCK_2_CAMLOCK(mpt);
TAILQ_INIT(&mpt->request_timeout_list);
- mpt->bus = 0;
maxq = (mpt->mpt_global_credits < MPT_MAX_REQUESTS(mpt))?
mpt->mpt_global_credits : MPT_MAX_REQUESTS(mpt);
handler.reply_handler = mpt_scsi_reply_handler;
error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler,
&scsi_io_handler_id);
- if (error != 0)
+ if (error != 0) {
goto cleanup;
+ }
+
handler.reply_handler = mpt_scsi_tmf_reply_handler;
error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler,
&scsi_tmf_handler_id);
- if (error != 0)
+ if (error != 0) {
goto cleanup;
+ }
+
+ /*
+ * We keep two requests reserved for ELS replies/responses
+ * if we're fibre channel and target mode.
+ */
+ if (mpt->is_fc && (mpt->role & MPT_ROLE_TARGET) != 0) {
+ request_t *req;
+ int i;
+
+ handler.reply_handler = mpt_fc_els_reply_handler;
+ error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler,
+ &fc_els_handler_id);
+ if (error != 0) {
+ goto cleanup;
+ }
+
+ /*
+ * Feed the chip some ELS buffer resources
+ */
+ for (i = 0; i < MPT_MAX_ELS; i++) {
+ req = mpt_get_request(mpt, FALSE);
+ if (req == NULL) {
+ break;
+ }
+ mpt_fc_add_els(mpt, req);
+ }
+ if (i == 0) {
+ mpt_prt(mpt, "Unable to add ELS buffer resources\n");
+ goto cleanup;
+ }
+ maxq -= i;
+ }
+
+ /*
+ * If we're in target mode, register a reply
+ * handler for it and add some commands.
+ */
+ if ((mpt->role & MPT_ROLE_TARGET) != 0) {
+ handler.reply_handler = mpt_scsi_tgt_reply_handler;
+ error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler,
+ &mpt->scsi_tgt_handler_id);
+ if (error != 0) {
+ goto cleanup;
+ }
+
+ /*
+ * Add some target command resources
+ */
+ mpt_add_target_commands(mpt);
+ }
/*
* We keep one request reserved for timeout TMF requests.
@@ -252,9 +321,9 @@ mpt_cam_attach(struct mpt_softc *mpt)
* Only register a second bus for RAID physical
* devices if the controller supports RAID.
*/
- if (mpt->ioc_page2 == NULL
- || mpt->ioc_page2->MaxPhysDisks == 0)
+ if (mpt->ioc_page2 == NULL || mpt->ioc_page2->MaxPhysDisks == 0) {
return (0);
+ }
/*
* Create a "bus" to export all hidden disks to CAM.
@@ -283,15 +352,412 @@ mpt_cam_attach(struct mpt_softc *mpt)
error = ENOMEM;
goto cleanup;
}
-
- CAMLOCK_2_MPTLOCK(mpt);
return (0);
cleanup:
- CAMLOCK_2_MPTLOCK(mpt);
mpt_cam_detach(mpt);
return (error);
}
+/*
+ * Read FC configuration information
+ */
+static int
+mpt_read_config_info_fc(struct mpt_softc *mpt)
+{
+ char *topology = NULL;
+ int rv, speed = 0;
+
+ rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_FC_PORT, 0,
+ 0, &mpt->mpt_fcport_page0.Header, FALSE, 5000);
+ if (rv) {
+ return (-1);
+ }
+ mpt_lprt(mpt, MPT_PRT_DEBUG, "FC Port Page 0 Header: %x %x %x %x\n",
+ mpt->mpt_fcport_page0.Header.PageVersion,
+ mpt->mpt_fcport_page0.Header.PageLength,
+ mpt->mpt_fcport_page0.Header.PageNumber,
+ mpt->mpt_fcport_page0.Header.PageType);
+
+
+ rv = mpt_read_cur_cfg_page(mpt, 0, &mpt->mpt_fcport_page0.Header,
+ sizeof(mpt->mpt_fcport_page0), FALSE, 5000);
+ if (rv) {
+ mpt_prt(mpt, "failed to read FC Port Page 0\n");
+ return (-1);
+ }
+
+ speed = mpt->mpt_fcport_page0.CurrentSpeed;
+
+ switch (mpt->mpt_fcport_page0.Flags &
+ MPI_FCPORTPAGE0_FLAGS_ATTACH_TYPE_MASK) {
+ case MPI_FCPORTPAGE0_FLAGS_ATTACH_NO_INIT:
+ speed = 0;
+ topology = "<NO LOOP>";
+ break;
+ case MPI_FCPORTPAGE0_FLAGS_ATTACH_POINT_TO_POINT:
+ topology = "N-Port";
+ break;
+ case MPI_FCPORTPAGE0_FLAGS_ATTACH_PRIVATE_LOOP:
+ topology = "NL-Port";
+ break;
+ case MPI_FCPORTPAGE0_FLAGS_ATTACH_FABRIC_DIRECT:
+ topology = "F-Port";
+ break;
+ case MPI_FCPORTPAGE0_FLAGS_ATTACH_PUBLIC_LOOP:
+ topology = "FL-Port";
+ break;
+ default:
+ speed = 0;
+ topology = "?";
+ break;
+ }
+
+ mpt_lprt(mpt, MPT_PRT_INFO,
+ "FC Port Page 0: Topology <%s> WWNN 0x%08x%08x WWPN 0x%08x%08x "
+ "Speed %u-Gbit\n", topology,
+ mpt->mpt_fcport_page0.WWNN.High,
+ mpt->mpt_fcport_page0.WWNN.Low,
+ mpt->mpt_fcport_page0.WWPN.High,
+ mpt->mpt_fcport_page0.WWPN.Low,
+ speed);
+
+ return (0);
+}
+
+/*
+ * Set FC configuration information.
+ */
+static int
+mpt_set_initial_config_fc(struct mpt_softc *mpt)
+{
+#if 0
+ CONFIG_PAGE_FC_PORT_1 fc;
+ U32 fl;
+ int r, doit = 0;
+
+ r = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_FC_PORT, 1, 0,
+ &fc.Header, FALSE, 5000);
+ if (r) {
+ return (mpt_fc_reset_link(mpt, 1));
+ }
+
+ r = mpt_read_cfg_page(mpt, MPI_CONFIG_ACTION_PAGE_READ_CURRENT, 0,
+ &fc.Header, sizeof (fc), FALSE, 5000);
+ if (r) {
+ return (mpt_fc_reset_link(mpt, 1));
+ }
+
+ fl = le32toh(fc.Flags);
+ if ((fl & MPI_FCPORTPAGE1_FLAGS_TARGET_MODE_OXID) == 0) {
+ fl |= MPI_FCPORTPAGE1_FLAGS_TARGET_MODE_OXID;
+ doit = 1;
+ }
+ if ((fl & MPI_FCPORTPAGE1_FLAGS_PROT_FCP_INIT) &&
+ (mpt->role & MPT_ROLE_INITIATOR) == 0) {
+ fl &= ~MPI_FCPORTPAGE1_FLAGS_PROT_FCP_INIT;
+ doit = 1;
+ }
+ if ((fl & MPI_FCPORTPAGE1_FLAGS_PROT_FCP_TARG) &&
+ (mpt->role & MPT_ROLE_TARGET) == 0) {
+ fl &= ~MPI_FCPORTPAGE1_FLAGS_PROT_FCP_TARG;
+ doit = 1;
+ }
+ if (doit) {
+ const char *cc;
+
+ mpt_lprt(mpt, MPT_PRT_INFO,
+ "FC Port Page 1: New Flags %x \n", fl);
+ fc.Flags = htole32(fl);
+ r = mpt_write_cfg_page(mpt,
+ MPI_CONFIG_ACTION_PAGE_WRITE_CURRENT, 0, &fc.Header,
+ sizeof(fc), FALSE, 5000);
+ if (r != 0) {
+ cc = "FC PORT PAGE1 UPDATE: FAILED\n";
+ } else {
+ cc = "FC PORT PAGE1 UPDATED: SYSTEM NEEDS RESET\n";
+ }
+ mpt_prt(mpt, cc);
+ }
+#endif
+ return (mpt_fc_reset_link(mpt, 1));
+}
+
+/*
+ * Read SAS configuration information. Nothing to do yet.
+ */
+static int
+mpt_read_config_info_sas(struct mpt_softc *mpt)
+{
+ return (0);
+}
+
+/*
+ * Set SAS configuration information. Nothing to do yet.
+ */
+static int
+mpt_set_initial_config_sas(struct mpt_softc *mpt)
+{
+ return (0);
+}
+
+/*
+ * Read SCSI configuration information
+ */
+static int
+mpt_read_config_info_spi(struct mpt_softc *mpt)
+{
+ int rv, i;
+
+ rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 0,
+ 0, &mpt->mpt_port_page0.Header,
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv)
+ return (-1);
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Port Page 0 Header: %x %x %x %x\n",
+ mpt->mpt_port_page0.Header.PageVersion,
+ mpt->mpt_port_page0.Header.PageLength,
+ mpt->mpt_port_page0.Header.PageNumber,
+ mpt->mpt_port_page0.Header.PageType);
+
+ rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 1,
+ 0, &mpt->mpt_port_page1.Header,
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv)
+ return (-1);
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG, "SPI Port Page 1 Header: %x %x %x %x\n",
+ mpt->mpt_port_page1.Header.PageVersion,
+ mpt->mpt_port_page1.Header.PageLength,
+ mpt->mpt_port_page1.Header.PageNumber,
+ mpt->mpt_port_page1.Header.PageType);
+
+ rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 2,
+ /*PageAddress*/0, &mpt->mpt_port_page2.Header,
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv)
+ return (-1);
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Port Page 2 Header: %x %x %x %x\n",
+ mpt->mpt_port_page1.Header.PageVersion,
+ mpt->mpt_port_page1.Header.PageLength,
+ mpt->mpt_port_page1.Header.PageNumber,
+ mpt->mpt_port_page1.Header.PageType);
+
+ for (i = 0; i < 16; i++) {
+ rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE,
+ 0, i, &mpt->mpt_dev_page0[i].Header,
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv)
+ return (-1);
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Target %d Device Page 0 Header: %x %x %x %x\n",
+ i, mpt->mpt_dev_page0[i].Header.PageVersion,
+ mpt->mpt_dev_page0[i].Header.PageLength,
+ mpt->mpt_dev_page0[i].Header.PageNumber,
+ mpt->mpt_dev_page0[i].Header.PageType);
+
+ rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE,
+ 1, i, &mpt->mpt_dev_page1[i].Header,
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv)
+ return (-1);
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Target %d Device Page 1 Header: %x %x %x %x\n",
+ i, mpt->mpt_dev_page1[i].Header.PageVersion,
+ mpt->mpt_dev_page1[i].Header.PageLength,
+ mpt->mpt_dev_page1[i].Header.PageNumber,
+ mpt->mpt_dev_page1[i].Header.PageType);
+ }
+
+ /*
+ * At this point, we don't *have* to fail. As long as we have
+ * valid config header information, we can (barely) lurch
+ * along.
+ */
+
+ rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
+ &mpt->mpt_port_page0.Header,
+ sizeof(mpt->mpt_port_page0),
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv) {
+ mpt_prt(mpt, "failed to read SPI Port Page 0\n");
+ } else {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Port Page 0: Capabilities %x PhysicalInterface %x\n",
+ mpt->mpt_port_page0.Capabilities,
+ mpt->mpt_port_page0.PhysicalInterface);
+ }
+
+ rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
+ &mpt->mpt_port_page1.Header,
+ sizeof(mpt->mpt_port_page1),
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv) {
+ mpt_prt(mpt, "failed to read SPI Port Page 1\n");
+ } else {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Port Page 1: Configuration %x OnBusTimerValue %x\n",
+ mpt->mpt_port_page1.Configuration,
+ mpt->mpt_port_page1.OnBusTimerValue);
+ }
+
+ rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
+ &mpt->mpt_port_page2.Header,
+ sizeof(mpt->mpt_port_page2),
+ /*sleep_ok*/FALSE, /*timeout_ms*/5000);
+ if (rv) {
+ mpt_prt(mpt, "failed to read SPI Port Page 2\n");
+ } else {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Port Page 2: Flags %x Settings %x\n",
+ mpt->mpt_port_page2.PortFlags,
+ mpt->mpt_port_page2.PortSettings);
+ for (i = 0; i < 16; i++) {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Port Page 2 Tgt %d: timo %x SF %x Flags %x\n",
+ i, mpt->mpt_port_page2.DeviceSettings[i].Timeout,
+ mpt->mpt_port_page2.DeviceSettings[i].SyncFactor,
+ mpt->mpt_port_page2.DeviceSettings[i].DeviceFlags);
+ }
+ }
+
+ for (i = 0; i < 16; i++) {
+ rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i,
+ &mpt->mpt_dev_page0[i].Header,
+ sizeof(*mpt->mpt_dev_page0),
+ /*sleep_ok*/FALSE,
+ /*timeout_ms*/5000);
+ if (rv) {
+ mpt_prt(mpt,
+ "cannot read SPI Tgt %d Device Page 0\n", i);
+ continue;
+ }
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Tgt %d Page 0: NParms %x Information %x",
+ i, mpt->mpt_dev_page0[i].NegotiatedParameters,
+ mpt->mpt_dev_page0[i].Information);
+
+ rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i,
+ &mpt->mpt_dev_page1[i].Header,
+ sizeof(*mpt->mpt_dev_page1),
+ /*sleep_ok*/FALSE,
+ /*timeout_ms*/5000);
+ if (rv) {
+ mpt_prt(mpt,
+ "cannot read SPI Tgt %d Device Page 1\n", i);
+ continue;
+ }
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Tgt %d Page 1: RParms %x Configuration %x\n",
+ i, mpt->mpt_dev_page1[i].RequestedParameters,
+ mpt->mpt_dev_page1[i].Configuration);
+ }
+ return (0);
+}
+
+/*
+ * Validate SPI configuration information.
+ *
+ * In particular, validate SPI Port Page 1.
+ */
+static int
+mpt_set_initial_config_spi(struct mpt_softc *mpt)
+{
+ int i, pp1val = ((1 << mpt->mpt_ini_id) << 16) | mpt->mpt_ini_id;
+ int error;
+
+ mpt->mpt_disc_enable = 0xff;
+ mpt->mpt_tag_enable = 0;
+
+ if (mpt->mpt_port_page1.Configuration != pp1val) {
+ CONFIG_PAGE_SCSI_PORT_1 tmp;
+
+ mpt_prt(mpt,
+ "SPI Port Page 1 Config value bad (%x)- should be %x\n",
+ mpt->mpt_port_page1.Configuration, pp1val);
+ tmp = mpt->mpt_port_page1;
+ tmp.Configuration = pp1val;
+ error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/0,
+ &tmp.Header, sizeof(tmp),
+ /*sleep_ok*/FALSE,
+ /*timeout_ms*/5000);
+ if (error)
+ return (-1);
+ error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0,
+ &tmp.Header, sizeof(tmp),
+ /*sleep_ok*/FALSE,
+ /*timeout_ms*/5000);
+ if (error)
+ return (-1);
+ if (tmp.Configuration != pp1val) {
+ mpt_prt(mpt,
+ "failed to reset SPI Port Page 1 Config value\n");
+ return (-1);
+ }
+ mpt->mpt_port_page1 = tmp;
+ }
+
+ for (i = 0; i < 16; i++) {
+ CONFIG_PAGE_SCSI_DEVICE_1 tmp;
+ tmp = mpt->mpt_dev_page1[i];
+ tmp.RequestedParameters = 0;
+ tmp.Configuration = 0;
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "Set Tgt %d SPI DevicePage 1 values to %x 0 %x\n",
+ i, tmp.RequestedParameters, tmp.Configuration);
+ error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/i,
+ &tmp.Header, sizeof(tmp),
+ /*sleep_ok*/FALSE,
+ /*timeout_ms*/5000);
+ if (error)
+ return (-1);
+ error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i,
+ &tmp.Header, sizeof(tmp),
+ /*sleep_ok*/FALSE,
+ /*timeout_ms*/5000);
+ if (error)
+ return (-1);
+ mpt->mpt_dev_page1[i] = tmp;
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "SPI Tgt %d Page 1: RParm %x Configuration %x\n", i,
+ mpt->mpt_dev_page1[i].RequestedParameters,
+ mpt->mpt_dev_page1[i].Configuration);
+ }
+ return (0);
+}
+
+int
+mpt_cam_enable(struct mpt_softc *mpt)
+{
+ if (mpt->is_fc) {
+ if (mpt_read_config_info_fc(mpt)) {
+ return (EIO);
+ }
+ if (mpt_set_initial_config_fc(mpt)) {
+ return (EIO);
+ }
+ } else if (mpt->is_sas) {
+ if (mpt_read_config_info_sas(mpt)) {
+ return (EIO);
+ }
+ if (mpt_set_initial_config_sas(mpt)) {
+ return (EIO);
+ }
+ } else {
+ if (mpt_read_config_info_spi(mpt)) {
+ return (EIO);
+ }
+ if (mpt_set_initial_config_spi(mpt)) {
+ return (EIO);
+ }
+ }
+ return (0);
+}
+
void
mpt_cam_detach(struct mpt_softc *mpt)
{
@@ -305,6 +771,12 @@ mpt_cam_detach(struct mpt_softc *mpt)
handler.reply_handler = mpt_scsi_tmf_reply_handler;
mpt_deregister_handler(mpt, MPT_HANDLER_REPLY, handler,
scsi_tmf_handler_id);
+ handler.reply_handler = mpt_fc_els_reply_handler;
+ mpt_deregister_handler(mpt, MPT_HANDLER_REPLY, handler,
+ fc_els_handler_id);
+ handler.reply_handler = mpt_scsi_tgt_reply_handler;
+ mpt_deregister_handler(mpt, MPT_HANDLER_REPLY, handler,
+ mpt->scsi_tgt_handler_id);
if (mpt->tmf_req != NULL) {
mpt_free_request(mpt, mpt->tmf_req);
@@ -326,8 +798,7 @@ mpt_cam_detach(struct mpt_softc *mpt)
}
}
-/* This routine is used after a system crash to dump core onto the
- * swap device.
+/* This routine is used after a system crash to dump core onto the swap device.
*/
static void
mpt_poll(struct cam_sim *sim)
@@ -361,7 +832,7 @@ mpt_timeout(void *arg)
MPT_LOCK(mpt);
req = ccb->ccb_h.ccb_req_ptr;
- mpt_prt(mpt, "Request %p:serno Timed out.\n", req, req->serno);
+ mpt_prt(mpt, "Request %p Timed out.\n", req);
if ((req->state & REQ_STATE_QUEUED) == REQ_STATE_QUEUED) {
TAILQ_REMOVE(&mpt->request_pending_list, req, links);
TAILQ_INSERT_TAIL(&mpt->request_timeout_list, req, links);
@@ -379,7 +850,7 @@ mpt_timeout(void *arg)
* aborted the transaction.
*/
static void
-mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
+mpt_execute_req_a64(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
{
request_t *req, *trq;
char *mpt_off;
@@ -387,8 +858,8 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
struct mpt_softc *mpt;
int seg, first_lim;
uint32_t flags, nxt_off;
- bus_dmasync_op_t op;
- MSG_SCSI_IO_REQUEST *mpt_req;
+ void *sglp;
+ MSG_REQUEST_HEADER *hdrp;
SGE_SIMPLE64 *se;
SGE_CHAIN64 *ce;
@@ -397,22 +868,31 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
mpt = ccb->ccb_h.ccb_mpt_ptr;
req = ccb->ccb_h.ccb_req_ptr;
- mpt_req = req->req_vbuf;
+
+ hdrp = req->req_vbuf;
mpt_off = req->req_vbuf;
+ if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) {
+ sglp = &((PTR_MSG_SCSI_IO_REQUEST)hdrp)->SGL;
+ } else /* if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) */ {
+ sglp = &((PTR_MSG_TARGET_ASSIST_REQUEST)hdrp)->SGL;
+ }
+
+
if (error == 0 && ((uint32_t)nseg) >= mpt->max_seg_cnt) {
error = EFBIG;
}
bad:
if (error != 0) {
- if (error != EFBIG && error != ENOMEM)
+ if (error != EFBIG && error != ENOMEM) {
mpt_prt(mpt, "mpt_execute_req: err %d\n", error);
- if (ccb->ccb_h.status == CAM_REQ_INPROG) {
- xpt_freeze_devq(ccb->ccb_h.path, 1);
- ccb->ccb_h.status = CAM_DEV_QFRZN;
+ }
+ if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) {
+ cam_status status;
+ mpt_freeze_ccb(ccb);
if (error == EFBIG) {
- ccb->ccb_h.status |= CAM_REQ_TOO_BIG;
+ status = CAM_REQ_TOO_BIG;
} else if (error == ENOMEM) {
if (mpt->outofbeer == 0) {
mpt->outofbeer = 1;
@@ -420,11 +900,21 @@ bad:
mpt_lprt(mpt, MPT_PRT_DEBUG,
"FREEZEQ\n");
}
- ccb->ccb_h.status |= CAM_REQUEUE_REQ;
- } else
- ccb->ccb_h.status |= CAM_REQ_CMP_ERR;
+ status = CAM_REQUEUE_REQ;
+ } else {
+ status = CAM_REQ_CMP_ERR;
+ }
+ mpt_set_ccb_status(ccb, status);
+ }
+ if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) {
+ request_t *cmd_req =
+ MPT_TAG_2_REQ(mpt, ccb->csio.tag_id);
+ MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM;
+ MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL;
+ MPT_TGT_STATE(mpt, cmd_req)->req = NULL;
}
ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__));
xpt_done(ccb);
CAMLOCK_2_MPTLOCK(mpt);
mpt_free_request(mpt, req);
@@ -438,29 +928,45 @@ bad:
*/
if (mpt->verbose >= MPT_PRT_DEBUG) {
- int tidx = ((char *)&mpt_req->SGL) - mpt_off;
+ int tidx = ((char *)sglp) - mpt_off;
memset(&mpt_off[tidx], 0xff, MPT_REQUEST_AREA - tidx);
}
if (nseg == 0) {
- SGE_SIMPLE32 *se1 = (SGE_SIMPLE32 *) &mpt_req->SGL;
+ SGE_SIMPLE32 *se1 = (SGE_SIMPLE32 *) sglp;
MPI_pSGE_SET_FLAGS(se1,
(MPI_SGE_FLAGS_LAST_ELEMENT | MPI_SGE_FLAGS_END_OF_BUFFER |
MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_END_OF_LIST));
goto out;
}
- mpt_req->DataLength = ccb->csio.dxfer_len;
- flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_64_BIT_ADDRESSING;
- if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT)
- flags |= MPI_SGE_FLAGS_HOST_TO_IOC;
- if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
- op = BUS_DMASYNC_PREREAD;
+ flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_64_BIT_ADDRESSING;
+ if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) {
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) {
+ flags |= MPI_SGE_FLAGS_HOST_TO_IOC;
+ }
} else {
- op = BUS_DMASYNC_PREWRITE;
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
+ flags |= MPI_SGE_FLAGS_HOST_TO_IOC;
+ }
}
+
if (!(ccb->ccb_h.flags & (CAM_SG_LIST_PHYS|CAM_DATA_PHYS))) {
+ bus_dmasync_op_t op;
+ if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) {
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
+ op = BUS_DMASYNC_PREREAD;
+ } else {
+ op = BUS_DMASYNC_PREWRITE;
+ }
+ } else {
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
+ op = BUS_DMASYNC_PREWRITE;
+ } else {
+ op = BUS_DMASYNC_PREREAD;
+ }
+ }
bus_dmamap_sync(mpt->buffer_dmat, req->dmap, op);
}
@@ -483,7 +989,7 @@ bad:
first_lim = MPT_NSGL_FIRST(mpt) - 1;
}
- se = (SGE_SIMPLE64 *) &mpt_req->SGL;
+ se = (SGE_SIMPLE64 *) sglp;
for (seg = 0; seg < first_lim; seg++, se++, dm_segs++) {
uint32_t tf;
@@ -511,7 +1017,7 @@ bad:
/*
* Tell the IOC where to find the first chain element.
*/
- mpt_req->ChainOffset = ((char *)se - (char *)mpt_req) >> 2;
+ hdrp->ChainOffset = ((char *)se - (char *)hdrp) >> 2;
nxt_off = MPT_RQSL(mpt);
trq = req;
@@ -650,7 +1156,6 @@ bad:
}
trq = nrq;
mpt_off = trq->req_vbuf;
- mpt_req = trq->req_vbuf;
if (mpt->verbose >= MPT_PRT_DEBUG) {
memset(mpt_off, 0xff, MPT_REQUEST_AREA);
}
@@ -662,18 +1167,393 @@ out:
/*
* Last time we need to check if this CCB needs to be aborted.
*/
- if (ccb->ccb_h.status != CAM_REQ_INPROG) {
- if (nseg && (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0)
+ if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_INPROG) {
+ if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) {
+ request_t *cmd_req =
+ MPT_TAG_2_REQ(mpt, ccb->csio.tag_id);
+ MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM;
+ MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL;
+ MPT_TGT_STATE(mpt, cmd_req)->req = NULL;
+ }
+ mpt_prt(mpt,
+ "mpt_execute_req_a64: I/O cancelled (status 0x%x)\n",
+ ccb->ccb_h.status & CAM_STATUS_MASK);
+ if (nseg && (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0) {
bus_dmamap_unload(mpt->buffer_dmat, req->dmap);
+ }
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__));
+ xpt_done(ccb);
CAMLOCK_2_MPTLOCK(mpt);
mpt_free_request(mpt, req);
MPTLOCK_2_CAMLOCK(mpt);
- xpt_done(ccb);
return;
}
ccb->ccb_h.status |= CAM_SIM_QUEUED;
+ if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) {
+ ccb->ccb_h.timeout_ch =
+ timeout(mpt_timeout, (caddr_t)ccb,
+ (ccb->ccb_h.timeout * hz) / 1000);
+ } else {
+ callout_handle_init(&ccb->ccb_h.timeout_ch);
+ }
+ if (mpt->verbose >= MPT_PRT_DEBUG) {
+ int nc = 0;
+ mpt_print_request(req->req_vbuf);
+ for (trq = req->chain; trq; trq = trq->chain) {
+ printf(" Additional Chain Area %d\n", nc++);
+ mpt_dump_sgl(trq->req_vbuf, 0);
+ }
+ }
+ if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) {
+ request_t *cmd_req = MPT_TAG_2_REQ(mpt, ccb->csio.tag_id);
+ mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, cmd_req);
+#ifdef WE_TRUST_AUTO_GOOD_STATUS
+ if ((ccb->ccb_h.flags & CAM_SEND_STATUS) &&
+ csio->scsi_status == SCSI_STATUS_OK && tgt->resid == 0) {
+ tgt->state = TGT_STATE_MOVING_DATA_AND_STATUS;
+ } else {
+ tgt->state = TGT_STATE_MOVING_DATA;
+ }
+#else
+ tgt->state = TGT_STATE_MOVING_DATA;
+#endif
+ }
CAMLOCK_2_MPTLOCK(mpt);
+ mpt_send_cmd(mpt, req);
+ MPTLOCK_2_CAMLOCK(mpt);
+}
+
+static void
+mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
+{
+ request_t *req, *trq;
+ char *mpt_off;
+ union ccb *ccb;
+ struct mpt_softc *mpt;
+ int seg, first_lim;
+ uint32_t flags, nxt_off;
+ void *sglp;
+ MSG_REQUEST_HEADER *hdrp;
+ SGE_SIMPLE32 *se;
+ SGE_CHAIN32 *ce;
+
+ req = (request_t *)arg;
+ ccb = req->ccb;
+
+ mpt = ccb->ccb_h.ccb_mpt_ptr;
+ req = ccb->ccb_h.ccb_req_ptr;
+
+ hdrp = req->req_vbuf;
+ mpt_off = req->req_vbuf;
+
+
+ if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) {
+ sglp = &((PTR_MSG_SCSI_IO_REQUEST)hdrp)->SGL;
+ } else /* if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) */ {
+ sglp = &((PTR_MSG_TARGET_ASSIST_REQUEST)hdrp)->SGL;
+ }
+
+
+ if (error == 0 && ((uint32_t)nseg) >= mpt->max_seg_cnt) {
+ error = EFBIG;
+ mpt_prt(mpt, "segment count %d too large (max %u)\n",
+ nseg, mpt->max_seg_cnt);
+ }
+
+bad:
+ if (error != 0) {
+ if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) {
+ request_t *cmd_req =
+ MPT_TAG_2_REQ(mpt, ccb->csio.tag_id);
+ MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM;
+ MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL;
+ MPT_TGT_STATE(mpt, cmd_req)->req = NULL;
+ }
+ if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) {
+ cam_status status;
+ mpt_freeze_ccb(ccb);
+ if (error == EFBIG) {
+ status = CAM_REQ_TOO_BIG;
+ } else if (error == ENOMEM) {
+ if (mpt->outofbeer == 0) {
+ mpt->outofbeer = 1;
+ xpt_freeze_simq(mpt->sim, 1);
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "FREEZEQ\n");
+ }
+ status = CAM_REQUEUE_REQ;
+ } else {
+ status = CAM_REQ_CMP_ERR;
+ }
+ mpt_set_ccb_status(ccb, status);
+ }
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__));
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ mpt_free_request(mpt, req);
+ MPTLOCK_2_CAMLOCK(mpt);
+ return;
+ }
+
+ /*
+ * No data to transfer?
+ * Just make a single simple SGL with zero length.
+ */
+
+ if (mpt->verbose >= MPT_PRT_DEBUG) {
+ int tidx = ((char *)sglp) - mpt_off;
+ memset(&mpt_off[tidx], 0xff, MPT_REQUEST_AREA - tidx);
+ }
+
+ if (nseg == 0) {
+ SGE_SIMPLE32 *se1 = (SGE_SIMPLE32 *) sglp;
+ MPI_pSGE_SET_FLAGS(se1,
+ (MPI_SGE_FLAGS_LAST_ELEMENT | MPI_SGE_FLAGS_END_OF_BUFFER |
+ MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_END_OF_LIST));
+ goto out;
+ }
+
+
+ flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT;
+ if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) {
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) {
+ flags |= MPI_SGE_FLAGS_HOST_TO_IOC;
+ }
+ } else {
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
+ flags |= MPI_SGE_FLAGS_HOST_TO_IOC;
+ }
+ }
+
+ if (!(ccb->ccb_h.flags & (CAM_SG_LIST_PHYS|CAM_DATA_PHYS))) {
+ bus_dmasync_op_t op;
+ if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) {
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
+ op = BUS_DMASYNC_PREREAD;
+ } else {
+ op = BUS_DMASYNC_PREWRITE;
+ }
+ } else {
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
+ op = BUS_DMASYNC_PREWRITE;
+ } else {
+ op = BUS_DMASYNC_PREREAD;
+ }
+ }
+ bus_dmamap_sync(mpt->buffer_dmat, req->dmap, op);
+ }
+
+ /*
+ * Okay, fill in what we can at the end of the command frame.
+ * If we have up to MPT_NSGL_FIRST, we can fit them all into
+ * the command frame.
+ *
+ * Otherwise, we fill up through MPT_NSGL_FIRST less one
+ * SIMPLE32 pointers and start doing CHAIN32 entries after
+ * that.
+ */
+
+ if (nseg < MPT_NSGL_FIRST(mpt)) {
+ first_lim = nseg;
+ } else {
+ /*
+ * Leave room for CHAIN element
+ */
+ first_lim = MPT_NSGL_FIRST(mpt) - 1;
+ }
+
+ se = (SGE_SIMPLE32 *) sglp;
+ for (seg = 0; seg < first_lim; seg++, se++, dm_segs++) {
+ uint32_t tf;
+
+ bzero(se, sizeof (*se));
+ se->Address = dm_segs->ds_addr;
+ MPI_pSGE_SET_LENGTH(se, dm_segs->ds_len);
+ tf = flags;
+ if (seg == first_lim - 1) {
+ tf |= MPI_SGE_FLAGS_LAST_ELEMENT;
+ }
+ if (seg == nseg - 1) {
+ tf |= MPI_SGE_FLAGS_END_OF_LIST |
+ MPI_SGE_FLAGS_END_OF_BUFFER;
+ }
+ MPI_pSGE_SET_FLAGS(se, tf);
+ }
+
+ if (seg == nseg) {
+ goto out;
+ }
+
+ /*
+ * Tell the IOC where to find the first chain element.
+ */
+ hdrp->ChainOffset = ((char *)se - (char *)hdrp) >> 2;
+ nxt_off = MPT_RQSL(mpt);
+ trq = req;
+
+ /*
+ * Make up the rest of the data segments out of a chain element
+ * (contiained in the current request frame) which points to
+ * SIMPLE32 elements in the next request frame, possibly ending
+ * with *another* chain element (if there's more).
+ */
+ while (seg < nseg) {
+ int this_seg_lim;
+ uint32_t tf, cur_off;
+ bus_addr_t chain_list_addr;
+
+ /*
+ * Point to the chain descriptor. Note that the chain
+ * descriptor is at the end of the *previous* list (whether
+ * chain or simple).
+ */
+ ce = (SGE_CHAIN32 *) se;
+
+ /*
+ * Before we change our current pointer, make sure we won't
+ * overflow the request area with this frame. Note that we
+ * test against 'greater than' here as it's okay in this case
+ * to have next offset be just outside the request area.
+ */
+ if ((nxt_off + MPT_RQSL(mpt)) > MPT_REQUEST_AREA) {
+ nxt_off = MPT_REQUEST_AREA;
+ goto next_chain;
+ }
+
+ /*
+ * Set our SGE element pointer to the beginning of the chain
+ * list and update our next chain list offset.
+ */
+ se = (SGE_SIMPLE32 *) &mpt_off[nxt_off];
+ cur_off = nxt_off;
+ nxt_off += MPT_RQSL(mpt);
+
+ /*
+ * Now initialized the chain descriptor.
+ */
+ bzero(ce, sizeof (SGE_CHAIN32));
+
+ /*
+ * Get the physical address of the chain list.
+ */
+ chain_list_addr = trq->req_pbuf;
+ chain_list_addr += cur_off;
+ ce->Address = chain_list_addr;
+ ce->Flags = MPI_SGE_FLAGS_CHAIN_ELEMENT;
+
+ /*
+ * If we have more than a frame's worth of segments left,
+ * set up the chain list to have the last element be another
+ * chain descriptor.
+ */
+ if ((nseg - seg) > MPT_NSGL(mpt)) {
+ this_seg_lim = seg + MPT_NSGL(mpt) - 1;
+ /*
+ * The length of the chain is the length in bytes of the
+ * number of segments plus the next chain element.
+ *
+ * The next chain descriptor offset is the length,
+ * in words, of the number of segments.
+ */
+ ce->Length = (this_seg_lim - seg) *
+ sizeof (SGE_SIMPLE32);
+ ce->NextChainOffset = ce->Length >> 2;
+ ce->Length += sizeof (SGE_CHAIN32);
+ } else {
+ this_seg_lim = nseg;
+ ce->Length = (this_seg_lim - seg) *
+ sizeof (SGE_SIMPLE32);
+ }
+
+ /*
+ * Fill in the chain list SGE elements with our segment data.
+ *
+ * If we're the last element in this chain list, set the last
+ * element flag. If we're the completely last element period,
+ * set the end of list and end of buffer flags.
+ */
+ while (seg < this_seg_lim) {
+ bzero(se, sizeof (*se));
+ se->Address = dm_segs->ds_addr;
+ MPI_pSGE_SET_LENGTH(se, dm_segs->ds_len);
+ tf = flags;
+ if (seg == this_seg_lim - 1) {
+ tf |= MPI_SGE_FLAGS_LAST_ELEMENT;
+ }
+ if (seg == nseg - 1) {
+ tf |= MPI_SGE_FLAGS_END_OF_LIST |
+ MPI_SGE_FLAGS_END_OF_BUFFER;
+ }
+ MPI_pSGE_SET_FLAGS(se, tf);
+ se++;
+ seg++;
+ dm_segs++;
+ }
+
+ next_chain:
+ /*
+ * If we have more segments to do and we've used up all of
+ * the space in a request area, go allocate another one
+ * and chain to that.
+ */
+ if (seg < nseg && nxt_off >= MPT_REQUEST_AREA) {
+ request_t *nrq = mpt_get_request(mpt, FALSE);
+
+ if (nrq == NULL) {
+ error = ENOMEM;
+ goto bad;
+ }
+
+ /*
+ * Append the new request area on the tail of our list.
+ */
+ if ((trq = req->chain) == NULL) {
+ req->chain = nrq;
+ } else {
+ while (trq->chain != NULL) {
+ trq = trq->chain;
+ }
+ trq->chain = nrq;
+ }
+ trq = nrq;
+ mpt_off = trq->req_vbuf;
+ if (mpt->verbose >= MPT_PRT_DEBUG) {
+ memset(mpt_off, 0xff, MPT_REQUEST_AREA);
+ }
+ nxt_off = 0;
+ }
+ }
+out:
+
+ /*
+ * Last time we need to check if this CCB needs to be aborted.
+ */
+ if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_INPROG) {
+ if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) {
+ request_t *cmd_req =
+ MPT_TAG_2_REQ(mpt, ccb->csio.tag_id);
+ MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM;
+ MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL;
+ MPT_TGT_STATE(mpt, cmd_req)->req = NULL;
+ }
+ mpt_prt(mpt, "mpt_execute_req: I/O cancelled (status 0x%x)\n",
+ ccb->ccb_h.status & CAM_STATUS_MASK);
+ if (nseg && (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0) {
+ bus_dmamap_unload(mpt->buffer_dmat, req->dmap);
+ }
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__));
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ mpt_free_request(mpt, req);
+ MPTLOCK_2_CAMLOCK(mpt);
+ return;
+ }
+
+ ccb->ccb_h.status |= CAM_SIM_QUEUED;
if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) {
ccb->ccb_h.timeout_ch =
timeout(mpt_timeout, (caddr_t)ccb,
@@ -683,12 +1563,27 @@ out:
}
if (mpt->verbose >= MPT_PRT_DEBUG) {
int nc = 0;
- mpt_print_scsi_io_request(req->req_vbuf);
+ mpt_print_request(req->req_vbuf);
for (trq = req->chain; trq; trq = trq->chain) {
printf(" Additional Chain Area %d\n", nc++);
mpt_dump_sgl(trq->req_vbuf, 0);
}
}
+ if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) {
+ request_t *cmd_req = MPT_TAG_2_REQ(mpt, ccb->csio.tag_id);
+ mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, cmd_req);
+#ifdef WE_TRUST_AUTO_GOOD_STATUS
+ if ((ccb->ccb_h.flags & CAM_SEND_STATUS) &&
+ csio->scsi_status == SCSI_STATUS_OK && tgt->resid == 0) {
+ tgt->state = TGT_STATE_MOVING_DATA_AND_STATUS;
+ } else {
+ tgt->state = TGT_STATE_MOVING_DATA;
+ }
+#else
+ tgt->state = TGT_STATE_MOVING_DATA;
+#endif
+ }
+ CAMLOCK_2_MPTLOCK(mpt);
mpt_send_cmd(mpt, req);
MPTLOCK_2_CAMLOCK(mpt);
}
@@ -701,6 +1596,7 @@ mpt_start(struct cam_sim *sim, union ccb *ccb)
MSG_SCSI_IO_REQUEST *mpt_req;
struct ccb_scsiio *csio = &ccb->csio;
struct ccb_hdr *ccbh = &ccb->ccb_h;
+ bus_dmamap_callback_t *cb;
int raid_passthru;
/* Get the pointer for the physical addapter */
@@ -715,14 +1611,20 @@ mpt_start(struct cam_sim *sim, union ccb *ccb)
xpt_freeze_simq(mpt->sim, 1);
mpt_lprt(mpt, MPT_PRT_DEBUG, "FREEZEQ\n");
}
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ);
MPTLOCK_2_CAMLOCK(mpt);
- ccb->ccb_h.status = CAM_REQUEUE_REQ;
xpt_done(ccb);
return;
}
-
MPTLOCK_2_CAMLOCK(mpt);
+ if (sizeof (bus_addr_t) > 4) {
+ cb = mpt_execute_req_a64;
+ } else {
+ cb = mpt_execute_req;
+ }
+
#if 0
COWWWWW
if (raid_passthru) {
@@ -743,10 +1645,11 @@ mpt_start(struct cam_sim *sim, union ccb *ccb)
bzero(mpt_req, sizeof *mpt_req);
mpt_req->Function = MPI_FUNCTION_SCSI_IO_REQUEST;
- if (raid_passthru)
+ if (raid_passthru) {
mpt_req->Function = MPI_FUNCTION_RAID_SCSI_IO_PASSTHROUGH;
+ }
- mpt_req->Bus = mpt->bus;
+ mpt_req->Bus = 0; /* we don't have multiport devices yet */
mpt_req->SenseBufferLength =
(csio->sense_len < MPT_SENSE_SIZE) ?
@@ -760,10 +1663,14 @@ mpt_start(struct cam_sim *sim, union ccb *ccb)
/* Which physical device to do the I/O on */
mpt_req->TargetID = ccb->ccb_h.target_id;
- /*
- * XXX Assumes Single level, Single byte, CAM LUN type.
- */
- mpt_req->LUN[1] = ccb->ccb_h.target_lun;
+
+ /* We assume a single level LUN type */
+ if (ccb->ccb_h.target_lun >= 256) {
+ mpt_req->LUN[0] = 0x40 | ((ccb->ccb_h.target_lun >> 8) & 0x3f);
+ mpt_req->LUN[1] = ccb->ccb_h.target_lun & 0xff;
+ } else {
+ mpt_req->LUN[1] = ccb->ccb_h.target_lun;
+ }
/* Set the direction of the transfer */
if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
@@ -829,10 +1736,11 @@ mpt_start(struct cam_sim *sim, union ccb *ccb)
* one or more physical address ranges.
*/
int error;
-
+ int s = splsoftvm();
error = bus_dmamap_load(mpt->buffer_dmat,
req->dmap, csio->data_ptr, csio->dxfer_len,
- mpt_execute_req, req, 0);
+ cb, req, 0);
+ splx(s);
if (error == EINPROGRESS) {
/*
* So as to maintain ordering,
@@ -852,7 +1760,7 @@ mpt_start(struct cam_sim *sim, union ccb *ccb)
seg.ds_addr =
(bus_addr_t)(vm_offset_t)csio->data_ptr;
seg.ds_len = csio->dxfer_len;
- mpt_execute_req(req, &seg, 1, 0);
+ (*cb)(req, &seg, 1, 0);
}
} else {
/*
@@ -863,15 +1771,15 @@ mpt_start(struct cam_sim *sim, union ccb *ccb)
*/
struct bus_dma_segment *segs;
if ((ccbh->flags & CAM_SG_LIST_PHYS) == 0) {
- mpt_execute_req(req, NULL, 0, EFAULT);
+ (*cb)(req, NULL, 0, EFAULT);
} else {
/* Just use the segments provided */
segs = (struct bus_dma_segment *)csio->data_ptr;
- mpt_execute_req(req, segs, csio->sglist_cnt, 0);
+ (*cb)(req, segs, csio->sglist_cnt, 0);
}
}
} else {
- mpt_execute_req(req, NULL, 0, 0);
+ (*cb)(req, NULL, 0, 0);
}
}
@@ -917,11 +1825,34 @@ mpt_bus_reset(struct mpt_softc *mpt, int sleep_ok)
}
static int
+mpt_fc_reset_link(struct mpt_softc *mpt, int dowait)
+{
+ int r = 0;
+ request_t *req;
+ PTR_MSG_FC_PRIMITIVE_SEND_REQUEST fc;
+
+ req = mpt_get_request(mpt, FALSE);
+ if (req == NULL) {
+ return (ENOMEM);
+ }
+ fc = req->req_vbuf;
+ memset(fc, 0, sizeof(*fc));
+ fc->SendFlags = MPI_FC_PRIM_SEND_FLAGS_RESET_LINK;
+ fc->Function = MPI_FUNCTION_FC_PRIMITIVE_SEND;
+ fc->MsgContext = htole32(req->index | fc_els_handler_id);
+ mpt_send_cmd(mpt, req);
+ if (dowait) {
+ r = mpt_wait_req(mpt, req, REQ_STATE_DONE,
+ REQ_STATE_DONE, FALSE, 60 * 1000);
+ mpt_free_request(mpt, req);
+ }
+ return (r);
+}
+
+static int
mpt_cam_event(struct mpt_softc *mpt, request_t *req,
MSG_EVENT_NOTIFY_REPLY *msg)
{
- mpt_lprt(mpt, MPT_PRT_ALWAYS, "mpt_cam_event: 0x%x\n",
- msg->Event & 0xFF);
switch(msg->Event & 0xFF) {
case MPI_EVENT_UNIT_ATTENTION:
mpt_prt(mpt, "Bus: 0x%02x TargetID: 0x%02x\n",
@@ -937,7 +1868,7 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req,
case MPI_EVENT_EXT_BUS_RESET:
/* Someone else generated a bus reset */
- mpt_prt(mpt, "Ext Bus Reset\n");
+ mpt_prt(mpt, "External Bus Reset Detected\n");
/*
* These replies don't return EventData like the MPI
* spec says they do
@@ -972,26 +1903,26 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req,
switch ((msg->Data[0] >> 8) & 0xff) {
case 0xF7:
if ((msg->Data[0] & 0xff) == 0xF7) {
- printf("Device needs AL_PA\n");
+ mpt_prt(mpt, "Device needs AL_PA\n");
} else {
- printf("Device %02x doesn't like "
+ mpt_prt(mpt, "Device %02x doesn't like "
"FC performance\n",
msg->Data[0] & 0xFF);
}
break;
case 0xF8:
if ((msg->Data[0] & 0xff) == 0xF7) {
- printf("Device had loop failure at its "
- "receiver prior to acquiring "
- "AL_PA\n");
+ mpt_prt(mpt, "Device had loop failure "
+ "at its receiver prior to acquiring"
+ " AL_PA\n");
} else {
- printf("Device %02x detected loop "
- "failure at its receiver\n",
+ mpt_prt(mpt, "Device %02x detected loop"
+ " failure at its receiver\n",
msg->Data[0] & 0xFF);
}
break;
default:
- printf("Device %02x requests that device "
+ mpt_prt(mpt, "Device %02x requests that device "
"%02x reset itself\n",
msg->Data[0] & 0xFF,
(msg->Data[0] >> 8) & 0xFF);
@@ -1038,6 +1969,8 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req,
"mpt_cam_event: MPI_EVENT_SAS_DEVICE_STATUS_CHANGE\n");
break;
default:
+ mpt_lprt(mpt, MPT_PRT_WARN, "mpt_cam_event: 0x%x\n",
+ msg->Event & 0xFF);
return (/*handled*/0);
}
return (/*handled*/1);
@@ -1055,7 +1988,7 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req,
*/
static int
mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req,
- MSG_DEFAULT_REPLY *reply_frame)
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
{
MSG_SCSI_IO_REQUEST *scsi_req;
union ccb *ccb;
@@ -1084,8 +2017,7 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req,
if (reply_frame == NULL) {
/*
- * Context only reply, completion
- * without error status.
+ * Context only reply, completion without error status.
*/
ccb->csio.resid = 0;
mpt_set_ccb_status(ccb, CAM_REQ_CMP);
@@ -1099,13 +2031,9 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req,
mpt->outofbeer = 0;
mpt_lprt(mpt, MPT_PRT_DEBUG, "THAWQ\n");
}
- ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
- MPTLOCK_2_CAMLOCK(mpt);
- if (scsi_req->Function == MPI_FUNCTION_RAID_SCSI_IO_PASSTHROUGH
- && scsi_req->CDB[0] == INQUIRY
- && (scsi_req->CDB[1] & SI_EVPD) == 0) {
+ if (scsi_req->Function == MPI_FUNCTION_RAID_SCSI_IO_PASSTHROUGH &&
+ scsi_req->CDB[0] == INQUIRY && (scsi_req->CDB[1] & SI_EVPD) == 0) {
struct scsi_inquiry_data *inq;
-
/*
* Fake out the device type so that only the
* pass-thru device will attach.
@@ -1114,13 +2042,16 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req,
inq->device &= ~0x1F;
inq->device |= T_NODEVICE;
}
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__));
+ MPTLOCK_2_CAMLOCK(mpt);
xpt_done(ccb);
CAMLOCK_2_MPTLOCK(mpt);
- if ((req->state & REQ_STATE_TIMEDOUT) == 0)
+ if ((req->state & REQ_STATE_TIMEDOUT) == 0) {
TAILQ_REMOVE(&mpt->request_pending_list, req, links);
- else
+ } else {
TAILQ_REMOVE(&mpt->request_timeout_list, req, links);
-
+ }
if ((req->state & REQ_STATE_NEED_WAKEUP) == 0) {
mpt_free_request(mpt, req);
return (/*free_reply*/TRUE);
@@ -1133,7 +2064,7 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req,
static int
mpt_scsi_tmf_reply_handler(struct mpt_softc *mpt, request_t *req,
- MSG_DEFAULT_REPLY *reply_frame)
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
{
MSG_SCSI_TASK_MGMT_REPLY *tmf_reply;
uint16_t status;
@@ -1158,6 +2089,884 @@ mpt_scsi_tmf_reply_handler(struct mpt_softc *mpt, request_t *req,
return (/*free_reply*/TRUE);
}
+
+/*
+ * XXX: Move to definitions file
+ */
+#define ELS 0x22
+#define FC4LS 0x32
+#define ABTS 0x81
+#define BA_ACC 0x84
+
+#define LS_RJT 0x01
+#define LS_ACC 0x02
+#define PLOGI 0x03
+#define LOGO 0x05
+#define SRR 0x14
+#define PRLI 0x20
+#define PRLO 0x21
+#define ADISC 0x52
+#define RSCN 0x61
+
+static void
+mpt_fc_els_send_response(struct mpt_softc *mpt, request_t *req,
+ PTR_MSG_LINK_SERVICE_BUFFER_POST_REPLY rp, U8 length)
+{
+ MSG_LINK_SERVICE_RSP_REQUEST tmp;
+ PTR_MSG_LINK_SERVICE_RSP_REQUEST rsp;
+
+ /*
+ * We are going to reuse the ELS request to send this response back.
+ */
+ rsp = &tmp;
+ memset(rsp, 0, sizeof(*rsp));
+
+#ifdef USE_IMMEDIATE_LINK_DATA
+ /*
+ * Apparently the IMMEDIATE stuff doesn't seem to work.
+ */
+ rsp->RspFlags = LINK_SERVICE_RSP_FLAGS_IMMEDIATE;
+#endif
+ rsp->RspLength = length;
+ rsp->Function = MPI_FUNCTION_FC_LINK_SRVC_RSP;
+ rsp->MsgContext = htole32(req->index | fc_els_handler_id);
+
+ /*
+ * Copy over information from the original reply frame to
+ * it's correct place in the response.
+ */
+ memcpy((U8 *)rsp + 0x0c, (U8 *)rp + 0x1c, 24);
+
+ /*
+ * And now copy back the temporary area to the original frame.
+ */
+ memcpy(req->req_vbuf, rsp, sizeof (MSG_LINK_SERVICE_RSP_REQUEST));
+ rsp = req->req_vbuf;
+
+#ifdef USE_IMMEDIATE_LINK_DATA
+ memcpy((U8 *)&rsp->SGL, &((U8 *)req->req_vbuf)[MPT_RQSL(mpt)], length);
+#else
+{
+ PTR_SGE_SIMPLE32 se = (PTR_SGE_SIMPLE32) &rsp->SGL;
+ bus_addr_t paddr = req->req_pbuf;
+ paddr += MPT_RQSL(mpt);
+
+ se->FlagsLength =
+ MPI_SGE_FLAGS_HOST_TO_IOC |
+ MPI_SGE_FLAGS_SIMPLE_ELEMENT |
+ MPI_SGE_FLAGS_LAST_ELEMENT |
+ MPI_SGE_FLAGS_END_OF_LIST |
+ MPI_SGE_FLAGS_END_OF_BUFFER;
+ se->FlagsLength <<= MPI_SGE_FLAGS_SHIFT;
+ se->FlagsLength |= (length);
+ se->Address = (uint32_t) paddr;
+}
+#endif
+
+ /*
+ * Send it on...
+ */
+ mpt_send_cmd(mpt, req);
+}
+
+static int
+mpt_fc_els_reply_handler(struct mpt_softc *mpt, request_t *req,
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
+{
+ PTR_MSG_LINK_SERVICE_BUFFER_POST_REPLY rp =
+ (PTR_MSG_LINK_SERVICE_BUFFER_POST_REPLY) reply_frame;
+ U8 rctl;
+ U8 type;
+ U8 cmd;
+ U16 status = le16toh(reply_frame->IOCStatus);
+ U32 *elsbuf;
+ int do_refresh = TRUE;
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG, "FC_ELS Complete: req %p:%u, reply %p\n",
+ req, req->serno, reply_frame);
+
+ if (status != MPI_IOCSTATUS_SUCCESS) {
+ mpt_prt(mpt, "ELS REPLY STATUS 0x%x for Function %x\n",
+ status, reply_frame->Function);
+ if (status == MPI_IOCSTATUS_INVALID_STATE) {
+ /*
+ * XXX: to get around shutdown issue
+ */
+ mpt->disabled = 1;
+ return (TRUE);
+ }
+ return (TRUE);
+ }
+
+ /*
+ * If the function of a link service response, we recycle the
+ * response to be a refresh for a new link service request.
+ */
+ if (rp->Function == MPI_FUNCTION_FC_LINK_SRVC_RSP) {
+ mpt_fc_add_els(mpt, req);
+ return (TRUE);
+ }
+
+ if (rp->Function == MPI_FUNCTION_FC_PRIMITIVE_SEND) {
+ req->state &= ~REQ_STATE_QUEUED;
+ req->state |= REQ_STATE_DONE;
+ if ((req->state & REQ_STATE_NEED_WAKEUP) == 0) {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "Async Primitive Send Complete\n");
+ mpt_free_request(mpt, req);
+ } else {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "Sync Primitive Send Complete\n");
+ wakeup(req);
+ }
+ return (TRUE);
+ }
+
+ if (rp->Function != MPI_FUNCTION_FC_LINK_SRVC_BUF_POST) {
+ mpt_prt(mpt, "unexpected ELS_REPLY: Function 0x%x Flags %x "
+ "Length %d Message Flags %x\n", rp->Function, rp->Flags,
+ rp->MsgLength, rp->MsgFlags);
+ return (TRUE);
+ }
+
+ if (rp->MsgLength <= 5) {
+ /*
+ * This is just a ack of an original ELS buffer post
+ */
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "Recv'd ACK of FC_ELS buf post %p:%u\n", req, req->serno);
+ return (TRUE);
+ }
+
+
+ rctl = (le32toh(rp->Rctl_Did) & MPI_FC_RCTL_MASK) >> MPI_FC_RCTL_SHIFT;
+ type = (le32toh(rp->Type_Fctl) & MPI_FC_TYPE_MASK) >> MPI_FC_TYPE_SHIFT;
+
+ elsbuf = &((U32 *)req->req_vbuf)[MPT_RQSL(mpt)/sizeof (U32)];
+ cmd = be32toh(elsbuf[0]) >> 24;
+
+ if (rp->Flags & MPI_LS_BUF_POST_REPLY_FLAG_NO_RSP_NEEDED) {
+ mpt_lprt(mpt, MPT_PRT_ALWAYS, "ELS_REPLY: response unneeded\n");
+ return (TRUE);
+ }
+
+
+ if (rctl == ELS && type == 1) {
+ switch (cmd) {
+ case PRLI:
+ /*
+ * Send back a PRLI ACC
+ */
+ mpt_prt(mpt, "PRLI from 0x%08x%08x\n",
+ le32toh(rp->Wwn.PortNameHigh),
+ le32toh(rp->Wwn.PortNameLow));
+ elsbuf[0] = htobe32(0x02100014);
+ elsbuf[1] |= htobe32(0x00000100);
+ elsbuf[4] = htobe32(0x00000002);
+ if (mpt->role & MPT_ROLE_TARGET)
+ elsbuf[4] |= htobe32(0x00000010);
+ if (mpt->role & MPT_ROLE_INITIATOR)
+ elsbuf[4] |= htobe32(0x00000020);
+ mpt_fc_els_send_response(mpt, req, rp, 20);
+ do_refresh = FALSE;
+ break;
+ case PRLO:
+ memset(elsbuf, 0, 5 * (sizeof (U32)));
+ elsbuf[0] = htobe32(0x02100014);
+ elsbuf[1] = htobe32(0x08000100);
+ mpt_prt(mpt, "PRLO from 0x%08x%08x\n",
+ le32toh(rp->Wwn.PortNameHigh),
+ le32toh(rp->Wwn.PortNameLow));
+ mpt_fc_els_send_response(mpt, req, rp, 20);
+ do_refresh = FALSE;
+ break;
+ default:
+ mpt_prt(mpt, "ELS TYPE 1 COMMAND: %x\n", cmd);
+ break;
+ }
+ } else if (rctl == ABTS && type == 0) {
+ uint16_t rx_id = le16toh(rp->Rxid);
+ uint16_t ox_id = le16toh(rp->Oxid);
+ request_t *tgt_req = NULL;
+
+ mpt_prt(mpt,
+ "ELS: ABTS OX_ID 0x%x RX_ID 0x%x from 0x%08x%08x\n",
+ ox_id, rx_id, le32toh(rp->Wwn.PortNameHigh),
+ le32toh(rp->Wwn.PortNameLow));
+ if (rx_id >= mpt->mpt_max_tgtcmds) {
+ mpt_prt(mpt, "Bad RX_ID 0x%x\n", rx_id);
+ } else if (mpt->tgt_cmd_ptrs == NULL) {
+ mpt_prt(mpt, "No TGT CMD PTRS\n");
+ } else {
+ tgt_req = mpt->tgt_cmd_ptrs[rx_id];
+ }
+ if (tgt_req) {
+ mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, tgt_req);
+ uint8_t *vbuf;
+ union ccb *ccb = tgt->ccb;
+ cam_status cs;
+ uint32_t ct_id;
+
+ vbuf = tgt_req->req_vbuf;
+ vbuf += MPT_RQSL(mpt);
+
+ /*
+ * Check to make sure we have the correct command
+ * The reply descriptor in the target state should
+ * should contain an IoIndex that should match the
+ * RX_ID.
+ *
+ * It'd be nice to have OX_ID to crosscheck with
+ * as well.
+ */
+ ct_id = GET_IO_INDEX(tgt->reply_desc);
+
+ if (ct_id != rx_id) {
+ mpt_lprt(mpt, MPT_PRT_ERROR, "ABORT Mismatch: "
+ "RX_ID received=0x%x; RX_ID in cmd=0x%x\n",
+ rx_id, ct_id);
+ goto skip;
+ }
+
+ ccb = tgt->ccb;
+ if (ccb) {
+ mpt_prt(mpt,
+ "CCB (%p): lun %u flags %x status %x\n",
+ ccb, ccb->ccb_h.target_lun,
+ ccb->ccb_h.flags, ccb->ccb_h.status);
+ }
+ mpt_prt(mpt, "target state 0x%x resid %u xfrd %u rpwrd "
+ "%x nxfers %x flags %x\n", tgt->state,
+ tgt->resid, tgt->bytes_xfered, tgt->reply_desc,
+ tgt->nxfers, tgt->flags);
+ skip:
+ cs = mpt_abort_target_cmd(mpt, tgt_req);
+ if (cs != CAM_REQ_INPROG) {
+ mpt_prt(mpt, "unable to do TargetAbort (%x)\n",
+ cs);
+ }
+ } else {
+ mpt_prt(mpt, "no back pointer for RX_ID 0x%x\n", rx_id);
+ }
+ memset(elsbuf, 0, 5 * (sizeof (U32)));
+ elsbuf[0] = htobe32(0);
+ elsbuf[1] = htobe32((ox_id << 16) | rx_id);
+ elsbuf[2] = htobe32(0x000ffff);
+ /*
+ * Dork with the reply frame so that the reponse to it
+ * will be correct.
+ */
+ rp->Rctl_Did += ((BA_ACC - ABTS) << MPI_FC_RCTL_SHIFT);
+ mpt_fc_els_send_response(mpt, req, rp, 12);
+ do_refresh = FALSE;
+ } else {
+ mpt_prt(mpt, "ELS: RCTL %x TYPE %x CMD %x\n", rctl, type, cmd);
+ }
+ if (do_refresh == TRUE) {
+ mpt_fc_add_els(mpt, req);
+ }
+ return (TRUE);
+}
+
+/*
+ * WE_TRUST_AUTO_GOOD_STATUS- I've found that setting
+ * TARGET_STATUS_SEND_FLAGS_AUTO_GOOD_STATUS leads the
+ * FC929 to set bogus FC_RSP fields (nonzero residuals
+ * but w/o RESID fields set). This causes QLogic initiators
+ * to think maybe that a frame was lost.
+ *
+ * WE_CAN_USE_AUTO_REPOST- we can't use AUTO_REPOST because
+ * we use allocated requests to do TARGET_ASSIST and we
+ * need to know when to release them.
+ */
+
+static void
+mpt_scsi_tgt_status(struct mpt_softc *mpt, union ccb *ccb, request_t *cmd_req,
+ uint8_t status, uint8_t const *sense_data)
+{
+ uint8_t *cmd_vbuf;
+ mpt_tgt_state_t *tgt;
+ PTR_MSG_TARGET_STATUS_SEND_REQUEST tp;
+ request_t *req;
+ bus_addr_t paddr;
+ int resplen = 0;
+
+ cmd_vbuf = cmd_req->req_vbuf;
+ cmd_vbuf += MPT_RQSL(mpt);
+ tgt = MPT_TGT_STATE(mpt, cmd_req);
+
+ req = mpt_get_request(mpt, FALSE);
+ if (req == NULL) {
+ if (ccb) {
+ ccb->ccb_h.status &= ~CAM_STATUS_MASK;
+ mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ);
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ } else {
+ /*
+ * XXX: put in deferred processing if we cannot allocate
+ */
+ mpt_prt(mpt,
+ "XXXX could not allocate status req- dropping\n");
+ }
+ return;
+ }
+ req->ccb = ccb;
+ if (ccb) {
+ ccb->ccb_h.ccb_mpt_ptr = mpt;
+ ccb->ccb_h.ccb_req_ptr = req;
+ }
+
+ /*
+ * Record the currently active ccb, if any, and the
+ * request for it in our target state area.
+ */
+ tgt->ccb = ccb;
+ tgt->req = req;
+ tgt->state = TGT_STATE_SENDING_STATUS;
+
+ tp = req->req_vbuf;
+ paddr = req->req_pbuf;
+ paddr += MPT_RQSL(mpt);
+
+ memset(tp, 0, sizeof (*tp));
+ tp->Function = MPI_FUNCTION_TARGET_STATUS_SEND;
+ if (mpt->is_fc) {
+ PTR_MPI_TARGET_FCP_CMD_BUFFER fc =
+ (PTR_MPI_TARGET_FCP_CMD_BUFFER) cmd_vbuf;
+ uint8_t *sts_vbuf;
+ uint32_t *rsp;
+
+ sts_vbuf = req->req_vbuf;
+ sts_vbuf += MPT_RQSL(mpt);
+ rsp = (uint32_t *) sts_vbuf;
+ memcpy(tp->LUN, fc->FcpLun, sizeof (tp->LUN));
+
+ /*
+ * The MPI_TARGET_FCP_RSP_BUFFER define is unfortunate.
+ * It has to be big-endian in memory and is organized
+ * in 32 bit words, which are much easier to deal with
+ * as words which are swizzled as needed.
+ *
+ * All we're filling here is the FC_RSP payload.
+ * We may just have the chip synthesize it if
+ * we have no residual and an OK status.
+ *
+ */
+ memset(rsp, 0, sizeof (MPI_TARGET_FCP_RSP_BUFFER));
+
+ rsp[2] = status;
+ if (tgt->resid) {
+ rsp[2] |= 0x800;
+ rsp[3] = htobe32(tgt->resid);
+#ifdef WE_TRUST_AUTO_GOOD_STATUS
+ resplen = sizeof (MPI_TARGET_FCP_RSP_BUFFER);
+#endif
+ }
+ if (status == SCSI_STATUS_CHECK_COND) {
+ int i;
+
+ rsp[2] |= 0x200;
+ rsp[4] = htobe32(MPT_SENSE_SIZE);
+ memcpy(&rsp[8], sense_data, MPT_SENSE_SIZE);
+ for (i = 8; i < (8 + (MPT_SENSE_SIZE >> 2)); i++) {
+ rsp[i] = htobe32(rsp[i]);
+ }
+#ifdef WE_TRUST_AUTO_GOOD_STATUS
+ resplen = sizeof (MPI_TARGET_FCP_RSP_BUFFER);
+#endif
+ }
+#ifndef WE_TRUST_AUTO_GOOD_STATUS
+ resplen = sizeof (MPI_TARGET_FCP_RSP_BUFFER);
+#endif
+ rsp[2] = htobe32(rsp[2]);
+ } else if (mpt->is_sas) {
+ PTR_MPI_TARGET_SSP_CMD_BUFFER ssp =
+ (PTR_MPI_TARGET_SSP_CMD_BUFFER) cmd_vbuf;
+ memcpy(tp->LUN, ssp->LogicalUnitNumber, sizeof (tp->LUN));
+ } else {
+ PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER sp =
+ (PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER) cmd_vbuf;
+ tp->StatusCode = status;
+ tp->QueueTag = htole16(sp->Tag);
+ memcpy(tp->LUN, sp->LogicalUnitNumber, sizeof (tp->LUN));
+ }
+
+ tp->ReplyWord = htole32(tgt->reply_desc);
+ tp->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id);
+
+#ifdef WE_CAN_USE_AUTO_REPOST
+ tp->MsgFlags = TARGET_STATUS_SEND_FLAGS_REPOST_CMD_BUFFER;
+#endif
+ if (status == SCSI_STATUS_OK && resplen == 0) {
+ tp->MsgFlags |= TARGET_STATUS_SEND_FLAGS_AUTO_GOOD_STATUS;
+ } else {
+ tp->StatusDataSGE.u.Address32 = (uint32_t) paddr;
+ tp->StatusDataSGE.FlagsLength =
+ MPI_SGE_FLAGS_HOST_TO_IOC |
+ MPI_SGE_FLAGS_SIMPLE_ELEMENT |
+ MPI_SGE_FLAGS_LAST_ELEMENT |
+ MPI_SGE_FLAGS_END_OF_LIST |
+ MPI_SGE_FLAGS_END_OF_BUFFER;
+ tp->StatusDataSGE.FlagsLength <<= MPI_SGE_FLAGS_SHIFT;
+ tp->StatusDataSGE.FlagsLength |= resplen;
+ }
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "STATUS_CCB %p (wit%s sense) tag %x req %p:%u resid %u\n",
+ ccb, sense_data?"h" : "hout", ccb? ccb->csio.tag_id : -1, req,
+ req->serno, tgt->resid);
+ if (ccb) {
+ ccb->ccb_h.status = CAM_SIM_QUEUED | CAM_REQ_INPROG;
+ ccb->ccb_h.timeout_ch = timeout(mpt_timeout, (caddr_t)ccb, hz);
+ }
+ mpt_send_cmd(mpt, req);
+}
+
+static void
+mpt_scsi_tgt_tsk_mgmt(struct mpt_softc *mpt, request_t *req, mpt_task_mgmt_t fc,
+ tgt_resource_t *trtp, int init_id)
+{
+ struct ccb_immed_notify *inot;
+ mpt_tgt_state_t *tgt;
+
+ tgt = MPT_TGT_STATE(mpt, req);
+ inot = (struct ccb_immed_notify *) STAILQ_FIRST(&trtp->inots);
+ if (inot == NULL) {
+ mpt_lprt(mpt, MPT_PRT_WARN, "no INOTSs- sending back BSY\n");
+ mpt_scsi_tgt_status(mpt, NULL, req, SCSI_STATUS_BUSY, NULL);
+ return;
+ }
+ STAILQ_REMOVE_HEAD(&trtp->inots, sim_links.stqe);
+ mpt_lprt(mpt, MPT_PRT_DEBUG1,
+ "Get FREE INOT %p lun %d\n", inot, inot->ccb_h.target_lun);
+
+ memset(&inot->sense_data, 0, sizeof (inot->sense_data));
+ inot->sense_len = 0;
+ memset(inot->message_args, 0, sizeof (inot->message_args));
+ inot->initiator_id = init_id; /* XXX */
+
+ /*
+ * This is a somewhat grotesque attempt to map from task management
+ * to old style SCSI messages. God help us all.
+ */
+ switch (fc) {
+ case MPT_ABORT_TASK_SET:
+ inot->message_args[0] = MSG_ABORT_TAG;
+ break;
+ case MPT_CLEAR_TASK_SET:
+ inot->message_args[0] = MSG_CLEAR_TASK_SET;
+ break;
+ case MPT_TARGET_RESET:
+ inot->message_args[0] = MSG_TARGET_RESET;
+ break;
+ case MPT_CLEAR_ACA:
+ inot->message_args[0] = MSG_CLEAR_ACA;
+ break;
+ case MPT_TERMINATE_TASK:
+ inot->message_args[0] = MSG_ABORT_TAG;
+ break;
+ default:
+ inot->message_args[0] = MSG_NOOP;
+ break;
+ }
+ tgt->ccb = (union ccb *) inot;
+ inot->ccb_h.status = CAM_MESSAGE_RECV|CAM_DEV_QFRZN;
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done((union ccb *)inot);
+ CAMLOCK_2_MPTLOCK(mpt);
+}
+
+static void
+mpt_scsi_tgt_atio(struct mpt_softc *mpt, request_t *req, uint32_t reply_desc)
+{
+ struct ccb_accept_tio *atiop;
+ lun_id_t lun;
+ int tag_action = 0;
+ mpt_tgt_state_t *tgt;
+ tgt_resource_t *trtp;
+ U8 *lunptr;
+ U8 *vbuf;
+ U16 itag;
+ U16 ioindex;
+ mpt_task_mgmt_t fct = MPT_NIL_TMT_VALUE;
+ uint8_t *cdbp;
+
+ /*
+ * First, DMA sync the received command- which is in the *request*
+ * phys area.
+ * XXX: We could optimize this for a range
+ */
+ bus_dmamap_sync(mpt->request_dmat, mpt->request_dmap,
+ BUS_DMASYNC_POSTREAD);
+
+ /*
+ * Stash info for the current command where we can get at it later.
+ */
+ vbuf = req->req_vbuf;
+ vbuf += MPT_RQSL(mpt);
+
+ /*
+ * Get our state pointer set up.
+ */
+ tgt = MPT_TGT_STATE(mpt, req);
+ KASSERT(tgt->state == TGT_STATE_LOADED,
+ ("bad target state %x in mpt_scsi_tgt_atio for req %p\n",
+ tgt->state, req));
+ memset(tgt, 0, sizeof (mpt_tgt_state_t));
+ tgt->state = TGT_STATE_IN_CAM;
+ tgt->reply_desc = reply_desc;
+ ioindex = GET_IO_INDEX(reply_desc);
+
+ if (mpt->is_fc) {
+ PTR_MPI_TARGET_FCP_CMD_BUFFER fc;
+ fc = (PTR_MPI_TARGET_FCP_CMD_BUFFER) vbuf;
+ if (fc->FcpCntl[2]) {
+ /*
+ * Task Management Request
+ */
+ switch (fc->FcpCntl[2]) {
+ case 0x2:
+ fct = MPT_ABORT_TASK_SET;
+ break;
+ case 0x4:
+ fct = MPT_CLEAR_TASK_SET;
+ break;
+ case 0x20:
+ fct = MPT_TARGET_RESET;
+ break;
+ case 0x40:
+ fct = MPT_CLEAR_ACA;
+ break;
+ case 0x80:
+ fct = MPT_TERMINATE_TASK;
+ break;
+ default:
+ mpt_prt(mpt, "CORRUPTED TASK MGMT BITS: 0x%x\n",
+ fc->FcpCntl[2]);
+ mpt_scsi_tgt_status(mpt, 0, req,
+ SCSI_STATUS_OK, 0);
+ return;
+ }
+ return;
+ }
+ switch (fc->FcpCntl[1]) {
+ case 0:
+ tag_action = MSG_SIMPLE_Q_TAG;
+ break;
+ case 1:
+ tag_action = MSG_HEAD_OF_Q_TAG;
+ break;
+ case 2:
+ tag_action = MSG_ORDERED_Q_TAG;
+ break;
+ default:
+ /*
+ * Bah. Ignore Untagged Queing and ACA
+ */
+ tag_action = MSG_SIMPLE_Q_TAG;
+ break;
+ }
+ tgt->resid = be32toh(fc->FcpDl);
+ cdbp = fc->FcpCdb;
+ lunptr = fc->FcpLun;
+ itag = be16toh(fc->OptionalOxid);
+ } else if (mpt->is_sas) {
+ PTR_MPI_TARGET_SSP_CMD_BUFFER ssp;
+ ssp = (PTR_MPI_TARGET_SSP_CMD_BUFFER) vbuf;
+ cdbp = ssp->CDB;
+ lunptr = ssp->LogicalUnitNumber;
+ itag = ssp->InitiatorTag;
+ } else {
+ PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER sp;
+ sp = (PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER) vbuf;
+ cdbp = sp->CDB;
+ lunptr = sp->LogicalUnitNumber;
+ itag = sp->Tag;
+ }
+
+ /*
+ * Generate a simple lun
+ */
+ switch (lunptr[0] & 0xc0) {
+ case 0x40:
+ lun = ((lunptr[0] & 0x3f) << 8) | lunptr[1];
+ break;
+ case 0:
+ lun = lunptr[1];
+ break;
+ default:
+ mpt_lprt(mpt, MPT_PRT_ERROR, "cannot handle this type lun\n");
+ lun = 0xffff;
+ break;
+ }
+
+ /*
+ * Deal with non-enabled or bad luns here.
+ */
+ if (lun >= MPT_MAX_LUNS || mpt->tenabled == 0 ||
+ mpt->trt[lun].enabled == 0) {
+ if (mpt->twildcard) {
+ trtp = &mpt->trt_wildcard;
+ } else {
+ const uint8_t sp[MPT_SENSE_SIZE] = {
+ 0xf0, 0, 0x5, 0, 0, 0, 0, 8, 0, 0, 0, 0, 0x25
+ };
+ mpt_scsi_tgt_status(mpt, NULL, req,
+ SCSI_STATUS_CHECK_COND, sp);
+ return;
+ }
+ } else {
+ trtp = &mpt->trt[lun];
+ }
+
+ if (fct != MPT_NIL_TMT_VALUE) {
+ /* undo any tgt residual settings */
+ tgt->resid = 0;
+ mpt_scsi_tgt_tsk_mgmt(mpt, req, fct, trtp,
+ GET_INITIATOR_INDEX(reply_desc));
+ return;
+ }
+
+ atiop = (struct ccb_accept_tio *) STAILQ_FIRST(&trtp->atios);
+ if (atiop == NULL) {
+ mpt_lprt(mpt, MPT_PRT_WARN,
+ "no ATIOs for lun %u- sending back %s\n", lun,
+ mpt->tenabled? "QUEUE FULL" : "BUSY");
+ mpt_scsi_tgt_status(mpt, NULL, req,
+ mpt->tenabled? SCSI_STATUS_QUEUE_FULL : SCSI_STATUS_BUSY,
+ NULL);
+ return;
+ }
+ STAILQ_REMOVE_HEAD(&trtp->atios, sim_links.stqe);
+ mpt_lprt(mpt, MPT_PRT_DEBUG1,
+ "Get FREE ATIO %p lun %d\n", atiop, atiop->ccb_h.target_lun);
+ atiop->ccb_h.ccb_mpt_ptr = mpt;
+ atiop->ccb_h.status = CAM_CDB_RECVD;
+ atiop->ccb_h.target_lun = lun;
+ atiop->sense_len = 0;
+ atiop->init_id = GET_INITIATOR_INDEX(reply_desc);
+ atiop->cdb_len = mpt_cdblen(cdbp[0], 16);
+ memcpy(atiop->cdb_io.cdb_bytes, cdbp, atiop->cdb_len);
+
+ /*
+ * The tag we construct here allows us to find the
+ * original request that the command came in with.
+ *
+ * This way we don't have to depend on anything but the
+ * tag to find things when CCBs show back up from CAM.
+ */
+ atiop->tag_id = MPT_MAKE_TAGID(mpt, req, ioindex);
+ if (tag_action) {
+ atiop->tag_action = tag_action;
+ atiop->ccb_h.flags = CAM_TAG_ACTION_VALID;
+ }
+ if (mpt->verbose >= MPT_PRT_DEBUG) {
+ int i;
+ mpt_prt(mpt, "START_CCB %p for lun %u CDB=<", atiop,
+ atiop->ccb_h.target_lun);
+ for (i = 0; i < atiop->cdb_len; i++) {
+ mpt_prtc(mpt, "%02x%c", cdbp[i] & 0xff,
+ (i == (atiop->cdb_len - 1))? '>' : ' ');
+ }
+ mpt_prtc(mpt, " itag %x tag %x rdesc %x dl=%u\n",
+ itag, atiop->tag_id, tgt->reply_desc, tgt->resid);
+ }
+ tgt->ccb = (union ccb *) atiop;
+
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done((union ccb *)atiop);
+ CAMLOCK_2_MPTLOCK(mpt);
+}
+
+static int
+mpt_scsi_tgt_reply_handler(struct mpt_softc *mpt, request_t *req,
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
+{
+ int dbg;
+ union ccb *ccb;
+ U16 status;
+
+ if (reply_frame == NULL) {
+ /*
+ * Figure out if this is a new command or a target assist
+ * completing.
+ */
+ mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, req);
+ char serno[8];
+
+ if (tgt->req) {
+ snprintf(serno, 8, "%u", tgt->req->serno);
+ } else {
+ strncpy(serno, "??", 8);
+ }
+
+ switch(tgt->state) {
+ case TGT_STATE_LOADED:
+ mpt_scsi_tgt_atio(mpt, req, reply_desc);
+ break;
+ case TGT_STATE_MOVING_DATA:
+ {
+ uint8_t *sp = NULL, sense[MPT_SENSE_SIZE];
+
+ ccb = tgt->ccb;
+ tgt->ccb = NULL;
+ tgt->nxfers++;
+ untimeout(mpt_timeout, ccb, ccb->ccb_h.timeout_ch);
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "TARGET_ASSIST %p (req %p:%s) done tag 0x%x\n",
+ ccb, tgt->req, serno, ccb->csio.tag_id);
+ /*
+ * Free the Target Assist Request
+ */
+ KASSERT(tgt->req && tgt->req->ccb == ccb,
+ ("tgt->req %p:%s tgt->req->ccb %p", tgt->req,
+ serno, tgt->req? tgt->req->ccb : NULL));
+ mpt_free_request(mpt, tgt->req);
+ tgt->req = NULL;
+ /*
+ * Do we need to send status now? That is, are
+ * we done with all our data transfers?
+ */
+ if ((ccb->ccb_h.flags & CAM_SEND_STATUS) == 0) {
+ mpt_set_ccb_status(ccb, CAM_REQ_CMP);
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ KASSERT(ccb->ccb_h.status,
+ ("zero ccb sts at %d\n", __LINE__));
+ tgt->state = TGT_STATE_IN_CAM;
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ break;
+ }
+ if (ccb->ccb_h.flags & CAM_SEND_SENSE) {
+ sp = sense;
+ memcpy(sp, &ccb->csio.sense_data,
+ min(ccb->csio.sense_len, MPT_SENSE_SIZE));
+ }
+ mpt_scsi_tgt_status(mpt, ccb, req,
+ ccb->csio.scsi_status, sp);
+ break;
+ }
+ case TGT_STATE_SENDING_STATUS:
+ case TGT_STATE_MOVING_DATA_AND_STATUS:
+ {
+ int ioindex;
+ ccb = tgt->ccb;
+
+ if (ccb) {
+ tgt->ccb = NULL;
+ tgt->nxfers++;
+ untimeout(mpt_timeout, ccb,
+ ccb->ccb_h.timeout_ch);
+ if (ccb->ccb_h.flags & CAM_SEND_SENSE) {
+ ccb->ccb_h.status |= CAM_SENT_SENSE;
+ }
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "TARGET_STATUS tag %x sts %x flgs %x req "
+ "%p\n", ccb->csio.tag_id, ccb->ccb_h.status,
+ ccb->ccb_h.flags, tgt->req);
+ /*
+ * Free the Target Send Status Request
+ */
+ KASSERT(tgt->req && tgt->req->ccb == ccb,
+ ("tgt->req %p:%s tgt->req->ccb %p",
+ tgt->req, serno,
+ tgt->req? tgt->req->ccb : NULL));
+ /*
+ * Notify CAM that we're done
+ */
+ mpt_set_ccb_status(ccb, CAM_REQ_CMP);
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ KASSERT(ccb->ccb_h.status,
+ ("ZERO ccb sts at %d\n", __LINE__));
+ tgt->ccb = NULL;
+ } else {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "TARGET_STATUS non-CAM for req %p:%s\n",
+ tgt->req, serno);
+ }
+ mpt_free_request(mpt, tgt->req);
+ tgt->req = NULL;
+
+ /*
+ * And re-post the Command Buffer.
+ */
+ ioindex = GET_IO_INDEX(reply_desc);
+ mpt_post_target_command(mpt, req, ioindex);
+
+ /*
+ * And post a done for anyone who cares
+ */
+ if (ccb) {
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ }
+ break;
+ }
+ case TGT_STATE_NIL: /* XXX This Never Happens XXX */
+ tgt->state = TGT_STATE_LOADED;
+ break;
+ default:
+ mpt_prt(mpt, "Unknown Target State 0x%x in Context "
+ "Reply Function\n", tgt->state);
+ }
+ return (TRUE);
+ }
+
+ status = le16toh(reply_frame->IOCStatus);
+ if (status != MPI_IOCSTATUS_SUCCESS) {
+ dbg = MPT_PRT_ERROR;
+ } else {
+ dbg = MPT_PRT_DEBUG1;
+ }
+
+ mpt_lprt(mpt, dbg,
+ "SCSI_TGT REPLY: req=%p:%u reply=%p func=%x IOCstatus 0x%x\n",
+ req, req->serno, reply_frame, reply_frame->Function, status);
+
+ switch (reply_frame->Function) {
+ case MPI_FUNCTION_TARGET_CMD_BUFFER_POST:
+ KASSERT(MPT_TGT_STATE(mpt,
+ req)->state == TGT_STATE_NIL,
+ ("bad state %x on reply to buffer post\n",
+ MPT_TGT_STATE(mpt, req)->state));
+ MPT_TGT_STATE(mpt, req)->state = TGT_STATE_LOADED;
+ break;
+ case MPI_FUNCTION_TARGET_ASSIST:
+ mpt_prt(mpt,
+ "TARGET_ASSIST err for request %p:%u (%x): status 0x%x\n",
+ req, req->serno, req->index, status);
+ mpt_free_request(mpt, req);
+ break;
+ case MPI_FUNCTION_TARGET_STATUS_SEND:
+ mpt_prt(mpt,
+ "TARGET_STATUS_SEND error for request %p:%u(%x): status "
+ "0x%x\n", req, req->serno, req->index, status);
+ mpt_free_request(mpt, req);
+ break;
+ case MPI_FUNCTION_TARGET_MODE_ABORT:
+ {
+ PTR_MSG_TARGET_MODE_ABORT_REPLY abtrp =
+ (PTR_MSG_TARGET_MODE_ABORT_REPLY) reply_frame;
+ PTR_MSG_TARGET_MODE_ABORT abtp =
+ (PTR_MSG_TARGET_MODE_ABORT) req->req_vbuf;
+ uint32_t cc = GET_IO_INDEX(le32toh(abtp->ReplyWord));
+ mpt_prt(mpt, "ABORT RX_ID 0x%x Complete; status 0x%x cnt %u\n",
+ cc, le16toh(abtrp->IOCStatus), le32toh(abtrp->AbortCount));
+ mpt_free_request(mpt, req);
+ break;
+ }
+ default:
+ mpt_prt(mpt, "Unknown Target Address Reply Function code: "
+ "0x%x\n", reply_frame->Function);
+ break;
+ }
+ return (TRUE);
+}
+
/*
* Clean up all SCSI Initiator personality state in response
* to a controller reset.
@@ -1354,12 +3163,12 @@ mpt_action(struct cam_sim *sim, union ccb *ccb)
raid_passthru = (sim == mpt->phydisk_sim);
tgt = ccb->ccb_h.target_id;
- if (raid_passthru
- && ccb->ccb_h.func_code != XPT_PATH_INQ
- && ccb->ccb_h.func_code != XPT_RESET_BUS) {
+ if (raid_passthru && ccb->ccb_h.func_code != XPT_PATH_INQ &&
+ ccb->ccb_h.func_code != XPT_RESET_BUS) {
CAMLOCK_2_MPTLOCK(mpt);
if (mpt_map_physdisk(mpt, ccb, &tgt) != 0) {
- ccb->ccb_h.status = CAM_DEV_NOT_THERE;
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_DEV_NOT_THERE);
MPTLOCK_2_CAMLOCK(mpt);
xpt_done(ccb);
return;
@@ -1376,7 +3185,8 @@ mpt_action(struct cam_sim *sim, union ccb *ccb)
*/
if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0) {
if ((ccb->ccb_h.flags & CAM_CDB_PHYS) != 0) {
- ccb->ccb_h.status = CAM_REQ_INVALID;
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_REQ_INVALID);
xpt_done(ccb);
break;
}
@@ -1385,7 +3195,8 @@ mpt_action(struct cam_sim *sim, union ccb *ccb)
/* XXX Unless we implement the new 32byte message type */
if (ccb->csio.cdb_len >
sizeof (((PTR_MSG_SCSI_IO_REQUEST)0)->CDB)) {
- ccb->ccb_h.status = CAM_REQ_INVALID;
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_REQ_INVALID);
xpt_done(ccb);
return;
}
@@ -1405,17 +3216,35 @@ mpt_action(struct cam_sim *sim, union ccb *ccb)
* will fall back to a hard reset should a bus
* reset attempt fail.
*/
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
mpt_set_ccb_status(ccb, CAM_REQ_CMP);
xpt_done(ccb);
break;
case XPT_ABORT:
- /*
- * XXX: Need to implement
- */
- ccb->ccb_h.status = CAM_UA_ABORT;
+ {
+ union ccb *accb = ccb->cab.abort_ccb;
+ CAMLOCK_2_MPTLOCK(mpt);
+ switch (accb->ccb_h.func_code) {
+ case XPT_ACCEPT_TARGET_IO:
+ case XPT_IMMED_NOTIFY:
+ ccb->ccb_h.status = mpt_abort_target_ccb(mpt, ccb);
+ break;
+ case XPT_CONT_TARGET_IO:
+ mpt_prt(mpt, "cannot abort active CTIOs yet\n");
+ ccb->ccb_h.status = CAM_UA_ABORT;
+ break;
+ case XPT_SCSI_IO:
+ ccb->ccb_h.status = CAM_UA_ABORT;
+ break;
+ default:
+ ccb->ccb_h.status = CAM_REQ_INVALID;
+ break;
+ }
+ MPTLOCK_2_CAMLOCK(mpt);
xpt_done(ccb);
break;
+ }
#ifdef CAM_NEW_TRAN_CODE
#define IS_CURRENT_SETTINGS(c) (c->type == CTS_TYPE_CURRENT_SETTINGS)
@@ -1440,7 +3269,8 @@ mpt_action(struct cam_sim *sim, union ccb *ccb)
cts = &ccb->cts;
if (!IS_CURRENT_SETTINGS(cts)) {
mpt_prt(mpt, "Attempt to set User settings\n");
- ccb->ccb_h.status = CAM_REQ_INVALID;
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_REQ_INVALID);
xpt_done(ccb);
break;
}
@@ -1521,8 +3351,10 @@ mpt_action(struct cam_sim *sim, union ccb *ccb)
}
if (dval & DP_WIDTH) {
if (mpt_setwidth(mpt, tgt, dval & DP_WIDE)) {
-mpt_prt(mpt, "Set width Failed!\n");
- ccb->ccb_h.status = CAM_REQ_CMP_ERR;
+ mpt_prt(mpt, "Set width Failed!\n");
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb,
+ CAM_REQ_CMP_ERR);
MPTLOCK_2_CAMLOCK(mpt);
xpt_done(ccb);
break;
@@ -1530,8 +3362,10 @@ mpt_prt(mpt, "Set width Failed!\n");
}
if (dval & DP_SYNC) {
if (mpt_setsync(mpt, tgt, period, offset)) {
-mpt_prt(mpt, "Set sync Failed!\n");
- ccb->ccb_h.status = CAM_REQ_CMP_ERR;
+ mpt_prt(mpt, "Set sync Failed!\n");
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb,
+ CAM_REQ_CMP_ERR);
MPTLOCK_2_CAMLOCK(mpt);
xpt_done(ccb);
break;
@@ -1542,7 +3376,8 @@ mpt_prt(mpt, "Set sync Failed!\n");
"SET tgt %d flags %x period %x off %x\n",
tgt, dval, period, offset);
}
- ccb->ccb_h.status = CAM_REQ_CMP;
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_REQ_CMP);
xpt_done(ccb);
break;
@@ -1722,7 +3557,8 @@ mpt_prt(mpt, "Set sync Failed!\n");
? "ACTIVE" : "NVRAM",
tgt, dval, pval, oval);
}
- ccb->ccb_h.status = CAM_REQ_CMP;
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_REQ_CMP);
xpt_done(ccb);
break;
@@ -1732,12 +3568,13 @@ mpt_prt(mpt, "Set sync Failed!\n");
ccg = &ccb->ccg;
if (ccg->block_size == 0) {
- ccb->ccb_h.status = CAM_REQ_INVALID;
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ mpt_set_ccb_status(ccb, CAM_REQ_INVALID);
xpt_done(ccb);
break;
}
-
mpt_calc_geometry(ccg, /*extended*/1);
+ KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__));
xpt_done(ccb);
break;
}
@@ -1766,16 +3603,16 @@ mpt_prt(mpt, "Set sync Failed!\n");
PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16;
}
} else if (mpt->is_fc) {
-/* XXX SHOULD BE BASED UPON IOC FACTS XXX */
+ /* XXX SHOULD BE BASED UPON IOC FACTS XXX XXX */
cpi->max_target = 255;
cpi->hba_misc = PIM_NOBUSRESET;
- cpi->initiator_id = cpi->max_target + 1;
+ cpi->initiator_id = mpt->mpt_ini_id;
cpi->base_transfer_speed = 100000;
cpi->hba_inquiry = PI_TAG_ABLE;
} else if (mpt->is_sas) {
cpi->max_target = 63; /* XXX */
cpi->hba_misc = PIM_NOBUSRESET;
- cpi->initiator_id = cpi->max_target;
+ cpi->initiator_id = mpt->mpt_ini_id;
cpi->base_transfer_speed = 300000;
cpi->hba_inquiry = PI_TAG_ABLE;
} else {
@@ -1786,6 +3623,15 @@ mpt_prt(mpt, "Set sync Failed!\n");
cpi->max_target = 15;
}
+ if ((mpt->role & MPT_ROLE_INITIATOR) == 0) {
+ cpi->hba_misc |= PIM_NOINITIATOR;
+ }
+ if ((mpt->role & MPT_ROLE_TARGET) != 0) {
+ cpi->target_sprt =
+ PIT_PROCESSOR | PIT_DISCONNECT | PIT_TERM_IO;
+ } else {
+ cpi->target_sprt = 0;
+ }
strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
strncpy(cpi->hba_vid, "LSI", HBA_IDLEN);
strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
@@ -1794,6 +3640,73 @@ mpt_prt(mpt, "Set sync Failed!\n");
xpt_done(ccb);
break;
}
+ case XPT_EN_LUN: /* Enable LUN as a target */
+ {
+ int result;
+
+ CAMLOCK_2_MPTLOCK(mpt);
+ if (ccb->cel.enable)
+ result = mpt_enable_lun(mpt,
+ ccb->ccb_h.target_id, ccb->ccb_h.target_lun);
+ else
+ result = mpt_disable_lun(mpt,
+ ccb->ccb_h.target_id, ccb->ccb_h.target_lun);
+ MPTLOCK_2_CAMLOCK(mpt);
+ if (result == 0) {
+ mpt_set_ccb_status(ccb, CAM_REQ_CMP);
+ } else {
+ mpt_set_ccb_status(ccb, CAM_REQ_CMP_ERR);
+ }
+ xpt_done(ccb);
+ break;
+ }
+ case XPT_NOTIFY_ACK: /* recycle notify ack */
+ case XPT_IMMED_NOTIFY: /* Add Immediate Notify Resource */
+ case XPT_ACCEPT_TARGET_IO: /* Add Accept Target IO Resource */
+ {
+ tgt_resource_t *trtp;
+ lun_id_t lun = ccb->ccb_h.target_lun;
+ ccb->ccb_h.sim_priv.entries[0].field = 0;
+ ccb->ccb_h.sim_priv.entries[1].ptr = mpt;
+ ccb->ccb_h.flags = 0;
+
+ if (lun == CAM_LUN_WILDCARD) {
+ if (ccb->ccb_h.target_id != CAM_TARGET_WILDCARD) {
+ mpt_set_ccb_status(ccb, CAM_REQ_INVALID);
+ xpt_done(ccb);
+ break;
+ }
+ trtp = &mpt->trt_wildcard;
+ } else if (lun >= MPT_MAX_LUNS) {
+ mpt_set_ccb_status(ccb, CAM_REQ_INVALID);
+ xpt_done(ccb);
+ break;
+ } else {
+ trtp = &mpt->trt[lun];
+ }
+ CAMLOCK_2_MPTLOCK(mpt);
+ if (ccb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) {
+ mpt_lprt(mpt, MPT_PRT_DEBUG1,
+ "Put FREE ATIO %p lun %d\n", ccb, lun);
+ STAILQ_INSERT_TAIL(&trtp->atios, &ccb->ccb_h,
+ sim_links.stqe);
+ } else if (ccb->ccb_h.func_code == XPT_IMMED_NOTIFY) {
+ mpt_lprt(mpt, MPT_PRT_DEBUG1,
+ "Put FREE INOT lun %d\n", lun);
+ STAILQ_INSERT_TAIL(&trtp->inots, &ccb->ccb_h,
+ sim_links.stqe);
+ } else {
+ mpt_lprt(mpt, MPT_PRT_ALWAYS, "Got Notify ACK\n");
+ }
+ mpt_set_ccb_status(ccb, CAM_REQ_INPROG);
+ MPTLOCK_2_CAMLOCK(mpt);
+ break;
+ }
+ case XPT_CONT_TARGET_IO:
+ CAMLOCK_2_MPTLOCK(mpt);
+ mpt_target_start_io(mpt, ccb);
+ MPTLOCK_2_CAMLOCK(mpt);
+ break;
default:
ccb->ccb_h.status = CAM_REQ_INVALID;
xpt_done(ccb);
@@ -1892,6 +3805,10 @@ mpt_calc_geometry(struct ccb_calc_geometry *ccg, int extended)
uint32_t size_mb;
uint32_t secs_per_cylinder;
+ if (ccg->block_size == 0) {
+ ccg->ccb_h.status = CAM_REQ_INVALID;
+ return;
+ }
size_mb = ccg->volume_size / ((1024L * 1024L) / ccg->block_size);
if (size_mb > 1024 && extended) {
ccg->heads = 255;
@@ -1918,16 +3835,10 @@ mpt_spawn_recovery_thread(struct mpt_softc *mpt)
return (error);
}
-/*
- * Lock is not held on entry.
- */
static void
mpt_terminate_recovery_thread(struct mpt_softc *mpt)
{
-
- MPT_LOCK(mpt);
if (mpt->recovery_thread == NULL) {
- MPT_UNLOCK(mpt);
return;
}
mpt->shutdwn_recovery = 1;
@@ -1937,7 +3848,6 @@ mpt_terminate_recovery_thread(struct mpt_softc *mpt)
* for this interlock just for added safety.
*/
mpt_sleep(mpt, &mpt->recovery_thread, PUSER, "thtrm", 0);
- MPT_UNLOCK(mpt);
}
static void
@@ -1956,12 +3866,10 @@ mpt_recovery_thread(void *arg)
&& mpt->shutdwn_recovery == 0)
mpt_sleep(mpt, mpt, PUSER, "idle", 0);
- if (mpt->shutdwn_recovery != 0)
+ if (mpt->shutdwn_recovery != 0) {
break;
-
- MPT_UNLOCK(mpt);
+ }
mpt_recover_commands(mpt);
- MPT_LOCK(mpt);
}
mpt->recovery_thread = NULL;
wakeup(&mpt->recovery_thread);
@@ -2016,12 +3924,443 @@ mpt_scsi_send_tmf(struct mpt_softc *mpt, u_int type,
if (mpt->verbose > MPT_PRT_DEBUG)
mpt_print_request(tmf_req);
- error = mpt_send_handshake_cmd(mpt, sizeof(*tmf_req), tmf_req);
- if (error != 0)
+ if (mpt->is_fc || mpt->is_sas) {
+ mpt_send_cmd(mpt, mpt->tmf_req);
+ error = MPT_OK;
+ } else {
+ error = mpt_send_handshake_cmd(mpt, sizeof(*tmf_req), tmf_req);
+ }
+ if (error != MPT_OK) {
mpt_reset(mpt, /*reinit*/TRUE);
+ }
return (error);
}
+static void
+mpt_fc_add_els(struct mpt_softc *mpt, request_t *req)
+{
+ MSG_LINK_SERVICE_BUFFER_POST_REQUEST *fc;
+ PTR_SGE_TRANSACTION32 tep;
+ PTR_SGE_SIMPLE32 se;
+ bus_addr_t paddr;
+
+ paddr = req->req_pbuf;
+ paddr += MPT_RQSL(mpt);
+
+ fc = req->req_vbuf;
+ memset(fc, 0, MPT_REQUEST_AREA);
+ fc->BufferCount = 1;
+ fc->Function = MPI_FUNCTION_FC_LINK_SRVC_BUF_POST;
+ fc->MsgContext = htole32(req->index | fc_els_handler_id);
+
+ /*
+ * Okay, set up ELS buffer pointers. ELS buffer pointers
+ * consist of a TE SGL element (with details length of zero)
+ * followe by a SIMPLE SGL element which holds the address
+ * of the buffer.
+ */
+
+ tep = (PTR_SGE_TRANSACTION32) &fc->SGL;
+
+ tep->ContextSize = 4;
+ tep->Flags = 0;
+ tep->TransactionContext[0] = htole32(req->index | fc_els_handler_id);
+
+ se = (PTR_SGE_SIMPLE32) &tep->TransactionDetails[0];
+ se->FlagsLength =
+ MPI_SGE_FLAGS_HOST_TO_IOC |
+ MPI_SGE_FLAGS_SIMPLE_ELEMENT |
+ MPI_SGE_FLAGS_LAST_ELEMENT |
+ MPI_SGE_FLAGS_END_OF_LIST |
+ MPI_SGE_FLAGS_END_OF_BUFFER;
+ se->FlagsLength <<= MPI_SGE_FLAGS_SHIFT;
+ se->FlagsLength |= (MPT_NRFM(mpt) - MPT_RQSL(mpt));
+ se->Address = (uint32_t) paddr;
+ mpt_check_doorbell(mpt);
+ mpt_send_cmd(mpt, req);
+}
+
+static void
+mpt_post_target_command(struct mpt_softc *mpt, request_t *req, int ioindex)
+{
+ PTR_MSG_TARGET_CMD_BUFFER_POST_REQUEST fc;
+ PTR_CMD_BUFFER_DESCRIPTOR cb;
+ bus_addr_t paddr;
+
+ paddr = req->req_pbuf;
+ paddr += MPT_RQSL(mpt);
+ memset(req->req_vbuf, 0, MPT_REQUEST_AREA);
+
+ fc = req->req_vbuf;
+ fc->BufferCount = 1;
+ fc->Function = MPI_FUNCTION_TARGET_CMD_BUFFER_POST;
+ fc->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id);
+
+ cb = &fc->Buffer[0];
+ cb->IoIndex = htole16(ioindex);
+ cb->u.PhysicalAddress32 = (U32) paddr;
+
+ mpt_check_doorbell(mpt);
+ mpt_send_cmd(mpt, req);
+}
+
+static void
+mpt_add_target_commands(struct mpt_softc *mpt)
+{
+ int i, max;
+
+ if (mpt->tgt_cmd_ptrs) {
+ return;
+ }
+
+ max = MPT_MAX_REQUESTS(mpt) >> 1;
+ if (max > mpt->mpt_max_tgtcmds) {
+ max = mpt->mpt_max_tgtcmds;
+ }
+ mpt->tgt_cmd_ptrs =
+ malloc(max * sizeof (void *), M_DEVBUF, M_NOWAIT | M_ZERO);
+ if (mpt->tgt_cmd_ptrs == NULL) {
+ mpt_prt(mpt, "could not allocate cmdptrs\n");
+ return;
+ }
+ mpt->tgt_cmds_allocated = max;
+
+ for (i = 0; i < max; i++) {
+ request_t *req;
+
+ req = mpt_get_request(mpt, FALSE);
+ if (req == NULL) {
+ break;
+ }
+ mpt->tgt_cmd_ptrs[i] = req;
+ mpt_post_target_command(mpt, req, i);
+ }
+
+ if (i == 0) {
+ mpt_lprt(mpt, MPT_PRT_ERROR, "could not add any target bufs\n");
+ free(mpt->tgt_cmd_ptrs, M_DEVBUF);
+ mpt->tgt_cmd_ptrs = NULL;
+ mpt->tgt_cmds_allocated = 0;
+ } else if (i < max) {
+ mpt_lprt(mpt, MPT_PRT_WARN, "added %d of %d target bufs\n",
+ i, max);
+ }
+}
+
+static int
+mpt_enable_lun(struct mpt_softc *mpt, target_id_t tgt, lun_id_t lun)
+{
+ if (tgt == CAM_TARGET_WILDCARD && lun == CAM_LUN_WILDCARD) {
+ mpt->twildcard = 1;
+ } else if (lun >= MPT_MAX_LUNS) {
+ return (EINVAL);
+ } else if (tgt != CAM_TARGET_WILDCARD && tgt != 0) {
+ return (EINVAL);
+ }
+ if (mpt->tenabled == 0) {
+#if 0
+ if (mpt->is_fc) {
+ (void) mpt_fc_reset_link(mpt, 0);
+ }
+#endif
+ mpt->tenabled = 1;
+ }
+ if (lun == CAM_LUN_WILDCARD) {
+ mpt->trt_wildcard.enabled = 1;
+ } else {
+ mpt->trt[lun].enabled = 1;
+ }
+ return (0);
+}
+
+static int
+mpt_disable_lun(struct mpt_softc *mpt, target_id_t tgt, lun_id_t lun)
+{
+ int i;
+ if (tgt == CAM_TARGET_WILDCARD && lun == CAM_LUN_WILDCARD) {
+ mpt->twildcard = 0;
+ } else if (lun >= MPT_MAX_LUNS) {
+ return (EINVAL);
+ } else if (tgt != CAM_TARGET_WILDCARD && tgt != 0) {
+ return (EINVAL);
+ }
+ if (lun == CAM_LUN_WILDCARD) {
+ mpt->trt_wildcard.enabled = 0;
+ } else {
+ mpt->trt[lun].enabled = 0;
+ }
+ for (i = 0; i < MPT_MAX_LUNS; i++) {
+ if (mpt->trt[lun].enabled) {
+ break;
+ }
+ }
+ if (i == MPT_MAX_LUNS && mpt->twildcard == 0) {
+ mpt->tenabled = 0;
+#if 0
+ if (mpt->is_fc) {
+ (void) mpt_fc_reset_link(mpt, 0);
+ }
+#endif
+ }
+ return (0);
+}
+
+/*
+ * Called with MPT lock held
+ */
+static void
+mpt_target_start_io(struct mpt_softc *mpt, union ccb *ccb)
+{
+ struct ccb_scsiio *csio = &ccb->csio;
+ request_t *cmd_req = MPT_TAG_2_REQ(mpt, csio->tag_id);
+ mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, cmd_req);
+
+
+ if (tgt->state != TGT_STATE_IN_CAM) {
+ mpt_prt(mpt, "tag 0x%08x in state %x when starting I/O\n",
+ csio->tag_id, tgt->state);
+ mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ);
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ return;
+ }
+
+ if (csio->dxfer_len) {
+ bus_dmamap_callback_t *cb;
+ PTR_MSG_TARGET_ASSIST_REQUEST ta;
+ request_t *req;
+
+ KASSERT((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE,
+ ("dxfer_len %u but direction is NONE\n", csio->dxfer_len));
+
+ req = mpt_get_request(mpt, FALSE);
+ if (req == NULL) {
+ mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ);
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ return;
+ }
+
+ ccb->ccb_h.status = CAM_SIM_QUEUED | CAM_REQ_INPROG;
+ if (sizeof (bus_addr_t) > 4) {
+ cb = mpt_execute_req_a64;
+ } else {
+ cb = mpt_execute_req;
+ }
+
+ req->ccb = ccb;
+ ccb->ccb_h.ccb_req_ptr = req;
+
+ /*
+ * Record the currently active ccb and the
+ * request for it in our target state area.
+ */
+ tgt->ccb = ccb;
+ tgt->req = req;
+
+ memset(req->req_vbuf, 0, MPT_RQSL(mpt));
+ ta = req->req_vbuf;
+
+ if (mpt->is_fc) {
+ ;
+ } else if (mpt->is_sas == 0) {
+ PTR_MPI_TARGET_SSP_CMD_BUFFER ssp =
+ cmd_req->req_vbuf;
+ ta->QueueTag = ssp->InitiatorTag;
+ } else {
+ PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER sp =
+ cmd_req->req_vbuf;
+ ta->QueueTag = sp->Tag;
+ }
+ ta->Function = MPI_FUNCTION_TARGET_ASSIST;
+ ta->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id);
+ ta->ReplyWord = htole32(tgt->reply_desc);
+ if (csio->ccb_h.target_lun > 256) {
+ ta->LUN[0] =
+ 0x40 | ((csio->ccb_h.target_lun >> 8) & 0x3f);
+ ta->LUN[1] = csio->ccb_h.target_lun & 0xff;
+ } else {
+ ta->LUN[1] = csio->ccb_h.target_lun;
+ }
+
+ ta->RelativeOffset = tgt->bytes_xfered;
+ ta->DataLength = ccb->csio.dxfer_len;
+ if (ta->DataLength > tgt->resid) {
+ ta->DataLength = tgt->resid;
+ }
+
+ /*
+ * XXX Should be done after data transfer completes?
+ */
+ tgt->resid -= csio->dxfer_len;
+ tgt->bytes_xfered += csio->dxfer_len;
+
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
+ ta->TargetAssistFlags |=
+ TARGET_ASSIST_FLAGS_DATA_DIRECTION;
+ }
+
+#ifdef WE_TRUST_AUTO_GOOD_STATUS
+ if ((ccb->ccb_h.flags & CAM_SEND_STATUS) &&
+ csio->scsi_status == SCSI_STATUS_OK && tgt->resid == 0) {
+ ta->TargetAssistFlags |=
+ TARGET_ASSIST_FLAGS_AUTO_STATUS;
+ }
+#endif
+ tgt->state = TGT_STATE_SETTING_UP_FOR_DATA;
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "DATA_CCB %p tag %x %u bytes %u resid flg %x req %p:%u "
+ "nxtstate=%d\n", csio, csio->tag_id, csio->dxfer_len,
+ tgt->resid, ccb->ccb_h.flags, req, req->serno, tgt->state);
+
+ MPTLOCK_2_CAMLOCK(mpt);
+ if ((ccb->ccb_h.flags & CAM_SCATTER_VALID) == 0) {
+ if ((ccb->ccb_h.flags & CAM_DATA_PHYS) == 0) {
+ int error;
+ int s = splsoftvm();
+ error = bus_dmamap_load(mpt->buffer_dmat,
+ req->dmap, csio->data_ptr, csio->dxfer_len,
+ cb, req, 0);
+ splx(s);
+ if (error == EINPROGRESS) {
+ xpt_freeze_simq(mpt->sim, 1);
+ ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
+ }
+ } else {
+ /*
+ * We have been given a pointer to single
+ * physical buffer.
+ */
+ struct bus_dma_segment seg;
+ seg.ds_addr = (bus_addr_t)
+ (vm_offset_t)csio->data_ptr;
+ seg.ds_len = csio->dxfer_len;
+ (*cb)(req, &seg, 1, 0);
+ }
+ } else {
+ /*
+ * We have been given a list of addresses.
+ * This case could be easily supported but they are not
+ * currently generated by the CAM subsystem so there
+ * is no point in wasting the time right now.
+ */
+ struct bus_dma_segment *sgs;
+ if ((ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0) {
+ (*cb)(req, NULL, 0, EFAULT);
+ } else {
+ /* Just use the segments provided */
+ sgs = (struct bus_dma_segment *)csio->data_ptr;
+ (*cb)(req, sgs, csio->sglist_cnt, 0);
+ }
+ }
+ CAMLOCK_2_MPTLOCK(mpt);
+ } else {
+ uint8_t *sp = NULL, sense[MPT_SENSE_SIZE];
+
+ /*
+ * XXX: I don't know why this seems to happen, but
+ * XXX: completing the CCB seems to make things happy.
+ * XXX: This seems to happen if the initiator requests
+ * XXX: enough data that we have to do multiple CTIOs.
+ */
+ if ((ccb->ccb_h.flags & CAM_SEND_STATUS) == 0) {
+ mpt_lprt(mpt, MPT_PRT_DEBUG,
+ "Meaningless STATUS CCB (%p): flags %x status %x "
+ "resid %d bytes_xfered %u\n", ccb, ccb->ccb_h.flags,
+ ccb->ccb_h.status, tgt->resid, tgt->bytes_xfered);
+ mpt_set_ccb_status(ccb, CAM_REQ_CMP);
+ ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
+ tgt->flags |= BOGUS_JO;
+ MPTLOCK_2_CAMLOCK(mpt);
+ xpt_done(ccb);
+ CAMLOCK_2_MPTLOCK(mpt);
+ return;
+ }
+ if (ccb->ccb_h.flags & CAM_SEND_SENSE) {
+ sp = sense;
+ memcpy(sp, &csio->sense_data,
+ min(csio->sense_len, MPT_SENSE_SIZE));
+ }
+ mpt_scsi_tgt_status(mpt, ccb, cmd_req, csio->scsi_status, sp);
+ }
+}
+
+/*
+ * Abort queued up CCBs
+ */
+static cam_status
+mpt_abort_target_ccb(struct mpt_softc *mpt, union ccb *ccb)
+{
+ struct mpt_hdr_stailq *lp;
+ struct ccb_hdr *srch;
+ int found = 0;
+ union ccb *accb = ccb->cab.abort_ccb;
+ tgt_resource_t *trtp;
+
+ mpt_lprt(mpt, MPT_PRT_DEBUG, "aborting ccb %p\n", accb);
+
+ if (ccb->ccb_h.target_lun == CAM_LUN_WILDCARD) {
+ trtp = &mpt->trt_wildcard;
+ } else {
+ trtp = &mpt->trt[ccb->ccb_h.target_lun];
+ }
+
+ if (accb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) {
+ lp = &trtp->atios;
+ } else if (accb->ccb_h.func_code == XPT_IMMED_NOTIFY) {
+ lp = &trtp->inots;
+ } else {
+ return (CAM_REQ_INVALID);
+ }
+
+ STAILQ_FOREACH(srch, lp, sim_links.stqe) {
+ if (srch == &accb->ccb_h) {
+ found = 1;
+ STAILQ_REMOVE(lp, srch, ccb_hdr, sim_links.stqe);
+ break;
+ }
+ }
+ if (found) {
+ accb->ccb_h.status = CAM_REQ_ABORTED;
+ xpt_done(accb);
+ return (CAM_REQ_CMP);
+ }
+ mpt_prt(mpt, "mpt_abort_tgt_ccb: CCB %p not found\n", ccb);
+ return (CAM_PATH_INVALID);
+}
+
+/*
+ * Ask the MPT to abort the current target command
+ */
+static cam_status
+mpt_abort_target_cmd(struct mpt_softc *mpt, request_t *cmd_req)
+{
+ int error;
+ request_t *req;
+ PTR_MSG_TARGET_MODE_ABORT abtp;
+
+ req = mpt_get_request(mpt, FALSE);
+ if (req == NULL) {
+ return (CAM_RESRC_UNAVAIL);
+ }
+ abtp = req->req_vbuf;
+ memset(abtp, 0, sizeof (*abtp));
+
+ abtp->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id);
+ abtp->AbortType = TARGET_MODE_ABORT_TYPE_EXACT_IO;
+ abtp->Function = MPI_FUNCTION_TARGET_MODE_ABORT;
+ abtp->ReplyWord = htole32(MPT_TGT_STATE(mpt, cmd_req)->reply_desc);
+ if (mpt->is_fc || mpt->is_sas) {
+ mpt_send_cmd(mpt, req);
+ } else {
+ error = mpt_send_handshake_cmd(mpt, sizeof(*req), req);
+ }
+ return (CAM_REQ_INPROG);
+}
+
/*
* When a command times out, it is placed on the requeust_timeout_list
* and we wake our recovery thread. The MPT-Fusion architecture supports
@@ -2037,13 +4376,11 @@ mpt_recover_commands(struct mpt_softc *mpt)
union ccb *ccb;
int error;
- MPT_LOCK(mpt);
if (TAILQ_EMPTY(&mpt->request_timeout_list) != 0) {
/*
* No work to do- leave.
*/
mpt_prt(mpt, "mpt_recover_commands: no requests.\n");
- MPT_UNLOCK(mpt);
return;
}
@@ -2064,7 +4401,6 @@ mpt_recover_commands(struct mpt_softc *mpt)
mpt_prt(mpt, "Timedout requests already complete. "
"Interrupts may not be functioning.\n");
mpt_enable_ints(mpt);
- MPT_UNLOCK(mpt);
return;
}
@@ -2075,20 +4411,18 @@ mpt_recover_commands(struct mpt_softc *mpt)
*/
while ((req = TAILQ_FIRST(&mpt->request_timeout_list)) != NULL) {
u_int status;
- u_int32_t serno = req->serno;
- mpt_prt(mpt, "Attempting to Abort Req %p:%u\n", req, serno);
+ mpt_prt(mpt, "Attempting to Abort Req %p\n", req);
+
ccb = req->ccb;
mpt_set_ccb_status(ccb, CAM_CMD_TIMEOUT);
error = mpt_scsi_send_tmf(mpt,
MPI_SCSITASKMGMT_TASKTYPE_ABORT_TASK,
- /*MsgFlags*/0, mpt->bus, ccb->ccb_h.target_id,
+ /*MsgFlags*/0, /*Bus*/0, ccb->ccb_h.target_id,
ccb->ccb_h.target_lun,
htole32(req->index | scsi_io_handler_id), /*sleep_ok*/TRUE);
if (error != 0) {
- mpt_prt(mpt, "Abort Req %p:%u failed to start TMF\n",
- req, serno);
/*
* mpt_scsi_send_tmf hard resets on failure, so no
* need to do so here. Our queue should be emptied
@@ -2107,8 +4441,8 @@ mpt_recover_commands(struct mpt_softc *mpt)
* If we've errored out and the transaction is still
* pending, reset the controller.
*/
- mpt_prt(mpt, "Abort Req %p:%d timed-out. "
- "Resetting controller\n", req, serno);
+ mpt_prt(mpt, "mpt_recover_commands: Abort timed-out. "
+ "Resetting controller\n");
mpt_reset(mpt, /*reinit*/TRUE);
continue;
}
@@ -2122,9 +4456,8 @@ mpt_recover_commands(struct mpt_softc *mpt)
continue;
mpt_lprt(mpt, MPT_PRT_DEBUG,
- "Abort Req %p: %u Failed "
- "with status 0x%x\n. Resetting bus.",
- req, serno, status);
+ "mpt_recover_commands: Abort Failed "
+ "with status 0x%x\n. Resetting bus", status);
/*
* If the abort attempt fails for any reason, reset the bus.
@@ -2133,6 +4466,4 @@ mpt_recover_commands(struct mpt_softc *mpt)
*/
mpt_bus_reset(mpt, /*sleep_ok*/TRUE);
}
-
- MPT_UNLOCK(mpt);
}
diff --git a/sys/dev/mpt/mpt_debug.c b/sys/dev/mpt/mpt_debug.c
index 270840afbc73..b9b6bc19b9a7 100644
--- a/sys/dev/mpt/mpt_debug.c
+++ b/sys/dev/mpt/mpt_debug.c
@@ -68,6 +68,7 @@ __FBSDID("$FreeBSD$");
#include <dev/mpt/mpilib/mpi_ioc.h>
#include <dev/mpt/mpilib/mpi_init.h>
#include <dev/mpt/mpilib/mpi_fc.h>
+#include <dev/mpt/mpilib/mpi_targ.h>
#include <cam/scsi/scsi_all.h>
@@ -574,6 +575,35 @@ mpt_print_scsi_tmf_request(MSG_SCSI_TASK_MGMT *msg)
printf("\tTaskMsgContext 0x%08x\n", msg->TaskMsgContext);
}
+
+static void
+mpt_print_scsi_target_assist_request(PTR_MSG_TARGET_ASSIST_REQUEST msg)
+{
+ mpt_print_request_hdr((MSG_REQUEST_HEADER *)msg);
+ printf("\tStatusCode 0x%02x\n", msg->StatusCode);
+ printf("\tTargetAssist 0x%02x\n", msg->TargetAssistFlags);
+ printf("\tQueueTag 0x%04x\n", msg->QueueTag);
+ printf("\tReplyWord 0x%08x\n", msg->ReplyWord);
+ printf("\tLun 0x%02x\n", msg->LUN[1]);
+ printf("\tRelativeOff 0x%08x\n", msg->RelativeOffset);
+ printf("\tDataLength 0x%08x\n", msg->DataLength);
+ mpt_dump_sgl(msg->SGL, 0);
+}
+
+static void
+mpt_print_scsi_target_status_send_request(MSG_TARGET_STATUS_SEND_REQUEST *msg)
+{
+ SGE_IO_UNION x;
+ mpt_print_request_hdr((MSG_REQUEST_HEADER *)msg);
+ printf("\tStatusCode 0x%02x\n", msg->StatusCode);
+ printf("\tStatusFlags 0x%02x\n", msg->StatusFlags);
+ printf("\tQueueTag 0x%04x\n", msg->QueueTag);
+ printf("\tReplyWord 0x%08x\n", msg->ReplyWord);
+ printf("\tLun 0x%02x\n", msg->LUN[1]);
+ x.u.Simple = msg->StatusDataSGE;
+ mpt_dump_sgl(&x, 0);
+}
+
void
mpt_print_request(void *vreq)
{
@@ -585,6 +615,15 @@ mpt_print_request(void *vreq)
break;
case MPI_FUNCTION_SCSI_TASK_MGMT:
mpt_print_scsi_tmf_request((MSG_SCSI_TASK_MGMT *)req);
+ break;
+ case MPI_FUNCTION_TARGET_ASSIST:
+ mpt_print_scsi_target_assist_request(
+ (PTR_MSG_TARGET_ASSIST_REQUEST)req);
+ break;
+ case MPI_FUNCTION_TARGET_STATUS_SEND:
+ mpt_print_scsi_target_status_send_request(
+ (MSG_TARGET_STATUS_SEND_REQUEST *)req);
+ break;
default:
mpt_print_request_hdr(req);
break;
@@ -754,6 +793,31 @@ mpt_dump_sgl(SGE_IO_UNION *su, int offset)
} while ((flags & MPI_SGE_FLAGS_END_OF_LIST) == 0 && nxtaddr < lim);
}
+#if __FreeBSD_version < 500000
+void
+mpt_lprt(struct mpt_softc *mpt, int level, const char *fmt, ...)
+{
+ va_list ap;
+ if (level <= mpt->verbose) {
+ printf("%s: ", device_get_nameunit(mpt->dev));
+ va_start(ap, fmt);
+ vprintf(fmt, ap);
+ va_end(ap);
+ }
+}
+
+void
+mpt_lprtc(struct mpt_softc *mpt, int level, const char *fmt, ...)
+{
+ va_list ap;
+ if (level <= mpt->verbose) {
+ va_start(ap, fmt);
+ vprintf(fmt, ap);
+ va_end(ap);
+ }
+}
+#endif
+
void
mpt_prt(struct mpt_softc *mpt, const char *fmt, ...)
{
diff --git a/sys/dev/mpt/mpt_pci.c b/sys/dev/mpt/mpt_pci.c
index a995618741ff..e29b421642a8 100644
--- a/sys/dev/mpt/mpt_pci.c
+++ b/sys/dev/mpt/mpt_pci.c
@@ -242,7 +242,7 @@ mpt_pci_probe(device_t dev)
return (0);
}
-#ifdef RELENG_4
+#if __FreeBSD_version < 500000
static void
mpt_set_options(struct mpt_softc *mpt)
{
@@ -261,7 +261,24 @@ mpt_set_options(struct mpt_softc *mpt)
mpt->verbose = MPT_PRT_DEBUG;
}
}
-
+ bitmap = 0;
+ if (getenv_int("mpt_target", &bitmap)) {
+ if (bitmap & (1 << mpt->unit)) {
+ mpt->role = MPT_ROLE_TARGET;
+ }
+ }
+ bitmap = 0;
+ if (getenv_int("mpt_none", &bitmap)) {
+ if (bitmap & (1 << mpt->unit)) {
+ mpt->role = MPT_ROLE_NONE;
+ }
+ }
+ bitmap = 0;
+ if (getenv_int("mpt_initiator", &bitmap)) {
+ if (bitmap & (1 << mpt->unit)) {
+ mpt->role = MPT_ROLE_INITIATOR;
+ }
+ }
}
#else
static void
@@ -279,6 +296,12 @@ mpt_set_options(struct mpt_softc *mpt)
device_get_unit(mpt->dev), "debug", &tval) == 0 && tval != 0) {
mpt->verbose += tval;
}
+ tval = 0;
+ if (resource_int_value(device_get_name(mpt->dev),
+ device_get_unit(mpt->dev), "role", &tval) == 0 && tval != 0 &&
+ tval <= 3) {
+ mpt->role = tval;
+ }
}
#endif
@@ -354,6 +377,7 @@ mpt_pci_attach(device_t dev)
mpt->raid_mwce_setting = MPT_RAID_MWCE_DEFAULT;
mpt->raid_queue_depth = MPT_RAID_QUEUE_DEPTH_DEFAULT;
mpt->verbose = MPT_PRT_NONE;
+ mpt->role = MPT_ROLE_DEFAULT;
mpt_set_options(mpt);
if (mpt->verbose == MPT_PRT_NONE) {
mpt->verbose = MPT_PRT_WARN;
@@ -474,16 +498,44 @@ mpt_pci_attach(device_t dev)
*/
pci_disable_io(dev, SYS_RES_IOPORT);
+ switch (mpt->role) {
+ case MPT_ROLE_TARGET:
+ break;
+ case MPT_ROLE_INITIATOR:
+ break;
+ case MPT_ROLE_TARGET|MPT_ROLE_INITIATOR:
+ mpt->disabled = 1;
+ mpt_prt(mpt, "dual roles unsupported\n");
+ goto bad;
+ case MPT_ROLE_NONE:
+ device_printf(dev, "role of NONE same as disable\n");
+ mpt->disabled = 1;
+ goto bad;
+ }
+
/* Initialize the hardware */
if (mpt->disabled == 0) {
-
MPT_LOCK(mpt);
if (mpt_attach(mpt) != 0) {
MPT_UNLOCK(mpt);
goto bad;
}
+ MPT_UNLOCK(mpt);
+ } else {
+ mpt_prt(mpt, "device disabled at user request\n");
+ goto bad;
}
+ mpt->eh = EVENTHANDLER_REGISTER(shutdown_post_sync, mpt_pci_shutdown,
+ dev, SHUTDOWN_PRI_DEFAULT);
+
+ if (mpt->eh == NULL) {
+ mpt_prt(mpt, "shutdown event registration failed\n");
+ MPT_LOCK(mpt);
+ (void) mpt_detach(mpt);
+ MPT_UNLOCK(mpt);
+ goto bad;
+ }
return (0);
bad:
@@ -535,24 +587,23 @@ mpt_pci_detach(device_t dev)
struct mpt_softc *mpt;
mpt = (struct mpt_softc*)device_get_softc(dev);
- mpt_prt(mpt, "mpt_detach\n");
if (mpt) {
+ MPT_LOCK(mpt);
mpt_disable_ints(mpt);
mpt_detach(mpt);
mpt_reset(mpt, /*reinit*/FALSE);
mpt_dma_mem_free(mpt);
mpt_free_bus_resources(mpt);
- if (mpt->raid_volumes != NULL
- && mpt->ioc_page2 != NULL) {
+ if (mpt->raid_volumes != NULL && mpt->ioc_page2 != NULL) {
int i;
-
for (i = 0; i < mpt->ioc_page2->MaxVolumes; i++) {
struct mpt_raid_volume *mpt_vol;
mpt_vol = &mpt->raid_volumes[i];
- if (mpt_vol->config_page)
+ if (mpt_vol->config_page) {
free(mpt_vol->config_page, M_DEVBUF);
+ }
}
}
if (mpt->ioc_page2 != NULL)
@@ -565,6 +616,7 @@ mpt_pci_detach(device_t dev)
free(mpt->raid_disks, M_DEVBUF);
if (mpt->eh != NULL)
EVENTHANDLER_DEREGISTER(shutdown_final, mpt->eh);
+ MPT_UNLOCK(mpt);
}
return(0);
}
@@ -572,7 +624,6 @@ mpt_pci_detach(device_t dev)
/*
* Disable the hardware
- * XXX - Called too early by New Bus!!! ???
*/
static int
mpt_pci_shutdown(device_t dev)
@@ -580,8 +631,13 @@ mpt_pci_shutdown(device_t dev)
struct mpt_softc *mpt;
mpt = (struct mpt_softc *)device_get_softc(dev);
- if (mpt)
- return (mpt_shutdown(mpt));
+ if (mpt) {
+ int r;
+ MPT_LOCK(mpt);
+ r = mpt_shutdown(mpt);
+ MPT_UNLOCK(mpt);
+ return (r);
+ }
return(0);
}
diff --git a/sys/dev/mpt/mpt_raid.c b/sys/dev/mpt/mpt_raid.c
index 447423ba2677..da25979e3eb5 100644
--- a/sys/dev/mpt/mpt_raid.c
+++ b/sys/dev/mpt/mpt_raid.c
@@ -46,6 +46,10 @@ __FBSDID("$FreeBSD$");
#include <cam/cam_sim.h>
#include <cam/cam_xpt_sim.h>
+#if __FreeBSD_version < 500000
+#include <sys/devicestat.h>
+#define GIANT_REQUIRED
+#endif
#include <cam/cam_periph.h>
#include <sys/callout.h>
@@ -241,9 +245,9 @@ mpt_raid_async(void *callback_arg, u_int32_t code,
int
mpt_raid_probe(struct mpt_softc *mpt)
{
- if (mpt->ioc_page2 == NULL
- || mpt->ioc_page2->MaxPhysDisks == 0)
+ if (mpt->ioc_page2 == NULL || mpt->ioc_page2->MaxPhysDisks == 0) {
return (ENODEV);
+ }
return (0);
}
@@ -446,8 +450,10 @@ mpt_raid_event(struct mpt_softc *mpt, request_t *req,
if (mpt_disk != NULL)
mpt_disk_prt(mpt, mpt_disk, "");
else
- mpt_prt(mpt, "Volume(%d:%d:%d: ");
- mpt_prtc(mpt, "ASC 0x%x, ASCQ 0x%x\n",
+ mpt_prt(mpt, "Volume(%d:%d:%d: ",
+ raid_event->VolumeBus, raid_event->VolumeID,
+ raid_event->PhysDiskNum);
+ mpt_prtc(mpt, "ASC 0x%x, ASCQ 0x%x)\n",
raid_event->ASC, raid_event->ASCQ);
}
@@ -460,19 +466,19 @@ mpt_raid_shutdown(struct mpt_softc *mpt)
{
struct mpt_raid_volume *mpt_vol;
- if (mpt->raid_mwce_setting != MPT_RAID_MWCE_REBUILD_ONLY)
+ if (mpt->raid_mwce_setting != MPT_RAID_MWCE_REBUILD_ONLY) {
return;
+ }
mpt->raid_mwce_setting = MPT_RAID_MWCE_OFF;
RAID_VOL_FOREACH(mpt, mpt_vol) {
-
mpt_verify_mwce(mpt, mpt_vol);
}
}
static int
mpt_raid_reply_handler(struct mpt_softc *mpt, request_t *req,
- MSG_DEFAULT_REPLY *reply_frame)
+ uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame)
{
int free_req;
@@ -607,16 +613,11 @@ mpt_spawn_raid_thread(struct mpt_softc *mpt)
return (error);
}
-/*
- * Lock is not held on entry.
- */
static void
mpt_terminate_raid_thread(struct mpt_softc *mpt)
{
- MPT_LOCK(mpt);
if (mpt->raid_thread == NULL) {
- MPT_UNLOCK(mpt);
return;
}
mpt->shutdwn_raid = 1;
@@ -626,7 +627,6 @@ mpt_terminate_raid_thread(struct mpt_softc *mpt)
* for this interlock just for added safety.
*/
mpt_sleep(mpt, &mpt->raid_thread, PUSER, "thtrm", 0);
- MPT_UNLOCK(mpt);
}
static void
@@ -845,10 +845,9 @@ mpt_verify_mwce(struct mpt_softc *mpt, struct mpt_raid_volume *mpt_vol)
*/
switch (mpt->raid_mwce_setting) {
case MPT_RAID_MWCE_REBUILD_ONLY:
- if ((resyncing && mwce)
- || (!resyncing && !mwce))
+ if ((resyncing && mwce) || (!resyncing && !mwce)) {
return;
-
+ }
mpt_vol->flags ^= MPT_RVF_WCE_CHANGED;
if ((mpt_vol->flags & MPT_RVF_WCE_CHANGED) == 0) {
/*
@@ -903,7 +902,6 @@ mpt_verify_mwce(struct mpt_softc *mpt, struct mpt_raid_volume *mpt_vol)
vol_pg->VolumeSettings.Settings ^=
MPI_RAIDVOL0_SETTING_WRITE_CACHING_ENABLE;
}
-
mpt_free_request(mpt, req);
}
@@ -1228,18 +1226,21 @@ mpt_refresh_raid_data(struct mpt_softc *mpt)
int i;
u_int nonopt_volumes;
- if (mpt->ioc_page2 == NULL || mpt->ioc_page3 == NULL)
+ if (mpt->ioc_page2 == NULL || mpt->ioc_page3 == NULL) {
return;
+ }
/*
- * Mark all items as unreferrened by the configuration.
+ * Mark all items as unreferenced by the configuration.
* This allows us to find, report, and discard stale
* entries.
*/
- for (i = 0; i < mpt->ioc_page2->MaxPhysDisks; i++)
+ for (i = 0; i < mpt->ioc_page2->MaxPhysDisks; i++) {
mpt->raid_disks[i].flags &= ~MPT_RDF_REFERENCED;
- for (i = 0; i < mpt->ioc_page2->MaxVolumes; i++)
+ }
+ for (i = 0; i < mpt->ioc_page2->MaxVolumes; i++) {
mpt->raid_volumes[i].flags &= ~MPT_RVF_REFERENCED;
+ }
/*
* Get Physical Disk information.
@@ -1386,9 +1387,15 @@ mpt_refresh_raid_data(struct mpt_softc *mpt)
mpt_vol_prt(mpt, mpt_vol, "%s Priority Re-Sync\n",
prio ? "High" : "Low");
}
+#if __FreeBSD_version >= 500000
mpt_vol_prt(mpt, mpt_vol, "%ju of %ju "
"blocks remaining\n", (uintmax_t)left,
(uintmax_t)total);
+#else
+ mpt_vol_prt(mpt, mpt_vol, "%llu of %llu "
+ "blocks remaining\n", (uint64_t)left,
+ (uint64_t)total);
+#endif
/* Periodically report on sync progress. */
mpt_schedule_raid_refresh(mpt);
@@ -1484,8 +1491,9 @@ mpt_raid_set_vol_resync_rate(struct mpt_softc *mpt, u_int rate)
MPT_LOCK(mpt);
mpt->raid_resync_rate = rate;
RAID_VOL_FOREACH(mpt, mpt_vol) {
- if ((mpt_vol->flags & MPT_RVF_ACTIVE) == 0)
+ if ((mpt_vol->flags & MPT_RVF_ACTIVE) == 0) {
continue;
+ }
mpt_verify_resync_rate(mpt, mpt_vol);
}
MPT_UNLOCK(mpt);
@@ -1497,8 +1505,7 @@ mpt_raid_set_vol_queue_depth(struct mpt_softc *mpt, u_int vol_queue_depth)
{
struct mpt_raid_volume *mpt_vol;
- if (vol_queue_depth > 255
- || vol_queue_depth < 1)
+ if (vol_queue_depth > 255 || vol_queue_depth < 1)
return (EINVAL);
MPT_LOCK(mpt);
@@ -1662,6 +1669,7 @@ mpt_raid_sysctl_vol_queue_depth(SYSCTL_HANDLER_ARGS)
static void
mpt_raid_sysctl_attach(struct mpt_softc *mpt)
{
+#if __FreeBSD_version >= 500000
struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(mpt->dev);
struct sysctl_oid *tree = device_get_sysctl_tree(mpt->dev);
@@ -1683,4 +1691,5 @@ mpt_raid_sysctl_attach(struct mpt_softc *mpt)
"nonoptimal_volumes", CTLFLAG_RD,
&mpt->raid_nonopt_volumes, 0,
"number of nonoptimal volumes");
+#endif
}
diff --git a/sys/dev/mpt/mpt_reg.h b/sys/dev/mpt/mpt_reg.h
index 23a34e07b54e..fc6d1f3d598d 100644
--- a/sys/dev/mpt/mpt_reg.h
+++ b/sys/dev/mpt/mpt_reg.h
@@ -125,7 +125,7 @@ enum _MPT_INTR_REQ_BITS {
#define MPT_DB_INTR(v) (((v) & MPT_INTR_DB_READY) != 0)
#define MPT_REPLY_INTR(v) (((v) & MPT_INTR_REPLY_READY) != 0)
-/* Function Maps for INTERRUPT make register */
+/* Function Maps for INTERRUPT mask register */
enum _MPT_INTR_MASK_BITS {
MPT_INTR_REPLY_MASK = 0x00000008,
MPT_INTR_DB_MASK = 0x00000001