File: [local] / sys / dev / ic / isp_openbsd.c (download)
Revision 1.1.1.1 (vendor branch), Tue Mar 4 16:10:40 2008 UTC (16 years, 6 months ago) by nbrk
Branch: OPENBSD_4_2_BASE, MAIN
CVS Tags: jornada-partial-support-wip, HEAD Changes since 1.1: +0 -0 lines
Import of OpenBSD 4.2 release kernel tree with initial code to support
Jornada 720/728, StrongARM 1110-based handheld PC.
At this point kernel roots on NFS and boots into vfs_mountroot() and traps.
What is supported:
- glass console, Jornada framebuffer (jfb) works in 16bpp direct color mode
(needs some palette tweaks for non black/white/blue colors, i think)
- saic, SA11x0 interrupt controller (needs cleanup)
- sacom, SA11x0 UART (supported only as boot console for now)
- SA11x0 GPIO controller fully supported (but can't handle multiple interrupt
handlers on one gpio pin)
- sassp, SSP port on SA11x0 that attaches spibus
- Jornada microcontroller (jmcu) to control kbd, battery, etc throught
the SPI bus (wskbd attaches on jmcu, but not tested)
- tod functions seem work
- initial code for SA-1111 (chip companion) : this is TODO
Next important steps, i think:
- gpio and intc on sa1111
- pcmcia support for sa11x0 (and sa1111 help logic)
- REAL root on nfs when we have PCMCIA support (we may use any of supported pccard NICs)
- root on wd0! (using already supported PCMCIA-ATA)
|
/* $OpenBSD: isp_openbsd.c,v 1.30 2007/04/10 17:47:55 miod Exp $ */
/*
* Platform (OpenBSD) dependent common attachment code for Qlogic adapters.
*
* Copyright (c) 1999, 2000, 2001 by Matthew Jacob
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice immediately at the beginning of the file, without modification,
* this list of conditions, and the following disclaimer.
* documentation and/or other materials provided with the distribution.
* 2. The name of the author may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
* ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* The author may be reached via electronic communications at
*
* mjacob@feral.com
*
* or, via United States Postal Address
*
* Matthew Jacob
* Feral Software
* PMB#825
* 5214-F Diamond Heights Blvd.
* San Francisco, CA, 94131
*/
/* expand expensive inline functions here. */
#define EXPENSIVE_INLINE
#include <dev/ic/isp_openbsd.h>
/*
* Set a timeout for the watchdogging of a command.
*
* The dimensional analysis is
*
* milliseconds * (seconds/millisecond) * (ticks/second) = ticks
*
* =
*
* (milliseconds / 1000) * hz = ticks
*
*
* For timeouts less than 1 second, we'll get zero. Because of this, and
* because we want to establish *our* timeout to be longer than what the
* firmware might do, we just add 3 seconds at the back end.
*/
#define _XT(xs) ((((xs)->timeout/1000) * hz) + (3 * hz))
static void ispminphys(struct buf *);
static int32_t ispcmd_slow(XS_T *);
static int32_t ispcmd(XS_T *);
static struct scsi_device isp_dev = { NULL, NULL, NULL, NULL };
static int isp_polled_cmd (struct ispsoftc *, XS_T *);
static void isp_wdog (void *);
static void isp_requeue(void *);
static void isp_trestart(void *);
static void isp_restart(struct ispsoftc *);
struct cfdriver isp_cd = {
NULL, "isp", DV_DULL
};
/*
* Complete attachment of hardware, include subdevices.
*/
void
isp_attach(struct ispsoftc *isp)
{
struct scsibus_attach_args saa;
struct scsi_link *lptr = &isp->isp_osinfo._link[0];
isp->isp_osinfo._adapter.scsi_minphys = ispminphys;
isp->isp_state = ISP_RUNSTATE;
/*
* We only manage a single wait queues for dual bus controllers.
* This is arguably broken.
*/
isp->isp_osinfo.wqf = isp->isp_osinfo.wqt = NULL;
lptr->adapter_softc = isp;
lptr->device = &isp_dev;
lptr->adapter = &isp->isp_osinfo._adapter;
lptr->openings = imin(isp->isp_maxcmds, MAXISPREQUEST(isp));
isp->isp_osinfo._adapter.scsi_cmd = ispcmd_slow;
if (IS_FC(isp)) {
lptr->adapter_buswidth = MAX_FC_TARG;
/* We can set max lun width here */
/* loopid set below */
} else {
sdparam *sdp = isp->isp_param;
lptr->adapter_buswidth = MAX_TARGETS;
/* We can set max lun width here */
lptr->adapter_target = sdp->isp_initiator_id;
isp->isp_osinfo.discovered[0] = 1 << sdp->isp_initiator_id;
if (IS_DUALBUS(isp)) {
struct scsi_link *lptrb = &isp->isp_osinfo._link[1];
lptrb->adapter_softc = isp;
lptrb->device = &isp_dev;
lptrb->adapter = &isp->isp_osinfo._adapter;
lptrb->openings = lptr->openings;
lptrb->adapter_buswidth = MAX_TARGETS;
lptrb->adapter_target = sdp->isp_initiator_id;
lptrb->flags = SDEV_2NDBUS;
isp->isp_osinfo.discovered[1] =
1 << (sdp+1)->isp_initiator_id;
}
}
/*
* Send a SCSI Bus Reset (used to be done as part of attach,
* but now left to the OS outer layers).
*
* We don't do 'bus resets' for FC because the LIP that occurs
* when we start the firmware does all that for us.
*/
if (IS_SCSI(isp)) {
int bus = 0;
ISP_LOCK(isp);
(void) isp_control(isp, ISPCTL_RESET_BUS, &bus);
if (IS_DUALBUS(isp)) {
bus++;
(void) isp_control(isp, ISPCTL_RESET_BUS, &bus);
}
ISP_UNLOCK(isp);
/*
* wait for the bus to settle.
*/
delay(4 * 1000000);
} else {
fcparam *fcp = isp->isp_param;
int defid = MAX_FC_TARG;
delay(2 * 1000000);
ISP_LOCK(isp);
isp_fc_runstate(isp, 10 * 1000000);
if (fcp->isp_fwstate == FW_READY &&
fcp->isp_loopstate >= LOOP_PDB_RCVD) {
defid = fcp->isp_loopid;
}
ISP_UNLOCK(isp);
lptr->adapter_target = defid;
}
bzero(&saa, sizeof(saa));
saa.saa_sc_link = lptr;
/*
* And attach children (if any).
*/
config_found((void *)isp, &saa, scsiprint);
if (IS_DUALBUS(isp)) {
lptr++;
bzero(&saa, sizeof(saa));
saa.saa_sc_link = lptr;
config_found((void *)isp, &saa, scsiprint);
}
}
/*
* minphys our xfers
*
* Unfortunately, the buffer pointer describes the target device- not the
* adapter device, so we can't use the pointer to find out what kind of
* adapter we are and adjust accordingly.
*/
static void
ispminphys(struct buf *bp)
{
/*
* XX: Only the 1020 has a 24 bit limit.
*/
if (bp->b_bcount >= (1 << 24)) {
bp->b_bcount = (1 << 24);
}
minphys(bp);
}
static int32_t
ispcmd_slow(XS_T *xs)
{
sdparam *sdp;
int tgt, chan;
u_int16_t f;
struct ispsoftc *isp = XS_ISP(xs);
if (IS_FC(isp)) {
if (cold == 0) {
isp->isp_osinfo.no_mbox_ints = 0;
isp->isp_osinfo._adapter.scsi_cmd = ispcmd;
}
return (ispcmd(xs));
}
/*
* Have we completed discovery for this target on this adapter?
*/
sdp = isp->isp_param;
tgt = XS_TGT(xs);
chan = XS_CHANNEL(xs);
sdp += chan;
if ((xs->flags & SCSI_POLL) != 0 ||
(isp->isp_osinfo.discovered[chan] & (1 << tgt)) != 0) {
return (ispcmd(xs));
}
if (cold == 0) {
isp->isp_osinfo.no_mbox_ints = 0;
}
f = DPARM_DEFAULT;
if (xs->sc_link->quirks & SDEV_NOSYNC) {
f &= ~DPARM_SYNC;
}
if (xs->sc_link->quirks & SDEV_NOWIDE) {
f &= ~DPARM_WIDE;
}
if (xs->sc_link->quirks & SDEV_NOTAGS) {
f &= ~DPARM_TQING;
}
/*
* Okay, we know about this device now,
* so mark parameters to be updated for it.
*/
sdp->isp_devparam[tgt].goal_flags = f;
sdp->isp_devparam[tgt].dev_update = 1;
isp->isp_update |= (1 << chan);
/*
* Now check to see whether we can get out of this checking mode now.
* XXX: WE CANNOT AS YET BECAUSE THERE IS NO MECHANISM TO TELL US
* XXX: WHEN WE'RE DONE DISCOVERY BECAUSE WE NEED ONE COMMAND AFTER
* XXX: DISCOVERY IS DONE FOR EACH TARGET TO TELL US THAT WE'RE DONE
* XXX: AND THAT DOESN'T HAPPEN HERE. AT BEST WE CAN MARK OURSELVES
* XXX: DONE WITH DISCOVERY FOR THIS TARGET AND SO SAVE MAYBE 20
* XXX: LINES OF C CODE.
*/
isp->isp_osinfo.discovered[chan] |= (1 << tgt);
/* do not bother with these lines- they'll never execute correctly */
#if 0
sdp = isp->isp_param;
for (chan = 0; chan < (IS_12X0(isp)? 2 : 1); chan++, sdp++) {
f = 0xffff & ~(1 << sdp->isp_initiator_id);
if (isp->isp_osinfo.discovered[chan] != f) {
break;
}
}
if (chan == (IS_12X0(isp)? 2 : 1)) {
isp->isp_osinfo._adapter.scsipi_cmd = ispcmd;
if (IS_12X0(isp))
isp->isp_update |= 2;
}
#endif
return (ispcmd(xs));
}
static INLINE void isp_add2_blocked_queue(struct ispsoftc *, XS_T *);
static INLINE void
isp_add2_blocked_queue(struct ispsoftc *isp, XS_T *xs)
{
if (isp->isp_osinfo.wqf != NULL) {
isp->isp_osinfo.wqt->free_list.le_next = xs;
} else {
isp->isp_osinfo.wqf = xs;
}
isp->isp_osinfo.wqt = xs;
xs->free_list.le_next = NULL;
}
static int32_t
ispcmd(XS_T *xs)
{
struct ispsoftc *isp;
int result;
/*
* Make sure that there's *some* kind of sane setting.
*/
isp = XS_ISP(xs);
timeout_set(&xs->stimeout, isp_wdog, isp);
if (XS_LUN(xs) >= isp->isp_maxluns) {
xs->error = XS_SELTIMEOUT;
return (COMPLETE);
}
ISP_LOCK(isp);
if (isp->isp_state < ISP_RUNSTATE) {
DISABLE_INTS(isp);
isp_init(isp);
if (isp->isp_state != ISP_INITSTATE) {
ENABLE_INTS(isp);
ISP_UNLOCK(isp);
XS_SETERR(xs, HBA_BOTCH);
return (CMD_COMPLETE);
}
isp->isp_state = ISP_RUNSTATE;
ENABLE_INTS(isp);
}
/*
* Check for queue blockage...
*/
if (isp->isp_osinfo.blocked) {
if (xs->flags & SCSI_POLL) {
ISP_UNLOCK(isp);
return (TRY_AGAIN_LATER);
}
if (isp->isp_osinfo.blocked == 2) {
isp_restart(isp);
}
if (isp->isp_osinfo.blocked) {
isp_add2_blocked_queue(isp, xs);
ISP_UNLOCK(isp);
isp_prt(isp, ISP_LOGDEBUG0, "added to blocked queue");
return (SUCCESSFULLY_QUEUED);
}
}
if (xs->flags & SCSI_POLL) {
volatile u_int8_t ombi = isp->isp_osinfo.no_mbox_ints;
isp->isp_osinfo.no_mbox_ints = 1;
result = isp_polled_cmd(isp, xs);
isp->isp_osinfo.no_mbox_ints = ombi;
ISP_UNLOCK(isp);
return (result);
}
result = isp_start(xs);
switch (result) {
case CMD_QUEUED:
result = SUCCESSFULLY_QUEUED;
if (xs->timeout) {
timeout_add(&xs->stimeout, _XT(xs));
XS_CMD_S_TIMER(xs);
}
if (isp->isp_osinfo.hiwater < isp->isp_nactive) {
isp->isp_osinfo.hiwater = isp->isp_nactive;
isp_prt(isp, ISP_LOGDEBUG0,
"Active Hiwater Mark=%d", isp->isp_nactive);
}
break;
case CMD_EAGAIN:
isp->isp_osinfo.blocked |= 2;
isp_prt(isp, ISP_LOGDEBUG0, "blocking queue");
isp_add2_blocked_queue(isp, xs);
result = SUCCESSFULLY_QUEUED;
break;
case CMD_RQLATER:
result = SUCCESSFULLY_QUEUED; /* Lie */
isp_prt(isp, ISP_LOGDEBUG1, "retrying later for %d.%d",
XS_TGT(xs), XS_LUN(xs));
timeout_set(&xs->stimeout, isp_requeue, xs);
timeout_add(&xs->stimeout, hz);
XS_CMD_S_TIMER(xs);
break;
case CMD_COMPLETE:
result = COMPLETE;
break;
}
ISP_UNLOCK(isp);
return (result);
}
static int
isp_polled_cmd(struct ispsoftc *isp, XS_T *xs)
{
int result;
int infinite = 0, mswait;
result = isp_start(xs);
switch (result) {
case CMD_QUEUED:
result = SUCCESSFULLY_QUEUED;
break;
case CMD_RQLATER:
case CMD_EAGAIN:
result = TRY_AGAIN_LATER;
break;
case CMD_COMPLETE:
result = COMPLETE;
break;
}
if (result != SUCCESSFULLY_QUEUED) {
return (result);
}
/*
* If we can't use interrupts, poll on completion.
*/
if ((mswait = XS_TIME(xs)) == 0)
infinite = 1;
while (mswait || infinite) {
u_int16_t isr, sema, mbox;
if (ISP_READ_ISR(isp, &isr, &sema, &mbox)) {
isp_intr(isp, isr, sema, mbox);
if (XS_CMD_DONE_P(xs)) {
break;
}
}
USEC_DELAY(1000);
mswait -= 1;
}
/*
* If no other error occurred but we didn't finish,
* something bad happened.
*/
if (XS_CMD_DONE_P(xs) == 0) {
if (isp_control(isp, ISPCTL_ABORT_CMD, xs)) {
isp_reinit(isp);
}
if (XS_NOERR(xs)) {
XS_SETERR(xs, HBA_BOTCH);
}
}
result = COMPLETE;
return (result);
}
void
isp_done(XS_T *xs)
{
XS_CMD_S_DONE(xs);
if (XS_CMD_WDOG_P(xs) == 0) {
struct ispsoftc *isp = XS_ISP(xs);
if (XS_CMD_TIMER_P(xs)) {
timeout_del(&xs->stimeout);
XS_CMD_C_TIMER(xs);
}
if (XS_CMD_GRACE_P(xs)) {
struct ispsoftc *isp = XS_ISP(xs);
isp_prt(isp, ISP_LOGWARN,
"finished command on borrowed time");
}
XS_CMD_S_CLEAR(xs);
scsi_done(xs);
if (isp->isp_osinfo.blocked == 2) {
isp->isp_osinfo.blocked = 0;
isp_prt(isp, ISP_LOGDEBUG0, "restarting blocked queue");
isp_restart(isp);
}
}
}
static void
isp_wdog(void *arg)
{
XS_T *xs = arg;
struct ispsoftc *isp = XS_ISP(xs);
u_int32_t handle;
/*
* We've decided this command is dead. Make sure we're not trying
* to kill a command that's already dead by getting its handle and
* and seeing whether it's still alive.
*/
ISP_LOCK(isp);
handle = isp_find_handle(isp, xs);
if (handle) {
u_int16_t isr, sema, mbox;
if (XS_CMD_DONE_P(xs)) {
isp_prt(isp, ISP_LOGDEBUG1,
"watchdog found done cmd (handle 0x%x)",
handle);
ISP_UNLOCK(isp);
return;
}
if (XS_CMD_WDOG_P(xs)) {
isp_prt(isp, ISP_LOGDEBUG1,
"recursive watchdog (handle 0x%x)",
handle);
ISP_UNLOCK(isp);
return;
}
XS_CMD_S_WDOG(xs);
if (ISP_READ_ISR(isp, &isr, &sema, &mbox)) {
isp_intr(isp, isr, sema, mbox);
}
if (XS_CMD_DONE_P(xs)) {
isp_prt(isp, ISP_LOGINFO,
"watchdog cleanup for handle 0x%x", handle);
XS_CMD_C_WDOG(xs);
isp_done(xs);
} else if (XS_CMD_GRACE_P(xs)) {
/*
* Make sure the command is *really* dead before we
* release the handle (and DMA resources) for reuse.
*/
(void) isp_control(isp, ISPCTL_ABORT_CMD, arg);
/*
* After this point, the command is really dead.
*/
if (XS_XFRLEN(xs)) {
ISP_DMAFREE(isp, xs, handle);
}
isp_prt(isp, ISP_LOGWARN,
"watchdog timeout on handle %x", handle);
isp_destroy_handle(isp, handle);
XS_SETERR(xs, XS_TIMEOUT);
XS_CMD_S_CLEAR(xs);
isp_done(xs);
} else {
u_int16_t nxti, optr;
ispreq_t local, *mp = &local, *qe;
isp_prt(isp, ISP_LOGWARN,
"possible command timeout on handle %x", handle);
XS_CMD_C_WDOG(xs);
timeout_add(&xs->stimeout, _XT(xs));
XS_CMD_S_TIMER(xs);
if (isp_getrqentry(isp, &nxti, &optr, (void **) &qe)) {
ISP_UNLOCK(isp);
return;
}
XS_CMD_S_GRACE(xs);
MEMZERO((void *) mp, sizeof (*mp));
mp->req_header.rqs_entry_count = 1;
mp->req_header.rqs_entry_type = RQSTYPE_MARKER;
mp->req_modifier = SYNC_ALL;
mp->req_target = XS_CHANNEL(xs) << 7;
isp_put_request(isp, mp, qe);
ISP_ADD_REQUEST(isp, nxti);
}
} else if (isp->isp_dblev) {
isp_prt(isp, ISP_LOGDEBUG1, "watchdog with no command");
}
ISP_UNLOCK(isp);
}
/*
* Free any associated resources prior to decommissioning and
* set the card to a known state (so it doesn't wake up and kick
* us when we aren't expecting it to).
*
* Locks are held before coming here.
*/
void
isp_uninit(struct ispsoftc *isp)
{
ISP_LOCK(isp);
/*
* Leave with interrupts disabled.
*/
DISABLE_INTS(isp);
ISP_UNLOCK(isp);
}
/*
* Restart function for a command to be requeued later.
*/
static void
isp_requeue(void *arg)
{
int r;
struct scsi_xfer *xs = arg;
struct ispsoftc *isp = XS_ISP(xs);
ISP_LOCK(isp);
r = isp_start(xs);
switch (r) {
case CMD_QUEUED:
isp_prt(isp, ISP_LOGDEBUG1, "restarted command for %d.%d",
XS_TGT(xs), XS_LUN(xs));
if (xs->timeout) {
timeout_set(&xs->stimeout, isp_wdog, isp);
timeout_add(&xs->stimeout, _XT(xs));
XS_CMD_S_TIMER(xs);
}
break;
case CMD_EAGAIN:
isp_prt(isp, ISP_LOGDEBUG0, "blocked cmd again");
isp->isp_osinfo.blocked |= 2;
isp_add2_blocked_queue(isp, xs);
break;
case CMD_RQLATER:
isp_prt(isp, ISP_LOGDEBUG0, "%s for %d.%d",
(r == CMD_EAGAIN)? "CMD_EAGAIN" : "CMD_RQLATER",
XS_TGT(xs), XS_LUN(xs));
timeout_set(&xs->stimeout, isp_requeue, xs);
timeout_add(&xs->stimeout, hz);
XS_CMD_S_TIMER(xs);
break;
case CMD_COMPLETE:
/* can only be an error */
if (XS_NOERR(xs))
XS_SETERR(xs, XS_DRIVER_STUFFUP);
isp_done(xs);
break;
}
ISP_UNLOCK(isp);
}
/*
* Restart function after a LOOP UP event or a command completing,
* sometimes done as a timeout for some hysteresis.
*/
static void
isp_trestart(void *arg)
{
struct ispsoftc *isp = arg;
struct scsi_xfer *list;
ISP_LOCK(isp);
isp->isp_osinfo.rtpend = 0;
list = isp->isp_osinfo.wqf;
if (isp->isp_osinfo.blocked == 0 && list != NULL) {
int nrestarted = 0;
isp->isp_osinfo.wqf = NULL;
ISP_UNLOCK(isp);
do {
struct scsi_xfer *xs = list;
list = xs->free_list.le_next;
xs->free_list.le_next = NULL;
isp_requeue(xs);
if (isp->isp_osinfo.wqf == NULL)
nrestarted++;
} while (list != NULL);
isp_prt(isp, ISP_LOGDEBUG0, "requeued %d commands", nrestarted);
} else {
ISP_UNLOCK(isp);
}
}
static void
isp_restart(struct ispsoftc *isp)
{
struct scsi_xfer *list;
list = isp->isp_osinfo.wqf;
if (isp->isp_osinfo.blocked == 0 && list != NULL) {
int nrestarted = 0;
isp->isp_osinfo.wqf = NULL;
do {
struct scsi_xfer *xs = list;
list = xs->free_list.le_next;
xs->free_list.le_next = NULL;
isp_requeue(xs);
if (isp->isp_osinfo.wqf == NULL)
nrestarted++;
} while (list != NULL);
isp_prt(isp, ISP_LOGDEBUG0, "requeued %d commands", nrestarted);
}
}
int
isp_async(struct ispsoftc *isp, ispasync_t cmd, void *arg)
{
int bus, tgt;
switch (cmd) {
case ISPASYNC_NEW_TGT_PARAMS:
if (IS_SCSI(isp) && isp->isp_dblev) {
sdparam *sdp = isp->isp_param;
char *wt;
int mhz, flags, period;
tgt = *((int *) arg);
bus = (tgt >> 16) & 0xffff;
tgt &= 0xffff;
sdp += bus;
flags = sdp->isp_devparam[tgt].actv_flags;
period = sdp->isp_devparam[tgt].actv_period;
if ((flags & DPARM_SYNC) && period &&
(sdp->isp_devparam[tgt].actv_offset) != 0) {
/*
* There's some ambiguity about our negotiated speed
* if we haven't detected LVD mode correctly (which
* seems to happen, unfortunately). If we're in LVD
* mode, then different rules apply about speed.
*/
if (sdp->isp_lvdmode || period < 0xc) {
switch (period) {
case 0x9:
mhz = 80;
break;
case 0xa:
mhz = 40;
break;
case 0xb:
mhz = 33;
break;
case 0xc:
mhz = 25;
break;
default:
mhz = 1000 / (period * 4);
break;
}
} else {
mhz = 1000 / (period * 4);
}
} else {
mhz = 0;
}
switch (flags & (DPARM_WIDE|DPARM_TQING)) {
case DPARM_WIDE:
wt = ", 16 bit wide";
break;
case DPARM_TQING:
wt = ", Tagged Queueing Enabled";
break;
case DPARM_WIDE|DPARM_TQING:
wt = ", 16 bit wide, Tagged Queueing Enabled";
break;
default:
wt = " ";
break;
}
if (mhz) {
isp_prt(isp, ISP_LOGINFO,
"Bus %d Target %d at %dMHz Max Offset %d%s",
bus, tgt, mhz, sdp->isp_devparam[tgt].actv_offset,
wt);
} else {
isp_prt(isp, ISP_LOGINFO,
"Bus %d Target %d Async Mode%s", bus, tgt, wt);
}
break;
}
case ISPASYNC_BUS_RESET:
if (arg)
bus = *((int *) arg);
else
bus = 0;
isp_prt(isp, ISP_LOGINFO, "SCSI bus %d reset detected", bus);
break;
case ISPASYNC_LOOP_DOWN:
/*
* Hopefully we get here in time to minimize the number
* of commands we are firing off that are sure to die.
*/
isp->isp_osinfo.blocked |= 1;
isp_prt(isp, ISP_LOGINFO, "Loop DOWN");
break;
case ISPASYNC_LOOP_UP:
isp->isp_osinfo.blocked &= ~1;
if (isp->isp_osinfo.rtpend == 0) {
timeout_set(&isp->isp_osinfo.rqt, isp_trestart, isp);
isp->isp_osinfo.rtpend = 1;
}
timeout_add(&isp->isp_osinfo.rqt, 1);
isp_prt(isp, ISP_LOGINFO, "Loop UP");
break;
case ISPASYNC_PROMENADE:
if (IS_FC(isp) && isp->isp_dblev) {
const char *fmt = "Target %d (Loop 0x%x) Port ID 0x%x "
"role %s %s\n Port WWN 0x%08x%08x\n Node WWN 0x%08x%08x";
const static char *roles[4] = {
"No", "Target", "Initiator", "Target/Initiator"
};
fcparam *fcp = isp->isp_param;
int tgt = *((int *) arg);
struct lportdb *lp = &fcp->portdb[tgt];
isp_prt(isp, ISP_LOGINFO, fmt, tgt, lp->loopid, lp->portid,
roles[lp->roles & 0x3],
(lp->valid)? "Arrived" : "Departed",
(u_int32_t) (lp->port_wwn >> 32),
(u_int32_t) (lp->port_wwn & 0xffffffffLL),
(u_int32_t) (lp->node_wwn >> 32),
(u_int32_t) (lp->node_wwn & 0xffffffffLL));
break;
}
case ISPASYNC_CHANGE_NOTIFY:
if (arg == (void *) 1) {
isp_prt(isp, ISP_LOGINFO,
"Name Server Database Changed");
} else {
isp_prt(isp, ISP_LOGINFO,
"Name Server Database Changed");
}
break;
case ISPASYNC_FABRIC_DEV:
{
int target, base, lim;
fcparam *fcp = isp->isp_param;
struct lportdb *lp = NULL;
struct lportdb *clp = (struct lportdb *) arg;
char *pt;
switch (clp->port_type) {
case 1:
pt = " N_Port";
break;
case 2:
pt = " NL_Port";
break;
case 3:
pt = "F/NL_Port";
break;
case 0x7f:
pt = " Nx_Port";
break;
case 0x81:
pt = " F_port";
break;
case 0x82:
pt = " FL_Port";
break;
case 0x84:
pt = " E_port";
break;
default:
pt = " ";
break;
}
isp_prt(isp, ISP_LOGINFO,
"%s Fabric Device @ PortID 0x%x", pt, clp->portid);
/*
* If we don't have an initiator role we bail.
*
* We just use ISPASYNC_FABRIC_DEV for announcement purposes.
*/
if ((isp->isp_role & ISP_ROLE_INITIATOR) == 0) {
break;
}
/*
* Is this entry for us? If so, we bail.
*/
if (fcp->isp_portid == clp->portid) {
break;
}
/*
* Else, the default policy is to find room for it in
* our local port database. Later, when we execute
* the call to isp_pdb_sync either this newly arrived
* or already logged in device will be (re)announced.
*/
if (fcp->isp_topo == TOPO_FL_PORT)
base = FC_SNS_ID+1;
else
base = 0;
if (fcp->isp_topo == TOPO_N_PORT)
lim = 1;
else
lim = MAX_FC_TARG;
/*
* Is it already in our list?
*/
for (target = base; target < lim; target++) {
if (target >= FL_PORT_ID && target <= FC_SNS_ID) {
continue;
}
lp = &fcp->portdb[target];
if (lp->port_wwn == clp->port_wwn &&
lp->node_wwn == clp->node_wwn) {
lp->fabric_dev = 1;
break;
}
}
if (target < lim) {
break;
}
for (target = base; target < lim; target++) {
if (target >= FL_PORT_ID && target <= FC_SNS_ID) {
continue;
}
lp = &fcp->portdb[target];
if (lp->port_wwn == 0) {
break;
}
}
if (target == lim) {
isp_prt(isp, ISP_LOGWARN,
"out of space for fabric devices");
break;
}
lp->port_type = clp->port_type;
lp->fc4_type = clp->fc4_type;
lp->node_wwn = clp->node_wwn;
lp->port_wwn = clp->port_wwn;
lp->portid = clp->portid;
lp->fabric_dev = 1;
break;
}
case ISPASYNC_FW_CRASH:
{
u_int16_t mbox1, mbox6;
mbox1 = ISP_READ(isp, OUTMAILBOX1);
if (IS_DUALBUS(isp)) {
mbox6 = ISP_READ(isp, OUTMAILBOX6);
} else {
mbox6 = 0;
}
isp_prt(isp, ISP_LOGERR,
"Internal Firmware Error on bus %d @ RISC Address 0x%x",
mbox6, mbox1);
#ifdef ISP_FW_CRASH_DUMP
if (IS_FC(isp)) {
isp->isp_osinfo.blocked |= 1;
isp_fw_dump(isp);
}
isp_reinit(isp);
isp_async(isp, ISPASYNC_FW_RESTART, NULL);
#endif
break;
}
default:
break;
}
return (0);
}
void
isp_prt(struct ispsoftc *isp, int level, const char *fmt, ...)
{
va_list ap;
if (level != ISP_LOGALL && (level & isp->isp_dblev) == 0) {
return;
}
printf("%s: ", isp->isp_name);
va_start(ap, fmt);
vprintf(fmt, ap);
va_end(ap);
printf("\n");
}