cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

aic7xxx_core.c (213981B)


      1/*
      2 * Core routines and tables shareable across OS platforms.
      3 *
      4 * Copyright (c) 1994-2002 Justin T. Gibbs.
      5 * Copyright (c) 2000-2002 Adaptec Inc.
      6 * All rights reserved.
      7 *
      8 * Redistribution and use in source and binary forms, with or without
      9 * modification, are permitted provided that the following conditions
     10 * are met:
     11 * 1. Redistributions of source code must retain the above copyright
     12 *    notice, this list of conditions, and the following disclaimer,
     13 *    without modification.
     14 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
     15 *    substantially similar to the "NO WARRANTY" disclaimer below
     16 *    ("Disclaimer") and any redistribution must be conditioned upon
     17 *    including a substantially similar Disclaimer requirement for further
     18 *    binary redistribution.
     19 * 3. Neither the names of the above-listed copyright holders nor the names
     20 *    of any contributors may be used to endorse or promote products derived
     21 *    from this software without specific prior written permission.
     22 *
     23 * Alternatively, this software may be distributed under the terms of the
     24 * GNU General Public License ("GPL") version 2 as published by the Free
     25 * Software Foundation.
     26 *
     27 * NO WARRANTY
     28 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     29 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     30 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR
     31 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
     32 * HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
     36 * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
     37 * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     38 * POSSIBILITY OF SUCH DAMAGES.
     39 *
     40 * $Id: //depot/aic7xxx/aic7xxx/aic7xxx.c#155 $
     41 */
     42
     43#include "aic7xxx_osm.h"
     44#include "aic7xxx_inline.h"
     45#include "aicasm/aicasm_insformat.h"
     46
     47/***************************** Lookup Tables **********************************/
     48static const char *const ahc_chip_names[] = {
     49	"NONE",
     50	"aic7770",
     51	"aic7850",
     52	"aic7855",
     53	"aic7859",
     54	"aic7860",
     55	"aic7870",
     56	"aic7880",
     57	"aic7895",
     58	"aic7895C",
     59	"aic7890/91",
     60	"aic7896/97",
     61	"aic7892",
     62	"aic7899"
     63};
     64
     65/*
     66 * Hardware error codes.
     67 */
     68struct ahc_hard_error_entry {
     69	uint8_t errno;
     70	const char *errmesg;
     71};
     72
     73static const struct ahc_hard_error_entry ahc_hard_errors[] = {
     74	{ ILLHADDR,	"Illegal Host Access" },
     75	{ ILLSADDR,	"Illegal Sequencer Address referenced" },
     76	{ ILLOPCODE,	"Illegal Opcode in sequencer program" },
     77	{ SQPARERR,	"Sequencer Parity Error" },
     78	{ DPARERR,	"Data-path Parity Error" },
     79	{ MPARERR,	"Scratch or SCB Memory Parity Error" },
     80	{ PCIERRSTAT,	"PCI Error detected" },
     81	{ CIOPARERR,	"CIOBUS Parity Error" },
     82};
     83static const u_int num_errors = ARRAY_SIZE(ahc_hard_errors);
     84
     85static const struct ahc_phase_table_entry ahc_phase_table[] =
     86{
     87	{ P_DATAOUT,	NOP,			"in Data-out phase"	},
     88	{ P_DATAIN,	INITIATOR_ERROR,	"in Data-in phase"	},
     89	{ P_DATAOUT_DT,	NOP,			"in DT Data-out phase"	},
     90	{ P_DATAIN_DT,	INITIATOR_ERROR,	"in DT Data-in phase"	},
     91	{ P_COMMAND,	NOP,			"in Command phase"	},
     92	{ P_MESGOUT,	NOP,			"in Message-out phase"	},
     93	{ P_STATUS,	INITIATOR_ERROR,	"in Status phase"	},
     94	{ P_MESGIN,	MSG_PARITY_ERROR,	"in Message-in phase"	},
     95	{ P_BUSFREE,	NOP,			"while idle"		},
     96	{ 0,		NOP,			"in unknown phase"	}
     97};
     98
     99/*
    100 * In most cases we only wish to itterate over real phases, so
    101 * exclude the last element from the count.
    102 */
    103static const u_int num_phases = ARRAY_SIZE(ahc_phase_table) - 1;
    104
    105/*
    106 * Valid SCSIRATE values.  (p. 3-17)
    107 * Provides a mapping of tranfer periods in ns to the proper value to
    108 * stick in the scsixfer reg.
    109 */
    110static const struct ahc_syncrate ahc_syncrates[] =
    111{
    112      /* ultra2    fast/ultra  period     rate */
    113	{ 0x42,      0x000,      9,      "80.0" },
    114	{ 0x03,      0x000,     10,      "40.0" },
    115	{ 0x04,      0x000,     11,      "33.0" },
    116	{ 0x05,      0x100,     12,      "20.0" },
    117	{ 0x06,      0x110,     15,      "16.0" },
    118	{ 0x07,      0x120,     18,      "13.4" },
    119	{ 0x08,      0x000,     25,      "10.0" },
    120	{ 0x19,      0x010,     31,      "8.0"  },
    121	{ 0x1a,      0x020,     37,      "6.67" },
    122	{ 0x1b,      0x030,     43,      "5.7"  },
    123	{ 0x1c,      0x040,     50,      "5.0"  },
    124	{ 0x00,      0x050,     56,      "4.4"  },
    125	{ 0x00,      0x060,     62,      "4.0"  },
    126	{ 0x00,      0x070,     68,      "3.6"  },
    127	{ 0x00,      0x000,      0,      NULL   }
    128};
    129
    130/* Our Sequencer Program */
    131#include "aic7xxx_seq.h"
    132
    133/**************************** Function Declarations ***************************/
    134static void		ahc_force_renegotiation(struct ahc_softc *ahc,
    135						struct ahc_devinfo *devinfo);
    136static struct ahc_tmode_tstate*
    137			ahc_alloc_tstate(struct ahc_softc *ahc,
    138					 u_int scsi_id, char channel);
    139#ifdef AHC_TARGET_MODE
    140static void		ahc_free_tstate(struct ahc_softc *ahc,
    141					u_int scsi_id, char channel, int force);
    142#endif
    143static const struct ahc_syncrate*
    144			ahc_devlimited_syncrate(struct ahc_softc *ahc,
    145						struct ahc_initiator_tinfo *,
    146						u_int *period,
    147						u_int *ppr_options,
    148						role_t role);
    149static void		ahc_update_pending_scbs(struct ahc_softc *ahc);
    150static void		ahc_fetch_devinfo(struct ahc_softc *ahc,
    151					  struct ahc_devinfo *devinfo);
    152static void		ahc_scb_devinfo(struct ahc_softc *ahc,
    153					struct ahc_devinfo *devinfo,
    154					struct scb *scb);
    155static void		ahc_assert_atn(struct ahc_softc *ahc);
    156static void		ahc_setup_initiator_msgout(struct ahc_softc *ahc,
    157						   struct ahc_devinfo *devinfo,
    158						   struct scb *scb);
    159static void		ahc_build_transfer_msg(struct ahc_softc *ahc,
    160					       struct ahc_devinfo *devinfo);
    161static void		ahc_construct_sdtr(struct ahc_softc *ahc,
    162					   struct ahc_devinfo *devinfo,
    163					   u_int period, u_int offset);
    164static void		ahc_construct_wdtr(struct ahc_softc *ahc,
    165					   struct ahc_devinfo *devinfo,
    166					   u_int bus_width);
    167static void		ahc_construct_ppr(struct ahc_softc *ahc,
    168					  struct ahc_devinfo *devinfo,
    169					  u_int period, u_int offset,
    170					  u_int bus_width, u_int ppr_options);
    171static void		ahc_clear_msg_state(struct ahc_softc *ahc);
    172static void		ahc_handle_proto_violation(struct ahc_softc *ahc);
    173static void		ahc_handle_message_phase(struct ahc_softc *ahc);
    174typedef enum {
    175	AHCMSG_1B,
    176	AHCMSG_2B,
    177	AHCMSG_EXT
    178} ahc_msgtype;
    179static int		ahc_sent_msg(struct ahc_softc *ahc, ahc_msgtype type,
    180				     u_int msgval, int full);
    181static int		ahc_parse_msg(struct ahc_softc *ahc,
    182				      struct ahc_devinfo *devinfo);
    183static int		ahc_handle_msg_reject(struct ahc_softc *ahc,
    184					      struct ahc_devinfo *devinfo);
    185static void		ahc_handle_ign_wide_residue(struct ahc_softc *ahc,
    186						struct ahc_devinfo *devinfo);
    187static void		ahc_reinitialize_dataptrs(struct ahc_softc *ahc);
    188static void		ahc_handle_devreset(struct ahc_softc *ahc,
    189					    struct ahc_devinfo *devinfo,
    190					    cam_status status, char *message,
    191					    int verbose_level);
    192#ifdef AHC_TARGET_MODE
    193static void		ahc_setup_target_msgin(struct ahc_softc *ahc,
    194					       struct ahc_devinfo *devinfo,
    195					       struct scb *scb);
    196#endif
    197
    198static bus_dmamap_callback_t	ahc_dmamap_cb;
    199static void		ahc_build_free_scb_list(struct ahc_softc *ahc);
    200static int		ahc_init_scbdata(struct ahc_softc *ahc);
    201static void		ahc_fini_scbdata(struct ahc_softc *ahc);
    202static void		ahc_qinfifo_requeue(struct ahc_softc *ahc,
    203					    struct scb *prev_scb,
    204					    struct scb *scb);
    205static int		ahc_qinfifo_count(struct ahc_softc *ahc);
    206static u_int		ahc_rem_scb_from_disc_list(struct ahc_softc *ahc,
    207						   u_int prev, u_int scbptr);
    208static void		ahc_add_curscb_to_free_list(struct ahc_softc *ahc);
    209static u_int		ahc_rem_wscb(struct ahc_softc *ahc,
    210				     u_int scbpos, u_int prev);
    211static void		ahc_reset_current_bus(struct ahc_softc *ahc);
    212#ifdef AHC_DUMP_SEQ
    213static void		ahc_dumpseq(struct ahc_softc *ahc);
    214#endif
    215static int		ahc_loadseq(struct ahc_softc *ahc);
    216static int		ahc_check_patch(struct ahc_softc *ahc,
    217					const struct patch **start_patch,
    218					u_int start_instr, u_int *skip_addr);
    219static void		ahc_download_instr(struct ahc_softc *ahc,
    220					   u_int instrptr, uint8_t *dconsts);
    221#ifdef AHC_TARGET_MODE
    222static void		ahc_queue_lstate_event(struct ahc_softc *ahc,
    223					       struct ahc_tmode_lstate *lstate,
    224					       u_int initiator_id,
    225					       u_int event_type,
    226					       u_int event_arg);
    227static void		ahc_update_scsiid(struct ahc_softc *ahc,
    228					  u_int targid_mask);
    229static int		ahc_handle_target_cmd(struct ahc_softc *ahc,
    230					      struct target_cmd *cmd);
    231#endif
    232
    233static u_int		ahc_index_busy_tcl(struct ahc_softc *ahc, u_int tcl);
    234static void		ahc_unbusy_tcl(struct ahc_softc *ahc, u_int tcl);
    235static void		ahc_busy_tcl(struct ahc_softc *ahc,
    236				     u_int tcl, u_int busyid);
    237
    238/************************** SCB and SCB queue management **********************/
    239static void		ahc_run_untagged_queues(struct ahc_softc *ahc);
    240static void		ahc_run_untagged_queue(struct ahc_softc *ahc,
    241					       struct scb_tailq *queue);
    242
    243/****************************** Initialization ********************************/
    244static void		 ahc_alloc_scbs(struct ahc_softc *ahc);
    245static void		 ahc_shutdown(void *arg);
    246
    247/*************************** Interrupt Services *******************************/
    248static void		ahc_clear_intstat(struct ahc_softc *ahc);
    249static void		ahc_run_qoutfifo(struct ahc_softc *ahc);
    250#ifdef AHC_TARGET_MODE
    251static void		ahc_run_tqinfifo(struct ahc_softc *ahc, int paused);
    252#endif
    253static void		ahc_handle_brkadrint(struct ahc_softc *ahc);
    254static void		ahc_handle_seqint(struct ahc_softc *ahc, u_int intstat);
    255static void		ahc_handle_scsiint(struct ahc_softc *ahc,
    256					   u_int intstat);
    257static void		ahc_clear_critical_section(struct ahc_softc *ahc);
    258
    259/***************************** Error Recovery *********************************/
    260static void		ahc_freeze_devq(struct ahc_softc *ahc, struct scb *scb);
    261static int		ahc_abort_scbs(struct ahc_softc *ahc, int target,
    262				       char channel, int lun, u_int tag,
    263				       role_t role, uint32_t status);
    264static void		ahc_calc_residual(struct ahc_softc *ahc,
    265					  struct scb *scb);
    266
    267/*********************** Untagged Transaction Routines ************************/
    268static inline void	ahc_freeze_untagged_queues(struct ahc_softc *ahc);
    269static inline void	ahc_release_untagged_queues(struct ahc_softc *ahc);
    270
    271/*
    272 * Block our completion routine from starting the next untagged
    273 * transaction for this target or target lun.
    274 */
    275static inline void
    276ahc_freeze_untagged_queues(struct ahc_softc *ahc)
    277{
    278	if ((ahc->flags & AHC_SCB_BTT) == 0)
    279		ahc->untagged_queue_lock++;
    280}
    281
    282/*
    283 * Allow the next untagged transaction for this target or target lun
    284 * to be executed.  We use a counting semaphore to allow the lock
    285 * to be acquired recursively.  Once the count drops to zero, the
    286 * transaction queues will be run.
    287 */
    288static inline void
    289ahc_release_untagged_queues(struct ahc_softc *ahc)
    290{
    291	if ((ahc->flags & AHC_SCB_BTT) == 0) {
    292		ahc->untagged_queue_lock--;
    293		if (ahc->untagged_queue_lock == 0)
    294			ahc_run_untagged_queues(ahc);
    295	}
    296}
    297
    298/************************* Sequencer Execution Control ************************/
    299/*
    300 * Work around any chip bugs related to halting sequencer execution.
    301 * On Ultra2 controllers, we must clear the CIOBUS stretch signal by
    302 * reading a register that will set this signal and deassert it.
    303 * Without this workaround, if the chip is paused, by an interrupt or
    304 * manual pause while accessing scb ram, accesses to certain registers
    305 * will hang the system (infinite pci retries).
    306 */
    307static void
    308ahc_pause_bug_fix(struct ahc_softc *ahc)
    309{
    310	if ((ahc->features & AHC_ULTRA2) != 0)
    311		(void)ahc_inb(ahc, CCSCBCTL);
    312}
    313
    314/*
    315 * Determine whether the sequencer has halted code execution.
    316 * Returns non-zero status if the sequencer is stopped.
    317 */
    318int
    319ahc_is_paused(struct ahc_softc *ahc)
    320{
    321	return ((ahc_inb(ahc, HCNTRL) & PAUSE) != 0);
    322}
    323
    324/*
    325 * Request that the sequencer stop and wait, indefinitely, for it
    326 * to stop.  The sequencer will only acknowledge that it is paused
    327 * once it has reached an instruction boundary and PAUSEDIS is
    328 * cleared in the SEQCTL register.  The sequencer may use PAUSEDIS
    329 * for critical sections.
    330 */
    331void
    332ahc_pause(struct ahc_softc *ahc)
    333{
    334	ahc_outb(ahc, HCNTRL, ahc->pause);
    335
    336	/*
    337	 * Since the sequencer can disable pausing in a critical section, we
    338	 * must loop until it actually stops.
    339	 */
    340	while (ahc_is_paused(ahc) == 0)
    341		;
    342
    343	ahc_pause_bug_fix(ahc);
    344}
    345
    346/*
    347 * Allow the sequencer to continue program execution.
    348 * We check here to ensure that no additional interrupt
    349 * sources that would cause the sequencer to halt have been
    350 * asserted.  If, for example, a SCSI bus reset is detected
    351 * while we are fielding a different, pausing, interrupt type,
    352 * we don't want to release the sequencer before going back
    353 * into our interrupt handler and dealing with this new
    354 * condition.
    355 */
    356void
    357ahc_unpause(struct ahc_softc *ahc)
    358{
    359	if ((ahc_inb(ahc, INTSTAT) & (SCSIINT | SEQINT | BRKADRINT)) == 0)
    360		ahc_outb(ahc, HCNTRL, ahc->unpause);
    361}
    362
    363/************************** Memory mapping routines ***************************/
    364static struct ahc_dma_seg *
    365ahc_sg_bus_to_virt(struct scb *scb, uint32_t sg_busaddr)
    366{
    367	int sg_index;
    368
    369	sg_index = (sg_busaddr - scb->sg_list_phys)/sizeof(struct ahc_dma_seg);
    370	/* sg_list_phys points to entry 1, not 0 */
    371	sg_index++;
    372
    373	return (&scb->sg_list[sg_index]);
    374}
    375
    376static uint32_t
    377ahc_sg_virt_to_bus(struct scb *scb, struct ahc_dma_seg *sg)
    378{
    379	int sg_index;
    380
    381	/* sg_list_phys points to entry 1, not 0 */
    382	sg_index = sg - &scb->sg_list[1];
    383
    384	return (scb->sg_list_phys + (sg_index * sizeof(*scb->sg_list)));
    385}
    386
    387static uint32_t
    388ahc_hscb_busaddr(struct ahc_softc *ahc, u_int index)
    389{
    390	return (ahc->scb_data->hscb_busaddr
    391		+ (sizeof(struct hardware_scb) * index));
    392}
    393
    394static void
    395ahc_sync_scb(struct ahc_softc *ahc, struct scb *scb, int op)
    396{
    397	ahc_dmamap_sync(ahc, ahc->scb_data->hscb_dmat,
    398			ahc->scb_data->hscb_dmamap,
    399			/*offset*/(scb->hscb - ahc->hscbs) * sizeof(*scb->hscb),
    400			/*len*/sizeof(*scb->hscb), op);
    401}
    402
    403void
    404ahc_sync_sglist(struct ahc_softc *ahc, struct scb *scb, int op)
    405{
    406	if (scb->sg_count == 0)
    407		return;
    408
    409	ahc_dmamap_sync(ahc, ahc->scb_data->sg_dmat, scb->sg_map->sg_dmamap,
    410			/*offset*/(scb->sg_list - scb->sg_map->sg_vaddr)
    411				* sizeof(struct ahc_dma_seg),
    412			/*len*/sizeof(struct ahc_dma_seg) * scb->sg_count, op);
    413}
    414
    415#ifdef AHC_TARGET_MODE
    416static uint32_t
    417ahc_targetcmd_offset(struct ahc_softc *ahc, u_int index)
    418{
    419	return (((uint8_t *)&ahc->targetcmds[index]) - ahc->qoutfifo);
    420}
    421#endif
    422
    423/*********************** Miscellaneous Support Functions ***********************/
    424/*
    425 * Determine whether the sequencer reported a residual
    426 * for this SCB/transaction.
    427 */
    428static void
    429ahc_update_residual(struct ahc_softc *ahc, struct scb *scb)
    430{
    431	uint32_t sgptr;
    432
    433	sgptr = ahc_le32toh(scb->hscb->sgptr);
    434	if ((sgptr & SG_RESID_VALID) != 0)
    435		ahc_calc_residual(ahc, scb);
    436}
    437
    438/*
    439 * Return pointers to the transfer negotiation information
    440 * for the specified our_id/remote_id pair.
    441 */
    442struct ahc_initiator_tinfo *
    443ahc_fetch_transinfo(struct ahc_softc *ahc, char channel, u_int our_id,
    444		    u_int remote_id, struct ahc_tmode_tstate **tstate)
    445{
    446	/*
    447	 * Transfer data structures are stored from the perspective
    448	 * of the target role.  Since the parameters for a connection
    449	 * in the initiator role to a given target are the same as
    450	 * when the roles are reversed, we pretend we are the target.
    451	 */
    452	if (channel == 'B')
    453		our_id += 8;
    454	*tstate = ahc->enabled_targets[our_id];
    455	return (&(*tstate)->transinfo[remote_id]);
    456}
    457
    458uint16_t
    459ahc_inw(struct ahc_softc *ahc, u_int port)
    460{
    461	uint16_t r = ahc_inb(ahc, port+1) << 8;
    462	return r | ahc_inb(ahc, port);
    463}
    464
    465void
    466ahc_outw(struct ahc_softc *ahc, u_int port, u_int value)
    467{
    468	ahc_outb(ahc, port, value & 0xFF);
    469	ahc_outb(ahc, port+1, (value >> 8) & 0xFF);
    470}
    471
    472uint32_t
    473ahc_inl(struct ahc_softc *ahc, u_int port)
    474{
    475	return ((ahc_inb(ahc, port))
    476	      | (ahc_inb(ahc, port+1) << 8)
    477	      | (ahc_inb(ahc, port+2) << 16)
    478	      | (ahc_inb(ahc, port+3) << 24));
    479}
    480
    481void
    482ahc_outl(struct ahc_softc *ahc, u_int port, uint32_t value)
    483{
    484	ahc_outb(ahc, port, (value) & 0xFF);
    485	ahc_outb(ahc, port+1, ((value) >> 8) & 0xFF);
    486	ahc_outb(ahc, port+2, ((value) >> 16) & 0xFF);
    487	ahc_outb(ahc, port+3, ((value) >> 24) & 0xFF);
    488}
    489
    490uint64_t
    491ahc_inq(struct ahc_softc *ahc, u_int port)
    492{
    493	return ((ahc_inb(ahc, port))
    494	      | (ahc_inb(ahc, port+1) << 8)
    495	      | (ahc_inb(ahc, port+2) << 16)
    496	      | (((uint64_t)ahc_inb(ahc, port+3)) << 24)
    497	      | (((uint64_t)ahc_inb(ahc, port+4)) << 32)
    498	      | (((uint64_t)ahc_inb(ahc, port+5)) << 40)
    499	      | (((uint64_t)ahc_inb(ahc, port+6)) << 48)
    500	      | (((uint64_t)ahc_inb(ahc, port+7)) << 56));
    501}
    502
    503void
    504ahc_outq(struct ahc_softc *ahc, u_int port, uint64_t value)
    505{
    506	ahc_outb(ahc, port, value & 0xFF);
    507	ahc_outb(ahc, port+1, (value >> 8) & 0xFF);
    508	ahc_outb(ahc, port+2, (value >> 16) & 0xFF);
    509	ahc_outb(ahc, port+3, (value >> 24) & 0xFF);
    510	ahc_outb(ahc, port+4, (value >> 32) & 0xFF);
    511	ahc_outb(ahc, port+5, (value >> 40) & 0xFF);
    512	ahc_outb(ahc, port+6, (value >> 48) & 0xFF);
    513	ahc_outb(ahc, port+7, (value >> 56) & 0xFF);
    514}
    515
    516/*
    517 * Get a free scb. If there are none, see if we can allocate a new SCB.
    518 */
    519struct scb *
    520ahc_get_scb(struct ahc_softc *ahc)
    521{
    522	struct scb *scb;
    523
    524	if ((scb = SLIST_FIRST(&ahc->scb_data->free_scbs)) == NULL) {
    525		ahc_alloc_scbs(ahc);
    526		scb = SLIST_FIRST(&ahc->scb_data->free_scbs);
    527		if (scb == NULL)
    528			return (NULL);
    529	}
    530	SLIST_REMOVE_HEAD(&ahc->scb_data->free_scbs, links.sle);
    531	return (scb);
    532}
    533
    534/*
    535 * Return an SCB resource to the free list.
    536 */
    537void
    538ahc_free_scb(struct ahc_softc *ahc, struct scb *scb)
    539{
    540	struct hardware_scb *hscb;
    541
    542	hscb = scb->hscb;
    543	/* Clean up for the next user */
    544	ahc->scb_data->scbindex[hscb->tag] = NULL;
    545	scb->flags = SCB_FREE;
    546	hscb->control = 0;
    547
    548	SLIST_INSERT_HEAD(&ahc->scb_data->free_scbs, scb, links.sle);
    549
    550	/* Notify the OSM that a resource is now available. */
    551	ahc_platform_scb_free(ahc, scb);
    552}
    553
    554struct scb *
    555ahc_lookup_scb(struct ahc_softc *ahc, u_int tag)
    556{
    557	struct scb* scb;
    558
    559	scb = ahc->scb_data->scbindex[tag];
    560	if (scb != NULL)
    561		ahc_sync_scb(ahc, scb,
    562			     BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
    563	return (scb);
    564}
    565
    566static void
    567ahc_swap_with_next_hscb(struct ahc_softc *ahc, struct scb *scb)
    568{
    569	struct hardware_scb *q_hscb;
    570	u_int  saved_tag;
    571
    572	/*
    573	 * Our queuing method is a bit tricky.  The card
    574	 * knows in advance which HSCB to download, and we
    575	 * can't disappoint it.  To achieve this, the next
    576	 * SCB to download is saved off in ahc->next_queued_scb.
    577	 * When we are called to queue "an arbitrary scb",
    578	 * we copy the contents of the incoming HSCB to the one
    579	 * the sequencer knows about, swap HSCB pointers and
    580	 * finally assign the SCB to the tag indexed location
    581	 * in the scb_array.  This makes sure that we can still
    582	 * locate the correct SCB by SCB_TAG.
    583	 */
    584	q_hscb = ahc->next_queued_scb->hscb;
    585	saved_tag = q_hscb->tag;
    586	memcpy(q_hscb, scb->hscb, sizeof(*scb->hscb));
    587	if ((scb->flags & SCB_CDB32_PTR) != 0) {
    588		q_hscb->shared_data.cdb_ptr =
    589		    ahc_htole32(ahc_hscb_busaddr(ahc, q_hscb->tag)
    590			      + offsetof(struct hardware_scb, cdb32));
    591	}
    592	q_hscb->tag = saved_tag;
    593	q_hscb->next = scb->hscb->tag;
    594
    595	/* Now swap HSCB pointers. */
    596	ahc->next_queued_scb->hscb = scb->hscb;
    597	scb->hscb = q_hscb;
    598
    599	/* Now define the mapping from tag to SCB in the scbindex */
    600	ahc->scb_data->scbindex[scb->hscb->tag] = scb;
    601}
    602
    603/*
    604 * Tell the sequencer about a new transaction to execute.
    605 */
    606void
    607ahc_queue_scb(struct ahc_softc *ahc, struct scb *scb)
    608{
    609	ahc_swap_with_next_hscb(ahc, scb);
    610
    611	if (scb->hscb->tag == SCB_LIST_NULL
    612	 || scb->hscb->next == SCB_LIST_NULL)
    613		panic("Attempt to queue invalid SCB tag %x:%x\n",
    614		      scb->hscb->tag, scb->hscb->next);
    615
    616	/*
    617	 * Setup data "oddness".
    618	 */
    619	scb->hscb->lun &= LID;
    620	if (ahc_get_transfer_length(scb) & 0x1)
    621		scb->hscb->lun |= SCB_XFERLEN_ODD;
    622
    623	/*
    624	 * Keep a history of SCBs we've downloaded in the qinfifo.
    625	 */
    626	ahc->qinfifo[ahc->qinfifonext++] = scb->hscb->tag;
    627
    628	/*
    629	 * Make sure our data is consistent from the
    630	 * perspective of the adapter.
    631	 */
    632	ahc_sync_scb(ahc, scb, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
    633
    634	/* Tell the adapter about the newly queued SCB */
    635	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
    636		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
    637	} else {
    638		if ((ahc->features & AHC_AUTOPAUSE) == 0)
    639			ahc_pause(ahc);
    640		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
    641		if ((ahc->features & AHC_AUTOPAUSE) == 0)
    642			ahc_unpause(ahc);
    643	}
    644}
    645
    646struct scsi_sense_data *
    647ahc_get_sense_buf(struct ahc_softc *ahc, struct scb *scb)
    648{
    649	int offset;
    650
    651	offset = scb - ahc->scb_data->scbarray;
    652	return (&ahc->scb_data->sense[offset]);
    653}
    654
    655static uint32_t
    656ahc_get_sense_bufaddr(struct ahc_softc *ahc, struct scb *scb)
    657{
    658	int offset;
    659
    660	offset = scb - ahc->scb_data->scbarray;
    661	return (ahc->scb_data->sense_busaddr
    662	      + (offset * sizeof(struct scsi_sense_data)));
    663}
    664
    665/************************** Interrupt Processing ******************************/
    666static void
    667ahc_sync_qoutfifo(struct ahc_softc *ahc, int op)
    668{
    669	ahc_dmamap_sync(ahc, ahc->shared_data_dmat, ahc->shared_data_dmamap,
    670			/*offset*/0, /*len*/256, op);
    671}
    672
    673static void
    674ahc_sync_tqinfifo(struct ahc_softc *ahc, int op)
    675{
    676#ifdef AHC_TARGET_MODE
    677	if ((ahc->flags & AHC_TARGETROLE) != 0) {
    678		ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
    679				ahc->shared_data_dmamap,
    680				ahc_targetcmd_offset(ahc, 0),
    681				sizeof(struct target_cmd) * AHC_TMODE_CMDS,
    682				op);
    683	}
    684#endif
    685}
    686
    687/*
    688 * See if the firmware has posted any completed commands
    689 * into our in-core command complete fifos.
    690 */
    691#define AHC_RUN_QOUTFIFO 0x1
    692#define AHC_RUN_TQINFIFO 0x2
    693static u_int
    694ahc_check_cmdcmpltqueues(struct ahc_softc *ahc)
    695{
    696	u_int retval;
    697
    698	retval = 0;
    699	ahc_dmamap_sync(ahc, ahc->shared_data_dmat, ahc->shared_data_dmamap,
    700			/*offset*/ahc->qoutfifonext, /*len*/1,
    701			BUS_DMASYNC_POSTREAD);
    702	if (ahc->qoutfifo[ahc->qoutfifonext] != SCB_LIST_NULL)
    703		retval |= AHC_RUN_QOUTFIFO;
    704#ifdef AHC_TARGET_MODE
    705	if ((ahc->flags & AHC_TARGETROLE) != 0
    706	 && (ahc->flags & AHC_TQINFIFO_BLOCKED) == 0) {
    707		ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
    708				ahc->shared_data_dmamap,
    709				ahc_targetcmd_offset(ahc, ahc->tqinfifofnext),
    710				/*len*/sizeof(struct target_cmd),
    711				BUS_DMASYNC_POSTREAD);
    712		if (ahc->targetcmds[ahc->tqinfifonext].cmd_valid != 0)
    713			retval |= AHC_RUN_TQINFIFO;
    714	}
    715#endif
    716	return (retval);
    717}
    718
    719/*
    720 * Catch an interrupt from the adapter
    721 */
    722int
    723ahc_intr(struct ahc_softc *ahc)
    724{
    725	u_int	intstat;
    726
    727	if ((ahc->pause & INTEN) == 0) {
    728		/*
    729		 * Our interrupt is not enabled on the chip
    730		 * and may be disabled for re-entrancy reasons,
    731		 * so just return.  This is likely just a shared
    732		 * interrupt.
    733		 */
    734		return (0);
    735	}
    736	/*
    737	 * Instead of directly reading the interrupt status register,
    738	 * infer the cause of the interrupt by checking our in-core
    739	 * completion queues.  This avoids a costly PCI bus read in
    740	 * most cases.
    741	 */
    742	if ((ahc->flags & (AHC_ALL_INTERRUPTS|AHC_EDGE_INTERRUPT)) == 0
    743	 && (ahc_check_cmdcmpltqueues(ahc) != 0))
    744		intstat = CMDCMPLT;
    745	else {
    746		intstat = ahc_inb(ahc, INTSTAT);
    747	}
    748
    749	if ((intstat & INT_PEND) == 0) {
    750#if AHC_PCI_CONFIG > 0
    751		if (ahc->unsolicited_ints > 500) {
    752			ahc->unsolicited_ints = 0;
    753			if ((ahc->chip & AHC_PCI) != 0
    754			 && (ahc_inb(ahc, ERROR) & PCIERRSTAT) != 0)
    755				ahc->bus_intr(ahc);
    756		}
    757#endif
    758		ahc->unsolicited_ints++;
    759		return (0);
    760	}
    761	ahc->unsolicited_ints = 0;
    762
    763	if (intstat & CMDCMPLT) {
    764		ahc_outb(ahc, CLRINT, CLRCMDINT);
    765
    766		/*
    767		 * Ensure that the chip sees that we've cleared
    768		 * this interrupt before we walk the output fifo.
    769		 * Otherwise, we may, due to posted bus writes,
    770		 * clear the interrupt after we finish the scan,
    771		 * and after the sequencer has added new entries
    772		 * and asserted the interrupt again.
    773		 */
    774		ahc_flush_device_writes(ahc);
    775		ahc_run_qoutfifo(ahc);
    776#ifdef AHC_TARGET_MODE
    777		if ((ahc->flags & AHC_TARGETROLE) != 0)
    778			ahc_run_tqinfifo(ahc, /*paused*/FALSE);
    779#endif
    780	}
    781
    782	/*
    783	 * Handle statuses that may invalidate our cached
    784	 * copy of INTSTAT separately.
    785	 */
    786	if (intstat == 0xFF && (ahc->features & AHC_REMOVABLE) != 0) {
    787		/* Hot eject.  Do nothing */
    788	} else if (intstat & BRKADRINT) {
    789		ahc_handle_brkadrint(ahc);
    790	} else if ((intstat & (SEQINT|SCSIINT)) != 0) {
    791
    792		ahc_pause_bug_fix(ahc);
    793
    794		if ((intstat & SEQINT) != 0)
    795			ahc_handle_seqint(ahc, intstat);
    796
    797		if ((intstat & SCSIINT) != 0)
    798			ahc_handle_scsiint(ahc, intstat);
    799	}
    800	return (1);
    801}
    802
    803/************************* Sequencer Execution Control ************************/
    804/*
    805 * Restart the sequencer program from address zero
    806 */
    807static void
    808ahc_restart(struct ahc_softc *ahc)
    809{
    810	uint8_t	sblkctl;
    811
    812	ahc_pause(ahc);
    813
    814	/* No more pending messages. */
    815	ahc_clear_msg_state(ahc);
    816
    817	ahc_outb(ahc, SCSISIGO, 0);		/* De-assert BSY */
    818	ahc_outb(ahc, MSG_OUT, NOP);	/* No message to send */
    819	ahc_outb(ahc, SXFRCTL1, ahc_inb(ahc, SXFRCTL1) & ~BITBUCKET);
    820	ahc_outb(ahc, LASTPHASE, P_BUSFREE);
    821	ahc_outb(ahc, SAVED_SCSIID, 0xFF);
    822	ahc_outb(ahc, SAVED_LUN, 0xFF);
    823
    824	/*
    825	 * Ensure that the sequencer's idea of TQINPOS
    826	 * matches our own.  The sequencer increments TQINPOS
    827	 * only after it sees a DMA complete and a reset could
    828	 * occur before the increment leaving the kernel to believe
    829	 * the command arrived but the sequencer to not.
    830	 */
    831	ahc_outb(ahc, TQINPOS, ahc->tqinfifonext);
    832
    833	/* Always allow reselection */
    834	ahc_outb(ahc, SCSISEQ,
    835		 ahc_inb(ahc, SCSISEQ_TEMPLATE) & (ENSELI|ENRSELI|ENAUTOATNP));
    836	if ((ahc->features & AHC_CMD_CHAN) != 0) {
    837		/* Ensure that no DMA operations are in progress */
    838		ahc_outb(ahc, CCSCBCNT, 0);
    839		ahc_outb(ahc, CCSGCTL, 0);
    840		ahc_outb(ahc, CCSCBCTL, 0);
    841	}
    842	/*
    843	 * If we were in the process of DMA'ing SCB data into
    844	 * an SCB, replace that SCB on the free list.  This prevents
    845	 * an SCB leak.
    846	 */
    847	if ((ahc_inb(ahc, SEQ_FLAGS2) & SCB_DMA) != 0) {
    848		ahc_add_curscb_to_free_list(ahc);
    849		ahc_outb(ahc, SEQ_FLAGS2,
    850			 ahc_inb(ahc, SEQ_FLAGS2) & ~SCB_DMA);
    851	}
    852
    853	/*
    854	 * Clear any pending sequencer interrupt.  It is no
    855	 * longer relevant since we're resetting the Program
    856	 * Counter.
    857	 */
    858	ahc_outb(ahc, CLRINT, CLRSEQINT);
    859
    860	ahc_outb(ahc, MWI_RESIDUAL, 0);
    861	ahc_outb(ahc, SEQCTL, ahc->seqctl);
    862	ahc_outb(ahc, SEQADDR0, 0);
    863	ahc_outb(ahc, SEQADDR1, 0);
    864
    865	/*
    866	 * Take the LED out of diagnostic mode on PM resume, too
    867	 */
    868	sblkctl = ahc_inb(ahc, SBLKCTL);
    869	ahc_outb(ahc, SBLKCTL, (sblkctl & ~(DIAGLEDEN|DIAGLEDON)));
    870
    871	ahc_unpause(ahc);
    872}
    873
    874/************************* Input/Output Queues ********************************/
    875static void
    876ahc_run_qoutfifo(struct ahc_softc *ahc)
    877{
    878	struct scb *scb;
    879	u_int  scb_index;
    880
    881	ahc_sync_qoutfifo(ahc, BUS_DMASYNC_POSTREAD);
    882	while (ahc->qoutfifo[ahc->qoutfifonext] != SCB_LIST_NULL) {
    883
    884		scb_index = ahc->qoutfifo[ahc->qoutfifonext];
    885		if ((ahc->qoutfifonext & 0x03) == 0x03) {
    886			u_int modnext;
    887
    888			/*
    889			 * Clear 32bits of QOUTFIFO at a time
    890			 * so that we don't clobber an incoming
    891			 * byte DMA to the array on architectures
    892			 * that only support 32bit load and store
    893			 * operations.
    894			 */
    895			modnext = ahc->qoutfifonext & ~0x3;
    896			*((uint32_t *)(&ahc->qoutfifo[modnext])) = 0xFFFFFFFFUL;
    897			ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
    898					ahc->shared_data_dmamap,
    899					/*offset*/modnext, /*len*/4,
    900					BUS_DMASYNC_PREREAD);
    901		}
    902		ahc->qoutfifonext++;
    903
    904		scb = ahc_lookup_scb(ahc, scb_index);
    905		if (scb == NULL) {
    906			printk("%s: WARNING no command for scb %d "
    907			       "(cmdcmplt)\nQOUTPOS = %d\n",
    908			       ahc_name(ahc), scb_index,
    909			       (ahc->qoutfifonext - 1) & 0xFF);
    910			continue;
    911		}
    912
    913		/*
    914		 * Save off the residual
    915		 * if there is one.
    916		 */
    917		ahc_update_residual(ahc, scb);
    918		ahc_done(ahc, scb);
    919	}
    920}
    921
    922static void
    923ahc_run_untagged_queues(struct ahc_softc *ahc)
    924{
    925	int i;
    926
    927	for (i = 0; i < 16; i++)
    928		ahc_run_untagged_queue(ahc, &ahc->untagged_queues[i]);
    929}
    930
    931static void
    932ahc_run_untagged_queue(struct ahc_softc *ahc, struct scb_tailq *queue)
    933{
    934	struct scb *scb;
    935
    936	if (ahc->untagged_queue_lock != 0)
    937		return;
    938
    939	if ((scb = TAILQ_FIRST(queue)) != NULL
    940	 && (scb->flags & SCB_ACTIVE) == 0) {
    941		scb->flags |= SCB_ACTIVE;
    942		ahc_queue_scb(ahc, scb);
    943	}
    944}
    945
    946/************************* Interrupt Handling *********************************/
    947static void
    948ahc_handle_brkadrint(struct ahc_softc *ahc)
    949{
    950	/*
    951	 * We upset the sequencer :-(
    952	 * Lookup the error message
    953	 */
    954	int i;
    955	int error;
    956
    957	error = ahc_inb(ahc, ERROR);
    958	for (i = 0; error != 1 && i < num_errors; i++)
    959		error >>= 1;
    960	printk("%s: brkadrint, %s at seqaddr = 0x%x\n",
    961	       ahc_name(ahc), ahc_hard_errors[i].errmesg,
    962	       ahc_inb(ahc, SEQADDR0) |
    963	       (ahc_inb(ahc, SEQADDR1) << 8));
    964
    965	ahc_dump_card_state(ahc);
    966
    967	/* Tell everyone that this HBA is no longer available */
    968	ahc_abort_scbs(ahc, CAM_TARGET_WILDCARD, ALL_CHANNELS,
    969		       CAM_LUN_WILDCARD, SCB_LIST_NULL, ROLE_UNKNOWN,
    970		       CAM_NO_HBA);
    971
    972	/* Disable all interrupt sources by resetting the controller */
    973	ahc_shutdown(ahc);
    974}
    975
    976static void
    977ahc_handle_seqint(struct ahc_softc *ahc, u_int intstat)
    978{
    979	struct scb *scb;
    980	struct ahc_devinfo devinfo;
    981
    982	ahc_fetch_devinfo(ahc, &devinfo);
    983
    984	/*
    985	 * Clear the upper byte that holds SEQINT status
    986	 * codes and clear the SEQINT bit. We will unpause
    987	 * the sequencer, if appropriate, after servicing
    988	 * the request.
    989	 */
    990	ahc_outb(ahc, CLRINT, CLRSEQINT);
    991	switch (intstat & SEQINT_MASK) {
    992	case BAD_STATUS:
    993	{
    994		u_int  scb_index;
    995		struct hardware_scb *hscb;
    996
    997		/*
    998		 * Set the default return value to 0 (don't
    999		 * send sense).  The sense code will change
   1000		 * this if needed.
   1001		 */
   1002		ahc_outb(ahc, RETURN_1, 0);
   1003
   1004		/*
   1005		 * The sequencer will notify us when a command
   1006		 * has an error that would be of interest to
   1007		 * the kernel.  This allows us to leave the sequencer
   1008		 * running in the common case of command completes
   1009		 * without error.  The sequencer will already have
   1010		 * dma'd the SCB back up to us, so we can reference
   1011		 * the in kernel copy directly.
   1012		 */
   1013		scb_index = ahc_inb(ahc, SCB_TAG);
   1014		scb = ahc_lookup_scb(ahc, scb_index);
   1015		if (scb == NULL) {
   1016			ahc_print_devinfo(ahc, &devinfo);
   1017			printk("ahc_intr - referenced scb "
   1018			       "not valid during seqint 0x%x scb(%d)\n",
   1019			       intstat, scb_index);
   1020			ahc_dump_card_state(ahc);
   1021			panic("for safety");
   1022			goto unpause;
   1023		}
   1024
   1025		hscb = scb->hscb;
   1026
   1027		/* Don't want to clobber the original sense code */
   1028		if ((scb->flags & SCB_SENSE) != 0) {
   1029			/*
   1030			 * Clear the SCB_SENSE Flag and have
   1031			 * the sequencer do a normal command
   1032			 * complete.
   1033			 */
   1034			scb->flags &= ~SCB_SENSE;
   1035			ahc_set_transaction_status(scb, CAM_AUTOSENSE_FAIL);
   1036			break;
   1037		}
   1038		ahc_set_transaction_status(scb, CAM_SCSI_STATUS_ERROR);
   1039		/* Freeze the queue until the client sees the error. */
   1040		ahc_freeze_devq(ahc, scb);
   1041		ahc_freeze_scb(scb);
   1042		ahc_set_scsi_status(scb, hscb->shared_data.status.scsi_status);
   1043		switch (hscb->shared_data.status.scsi_status) {
   1044		case SAM_STAT_GOOD:
   1045			printk("%s: Interrupted for status of 0???\n",
   1046			       ahc_name(ahc));
   1047			break;
   1048		case SAM_STAT_COMMAND_TERMINATED:
   1049		case SAM_STAT_CHECK_CONDITION:
   1050		{
   1051			struct ahc_dma_seg *sg;
   1052			struct scsi_sense *sc;
   1053			struct ahc_initiator_tinfo *targ_info;
   1054			struct ahc_tmode_tstate *tstate;
   1055			struct ahc_transinfo *tinfo;
   1056#ifdef AHC_DEBUG
   1057			if (ahc_debug & AHC_SHOW_SENSE) {
   1058				ahc_print_path(ahc, scb);
   1059				printk("SCB %d: requests Check Status\n",
   1060				       scb->hscb->tag);
   1061			}
   1062#endif
   1063
   1064			if (ahc_perform_autosense(scb) == 0)
   1065				break;
   1066
   1067			targ_info = ahc_fetch_transinfo(ahc,
   1068							devinfo.channel,
   1069							devinfo.our_scsiid,
   1070							devinfo.target,
   1071							&tstate);
   1072			tinfo = &targ_info->curr;
   1073			sg = scb->sg_list;
   1074			sc = (struct scsi_sense *)(&hscb->shared_data.cdb);
   1075			/*
   1076			 * Save off the residual if there is one.
   1077			 */
   1078			ahc_update_residual(ahc, scb);
   1079#ifdef AHC_DEBUG
   1080			if (ahc_debug & AHC_SHOW_SENSE) {
   1081				ahc_print_path(ahc, scb);
   1082				printk("Sending Sense\n");
   1083			}
   1084#endif
   1085			sg->addr = ahc_get_sense_bufaddr(ahc, scb);
   1086			sg->len = ahc_get_sense_bufsize(ahc, scb);
   1087			sg->len |= AHC_DMA_LAST_SEG;
   1088
   1089			/* Fixup byte order */
   1090			sg->addr = ahc_htole32(sg->addr);
   1091			sg->len = ahc_htole32(sg->len);
   1092
   1093			sc->opcode = REQUEST_SENSE;
   1094			sc->byte2 = 0;
   1095			if (tinfo->protocol_version <= SCSI_REV_2
   1096			 && SCB_GET_LUN(scb) < 8)
   1097				sc->byte2 = SCB_GET_LUN(scb) << 5;
   1098			sc->unused[0] = 0;
   1099			sc->unused[1] = 0;
   1100			sc->length = sg->len;
   1101			sc->control = 0;
   1102
   1103			/*
   1104			 * We can't allow the target to disconnect.
   1105			 * This will be an untagged transaction and
   1106			 * having the target disconnect will make this
   1107			 * transaction indestinguishable from outstanding
   1108			 * tagged transactions.
   1109			 */
   1110			hscb->control = 0;
   1111
   1112			/*
   1113			 * This request sense could be because the
   1114			 * the device lost power or in some other
   1115			 * way has lost our transfer negotiations.
   1116			 * Renegotiate if appropriate.  Unit attention
   1117			 * errors will be reported before any data
   1118			 * phases occur.
   1119			 */
   1120			if (ahc_get_residual(scb)
   1121			    == ahc_get_transfer_length(scb)) {
   1122				ahc_update_neg_request(ahc, &devinfo,
   1123						       tstate, targ_info,
   1124						       AHC_NEG_IF_NON_ASYNC);
   1125			}
   1126			if (tstate->auto_negotiate & devinfo.target_mask) {
   1127				hscb->control |= MK_MESSAGE;
   1128				scb->flags &= ~SCB_NEGOTIATE;
   1129				scb->flags |= SCB_AUTO_NEGOTIATE;
   1130			}
   1131			hscb->cdb_len = sizeof(*sc);
   1132			hscb->dataptr = sg->addr;
   1133			hscb->datacnt = sg->len;
   1134			hscb->sgptr = scb->sg_list_phys | SG_FULL_RESID;
   1135			hscb->sgptr = ahc_htole32(hscb->sgptr);
   1136			scb->sg_count = 1;
   1137			scb->flags |= SCB_SENSE;
   1138			ahc_qinfifo_requeue_tail(ahc, scb);
   1139			ahc_outb(ahc, RETURN_1, SEND_SENSE);
   1140			/*
   1141			 * Ensure we have enough time to actually
   1142			 * retrieve the sense.
   1143			 */
   1144			ahc_scb_timer_reset(scb, 5 * 1000000);
   1145			break;
   1146		}
   1147		default:
   1148			break;
   1149		}
   1150		break;
   1151	}
   1152	case NO_MATCH:
   1153	{
   1154		/* Ensure we don't leave the selection hardware on */
   1155		ahc_outb(ahc, SCSISEQ,
   1156			 ahc_inb(ahc, SCSISEQ) & (ENSELI|ENRSELI|ENAUTOATNP));
   1157
   1158		printk("%s:%c:%d: no active SCB for reconnecting "
   1159		       "target - issuing BUS DEVICE RESET\n",
   1160		       ahc_name(ahc), devinfo.channel, devinfo.target);
   1161		printk("SAVED_SCSIID == 0x%x, SAVED_LUN == 0x%x, "
   1162		       "ARG_1 == 0x%x ACCUM = 0x%x\n",
   1163		       ahc_inb(ahc, SAVED_SCSIID), ahc_inb(ahc, SAVED_LUN),
   1164		       ahc_inb(ahc, ARG_1), ahc_inb(ahc, ACCUM));
   1165		printk("SEQ_FLAGS == 0x%x, SCBPTR == 0x%x, BTT == 0x%x, "
   1166		       "SINDEX == 0x%x\n",
   1167		       ahc_inb(ahc, SEQ_FLAGS), ahc_inb(ahc, SCBPTR),
   1168		       ahc_index_busy_tcl(ahc,
   1169			    BUILD_TCL(ahc_inb(ahc, SAVED_SCSIID),
   1170				      ahc_inb(ahc, SAVED_LUN))),
   1171		       ahc_inb(ahc, SINDEX));
   1172		printk("SCSIID == 0x%x, SCB_SCSIID == 0x%x, SCB_LUN == 0x%x, "
   1173		       "SCB_TAG == 0x%x, SCB_CONTROL == 0x%x\n",
   1174		       ahc_inb(ahc, SCSIID), ahc_inb(ahc, SCB_SCSIID),
   1175		       ahc_inb(ahc, SCB_LUN), ahc_inb(ahc, SCB_TAG),
   1176		       ahc_inb(ahc, SCB_CONTROL));
   1177		printk("SCSIBUSL == 0x%x, SCSISIGI == 0x%x\n",
   1178		       ahc_inb(ahc, SCSIBUSL), ahc_inb(ahc, SCSISIGI));
   1179		printk("SXFRCTL0 == 0x%x\n", ahc_inb(ahc, SXFRCTL0));
   1180		printk("SEQCTL == 0x%x\n", ahc_inb(ahc, SEQCTL));
   1181		ahc_dump_card_state(ahc);
   1182		ahc->msgout_buf[0] = TARGET_RESET;
   1183		ahc->msgout_len = 1;
   1184		ahc->msgout_index = 0;
   1185		ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
   1186		ahc_outb(ahc, MSG_OUT, HOST_MSG);
   1187		ahc_assert_atn(ahc);
   1188		break;
   1189	}
   1190	case SEND_REJECT:
   1191	{
   1192		u_int rejbyte = ahc_inb(ahc, ACCUM);
   1193		printk("%s:%c:%d: Warning - unknown message received from "
   1194		       "target (0x%x).  Rejecting\n",
   1195		       ahc_name(ahc), devinfo.channel, devinfo.target, rejbyte);
   1196		break;
   1197	}
   1198	case PROTO_VIOLATION:
   1199	{
   1200		ahc_handle_proto_violation(ahc);
   1201		break;
   1202	}
   1203	case IGN_WIDE_RES:
   1204		ahc_handle_ign_wide_residue(ahc, &devinfo);
   1205		break;
   1206	case PDATA_REINIT:
   1207		ahc_reinitialize_dataptrs(ahc);
   1208		break;
   1209	case BAD_PHASE:
   1210	{
   1211		u_int lastphase;
   1212
   1213		lastphase = ahc_inb(ahc, LASTPHASE);
   1214		printk("%s:%c:%d: unknown scsi bus phase %x, "
   1215		       "lastphase = 0x%x.  Attempting to continue\n",
   1216		       ahc_name(ahc), devinfo.channel, devinfo.target,
   1217		       lastphase, ahc_inb(ahc, SCSISIGI));
   1218		break;
   1219	}
   1220	case MISSED_BUSFREE:
   1221	{
   1222		u_int lastphase;
   1223
   1224		lastphase = ahc_inb(ahc, LASTPHASE);
   1225		printk("%s:%c:%d: Missed busfree. "
   1226		       "Lastphase = 0x%x, Curphase = 0x%x\n",
   1227		       ahc_name(ahc), devinfo.channel, devinfo.target,
   1228		       lastphase, ahc_inb(ahc, SCSISIGI));
   1229		ahc_restart(ahc);
   1230		return;
   1231	}
   1232	case HOST_MSG_LOOP:
   1233	{
   1234		/*
   1235		 * The sequencer has encountered a message phase
   1236		 * that requires host assistance for completion.
   1237		 * While handling the message phase(s), we will be
   1238		 * notified by the sequencer after each byte is
   1239		 * transferred so we can track bus phase changes.
   1240		 *
   1241		 * If this is the first time we've seen a HOST_MSG_LOOP
   1242		 * interrupt, initialize the state of the host message
   1243		 * loop.
   1244		 */
   1245		if (ahc->msg_type == MSG_TYPE_NONE) {
   1246			struct scb *scb;
   1247			u_int scb_index;
   1248			u_int bus_phase;
   1249
   1250			bus_phase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
   1251			if (bus_phase != P_MESGIN
   1252			 && bus_phase != P_MESGOUT) {
   1253				printk("ahc_intr: HOST_MSG_LOOP bad "
   1254				       "phase 0x%x\n",
   1255				      bus_phase);
   1256				/*
   1257				 * Probably transitioned to bus free before
   1258				 * we got here.  Just punt the message.
   1259				 */
   1260				ahc_clear_intstat(ahc);
   1261				ahc_restart(ahc);
   1262				return;
   1263			}
   1264
   1265			scb_index = ahc_inb(ahc, SCB_TAG);
   1266			scb = ahc_lookup_scb(ahc, scb_index);
   1267			if (devinfo.role == ROLE_INITIATOR) {
   1268				if (bus_phase == P_MESGOUT) {
   1269					if (scb == NULL)
   1270						panic("HOST_MSG_LOOP with "
   1271						      "invalid SCB %x\n",
   1272						      scb_index);
   1273
   1274					ahc_setup_initiator_msgout(ahc,
   1275								   &devinfo,
   1276								   scb);
   1277				} else {
   1278					ahc->msg_type =
   1279					    MSG_TYPE_INITIATOR_MSGIN;
   1280					ahc->msgin_index = 0;
   1281				}
   1282			}
   1283#ifdef AHC_TARGET_MODE
   1284			else {
   1285				if (bus_phase == P_MESGOUT) {
   1286					ahc->msg_type =
   1287					    MSG_TYPE_TARGET_MSGOUT;
   1288					ahc->msgin_index = 0;
   1289				} else
   1290					ahc_setup_target_msgin(ahc,
   1291							       &devinfo,
   1292							       scb);
   1293			}
   1294#endif
   1295		}
   1296
   1297		ahc_handle_message_phase(ahc);
   1298		break;
   1299	}
   1300	case PERR_DETECTED:
   1301	{
   1302		/*
   1303		 * If we've cleared the parity error interrupt
   1304		 * but the sequencer still believes that SCSIPERR
   1305		 * is true, it must be that the parity error is
   1306		 * for the currently presented byte on the bus,
   1307		 * and we are not in a phase (data-in) where we will
   1308		 * eventually ack this byte.  Ack the byte and
   1309		 * throw it away in the hope that the target will
   1310		 * take us to message out to deliver the appropriate
   1311		 * error message.
   1312		 */
   1313		if ((intstat & SCSIINT) == 0
   1314		 && (ahc_inb(ahc, SSTAT1) & SCSIPERR) != 0) {
   1315
   1316			if ((ahc->features & AHC_DT) == 0) {
   1317				u_int curphase;
   1318
   1319				/*
   1320				 * The hardware will only let you ack bytes
   1321				 * if the expected phase in SCSISIGO matches
   1322				 * the current phase.  Make sure this is
   1323				 * currently the case.
   1324				 */
   1325				curphase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
   1326				ahc_outb(ahc, LASTPHASE, curphase);
   1327				ahc_outb(ahc, SCSISIGO, curphase);
   1328			}
   1329			if ((ahc_inb(ahc, SCSISIGI) & (CDI|MSGI)) == 0) {
   1330				int wait;
   1331
   1332				/*
   1333				 * In a data phase.  Faster to bitbucket
   1334				 * the data than to individually ack each
   1335				 * byte.  This is also the only strategy
   1336				 * that will work with AUTOACK enabled.
   1337				 */
   1338				ahc_outb(ahc, SXFRCTL1,
   1339					 ahc_inb(ahc, SXFRCTL1) | BITBUCKET);
   1340				wait = 5000;
   1341				while (--wait != 0) {
   1342					if ((ahc_inb(ahc, SCSISIGI)
   1343					  & (CDI|MSGI)) != 0)
   1344						break;
   1345					ahc_delay(100);
   1346				}
   1347				ahc_outb(ahc, SXFRCTL1,
   1348					 ahc_inb(ahc, SXFRCTL1) & ~BITBUCKET);
   1349				if (wait == 0) {
   1350					struct	scb *scb;
   1351					u_int	scb_index;
   1352
   1353					ahc_print_devinfo(ahc, &devinfo);
   1354					printk("Unable to clear parity error.  "
   1355					       "Resetting bus.\n");
   1356					scb_index = ahc_inb(ahc, SCB_TAG);
   1357					scb = ahc_lookup_scb(ahc, scb_index);
   1358					if (scb != NULL)
   1359						ahc_set_transaction_status(scb,
   1360						    CAM_UNCOR_PARITY);
   1361					ahc_reset_channel(ahc, devinfo.channel,
   1362							  /*init reset*/TRUE);
   1363				}
   1364			} else {
   1365				ahc_inb(ahc, SCSIDATL);
   1366			}
   1367		}
   1368		break;
   1369	}
   1370	case DATA_OVERRUN:
   1371	{
   1372		/*
   1373		 * When the sequencer detects an overrun, it
   1374		 * places the controller in "BITBUCKET" mode
   1375		 * and allows the target to complete its transfer.
   1376		 * Unfortunately, none of the counters get updated
   1377		 * when the controller is in this mode, so we have
   1378		 * no way of knowing how large the overrun was.
   1379		 */
   1380		u_int scbindex = ahc_inb(ahc, SCB_TAG);
   1381		u_int lastphase = ahc_inb(ahc, LASTPHASE);
   1382		u_int i;
   1383
   1384		scb = ahc_lookup_scb(ahc, scbindex);
   1385		for (i = 0; i < num_phases; i++) {
   1386			if (lastphase == ahc_phase_table[i].phase)
   1387				break;
   1388		}
   1389		ahc_print_path(ahc, scb);
   1390		printk("data overrun detected %s."
   1391		       "  Tag == 0x%x.\n",
   1392		       ahc_phase_table[i].phasemsg,
   1393		       scb->hscb->tag);
   1394		ahc_print_path(ahc, scb);
   1395		printk("%s seen Data Phase.  Length = %ld.  NumSGs = %d.\n",
   1396		       ahc_inb(ahc, SEQ_FLAGS) & DPHASE ? "Have" : "Haven't",
   1397		       ahc_get_transfer_length(scb), scb->sg_count);
   1398		if (scb->sg_count > 0) {
   1399			for (i = 0; i < scb->sg_count; i++) {
   1400
   1401				printk("sg[%d] - Addr 0x%x%x : Length %d\n",
   1402				       i,
   1403				       (ahc_le32toh(scb->sg_list[i].len) >> 24
   1404					& SG_HIGH_ADDR_BITS),
   1405				       ahc_le32toh(scb->sg_list[i].addr),
   1406				       ahc_le32toh(scb->sg_list[i].len)
   1407				       & AHC_SG_LEN_MASK);
   1408			}
   1409		}
   1410		/*
   1411		 * Set this and it will take effect when the
   1412		 * target does a command complete.
   1413		 */
   1414		ahc_freeze_devq(ahc, scb);
   1415		if ((scb->flags & SCB_SENSE) == 0) {
   1416			ahc_set_transaction_status(scb, CAM_DATA_RUN_ERR);
   1417		} else {
   1418			scb->flags &= ~SCB_SENSE;
   1419			ahc_set_transaction_status(scb, CAM_AUTOSENSE_FAIL);
   1420		}
   1421		ahc_freeze_scb(scb);
   1422
   1423		if ((ahc->features & AHC_ULTRA2) != 0) {
   1424			/*
   1425			 * Clear the channel in case we return
   1426			 * to data phase later.
   1427			 */
   1428			ahc_outb(ahc, SXFRCTL0,
   1429				 ahc_inb(ahc, SXFRCTL0) | CLRSTCNT|CLRCHN);
   1430			ahc_outb(ahc, SXFRCTL0,
   1431				 ahc_inb(ahc, SXFRCTL0) | CLRSTCNT|CLRCHN);
   1432		}
   1433		if ((ahc->flags & AHC_39BIT_ADDRESSING) != 0) {
   1434			u_int dscommand1;
   1435
   1436			/* Ensure HHADDR is 0 for future DMA operations. */
   1437			dscommand1 = ahc_inb(ahc, DSCOMMAND1);
   1438			ahc_outb(ahc, DSCOMMAND1, dscommand1 | HADDLDSEL0);
   1439			ahc_outb(ahc, HADDR, 0);
   1440			ahc_outb(ahc, DSCOMMAND1, dscommand1);
   1441		}
   1442		break;
   1443	}
   1444	case MKMSG_FAILED:
   1445	{
   1446		u_int scbindex;
   1447
   1448		printk("%s:%c:%d:%d: Attempt to issue message failed\n",
   1449		       ahc_name(ahc), devinfo.channel, devinfo.target,
   1450		       devinfo.lun);
   1451		scbindex = ahc_inb(ahc, SCB_TAG);
   1452		scb = ahc_lookup_scb(ahc, scbindex);
   1453		if (scb != NULL
   1454		 && (scb->flags & SCB_RECOVERY_SCB) != 0)
   1455			/*
   1456			 * Ensure that we didn't put a second instance of this
   1457			 * SCB into the QINFIFO.
   1458			 */
   1459			ahc_search_qinfifo(ahc, SCB_GET_TARGET(ahc, scb),
   1460					   SCB_GET_CHANNEL(ahc, scb),
   1461					   SCB_GET_LUN(scb), scb->hscb->tag,
   1462					   ROLE_INITIATOR, /*status*/0,
   1463					   SEARCH_REMOVE);
   1464		break;
   1465	}
   1466	case NO_FREE_SCB:
   1467	{
   1468		printk("%s: No free or disconnected SCBs\n", ahc_name(ahc));
   1469		ahc_dump_card_state(ahc);
   1470		panic("for safety");
   1471		break;
   1472	}
   1473	case SCB_MISMATCH:
   1474	{
   1475		u_int scbptr;
   1476
   1477		scbptr = ahc_inb(ahc, SCBPTR);
   1478		printk("Bogus TAG after DMA.  SCBPTR %d, tag %d, our tag %d\n",
   1479		       scbptr, ahc_inb(ahc, ARG_1),
   1480		       ahc->scb_data->hscbs[scbptr].tag);
   1481		ahc_dump_card_state(ahc);
   1482		panic("for safety");
   1483		break;
   1484	}
   1485	case OUT_OF_RANGE:
   1486	{
   1487		printk("%s: BTT calculation out of range\n", ahc_name(ahc));
   1488		printk("SAVED_SCSIID == 0x%x, SAVED_LUN == 0x%x, "
   1489		       "ARG_1 == 0x%x ACCUM = 0x%x\n",
   1490		       ahc_inb(ahc, SAVED_SCSIID), ahc_inb(ahc, SAVED_LUN),
   1491		       ahc_inb(ahc, ARG_1), ahc_inb(ahc, ACCUM));
   1492		printk("SEQ_FLAGS == 0x%x, SCBPTR == 0x%x, BTT == 0x%x, "
   1493		       "SINDEX == 0x%x\n, A == 0x%x\n",
   1494		       ahc_inb(ahc, SEQ_FLAGS), ahc_inb(ahc, SCBPTR),
   1495		       ahc_index_busy_tcl(ahc,
   1496			    BUILD_TCL(ahc_inb(ahc, SAVED_SCSIID),
   1497				      ahc_inb(ahc, SAVED_LUN))),
   1498		       ahc_inb(ahc, SINDEX),
   1499		       ahc_inb(ahc, ACCUM));
   1500		printk("SCSIID == 0x%x, SCB_SCSIID == 0x%x, SCB_LUN == 0x%x, "
   1501		       "SCB_TAG == 0x%x, SCB_CONTROL == 0x%x\n",
   1502		       ahc_inb(ahc, SCSIID), ahc_inb(ahc, SCB_SCSIID),
   1503		       ahc_inb(ahc, SCB_LUN), ahc_inb(ahc, SCB_TAG),
   1504		       ahc_inb(ahc, SCB_CONTROL));
   1505		printk("SCSIBUSL == 0x%x, SCSISIGI == 0x%x\n",
   1506		       ahc_inb(ahc, SCSIBUSL), ahc_inb(ahc, SCSISIGI));
   1507		ahc_dump_card_state(ahc);
   1508		panic("for safety");
   1509		break;
   1510	}
   1511	default:
   1512		printk("ahc_intr: seqint, "
   1513		       "intstat == 0x%x, scsisigi = 0x%x\n",
   1514		       intstat, ahc_inb(ahc, SCSISIGI));
   1515		break;
   1516	}
   1517unpause:
   1518	/*
   1519	 *  The sequencer is paused immediately on
   1520	 *  a SEQINT, so we should restart it when
   1521	 *  we're done.
   1522	 */
   1523	ahc_unpause(ahc);
   1524}
   1525
   1526static void
   1527ahc_handle_scsiint(struct ahc_softc *ahc, u_int intstat)
   1528{
   1529	u_int	scb_index;
   1530	u_int	status0;
   1531	u_int	status;
   1532	struct	scb *scb;
   1533	char	cur_channel;
   1534	char	intr_channel;
   1535
   1536	if ((ahc->features & AHC_TWIN) != 0
   1537	 && ((ahc_inb(ahc, SBLKCTL) & SELBUSB) != 0))
   1538		cur_channel = 'B';
   1539	else
   1540		cur_channel = 'A';
   1541	intr_channel = cur_channel;
   1542
   1543	if ((ahc->features & AHC_ULTRA2) != 0)
   1544		status0 = ahc_inb(ahc, SSTAT0) & IOERR;
   1545	else
   1546		status0 = 0;
   1547	status = ahc_inb(ahc, SSTAT1) & (SELTO|SCSIRSTI|BUSFREE|SCSIPERR);
   1548	if (status == 0 && status0 == 0) {
   1549		if ((ahc->features & AHC_TWIN) != 0) {
   1550			/* Try the other channel */
   1551			ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) ^ SELBUSB);
   1552			status = ahc_inb(ahc, SSTAT1)
   1553			       & (SELTO|SCSIRSTI|BUSFREE|SCSIPERR);
   1554			intr_channel = (cur_channel == 'A') ? 'B' : 'A';
   1555		}
   1556		if (status == 0) {
   1557			printk("%s: Spurious SCSI interrupt\n", ahc_name(ahc));
   1558			ahc_outb(ahc, CLRINT, CLRSCSIINT);
   1559			ahc_unpause(ahc);
   1560			return;
   1561		}
   1562	}
   1563
   1564	/* Make sure the sequencer is in a safe location. */
   1565	ahc_clear_critical_section(ahc);
   1566
   1567	scb_index = ahc_inb(ahc, SCB_TAG);
   1568	scb = ahc_lookup_scb(ahc, scb_index);
   1569	if (scb != NULL
   1570	 && (ahc_inb(ahc, SEQ_FLAGS) & NOT_IDENTIFIED) != 0)
   1571		scb = NULL;
   1572
   1573	if ((ahc->features & AHC_ULTRA2) != 0
   1574	 && (status0 & IOERR) != 0) {
   1575		int now_lvd;
   1576
   1577		now_lvd = ahc_inb(ahc, SBLKCTL) & ENAB40;
   1578		printk("%s: Transceiver State Has Changed to %s mode\n",
   1579		       ahc_name(ahc), now_lvd ? "LVD" : "SE");
   1580		ahc_outb(ahc, CLRSINT0, CLRIOERR);
   1581		/*
   1582		 * When transitioning to SE mode, the reset line
   1583		 * glitches, triggering an arbitration bug in some
   1584		 * Ultra2 controllers.  This bug is cleared when we
   1585		 * assert the reset line.  Since a reset glitch has
   1586		 * already occurred with this transition and a
   1587		 * transceiver state change is handled just like
   1588		 * a bus reset anyway, asserting the reset line
   1589		 * ourselves is safe.
   1590		 */
   1591		ahc_reset_channel(ahc, intr_channel,
   1592				 /*Initiate Reset*/now_lvd == 0);
   1593	} else if ((status & SCSIRSTI) != 0) {
   1594		printk("%s: Someone reset channel %c\n",
   1595			ahc_name(ahc), intr_channel);
   1596		if (intr_channel != cur_channel)
   1597			ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) ^ SELBUSB);
   1598		ahc_reset_channel(ahc, intr_channel, /*Initiate Reset*/FALSE);
   1599	} else if ((status & SCSIPERR) != 0) {
   1600		/*
   1601		 * Determine the bus phase and queue an appropriate message.
   1602		 * SCSIPERR is latched true as soon as a parity error
   1603		 * occurs.  If the sequencer acked the transfer that
   1604		 * caused the parity error and the currently presented
   1605		 * transfer on the bus has correct parity, SCSIPERR will
   1606		 * be cleared by CLRSCSIPERR.  Use this to determine if
   1607		 * we should look at the last phase the sequencer recorded,
   1608		 * or the current phase presented on the bus.
   1609		 */
   1610		struct	ahc_devinfo devinfo;
   1611		u_int	mesg_out;
   1612		u_int	curphase;
   1613		u_int	errorphase;
   1614		u_int	lastphase;
   1615		u_int	scsirate;
   1616		u_int	i;
   1617		u_int	sstat2;
   1618		int	silent;
   1619
   1620		lastphase = ahc_inb(ahc, LASTPHASE);
   1621		curphase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
   1622		sstat2 = ahc_inb(ahc, SSTAT2);
   1623		ahc_outb(ahc, CLRSINT1, CLRSCSIPERR);
   1624		/*
   1625		 * For all phases save DATA, the sequencer won't
   1626		 * automatically ack a byte that has a parity error
   1627		 * in it.  So the only way that the current phase
   1628		 * could be 'data-in' is if the parity error is for
   1629		 * an already acked byte in the data phase.  During
   1630		 * synchronous data-in transfers, we may actually
   1631		 * ack bytes before latching the current phase in
   1632		 * LASTPHASE, leading to the discrepancy between
   1633		 * curphase and lastphase.
   1634		 */
   1635		if ((ahc_inb(ahc, SSTAT1) & SCSIPERR) != 0
   1636		 || curphase == P_DATAIN || curphase == P_DATAIN_DT)
   1637			errorphase = curphase;
   1638		else
   1639			errorphase = lastphase;
   1640
   1641		for (i = 0; i < num_phases; i++) {
   1642			if (errorphase == ahc_phase_table[i].phase)
   1643				break;
   1644		}
   1645		mesg_out = ahc_phase_table[i].mesg_out;
   1646		silent = FALSE;
   1647		if (scb != NULL) {
   1648			if (SCB_IS_SILENT(scb))
   1649				silent = TRUE;
   1650			else
   1651				ahc_print_path(ahc, scb);
   1652			scb->flags |= SCB_TRANSMISSION_ERROR;
   1653		} else
   1654			printk("%s:%c:%d: ", ahc_name(ahc), intr_channel,
   1655			       SCSIID_TARGET(ahc, ahc_inb(ahc, SAVED_SCSIID)));
   1656		scsirate = ahc_inb(ahc, SCSIRATE);
   1657		if (silent == FALSE) {
   1658			printk("parity error detected %s. "
   1659			       "SEQADDR(0x%x) SCSIRATE(0x%x)\n",
   1660			       ahc_phase_table[i].phasemsg,
   1661			       ahc_inw(ahc, SEQADDR0),
   1662			       scsirate);
   1663			if ((ahc->features & AHC_DT) != 0) {
   1664				if ((sstat2 & CRCVALERR) != 0)
   1665					printk("\tCRC Value Mismatch\n");
   1666				if ((sstat2 & CRCENDERR) != 0)
   1667					printk("\tNo terminal CRC packet "
   1668					       "received\n");
   1669				if ((sstat2 & CRCREQERR) != 0)
   1670					printk("\tIllegal CRC packet "
   1671					       "request\n");
   1672				if ((sstat2 & DUAL_EDGE_ERR) != 0)
   1673					printk("\tUnexpected %sDT Data Phase\n",
   1674					       (scsirate & SINGLE_EDGE)
   1675					     ? "" : "non-");
   1676			}
   1677		}
   1678
   1679		if ((ahc->features & AHC_DT) != 0
   1680		 && (sstat2 & DUAL_EDGE_ERR) != 0) {
   1681			/*
   1682			 * This error applies regardless of
   1683			 * data direction, so ignore the value
   1684			 * in the phase table.
   1685			 */
   1686			mesg_out = INITIATOR_ERROR;
   1687		}
   1688
   1689		/*
   1690		 * We've set the hardware to assert ATN if we
   1691		 * get a parity error on "in" phases, so all we
   1692		 * need to do is stuff the message buffer with
   1693		 * the appropriate message.  "In" phases have set
   1694		 * mesg_out to something other than MSG_NOP.
   1695		 */
   1696		if (mesg_out != NOP) {
   1697			if (ahc->msg_type != MSG_TYPE_NONE)
   1698				ahc->send_msg_perror = TRUE;
   1699			else
   1700				ahc_outb(ahc, MSG_OUT, mesg_out);
   1701		}
   1702		/*
   1703		 * Force a renegotiation with this target just in
   1704		 * case we are out of sync for some external reason
   1705		 * unknown (or unreported) by the target.
   1706		 */
   1707		ahc_fetch_devinfo(ahc, &devinfo);
   1708		ahc_force_renegotiation(ahc, &devinfo);
   1709
   1710		ahc_outb(ahc, CLRINT, CLRSCSIINT);
   1711		ahc_unpause(ahc);
   1712	} else if ((status & SELTO) != 0) {
   1713		u_int	scbptr;
   1714
   1715		/* Stop the selection */
   1716		ahc_outb(ahc, SCSISEQ, 0);
   1717
   1718		/* No more pending messages */
   1719		ahc_clear_msg_state(ahc);
   1720
   1721		/* Clear interrupt state */
   1722		ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) & ~ENBUSFREE);
   1723		ahc_outb(ahc, CLRSINT1, CLRSELTIMEO|CLRBUSFREE|CLRSCSIPERR);
   1724
   1725		/*
   1726		 * Although the driver does not care about the
   1727		 * 'Selection in Progress' status bit, the busy
   1728		 * LED does.  SELINGO is only cleared by a successful
   1729		 * selection, so we must manually clear it to insure
   1730		 * the LED turns off just incase no future successful
   1731		 * selections occur (e.g. no devices on the bus).
   1732		 */
   1733		ahc_outb(ahc, CLRSINT0, CLRSELINGO);
   1734
   1735		scbptr = ahc_inb(ahc, WAITING_SCBH);
   1736		ahc_outb(ahc, SCBPTR, scbptr);
   1737		scb_index = ahc_inb(ahc, SCB_TAG);
   1738
   1739		scb = ahc_lookup_scb(ahc, scb_index);
   1740		if (scb == NULL) {
   1741			printk("%s: ahc_intr - referenced scb not "
   1742			       "valid during SELTO scb(%d, %d)\n",
   1743			       ahc_name(ahc), scbptr, scb_index);
   1744			ahc_dump_card_state(ahc);
   1745		} else {
   1746			struct ahc_devinfo devinfo;
   1747#ifdef AHC_DEBUG
   1748			if ((ahc_debug & AHC_SHOW_SELTO) != 0) {
   1749				ahc_print_path(ahc, scb);
   1750				printk("Saw Selection Timeout for SCB 0x%x\n",
   1751				       scb_index);
   1752			}
   1753#endif
   1754			ahc_scb_devinfo(ahc, &devinfo, scb);
   1755			ahc_set_transaction_status(scb, CAM_SEL_TIMEOUT);
   1756			ahc_freeze_devq(ahc, scb);
   1757
   1758			/*
   1759			 * Cancel any pending transactions on the device
   1760			 * now that it seems to be missing.  This will
   1761			 * also revert us to async/narrow transfers until
   1762			 * we can renegotiate with the device.
   1763			 */
   1764			ahc_handle_devreset(ahc, &devinfo,
   1765					    CAM_SEL_TIMEOUT,
   1766					    "Selection Timeout",
   1767					    /*verbose_level*/1);
   1768		}
   1769		ahc_outb(ahc, CLRINT, CLRSCSIINT);
   1770		ahc_restart(ahc);
   1771	} else if ((status & BUSFREE) != 0
   1772		&& (ahc_inb(ahc, SIMODE1) & ENBUSFREE) != 0) {
   1773		struct	ahc_devinfo devinfo;
   1774		u_int	lastphase;
   1775		u_int	saved_scsiid;
   1776		u_int	saved_lun;
   1777		u_int	target;
   1778		u_int	initiator_role_id;
   1779		char	channel;
   1780		int	printerror;
   1781
   1782		/*
   1783		 * Clear our selection hardware as soon as possible.
   1784		 * We may have an entry in the waiting Q for this target,
   1785		 * that is affected by this busfree and we don't want to
   1786		 * go about selecting the target while we handle the event.
   1787		 */
   1788		ahc_outb(ahc, SCSISEQ,
   1789			 ahc_inb(ahc, SCSISEQ) & (ENSELI|ENRSELI|ENAUTOATNP));
   1790
   1791		/*
   1792		 * Disable busfree interrupts and clear the busfree
   1793		 * interrupt status.  We do this here so that several
   1794		 * bus transactions occur prior to clearing the SCSIINT
   1795		 * latch.  It can take a bit for the clearing to take effect.
   1796		 */
   1797		ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) & ~ENBUSFREE);
   1798		ahc_outb(ahc, CLRSINT1, CLRBUSFREE|CLRSCSIPERR);
   1799
   1800		/*
   1801		 * Look at what phase we were last in.
   1802		 * If its message out, chances are pretty good
   1803		 * that the busfree was in response to one of
   1804		 * our abort requests.
   1805		 */
   1806		lastphase = ahc_inb(ahc, LASTPHASE);
   1807		saved_scsiid = ahc_inb(ahc, SAVED_SCSIID);
   1808		saved_lun = ahc_inb(ahc, SAVED_LUN);
   1809		target = SCSIID_TARGET(ahc, saved_scsiid);
   1810		initiator_role_id = SCSIID_OUR_ID(saved_scsiid);
   1811		channel = SCSIID_CHANNEL(ahc, saved_scsiid);
   1812		ahc_compile_devinfo(&devinfo, initiator_role_id,
   1813				    target, saved_lun, channel, ROLE_INITIATOR);
   1814		printerror = 1;
   1815
   1816		if (lastphase == P_MESGOUT) {
   1817			u_int tag;
   1818
   1819			tag = SCB_LIST_NULL;
   1820			if (ahc_sent_msg(ahc, AHCMSG_1B, ABORT_TASK, TRUE)
   1821			 || ahc_sent_msg(ahc, AHCMSG_1B, ABORT_TASK_SET, TRUE)) {
   1822				if (ahc->msgout_buf[ahc->msgout_index - 1]
   1823				 == ABORT_TASK)
   1824					tag = scb->hscb->tag;
   1825				ahc_print_path(ahc, scb);
   1826				printk("SCB %d - Abort%s Completed.\n",
   1827				       scb->hscb->tag, tag == SCB_LIST_NULL ?
   1828				       "" : " Tag");
   1829				ahc_abort_scbs(ahc, target, channel,
   1830					       saved_lun, tag,
   1831					       ROLE_INITIATOR,
   1832					       CAM_REQ_ABORTED);
   1833				printerror = 0;
   1834			} else if (ahc_sent_msg(ahc, AHCMSG_1B,
   1835						TARGET_RESET, TRUE)) {
   1836				ahc_compile_devinfo(&devinfo,
   1837						    initiator_role_id,
   1838						    target,
   1839						    CAM_LUN_WILDCARD,
   1840						    channel,
   1841						    ROLE_INITIATOR);
   1842				ahc_handle_devreset(ahc, &devinfo,
   1843						    CAM_BDR_SENT,
   1844						    "Bus Device Reset",
   1845						    /*verbose_level*/0);
   1846				printerror = 0;
   1847			} else if (ahc_sent_msg(ahc, AHCMSG_EXT,
   1848						EXTENDED_PPR, FALSE)) {
   1849				struct ahc_initiator_tinfo *tinfo;
   1850				struct ahc_tmode_tstate *tstate;
   1851
   1852				/*
   1853				 * PPR Rejected.  Try non-ppr negotiation
   1854				 * and retry command.
   1855				 */
   1856				tinfo = ahc_fetch_transinfo(ahc,
   1857							    devinfo.channel,
   1858							    devinfo.our_scsiid,
   1859							    devinfo.target,
   1860							    &tstate);
   1861				tinfo->curr.transport_version = 2;
   1862				tinfo->goal.transport_version = 2;
   1863				tinfo->goal.ppr_options = 0;
   1864				ahc_qinfifo_requeue_tail(ahc, scb);
   1865				printerror = 0;
   1866			} else if (ahc_sent_msg(ahc, AHCMSG_EXT,
   1867						EXTENDED_WDTR, FALSE)) {
   1868				/*
   1869				 * Negotiation Rejected.  Go-narrow and
   1870				 * retry command.
   1871				 */
   1872				ahc_set_width(ahc, &devinfo,
   1873					      MSG_EXT_WDTR_BUS_8_BIT,
   1874					      AHC_TRANS_CUR|AHC_TRANS_GOAL,
   1875					      /*paused*/TRUE);
   1876				ahc_qinfifo_requeue_tail(ahc, scb);
   1877				printerror = 0;
   1878			} else if (ahc_sent_msg(ahc, AHCMSG_EXT,
   1879						EXTENDED_SDTR, FALSE)) {
   1880				/*
   1881				 * Negotiation Rejected.  Go-async and
   1882				 * retry command.
   1883				 */
   1884				ahc_set_syncrate(ahc, &devinfo,
   1885						/*syncrate*/NULL,
   1886						/*period*/0, /*offset*/0,
   1887						/*ppr_options*/0,
   1888						AHC_TRANS_CUR|AHC_TRANS_GOAL,
   1889						/*paused*/TRUE);
   1890				ahc_qinfifo_requeue_tail(ahc, scb);
   1891				printerror = 0;
   1892			}
   1893		}
   1894		if (printerror != 0) {
   1895			u_int i;
   1896
   1897			if (scb != NULL) {
   1898				u_int tag;
   1899
   1900				if ((scb->hscb->control & TAG_ENB) != 0)
   1901					tag = scb->hscb->tag;
   1902				else
   1903					tag = SCB_LIST_NULL;
   1904				ahc_print_path(ahc, scb);
   1905				ahc_abort_scbs(ahc, target, channel,
   1906					       SCB_GET_LUN(scb), tag,
   1907					       ROLE_INITIATOR,
   1908					       CAM_UNEXP_BUSFREE);
   1909			} else {
   1910				/*
   1911				 * We had not fully identified this connection,
   1912				 * so we cannot abort anything.
   1913				 */
   1914				printk("%s: ", ahc_name(ahc));
   1915			}
   1916			for (i = 0; i < num_phases; i++) {
   1917				if (lastphase == ahc_phase_table[i].phase)
   1918					break;
   1919			}
   1920			if (lastphase != P_BUSFREE) {
   1921				/*
   1922				 * Renegotiate with this device at the
   1923				 * next opportunity just in case this busfree
   1924				 * is due to a negotiation mismatch with the
   1925				 * device.
   1926				 */
   1927				ahc_force_renegotiation(ahc, &devinfo);
   1928			}
   1929			printk("Unexpected busfree %s\n"
   1930			       "SEQADDR == 0x%x\n",
   1931			       ahc_phase_table[i].phasemsg,
   1932			       ahc_inb(ahc, SEQADDR0)
   1933				| (ahc_inb(ahc, SEQADDR1) << 8));
   1934		}
   1935		ahc_outb(ahc, CLRINT, CLRSCSIINT);
   1936		ahc_restart(ahc);
   1937	} else {
   1938		printk("%s: Missing case in ahc_handle_scsiint. status = %x\n",
   1939		       ahc_name(ahc), status);
   1940		ahc_outb(ahc, CLRINT, CLRSCSIINT);
   1941	}
   1942}
   1943
   1944/*
   1945 * Force renegotiation to occur the next time we initiate
   1946 * a command to the current device.
   1947 */
   1948static void
   1949ahc_force_renegotiation(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
   1950{
   1951	struct	ahc_initiator_tinfo *targ_info;
   1952	struct	ahc_tmode_tstate *tstate;
   1953
   1954	targ_info = ahc_fetch_transinfo(ahc,
   1955					devinfo->channel,
   1956					devinfo->our_scsiid,
   1957					devinfo->target,
   1958					&tstate);
   1959	ahc_update_neg_request(ahc, devinfo, tstate,
   1960			       targ_info, AHC_NEG_IF_NON_ASYNC);
   1961}
   1962
   1963#define AHC_MAX_STEPS 2000
   1964static void
   1965ahc_clear_critical_section(struct ahc_softc *ahc)
   1966{
   1967	int	stepping;
   1968	int	steps;
   1969	u_int	simode0;
   1970	u_int	simode1;
   1971
   1972	if (ahc->num_critical_sections == 0)
   1973		return;
   1974
   1975	stepping = FALSE;
   1976	steps = 0;
   1977	simode0 = 0;
   1978	simode1 = 0;
   1979	for (;;) {
   1980		struct	cs *cs;
   1981		u_int	seqaddr;
   1982		u_int	i;
   1983
   1984		seqaddr = ahc_inb(ahc, SEQADDR0)
   1985			| (ahc_inb(ahc, SEQADDR1) << 8);
   1986
   1987		/*
   1988		 * Seqaddr represents the next instruction to execute,
   1989		 * so we are really executing the instruction just
   1990		 * before it.
   1991		 */
   1992		if (seqaddr != 0)
   1993			seqaddr -= 1;
   1994		cs = ahc->critical_sections;
   1995		for (i = 0; i < ahc->num_critical_sections; i++, cs++) {
   1996			if (cs->begin < seqaddr && cs->end >= seqaddr)
   1997				break;
   1998		}
   1999
   2000		if (i == ahc->num_critical_sections)
   2001			break;
   2002
   2003		if (steps > AHC_MAX_STEPS) {
   2004			printk("%s: Infinite loop in critical section\n",
   2005			       ahc_name(ahc));
   2006			ahc_dump_card_state(ahc);
   2007			panic("critical section loop");
   2008		}
   2009
   2010		steps++;
   2011		if (stepping == FALSE) {
   2012
   2013			/*
   2014			 * Disable all interrupt sources so that the
   2015			 * sequencer will not be stuck by a pausing
   2016			 * interrupt condition while we attempt to
   2017			 * leave a critical section.
   2018			 */
   2019			simode0 = ahc_inb(ahc, SIMODE0);
   2020			ahc_outb(ahc, SIMODE0, 0);
   2021			simode1 = ahc_inb(ahc, SIMODE1);
   2022			if ((ahc->features & AHC_DT) != 0)
   2023				/*
   2024				 * On DT class controllers, we
   2025				 * use the enhanced busfree logic.
   2026				 * Unfortunately we cannot re-enable
   2027				 * busfree detection within the
   2028				 * current connection, so we must
   2029				 * leave it on while single stepping.
   2030				 */
   2031				ahc_outb(ahc, SIMODE1, simode1 & ENBUSFREE);
   2032			else
   2033				ahc_outb(ahc, SIMODE1, 0);
   2034			ahc_outb(ahc, CLRINT, CLRSCSIINT);
   2035			ahc_outb(ahc, SEQCTL, ahc->seqctl | STEP);
   2036			stepping = TRUE;
   2037		}
   2038		if ((ahc->features & AHC_DT) != 0) {
   2039			ahc_outb(ahc, CLRSINT1, CLRBUSFREE);
   2040			ahc_outb(ahc, CLRINT, CLRSCSIINT);
   2041		}
   2042		ahc_outb(ahc, HCNTRL, ahc->unpause);
   2043		while (!ahc_is_paused(ahc))
   2044			ahc_delay(200);
   2045	}
   2046	if (stepping) {
   2047		ahc_outb(ahc, SIMODE0, simode0);
   2048		ahc_outb(ahc, SIMODE1, simode1);
   2049		ahc_outb(ahc, SEQCTL, ahc->seqctl);
   2050	}
   2051}
   2052
   2053/*
   2054 * Clear any pending interrupt status.
   2055 */
   2056static void
   2057ahc_clear_intstat(struct ahc_softc *ahc)
   2058{
   2059	/* Clear any interrupt conditions this may have caused */
   2060	ahc_outb(ahc, CLRSINT1, CLRSELTIMEO|CLRATNO|CLRSCSIRSTI
   2061				|CLRBUSFREE|CLRSCSIPERR|CLRPHASECHG|
   2062				CLRREQINIT);
   2063	ahc_flush_device_writes(ahc);
   2064	ahc_outb(ahc, CLRSINT0, CLRSELDO|CLRSELDI|CLRSELINGO);
   2065	ahc_flush_device_writes(ahc);
   2066	ahc_outb(ahc, CLRINT, CLRSCSIINT);
   2067	ahc_flush_device_writes(ahc);
   2068}
   2069
   2070/**************************** Debugging Routines ******************************/
   2071#ifdef AHC_DEBUG
   2072uint32_t ahc_debug = AHC_DEBUG_OPTS;
   2073#endif
   2074
   2075#if 0 /* unused */
   2076static void
   2077ahc_print_scb(struct scb *scb)
   2078{
   2079	int i;
   2080
   2081	struct hardware_scb *hscb = scb->hscb;
   2082
   2083	printk("scb:%p control:0x%x scsiid:0x%x lun:%d cdb_len:%d\n",
   2084	       (void *)scb,
   2085	       hscb->control,
   2086	       hscb->scsiid,
   2087	       hscb->lun,
   2088	       hscb->cdb_len);
   2089	printk("Shared Data: ");
   2090	for (i = 0; i < sizeof(hscb->shared_data.cdb); i++)
   2091		printk("%#02x", hscb->shared_data.cdb[i]);
   2092	printk("        dataptr:%#x datacnt:%#x sgptr:%#x tag:%#x\n",
   2093		ahc_le32toh(hscb->dataptr),
   2094		ahc_le32toh(hscb->datacnt),
   2095		ahc_le32toh(hscb->sgptr),
   2096		hscb->tag);
   2097	if (scb->sg_count > 0) {
   2098		for (i = 0; i < scb->sg_count; i++) {
   2099			printk("sg[%d] - Addr 0x%x%x : Length %d\n",
   2100			       i,
   2101			       (ahc_le32toh(scb->sg_list[i].len) >> 24
   2102				& SG_HIGH_ADDR_BITS),
   2103			       ahc_le32toh(scb->sg_list[i].addr),
   2104			       ahc_le32toh(scb->sg_list[i].len));
   2105		}
   2106	}
   2107}
   2108#endif
   2109
   2110/************************* Transfer Negotiation *******************************/
   2111/*
   2112 * Allocate per target mode instance (ID we respond to as a target)
   2113 * transfer negotiation data structures.
   2114 */
   2115static struct ahc_tmode_tstate *
   2116ahc_alloc_tstate(struct ahc_softc *ahc, u_int scsi_id, char channel)
   2117{
   2118	struct ahc_tmode_tstate *master_tstate;
   2119	struct ahc_tmode_tstate *tstate;
   2120	int i;
   2121
   2122	master_tstate = ahc->enabled_targets[ahc->our_id];
   2123	if (channel == 'B') {
   2124		scsi_id += 8;
   2125		master_tstate = ahc->enabled_targets[ahc->our_id_b + 8];
   2126	}
   2127	if (ahc->enabled_targets[scsi_id] != NULL
   2128	 && ahc->enabled_targets[scsi_id] != master_tstate)
   2129		panic("%s: ahc_alloc_tstate - Target already allocated",
   2130		      ahc_name(ahc));
   2131	tstate = kmalloc(sizeof(*tstate), GFP_ATOMIC);
   2132	if (tstate == NULL)
   2133		return (NULL);
   2134
   2135	/*
   2136	 * If we have allocated a master tstate, copy user settings from
   2137	 * the master tstate (taken from SRAM or the EEPROM) for this
   2138	 * channel, but reset our current and goal settings to async/narrow
   2139	 * until an initiator talks to us.
   2140	 */
   2141	if (master_tstate != NULL) {
   2142		memcpy(tstate, master_tstate, sizeof(*tstate));
   2143		memset(tstate->enabled_luns, 0, sizeof(tstate->enabled_luns));
   2144		tstate->ultraenb = 0;
   2145		for (i = 0; i < AHC_NUM_TARGETS; i++) {
   2146			memset(&tstate->transinfo[i].curr, 0,
   2147			      sizeof(tstate->transinfo[i].curr));
   2148			memset(&tstate->transinfo[i].goal, 0,
   2149			      sizeof(tstate->transinfo[i].goal));
   2150		}
   2151	} else
   2152		memset(tstate, 0, sizeof(*tstate));
   2153	ahc->enabled_targets[scsi_id] = tstate;
   2154	return (tstate);
   2155}
   2156
   2157#ifdef AHC_TARGET_MODE
   2158/*
   2159 * Free per target mode instance (ID we respond to as a target)
   2160 * transfer negotiation data structures.
   2161 */
   2162static void
   2163ahc_free_tstate(struct ahc_softc *ahc, u_int scsi_id, char channel, int force)
   2164{
   2165	struct ahc_tmode_tstate *tstate;
   2166
   2167	/*
   2168	 * Don't clean up our "master" tstate.
   2169	 * It has our default user settings.
   2170	 */
   2171	if (((channel == 'B' && scsi_id == ahc->our_id_b)
   2172	  || (channel == 'A' && scsi_id == ahc->our_id))
   2173	 && force == FALSE)
   2174		return;
   2175
   2176	if (channel == 'B')
   2177		scsi_id += 8;
   2178	tstate = ahc->enabled_targets[scsi_id];
   2179	kfree(tstate);
   2180	ahc->enabled_targets[scsi_id] = NULL;
   2181}
   2182#endif
   2183
   2184/*
   2185 * Called when we have an active connection to a target on the bus,
   2186 * this function finds the nearest syncrate to the input period limited
   2187 * by the capabilities of the bus connectivity of and sync settings for
   2188 * the target.
   2189 */
   2190static const struct ahc_syncrate *
   2191ahc_devlimited_syncrate(struct ahc_softc *ahc,
   2192			struct ahc_initiator_tinfo *tinfo,
   2193			u_int *period, u_int *ppr_options, role_t role)
   2194{
   2195	struct	ahc_transinfo *transinfo;
   2196	u_int	maxsync;
   2197
   2198	if ((ahc->features & AHC_ULTRA2) != 0) {
   2199		if ((ahc_inb(ahc, SBLKCTL) & ENAB40) != 0
   2200		 && (ahc_inb(ahc, SSTAT2) & EXP_ACTIVE) == 0) {
   2201			maxsync = AHC_SYNCRATE_DT;
   2202		} else {
   2203			maxsync = AHC_SYNCRATE_ULTRA;
   2204			/* Can't do DT on an SE bus */
   2205			*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
   2206		}
   2207	} else if ((ahc->features & AHC_ULTRA) != 0) {
   2208		maxsync = AHC_SYNCRATE_ULTRA;
   2209	} else {
   2210		maxsync = AHC_SYNCRATE_FAST;
   2211	}
   2212	/*
   2213	 * Never allow a value higher than our current goal
   2214	 * period otherwise we may allow a target initiated
   2215	 * negotiation to go above the limit as set by the
   2216	 * user.  In the case of an initiator initiated
   2217	 * sync negotiation, we limit based on the user
   2218	 * setting.  This allows the system to still accept
   2219	 * incoming negotiations even if target initiated
   2220	 * negotiation is not performed.
   2221	 */
   2222	if (role == ROLE_TARGET)
   2223		transinfo = &tinfo->user;
   2224	else
   2225		transinfo = &tinfo->goal;
   2226	*ppr_options &= transinfo->ppr_options;
   2227	if (transinfo->width == MSG_EXT_WDTR_BUS_8_BIT) {
   2228		maxsync = max(maxsync, (u_int)AHC_SYNCRATE_ULTRA2);
   2229		*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
   2230	}
   2231	if (transinfo->period == 0) {
   2232		*period = 0;
   2233		*ppr_options = 0;
   2234		return (NULL);
   2235	}
   2236	*period = max(*period, (u_int)transinfo->period);
   2237	return (ahc_find_syncrate(ahc, period, ppr_options, maxsync));
   2238}
   2239
   2240/*
   2241 * Look up the valid period to SCSIRATE conversion in our table.
   2242 * Return the period and offset that should be sent to the target
   2243 * if this was the beginning of an SDTR.
   2244 */
   2245const struct ahc_syncrate *
   2246ahc_find_syncrate(struct ahc_softc *ahc, u_int *period,
   2247		  u_int *ppr_options, u_int maxsync)
   2248{
   2249	const struct ahc_syncrate *syncrate;
   2250
   2251	if ((ahc->features & AHC_DT) == 0)
   2252		*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
   2253
   2254	/* Skip all DT only entries if DT is not available */
   2255	if ((*ppr_options & MSG_EXT_PPR_DT_REQ) == 0
   2256	 && maxsync < AHC_SYNCRATE_ULTRA2)
   2257		maxsync = AHC_SYNCRATE_ULTRA2;
   2258
   2259	/* Now set the maxsync based on the card capabilities
   2260	 * DT is already done above */
   2261	if ((ahc->features & (AHC_DT | AHC_ULTRA2)) == 0
   2262	    && maxsync < AHC_SYNCRATE_ULTRA)
   2263		maxsync = AHC_SYNCRATE_ULTRA;
   2264	if ((ahc->features & (AHC_DT | AHC_ULTRA2 | AHC_ULTRA)) == 0
   2265	    && maxsync < AHC_SYNCRATE_FAST)
   2266		maxsync = AHC_SYNCRATE_FAST;
   2267
   2268	for (syncrate = &ahc_syncrates[maxsync];
   2269	     syncrate->rate != NULL;
   2270	     syncrate++) {
   2271
   2272		/*
   2273		 * The Ultra2 table doesn't go as low
   2274		 * as for the Fast/Ultra cards.
   2275		 */
   2276		if ((ahc->features & AHC_ULTRA2) != 0
   2277		 && (syncrate->sxfr_u2 == 0))
   2278			break;
   2279
   2280		if (*period <= syncrate->period) {
   2281			/*
   2282			 * When responding to a target that requests
   2283			 * sync, the requested rate may fall between
   2284			 * two rates that we can output, but still be
   2285			 * a rate that we can receive.  Because of this,
   2286			 * we want to respond to the target with
   2287			 * the same rate that it sent to us even
   2288			 * if the period we use to send data to it
   2289			 * is lower.  Only lower the response period
   2290			 * if we must.
   2291			 */
   2292			if (syncrate == &ahc_syncrates[maxsync])
   2293				*period = syncrate->period;
   2294
   2295			/*
   2296			 * At some speeds, we only support
   2297			 * ST transfers.
   2298			 */
   2299			if ((syncrate->sxfr_u2 & ST_SXFR) != 0)
   2300				*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
   2301			break;
   2302		}
   2303	}
   2304
   2305	if ((*period == 0)
   2306	 || (syncrate->rate == NULL)
   2307	 || ((ahc->features & AHC_ULTRA2) != 0
   2308	  && (syncrate->sxfr_u2 == 0))) {
   2309		/* Use asynchronous transfers. */
   2310		*period = 0;
   2311		syncrate = NULL;
   2312		*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
   2313	}
   2314	return (syncrate);
   2315}
   2316
   2317/*
   2318 * Convert from an entry in our syncrate table to the SCSI equivalent
   2319 * sync "period" factor.
   2320 */
   2321u_int
   2322ahc_find_period(struct ahc_softc *ahc, u_int scsirate, u_int maxsync)
   2323{
   2324	const struct ahc_syncrate *syncrate;
   2325
   2326	if ((ahc->features & AHC_ULTRA2) != 0)
   2327		scsirate &= SXFR_ULTRA2;
   2328	else
   2329		scsirate &= SXFR;
   2330
   2331	/* now set maxsync based on card capabilities */
   2332	if ((ahc->features & AHC_DT) == 0 && maxsync < AHC_SYNCRATE_ULTRA2)
   2333		maxsync = AHC_SYNCRATE_ULTRA2;
   2334	if ((ahc->features & (AHC_DT | AHC_ULTRA2)) == 0
   2335	    && maxsync < AHC_SYNCRATE_ULTRA)
   2336		maxsync = AHC_SYNCRATE_ULTRA;
   2337	if ((ahc->features & (AHC_DT | AHC_ULTRA2 | AHC_ULTRA)) == 0
   2338	    && maxsync < AHC_SYNCRATE_FAST)
   2339		maxsync = AHC_SYNCRATE_FAST;
   2340
   2341
   2342	syncrate = &ahc_syncrates[maxsync];
   2343	while (syncrate->rate != NULL) {
   2344
   2345		if ((ahc->features & AHC_ULTRA2) != 0) {
   2346			if (syncrate->sxfr_u2 == 0)
   2347				break;
   2348			else if (scsirate == (syncrate->sxfr_u2 & SXFR_ULTRA2))
   2349				return (syncrate->period);
   2350		} else if (scsirate == (syncrate->sxfr & SXFR)) {
   2351				return (syncrate->period);
   2352		}
   2353		syncrate++;
   2354	}
   2355	return (0); /* async */
   2356}
   2357
   2358/*
   2359 * Truncate the given synchronous offset to a value the
   2360 * current adapter type and syncrate are capable of.
   2361 */
   2362static void
   2363ahc_validate_offset(struct ahc_softc *ahc,
   2364		    struct ahc_initiator_tinfo *tinfo,
   2365		    const struct ahc_syncrate *syncrate,
   2366		    u_int *offset, int wide, role_t role)
   2367{
   2368	u_int maxoffset;
   2369
   2370	/* Limit offset to what we can do */
   2371	if (syncrate == NULL) {
   2372		maxoffset = 0;
   2373	} else if ((ahc->features & AHC_ULTRA2) != 0) {
   2374		maxoffset = MAX_OFFSET_ULTRA2;
   2375	} else {
   2376		if (wide)
   2377			maxoffset = MAX_OFFSET_16BIT;
   2378		else
   2379			maxoffset = MAX_OFFSET_8BIT;
   2380	}
   2381	*offset = min(*offset, maxoffset);
   2382	if (tinfo != NULL) {
   2383		if (role == ROLE_TARGET)
   2384			*offset = min(*offset, (u_int)tinfo->user.offset);
   2385		else
   2386			*offset = min(*offset, (u_int)tinfo->goal.offset);
   2387	}
   2388}
   2389
   2390/*
   2391 * Truncate the given transfer width parameter to a value the
   2392 * current adapter type is capable of.
   2393 */
   2394static void
   2395ahc_validate_width(struct ahc_softc *ahc, struct ahc_initiator_tinfo *tinfo,
   2396		   u_int *bus_width, role_t role)
   2397{
   2398	switch (*bus_width) {
   2399	default:
   2400		if (ahc->features & AHC_WIDE) {
   2401			/* Respond Wide */
   2402			*bus_width = MSG_EXT_WDTR_BUS_16_BIT;
   2403			break;
   2404		}
   2405		fallthrough;
   2406	case MSG_EXT_WDTR_BUS_8_BIT:
   2407		*bus_width = MSG_EXT_WDTR_BUS_8_BIT;
   2408		break;
   2409	}
   2410	if (tinfo != NULL) {
   2411		if (role == ROLE_TARGET)
   2412			*bus_width = min((u_int)tinfo->user.width, *bus_width);
   2413		else
   2414			*bus_width = min((u_int)tinfo->goal.width, *bus_width);
   2415	}
   2416}
   2417
   2418/*
   2419 * Update the bitmask of targets for which the controller should
   2420 * negotiate with at the next convenient opportunity.  This currently
   2421 * means the next time we send the initial identify messages for
   2422 * a new transaction.
   2423 */
   2424int
   2425ahc_update_neg_request(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   2426		       struct ahc_tmode_tstate *tstate,
   2427		       struct ahc_initiator_tinfo *tinfo, ahc_neg_type neg_type)
   2428{
   2429	u_int auto_negotiate_orig;
   2430
   2431	auto_negotiate_orig = tstate->auto_negotiate;
   2432	if (neg_type == AHC_NEG_ALWAYS) {
   2433		/*
   2434		 * Force our "current" settings to be
   2435		 * unknown so that unless a bus reset
   2436		 * occurs the need to renegotiate is
   2437		 * recorded persistently.
   2438		 */
   2439		if ((ahc->features & AHC_WIDE) != 0)
   2440			tinfo->curr.width = AHC_WIDTH_UNKNOWN;
   2441		tinfo->curr.period = AHC_PERIOD_UNKNOWN;
   2442		tinfo->curr.offset = AHC_OFFSET_UNKNOWN;
   2443	}
   2444	if (tinfo->curr.period != tinfo->goal.period
   2445	 || tinfo->curr.width != tinfo->goal.width
   2446	 || tinfo->curr.offset != tinfo->goal.offset
   2447	 || tinfo->curr.ppr_options != tinfo->goal.ppr_options
   2448	 || (neg_type == AHC_NEG_IF_NON_ASYNC
   2449	  && (tinfo->goal.offset != 0
   2450	   || tinfo->goal.width != MSG_EXT_WDTR_BUS_8_BIT
   2451	   || tinfo->goal.ppr_options != 0)))
   2452		tstate->auto_negotiate |= devinfo->target_mask;
   2453	else
   2454		tstate->auto_negotiate &= ~devinfo->target_mask;
   2455
   2456	return (auto_negotiate_orig != tstate->auto_negotiate);
   2457}
   2458
   2459/*
   2460 * Update the user/goal/curr tables of synchronous negotiation
   2461 * parameters as well as, in the case of a current or active update,
   2462 * any data structures on the host controller.  In the case of an
   2463 * active update, the specified target is currently talking to us on
   2464 * the bus, so the transfer parameter update must take effect
   2465 * immediately.
   2466 */
   2467void
   2468ahc_set_syncrate(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   2469		 const struct ahc_syncrate *syncrate, u_int period,
   2470		 u_int offset, u_int ppr_options, u_int type, int paused)
   2471{
   2472	struct	ahc_initiator_tinfo *tinfo;
   2473	struct	ahc_tmode_tstate *tstate;
   2474	u_int	old_period;
   2475	u_int	old_offset;
   2476	u_int	old_ppr;
   2477	int	active;
   2478	int	update_needed;
   2479
   2480	active = (type & AHC_TRANS_ACTIVE) == AHC_TRANS_ACTIVE;
   2481	update_needed = 0;
   2482
   2483	if (syncrate == NULL) {
   2484		period = 0;
   2485		offset = 0;
   2486	}
   2487
   2488	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
   2489				    devinfo->target, &tstate);
   2490
   2491	if ((type & AHC_TRANS_USER) != 0) {
   2492		tinfo->user.period = period;
   2493		tinfo->user.offset = offset;
   2494		tinfo->user.ppr_options = ppr_options;
   2495	}
   2496
   2497	if ((type & AHC_TRANS_GOAL) != 0) {
   2498		tinfo->goal.period = period;
   2499		tinfo->goal.offset = offset;
   2500		tinfo->goal.ppr_options = ppr_options;
   2501	}
   2502
   2503	old_period = tinfo->curr.period;
   2504	old_offset = tinfo->curr.offset;
   2505	old_ppr	   = tinfo->curr.ppr_options;
   2506
   2507	if ((type & AHC_TRANS_CUR) != 0
   2508	 && (old_period != period
   2509	  || old_offset != offset
   2510	  || old_ppr != ppr_options)) {
   2511		u_int	scsirate;
   2512
   2513		update_needed++;
   2514		scsirate = tinfo->scsirate;
   2515		if ((ahc->features & AHC_ULTRA2) != 0) {
   2516
   2517			scsirate &= ~(SXFR_ULTRA2|SINGLE_EDGE|ENABLE_CRC);
   2518			if (syncrate != NULL) {
   2519				scsirate |= syncrate->sxfr_u2;
   2520				if ((ppr_options & MSG_EXT_PPR_DT_REQ) != 0)
   2521					scsirate |= ENABLE_CRC;
   2522				else
   2523					scsirate |= SINGLE_EDGE;
   2524			}
   2525		} else {
   2526
   2527			scsirate &= ~(SXFR|SOFS);
   2528			/*
   2529			 * Ensure Ultra mode is set properly for
   2530			 * this target.
   2531			 */
   2532			tstate->ultraenb &= ~devinfo->target_mask;
   2533			if (syncrate != NULL) {
   2534				if (syncrate->sxfr & ULTRA_SXFR) {
   2535					tstate->ultraenb |=
   2536						devinfo->target_mask;
   2537				}
   2538				scsirate |= syncrate->sxfr & SXFR;
   2539				scsirate |= offset & SOFS;
   2540			}
   2541			if (active) {
   2542				u_int sxfrctl0;
   2543
   2544				sxfrctl0 = ahc_inb(ahc, SXFRCTL0);
   2545				sxfrctl0 &= ~FAST20;
   2546				if (tstate->ultraenb & devinfo->target_mask)
   2547					sxfrctl0 |= FAST20;
   2548				ahc_outb(ahc, SXFRCTL0, sxfrctl0);
   2549			}
   2550		}
   2551		if (active) {
   2552			ahc_outb(ahc, SCSIRATE, scsirate);
   2553			if ((ahc->features & AHC_ULTRA2) != 0)
   2554				ahc_outb(ahc, SCSIOFFSET, offset);
   2555		}
   2556
   2557		tinfo->scsirate = scsirate;
   2558		tinfo->curr.period = period;
   2559		tinfo->curr.offset = offset;
   2560		tinfo->curr.ppr_options = ppr_options;
   2561
   2562		ahc_send_async(ahc, devinfo->channel, devinfo->target,
   2563			       CAM_LUN_WILDCARD, AC_TRANSFER_NEG);
   2564		if (bootverbose) {
   2565			if (offset != 0) {
   2566				printk("%s: target %d synchronous at %sMHz%s, "
   2567				       "offset = 0x%x\n", ahc_name(ahc),
   2568				       devinfo->target, syncrate->rate,
   2569				       (ppr_options & MSG_EXT_PPR_DT_REQ)
   2570				       ? " DT" : "", offset);
   2571			} else {
   2572				printk("%s: target %d using "
   2573				       "asynchronous transfers\n",
   2574				       ahc_name(ahc), devinfo->target);
   2575			}
   2576		}
   2577	}
   2578
   2579	update_needed += ahc_update_neg_request(ahc, devinfo, tstate,
   2580						tinfo, AHC_NEG_TO_GOAL);
   2581
   2582	if (update_needed)
   2583		ahc_update_pending_scbs(ahc);
   2584}
   2585
   2586/*
   2587 * Update the user/goal/curr tables of wide negotiation
   2588 * parameters as well as, in the case of a current or active update,
   2589 * any data structures on the host controller.  In the case of an
   2590 * active update, the specified target is currently talking to us on
   2591 * the bus, so the transfer parameter update must take effect
   2592 * immediately.
   2593 */
   2594void
   2595ahc_set_width(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   2596	      u_int width, u_int type, int paused)
   2597{
   2598	struct	ahc_initiator_tinfo *tinfo;
   2599	struct	ahc_tmode_tstate *tstate;
   2600	u_int	oldwidth;
   2601	int	active;
   2602	int	update_needed;
   2603
   2604	active = (type & AHC_TRANS_ACTIVE) == AHC_TRANS_ACTIVE;
   2605	update_needed = 0;
   2606	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
   2607				    devinfo->target, &tstate);
   2608
   2609	if ((type & AHC_TRANS_USER) != 0)
   2610		tinfo->user.width = width;
   2611
   2612	if ((type & AHC_TRANS_GOAL) != 0)
   2613		tinfo->goal.width = width;
   2614
   2615	oldwidth = tinfo->curr.width;
   2616	if ((type & AHC_TRANS_CUR) != 0 && oldwidth != width) {
   2617		u_int	scsirate;
   2618
   2619		update_needed++;
   2620		scsirate =  tinfo->scsirate;
   2621		scsirate &= ~WIDEXFER;
   2622		if (width == MSG_EXT_WDTR_BUS_16_BIT)
   2623			scsirate |= WIDEXFER;
   2624
   2625		tinfo->scsirate = scsirate;
   2626
   2627		if (active)
   2628			ahc_outb(ahc, SCSIRATE, scsirate);
   2629
   2630		tinfo->curr.width = width;
   2631
   2632		ahc_send_async(ahc, devinfo->channel, devinfo->target,
   2633			       CAM_LUN_WILDCARD, AC_TRANSFER_NEG);
   2634		if (bootverbose) {
   2635			printk("%s: target %d using %dbit transfers\n",
   2636			       ahc_name(ahc), devinfo->target,
   2637			       8 * (0x01 << width));
   2638		}
   2639	}
   2640
   2641	update_needed += ahc_update_neg_request(ahc, devinfo, tstate,
   2642						tinfo, AHC_NEG_TO_GOAL);
   2643	if (update_needed)
   2644		ahc_update_pending_scbs(ahc);
   2645}
   2646
   2647/*
   2648 * Update the current state of tagged queuing for a given target.
   2649 */
   2650static void
   2651ahc_set_tags(struct ahc_softc *ahc, struct scsi_cmnd *cmd,
   2652	     struct ahc_devinfo *devinfo, ahc_queue_alg alg)
   2653{
   2654	struct scsi_device *sdev = cmd->device;
   2655
   2656	ahc_platform_set_tags(ahc, sdev, devinfo, alg);
   2657	ahc_send_async(ahc, devinfo->channel, devinfo->target,
   2658		       devinfo->lun, AC_TRANSFER_NEG);
   2659}
   2660
   2661/*
   2662 * When the transfer settings for a connection change, update any
   2663 * in-transit SCBs to contain the new data so the hardware will
   2664 * be set correctly during future (re)selections.
   2665 */
   2666static void
   2667ahc_update_pending_scbs(struct ahc_softc *ahc)
   2668{
   2669	struct	scb *pending_scb;
   2670	int	pending_scb_count;
   2671	int	i;
   2672	int	paused;
   2673	u_int	saved_scbptr;
   2674
   2675	/*
   2676	 * Traverse the pending SCB list and ensure that all of the
   2677	 * SCBs there have the proper settings.
   2678	 */
   2679	pending_scb_count = 0;
   2680	LIST_FOREACH(pending_scb, &ahc->pending_scbs, pending_links) {
   2681		struct ahc_devinfo devinfo;
   2682		struct hardware_scb *pending_hscb;
   2683		struct ahc_initiator_tinfo *tinfo;
   2684		struct ahc_tmode_tstate *tstate;
   2685
   2686		ahc_scb_devinfo(ahc, &devinfo, pending_scb);
   2687		tinfo = ahc_fetch_transinfo(ahc, devinfo.channel,
   2688					    devinfo.our_scsiid,
   2689					    devinfo.target, &tstate);
   2690		pending_hscb = pending_scb->hscb;
   2691		pending_hscb->control &= ~ULTRAENB;
   2692		if ((tstate->ultraenb & devinfo.target_mask) != 0)
   2693			pending_hscb->control |= ULTRAENB;
   2694		pending_hscb->scsirate = tinfo->scsirate;
   2695		pending_hscb->scsioffset = tinfo->curr.offset;
   2696		if ((tstate->auto_negotiate & devinfo.target_mask) == 0
   2697		 && (pending_scb->flags & SCB_AUTO_NEGOTIATE) != 0) {
   2698			pending_scb->flags &= ~SCB_AUTO_NEGOTIATE;
   2699			pending_hscb->control &= ~MK_MESSAGE;
   2700		}
   2701		ahc_sync_scb(ahc, pending_scb,
   2702			     BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
   2703		pending_scb_count++;
   2704	}
   2705
   2706	if (pending_scb_count == 0)
   2707		return;
   2708
   2709	if (ahc_is_paused(ahc)) {
   2710		paused = 1;
   2711	} else {
   2712		paused = 0;
   2713		ahc_pause(ahc);
   2714	}
   2715
   2716	saved_scbptr = ahc_inb(ahc, SCBPTR);
   2717	/* Ensure that the hscbs down on the card match the new information */
   2718	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
   2719		struct	hardware_scb *pending_hscb;
   2720		u_int	control;
   2721		u_int	scb_tag;
   2722
   2723		ahc_outb(ahc, SCBPTR, i);
   2724		scb_tag = ahc_inb(ahc, SCB_TAG);
   2725		pending_scb = ahc_lookup_scb(ahc, scb_tag);
   2726		if (pending_scb == NULL)
   2727			continue;
   2728
   2729		pending_hscb = pending_scb->hscb;
   2730		control = ahc_inb(ahc, SCB_CONTROL);
   2731		control &= ~(ULTRAENB|MK_MESSAGE);
   2732		control |= pending_hscb->control & (ULTRAENB|MK_MESSAGE);
   2733		ahc_outb(ahc, SCB_CONTROL, control);
   2734		ahc_outb(ahc, SCB_SCSIRATE, pending_hscb->scsirate);
   2735		ahc_outb(ahc, SCB_SCSIOFFSET, pending_hscb->scsioffset);
   2736	}
   2737	ahc_outb(ahc, SCBPTR, saved_scbptr);
   2738
   2739	if (paused == 0)
   2740		ahc_unpause(ahc);
   2741}
   2742
   2743/**************************** Pathing Information *****************************/
   2744static void
   2745ahc_fetch_devinfo(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
   2746{
   2747	u_int	saved_scsiid;
   2748	role_t	role;
   2749	int	our_id;
   2750
   2751	if (ahc_inb(ahc, SSTAT0) & TARGET)
   2752		role = ROLE_TARGET;
   2753	else
   2754		role = ROLE_INITIATOR;
   2755
   2756	if (role == ROLE_TARGET
   2757	    && (ahc->features & AHC_MULTI_TID) != 0
   2758	    && (ahc_inb(ahc, SEQ_FLAGS)
   2759	     & (CMDPHASE_PENDING|TARG_CMD_PENDING|NO_DISCONNECT)) != 0) {
   2760		/* We were selected, so pull our id from TARGIDIN */
   2761		our_id = ahc_inb(ahc, TARGIDIN) & OID;
   2762	} else if ((ahc->features & AHC_ULTRA2) != 0)
   2763		our_id = ahc_inb(ahc, SCSIID_ULTRA2) & OID;
   2764	else
   2765		our_id = ahc_inb(ahc, SCSIID) & OID;
   2766
   2767	saved_scsiid = ahc_inb(ahc, SAVED_SCSIID);
   2768	ahc_compile_devinfo(devinfo,
   2769			    our_id,
   2770			    SCSIID_TARGET(ahc, saved_scsiid),
   2771			    ahc_inb(ahc, SAVED_LUN),
   2772			    SCSIID_CHANNEL(ahc, saved_scsiid),
   2773			    role);
   2774}
   2775
   2776static const struct ahc_phase_table_entry*
   2777ahc_lookup_phase_entry(int phase)
   2778{
   2779	const struct ahc_phase_table_entry *entry;
   2780	const struct ahc_phase_table_entry *last_entry;
   2781
   2782	/*
   2783	 * num_phases doesn't include the default entry which
   2784	 * will be returned if the phase doesn't match.
   2785	 */
   2786	last_entry = &ahc_phase_table[num_phases];
   2787	for (entry = ahc_phase_table; entry < last_entry; entry++) {
   2788		if (phase == entry->phase)
   2789			break;
   2790	}
   2791	return (entry);
   2792}
   2793
   2794void
   2795ahc_compile_devinfo(struct ahc_devinfo *devinfo, u_int our_id, u_int target,
   2796		    u_int lun, char channel, role_t role)
   2797{
   2798	devinfo->our_scsiid = our_id;
   2799	devinfo->target = target;
   2800	devinfo->lun = lun;
   2801	devinfo->target_offset = target;
   2802	devinfo->channel = channel;
   2803	devinfo->role = role;
   2804	if (channel == 'B')
   2805		devinfo->target_offset += 8;
   2806	devinfo->target_mask = (0x01 << devinfo->target_offset);
   2807}
   2808
   2809void
   2810ahc_print_devinfo(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
   2811{
   2812	printk("%s:%c:%d:%d: ", ahc_name(ahc), devinfo->channel,
   2813	       devinfo->target, devinfo->lun);
   2814}
   2815
   2816static void
   2817ahc_scb_devinfo(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   2818		struct scb *scb)
   2819{
   2820	role_t	role;
   2821	int	our_id;
   2822
   2823	our_id = SCSIID_OUR_ID(scb->hscb->scsiid);
   2824	role = ROLE_INITIATOR;
   2825	if ((scb->flags & SCB_TARGET_SCB) != 0)
   2826		role = ROLE_TARGET;
   2827	ahc_compile_devinfo(devinfo, our_id, SCB_GET_TARGET(ahc, scb),
   2828			    SCB_GET_LUN(scb), SCB_GET_CHANNEL(ahc, scb), role);
   2829}
   2830
   2831
   2832/************************ Message Phase Processing ****************************/
   2833static void
   2834ahc_assert_atn(struct ahc_softc *ahc)
   2835{
   2836	u_int scsisigo;
   2837
   2838	scsisigo = ATNO;
   2839	if ((ahc->features & AHC_DT) == 0)
   2840		scsisigo |= ahc_inb(ahc, SCSISIGI);
   2841	ahc_outb(ahc, SCSISIGO, scsisigo);
   2842}
   2843
   2844/*
   2845 * When an initiator transaction with the MK_MESSAGE flag either reconnects
   2846 * or enters the initial message out phase, we are interrupted.  Fill our
   2847 * outgoing message buffer with the appropriate message and beging handing
   2848 * the message phase(s) manually.
   2849 */
   2850static void
   2851ahc_setup_initiator_msgout(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   2852			   struct scb *scb)
   2853{
   2854	/*
   2855	 * To facilitate adding multiple messages together,
   2856	 * each routine should increment the index and len
   2857	 * variables instead of setting them explicitly.
   2858	 */
   2859	ahc->msgout_index = 0;
   2860	ahc->msgout_len = 0;
   2861
   2862	if ((scb->flags & SCB_DEVICE_RESET) == 0
   2863	 && ahc_inb(ahc, MSG_OUT) == MSG_IDENTIFYFLAG) {
   2864		u_int identify_msg;
   2865
   2866		identify_msg = MSG_IDENTIFYFLAG | SCB_GET_LUN(scb);
   2867		if ((scb->hscb->control & DISCENB) != 0)
   2868			identify_msg |= MSG_IDENTIFY_DISCFLAG;
   2869		ahc->msgout_buf[ahc->msgout_index++] = identify_msg;
   2870		ahc->msgout_len++;
   2871
   2872		if ((scb->hscb->control & TAG_ENB) != 0) {
   2873			ahc->msgout_buf[ahc->msgout_index++] =
   2874			    scb->hscb->control & (TAG_ENB|SCB_TAG_TYPE);
   2875			ahc->msgout_buf[ahc->msgout_index++] = scb->hscb->tag;
   2876			ahc->msgout_len += 2;
   2877		}
   2878	}
   2879
   2880	if (scb->flags & SCB_DEVICE_RESET) {
   2881		ahc->msgout_buf[ahc->msgout_index++] = TARGET_RESET;
   2882		ahc->msgout_len++;
   2883		ahc_print_path(ahc, scb);
   2884		printk("Bus Device Reset Message Sent\n");
   2885		/*
   2886		 * Clear our selection hardware in advance of
   2887		 * the busfree.  We may have an entry in the waiting
   2888		 * Q for this target, and we don't want to go about
   2889		 * selecting while we handle the busfree and blow it
   2890		 * away.
   2891		 */
   2892		ahc_outb(ahc, SCSISEQ, (ahc_inb(ahc, SCSISEQ) & ~ENSELO));
   2893	} else if ((scb->flags & SCB_ABORT) != 0) {
   2894		if ((scb->hscb->control & TAG_ENB) != 0)
   2895			ahc->msgout_buf[ahc->msgout_index++] = ABORT_TASK;
   2896		else
   2897			ahc->msgout_buf[ahc->msgout_index++] = ABORT_TASK_SET;
   2898		ahc->msgout_len++;
   2899		ahc_print_path(ahc, scb);
   2900		printk("Abort%s Message Sent\n",
   2901		       (scb->hscb->control & TAG_ENB) != 0 ? " Tag" : "");
   2902		/*
   2903		 * Clear our selection hardware in advance of
   2904		 * the busfree.  We may have an entry in the waiting
   2905		 * Q for this target, and we don't want to go about
   2906		 * selecting while we handle the busfree and blow it
   2907		 * away.
   2908		 */
   2909		ahc_outb(ahc, SCSISEQ, (ahc_inb(ahc, SCSISEQ) & ~ENSELO));
   2910	} else if ((scb->flags & (SCB_AUTO_NEGOTIATE|SCB_NEGOTIATE)) != 0) {
   2911		ahc_build_transfer_msg(ahc, devinfo);
   2912	} else {
   2913		printk("ahc_intr: AWAITING_MSG for an SCB that "
   2914		       "does not have a waiting message\n");
   2915		printk("SCSIID = %x, target_mask = %x\n", scb->hscb->scsiid,
   2916		       devinfo->target_mask);
   2917		panic("SCB = %d, SCB Control = %x, MSG_OUT = %x "
   2918		      "SCB flags = %x", scb->hscb->tag, scb->hscb->control,
   2919		      ahc_inb(ahc, MSG_OUT), scb->flags);
   2920	}
   2921
   2922	/*
   2923	 * Clear the MK_MESSAGE flag from the SCB so we aren't
   2924	 * asked to send this message again.
   2925	 */
   2926	ahc_outb(ahc, SCB_CONTROL, ahc_inb(ahc, SCB_CONTROL) & ~MK_MESSAGE);
   2927	scb->hscb->control &= ~MK_MESSAGE;
   2928	ahc->msgout_index = 0;
   2929	ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
   2930}
   2931
   2932/*
   2933 * Build an appropriate transfer negotiation message for the
   2934 * currently active target.
   2935 */
   2936static void
   2937ahc_build_transfer_msg(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
   2938{
   2939	/*
   2940	 * We need to initiate transfer negotiations.
   2941	 * If our current and goal settings are identical,
   2942	 * we want to renegotiate due to a check condition.
   2943	 */
   2944	struct	ahc_initiator_tinfo *tinfo;
   2945	struct	ahc_tmode_tstate *tstate;
   2946	const struct ahc_syncrate *rate;
   2947	int	dowide;
   2948	int	dosync;
   2949	int	doppr;
   2950	u_int	period;
   2951	u_int	ppr_options;
   2952	u_int	offset;
   2953
   2954	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
   2955				    devinfo->target, &tstate);
   2956	/*
   2957	 * Filter our period based on the current connection.
   2958	 * If we can't perform DT transfers on this segment (not in LVD
   2959	 * mode for instance), then our decision to issue a PPR message
   2960	 * may change.
   2961	 */
   2962	period = tinfo->goal.period;
   2963	offset = tinfo->goal.offset;
   2964	ppr_options = tinfo->goal.ppr_options;
   2965	/* Target initiated PPR is not allowed in the SCSI spec */
   2966	if (devinfo->role == ROLE_TARGET)
   2967		ppr_options = 0;
   2968	rate = ahc_devlimited_syncrate(ahc, tinfo, &period,
   2969				       &ppr_options, devinfo->role);
   2970	dowide = tinfo->curr.width != tinfo->goal.width;
   2971	dosync = tinfo->curr.offset != offset || tinfo->curr.period != period;
   2972	/*
   2973	 * Only use PPR if we have options that need it, even if the device
   2974	 * claims to support it.  There might be an expander in the way
   2975	 * that doesn't.
   2976	 */
   2977	doppr = ppr_options != 0;
   2978
   2979	if (!dowide && !dosync && !doppr) {
   2980		dowide = tinfo->goal.width != MSG_EXT_WDTR_BUS_8_BIT;
   2981		dosync = tinfo->goal.offset != 0;
   2982	}
   2983
   2984	if (!dowide && !dosync && !doppr) {
   2985		/*
   2986		 * Force async with a WDTR message if we have a wide bus,
   2987		 * or just issue an SDTR with a 0 offset.
   2988		 */
   2989		if ((ahc->features & AHC_WIDE) != 0)
   2990			dowide = 1;
   2991		else
   2992			dosync = 1;
   2993
   2994		if (bootverbose) {
   2995			ahc_print_devinfo(ahc, devinfo);
   2996			printk("Ensuring async\n");
   2997		}
   2998	}
   2999
   3000	/* Target initiated PPR is not allowed in the SCSI spec */
   3001	if (devinfo->role == ROLE_TARGET)
   3002		doppr = 0;
   3003
   3004	/*
   3005	 * Both the PPR message and SDTR message require the
   3006	 * goal syncrate to be limited to what the target device
   3007	 * is capable of handling (based on whether an LVD->SE
   3008	 * expander is on the bus), so combine these two cases.
   3009	 * Regardless, guarantee that if we are using WDTR and SDTR
   3010	 * messages that WDTR comes first.
   3011	 */
   3012	if (doppr || (dosync && !dowide)) {
   3013
   3014		offset = tinfo->goal.offset;
   3015		ahc_validate_offset(ahc, tinfo, rate, &offset,
   3016				    doppr ? tinfo->goal.width
   3017					  : tinfo->curr.width,
   3018				    devinfo->role);
   3019		if (doppr) {
   3020			ahc_construct_ppr(ahc, devinfo, period, offset,
   3021					  tinfo->goal.width, ppr_options);
   3022		} else {
   3023			ahc_construct_sdtr(ahc, devinfo, period, offset);
   3024		}
   3025	} else {
   3026		ahc_construct_wdtr(ahc, devinfo, tinfo->goal.width);
   3027	}
   3028}
   3029
   3030/*
   3031 * Build a synchronous negotiation message in our message
   3032 * buffer based on the input parameters.
   3033 */
   3034static void
   3035ahc_construct_sdtr(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   3036		   u_int period, u_int offset)
   3037{
   3038	if (offset == 0)
   3039		period = AHC_ASYNC_XFER_PERIOD;
   3040	ahc->msgout_index += spi_populate_sync_msg(
   3041			ahc->msgout_buf + ahc->msgout_index, period, offset);
   3042	ahc->msgout_len += 5;
   3043	if (bootverbose) {
   3044		printk("(%s:%c:%d:%d): Sending SDTR period %x, offset %x\n",
   3045		       ahc_name(ahc), devinfo->channel, devinfo->target,
   3046		       devinfo->lun, period, offset);
   3047	}
   3048}
   3049
   3050/*
   3051 * Build a wide negotiation message in our message
   3052 * buffer based on the input parameters.
   3053 */
   3054static void
   3055ahc_construct_wdtr(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   3056		   u_int bus_width)
   3057{
   3058	ahc->msgout_index += spi_populate_width_msg(
   3059			ahc->msgout_buf + ahc->msgout_index, bus_width);
   3060	ahc->msgout_len += 4;
   3061	if (bootverbose) {
   3062		printk("(%s:%c:%d:%d): Sending WDTR %x\n",
   3063		       ahc_name(ahc), devinfo->channel, devinfo->target,
   3064		       devinfo->lun, bus_width);
   3065	}
   3066}
   3067
   3068/*
   3069 * Build a parallel protocol request message in our message
   3070 * buffer based on the input parameters.
   3071 */
   3072static void
   3073ahc_construct_ppr(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   3074		  u_int period, u_int offset, u_int bus_width,
   3075		  u_int ppr_options)
   3076{
   3077	if (offset == 0)
   3078		period = AHC_ASYNC_XFER_PERIOD;
   3079	ahc->msgout_index += spi_populate_ppr_msg(
   3080			ahc->msgout_buf + ahc->msgout_index, period, offset,
   3081			bus_width, ppr_options);
   3082	ahc->msgout_len += 8;
   3083	if (bootverbose) {
   3084		printk("(%s:%c:%d:%d): Sending PPR bus_width %x, period %x, "
   3085		       "offset %x, ppr_options %x\n", ahc_name(ahc),
   3086		       devinfo->channel, devinfo->target, devinfo->lun,
   3087		       bus_width, period, offset, ppr_options);
   3088	}
   3089}
   3090
   3091/*
   3092 * Clear any active message state.
   3093 */
   3094static void
   3095ahc_clear_msg_state(struct ahc_softc *ahc)
   3096{
   3097	ahc->msgout_len = 0;
   3098	ahc->msgin_index = 0;
   3099	ahc->msg_type = MSG_TYPE_NONE;
   3100	if ((ahc_inb(ahc, SCSISIGI) & ATNI) != 0) {
   3101		/*
   3102		 * The target didn't care to respond to our
   3103		 * message request, so clear ATN.
   3104		 */
   3105		ahc_outb(ahc, CLRSINT1, CLRATNO);
   3106	}
   3107	ahc_outb(ahc, MSG_OUT, NOP);
   3108	ahc_outb(ahc, SEQ_FLAGS2,
   3109		 ahc_inb(ahc, SEQ_FLAGS2) & ~TARGET_MSG_PENDING);
   3110}
   3111
   3112static void
   3113ahc_handle_proto_violation(struct ahc_softc *ahc)
   3114{
   3115	struct	ahc_devinfo devinfo;
   3116	struct	scb *scb;
   3117	u_int	scbid;
   3118	u_int	seq_flags;
   3119	u_int	curphase;
   3120	u_int	lastphase;
   3121	int	found;
   3122
   3123	ahc_fetch_devinfo(ahc, &devinfo);
   3124	scbid = ahc_inb(ahc, SCB_TAG);
   3125	scb = ahc_lookup_scb(ahc, scbid);
   3126	seq_flags = ahc_inb(ahc, SEQ_FLAGS);
   3127	curphase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
   3128	lastphase = ahc_inb(ahc, LASTPHASE);
   3129	if ((seq_flags & NOT_IDENTIFIED) != 0) {
   3130
   3131		/*
   3132		 * The reconnecting target either did not send an
   3133		 * identify message, or did, but we didn't find an SCB
   3134		 * to match.
   3135		 */
   3136		ahc_print_devinfo(ahc, &devinfo);
   3137		printk("Target did not send an IDENTIFY message. "
   3138		       "LASTPHASE = 0x%x.\n", lastphase);
   3139		scb = NULL;
   3140	} else if (scb == NULL) {
   3141		/*
   3142		 * We don't seem to have an SCB active for this
   3143		 * transaction.  Print an error and reset the bus.
   3144		 */
   3145		ahc_print_devinfo(ahc, &devinfo);
   3146		printk("No SCB found during protocol violation\n");
   3147		goto proto_violation_reset;
   3148	} else {
   3149		ahc_set_transaction_status(scb, CAM_SEQUENCE_FAIL);
   3150		if ((seq_flags & NO_CDB_SENT) != 0) {
   3151			ahc_print_path(ahc, scb);
   3152			printk("No or incomplete CDB sent to device.\n");
   3153		} else if ((ahc_inb(ahc, SCB_CONTROL) & STATUS_RCVD) == 0) {
   3154			/*
   3155			 * The target never bothered to provide status to
   3156			 * us prior to completing the command.  Since we don't
   3157			 * know the disposition of this command, we must attempt
   3158			 * to abort it.  Assert ATN and prepare to send an abort
   3159			 * message.
   3160			 */
   3161			ahc_print_path(ahc, scb);
   3162			printk("Completed command without status.\n");
   3163		} else {
   3164			ahc_print_path(ahc, scb);
   3165			printk("Unknown protocol violation.\n");
   3166			ahc_dump_card_state(ahc);
   3167		}
   3168	}
   3169	if ((lastphase & ~P_DATAIN_DT) == 0
   3170	 || lastphase == P_COMMAND) {
   3171proto_violation_reset:
   3172		/*
   3173		 * Target either went directly to data/command
   3174		 * phase or didn't respond to our ATN.
   3175		 * The only safe thing to do is to blow
   3176		 * it away with a bus reset.
   3177		 */
   3178		found = ahc_reset_channel(ahc, 'A', TRUE);
   3179		printk("%s: Issued Channel %c Bus Reset. "
   3180		       "%d SCBs aborted\n", ahc_name(ahc), 'A', found);
   3181	} else {
   3182		/*
   3183		 * Leave the selection hardware off in case
   3184		 * this abort attempt will affect yet to
   3185		 * be sent commands.
   3186		 */
   3187		ahc_outb(ahc, SCSISEQ,
   3188			 ahc_inb(ahc, SCSISEQ) & ~ENSELO);
   3189		ahc_assert_atn(ahc);
   3190		ahc_outb(ahc, MSG_OUT, HOST_MSG);
   3191		if (scb == NULL) {
   3192			ahc_print_devinfo(ahc, &devinfo);
   3193			ahc->msgout_buf[0] = ABORT_TASK;
   3194			ahc->msgout_len = 1;
   3195			ahc->msgout_index = 0;
   3196			ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
   3197		} else {
   3198			ahc_print_path(ahc, scb);
   3199			scb->flags |= SCB_ABORT;
   3200		}
   3201		printk("Protocol violation %s.  Attempting to abort.\n",
   3202		       ahc_lookup_phase_entry(curphase)->phasemsg);
   3203	}
   3204}
   3205
   3206/*
   3207 * Manual message loop handler.
   3208 */
   3209static void
   3210ahc_handle_message_phase(struct ahc_softc *ahc)
   3211{
   3212	struct	ahc_devinfo devinfo;
   3213	u_int	bus_phase;
   3214	int	end_session;
   3215
   3216	ahc_fetch_devinfo(ahc, &devinfo);
   3217	end_session = FALSE;
   3218	bus_phase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
   3219
   3220reswitch:
   3221	switch (ahc->msg_type) {
   3222	case MSG_TYPE_INITIATOR_MSGOUT:
   3223	{
   3224		int lastbyte;
   3225		int phasemis;
   3226		int msgdone;
   3227
   3228		if (ahc->msgout_len == 0)
   3229			panic("HOST_MSG_LOOP interrupt with no active message");
   3230
   3231#ifdef AHC_DEBUG
   3232		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
   3233			ahc_print_devinfo(ahc, &devinfo);
   3234			printk("INITIATOR_MSG_OUT");
   3235		}
   3236#endif
   3237		phasemis = bus_phase != P_MESGOUT;
   3238		if (phasemis) {
   3239#ifdef AHC_DEBUG
   3240			if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
   3241				printk(" PHASEMIS %s\n",
   3242				       ahc_lookup_phase_entry(bus_phase)
   3243							     ->phasemsg);
   3244			}
   3245#endif
   3246			if (bus_phase == P_MESGIN) {
   3247				/*
   3248				 * Change gears and see if
   3249				 * this messages is of interest to
   3250				 * us or should be passed back to
   3251				 * the sequencer.
   3252				 */
   3253				ahc_outb(ahc, CLRSINT1, CLRATNO);
   3254				ahc->send_msg_perror = FALSE;
   3255				ahc->msg_type = MSG_TYPE_INITIATOR_MSGIN;
   3256				ahc->msgin_index = 0;
   3257				goto reswitch;
   3258			}
   3259			end_session = TRUE;
   3260			break;
   3261		}
   3262
   3263		if (ahc->send_msg_perror) {
   3264			ahc_outb(ahc, CLRSINT1, CLRATNO);
   3265			ahc_outb(ahc, CLRSINT1, CLRREQINIT);
   3266#ifdef AHC_DEBUG
   3267			if ((ahc_debug & AHC_SHOW_MESSAGES) != 0)
   3268				printk(" byte 0x%x\n", ahc->send_msg_perror);
   3269#endif
   3270			ahc_outb(ahc, SCSIDATL, MSG_PARITY_ERROR);
   3271			break;
   3272		}
   3273
   3274		msgdone	= ahc->msgout_index == ahc->msgout_len;
   3275		if (msgdone) {
   3276			/*
   3277			 * The target has requested a retry.
   3278			 * Re-assert ATN, reset our message index to
   3279			 * 0, and try again.
   3280			 */
   3281			ahc->msgout_index = 0;
   3282			ahc_assert_atn(ahc);
   3283		}
   3284
   3285		lastbyte = ahc->msgout_index == (ahc->msgout_len - 1);
   3286		if (lastbyte) {
   3287			/* Last byte is signified by dropping ATN */
   3288			ahc_outb(ahc, CLRSINT1, CLRATNO);
   3289		}
   3290
   3291		/*
   3292		 * Clear our interrupt status and present
   3293		 * the next byte on the bus.
   3294		 */
   3295		ahc_outb(ahc, CLRSINT1, CLRREQINIT);
   3296#ifdef AHC_DEBUG
   3297		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0)
   3298			printk(" byte 0x%x\n",
   3299			       ahc->msgout_buf[ahc->msgout_index]);
   3300#endif
   3301		ahc_outb(ahc, SCSIDATL, ahc->msgout_buf[ahc->msgout_index++]);
   3302		break;
   3303	}
   3304	case MSG_TYPE_INITIATOR_MSGIN:
   3305	{
   3306		int phasemis;
   3307		int message_done;
   3308
   3309#ifdef AHC_DEBUG
   3310		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
   3311			ahc_print_devinfo(ahc, &devinfo);
   3312			printk("INITIATOR_MSG_IN");
   3313		}
   3314#endif
   3315		phasemis = bus_phase != P_MESGIN;
   3316		if (phasemis) {
   3317#ifdef AHC_DEBUG
   3318			if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
   3319				printk(" PHASEMIS %s\n",
   3320				       ahc_lookup_phase_entry(bus_phase)
   3321							     ->phasemsg);
   3322			}
   3323#endif
   3324			ahc->msgin_index = 0;
   3325			if (bus_phase == P_MESGOUT
   3326			 && (ahc->send_msg_perror == TRUE
   3327			  || (ahc->msgout_len != 0
   3328			   && ahc->msgout_index == 0))) {
   3329				ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
   3330				goto reswitch;
   3331			}
   3332			end_session = TRUE;
   3333			break;
   3334		}
   3335
   3336		/* Pull the byte in without acking it */
   3337		ahc->msgin_buf[ahc->msgin_index] = ahc_inb(ahc, SCSIBUSL);
   3338#ifdef AHC_DEBUG
   3339		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0)
   3340			printk(" byte 0x%x\n",
   3341			       ahc->msgin_buf[ahc->msgin_index]);
   3342#endif
   3343
   3344		message_done = ahc_parse_msg(ahc, &devinfo);
   3345
   3346		if (message_done) {
   3347			/*
   3348			 * Clear our incoming message buffer in case there
   3349			 * is another message following this one.
   3350			 */
   3351			ahc->msgin_index = 0;
   3352
   3353			/*
   3354			 * If this message illicited a response,
   3355			 * assert ATN so the target takes us to the
   3356			 * message out phase.
   3357			 */
   3358			if (ahc->msgout_len != 0) {
   3359#ifdef AHC_DEBUG
   3360				if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
   3361					ahc_print_devinfo(ahc, &devinfo);
   3362					printk("Asserting ATN for response\n");
   3363				}
   3364#endif
   3365				ahc_assert_atn(ahc);
   3366			}
   3367		} else
   3368			ahc->msgin_index++;
   3369
   3370		if (message_done == MSGLOOP_TERMINATED) {
   3371			end_session = TRUE;
   3372		} else {
   3373			/* Ack the byte */
   3374			ahc_outb(ahc, CLRSINT1, CLRREQINIT);
   3375			ahc_inb(ahc, SCSIDATL);
   3376		}
   3377		break;
   3378	}
   3379	case MSG_TYPE_TARGET_MSGIN:
   3380	{
   3381		int msgdone;
   3382		int msgout_request;
   3383
   3384		if (ahc->msgout_len == 0)
   3385			panic("Target MSGIN with no active message");
   3386
   3387		/*
   3388		 * If we interrupted a mesgout session, the initiator
   3389		 * will not know this until our first REQ.  So, we
   3390		 * only honor mesgout requests after we've sent our
   3391		 * first byte.
   3392		 */
   3393		if ((ahc_inb(ahc, SCSISIGI) & ATNI) != 0
   3394		 && ahc->msgout_index > 0)
   3395			msgout_request = TRUE;
   3396		else
   3397			msgout_request = FALSE;
   3398
   3399		if (msgout_request) {
   3400
   3401			/*
   3402			 * Change gears and see if
   3403			 * this messages is of interest to
   3404			 * us or should be passed back to
   3405			 * the sequencer.
   3406			 */
   3407			ahc->msg_type = MSG_TYPE_TARGET_MSGOUT;
   3408			ahc_outb(ahc, SCSISIGO, P_MESGOUT | BSYO);
   3409			ahc->msgin_index = 0;
   3410			/* Dummy read to REQ for first byte */
   3411			ahc_inb(ahc, SCSIDATL);
   3412			ahc_outb(ahc, SXFRCTL0,
   3413				 ahc_inb(ahc, SXFRCTL0) | SPIOEN);
   3414			break;
   3415		}
   3416
   3417		msgdone = ahc->msgout_index == ahc->msgout_len;
   3418		if (msgdone) {
   3419			ahc_outb(ahc, SXFRCTL0,
   3420				 ahc_inb(ahc, SXFRCTL0) & ~SPIOEN);
   3421			end_session = TRUE;
   3422			break;
   3423		}
   3424
   3425		/*
   3426		 * Present the next byte on the bus.
   3427		 */
   3428		ahc_outb(ahc, SXFRCTL0, ahc_inb(ahc, SXFRCTL0) | SPIOEN);
   3429		ahc_outb(ahc, SCSIDATL, ahc->msgout_buf[ahc->msgout_index++]);
   3430		break;
   3431	}
   3432	case MSG_TYPE_TARGET_MSGOUT:
   3433	{
   3434		int lastbyte;
   3435		int msgdone;
   3436
   3437		/*
   3438		 * The initiator signals that this is
   3439		 * the last byte by dropping ATN.
   3440		 */
   3441		lastbyte = (ahc_inb(ahc, SCSISIGI) & ATNI) == 0;
   3442
   3443		/*
   3444		 * Read the latched byte, but turn off SPIOEN first
   3445		 * so that we don't inadvertently cause a REQ for the
   3446		 * next byte.
   3447		 */
   3448		ahc_outb(ahc, SXFRCTL0, ahc_inb(ahc, SXFRCTL0) & ~SPIOEN);
   3449		ahc->msgin_buf[ahc->msgin_index] = ahc_inb(ahc, SCSIDATL);
   3450		msgdone = ahc_parse_msg(ahc, &devinfo);
   3451		if (msgdone == MSGLOOP_TERMINATED) {
   3452			/*
   3453			 * The message is *really* done in that it caused
   3454			 * us to go to bus free.  The sequencer has already
   3455			 * been reset at this point, so pull the ejection
   3456			 * handle.
   3457			 */
   3458			return;
   3459		}
   3460
   3461		ahc->msgin_index++;
   3462
   3463		/*
   3464		 * XXX Read spec about initiator dropping ATN too soon
   3465		 *     and use msgdone to detect it.
   3466		 */
   3467		if (msgdone == MSGLOOP_MSGCOMPLETE) {
   3468			ahc->msgin_index = 0;
   3469
   3470			/*
   3471			 * If this message illicited a response, transition
   3472			 * to the Message in phase and send it.
   3473			 */
   3474			if (ahc->msgout_len != 0) {
   3475				ahc_outb(ahc, SCSISIGO, P_MESGIN | BSYO);
   3476				ahc_outb(ahc, SXFRCTL0,
   3477					 ahc_inb(ahc, SXFRCTL0) | SPIOEN);
   3478				ahc->msg_type = MSG_TYPE_TARGET_MSGIN;
   3479				ahc->msgin_index = 0;
   3480				break;
   3481			}
   3482		}
   3483
   3484		if (lastbyte)
   3485			end_session = TRUE;
   3486		else {
   3487			/* Ask for the next byte. */
   3488			ahc_outb(ahc, SXFRCTL0,
   3489				 ahc_inb(ahc, SXFRCTL0) | SPIOEN);
   3490		}
   3491
   3492		break;
   3493	}
   3494	default:
   3495		panic("Unknown REQINIT message type");
   3496	}
   3497
   3498	if (end_session) {
   3499		ahc_clear_msg_state(ahc);
   3500		ahc_outb(ahc, RETURN_1, EXIT_MSG_LOOP);
   3501	} else
   3502		ahc_outb(ahc, RETURN_1, CONT_MSG_LOOP);
   3503}
   3504
   3505/*
   3506 * See if we sent a particular extended message to the target.
   3507 * If "full" is true, return true only if the target saw the full
   3508 * message.  If "full" is false, return true if the target saw at
   3509 * least the first byte of the message.
   3510 */
   3511static int
   3512ahc_sent_msg(struct ahc_softc *ahc, ahc_msgtype type, u_int msgval, int full)
   3513{
   3514	int found;
   3515	u_int index;
   3516
   3517	found = FALSE;
   3518	index = 0;
   3519
   3520	while (index < ahc->msgout_len) {
   3521		if (ahc->msgout_buf[index] == EXTENDED_MESSAGE) {
   3522			u_int end_index;
   3523
   3524			end_index = index + 1 + ahc->msgout_buf[index + 1];
   3525			if (ahc->msgout_buf[index+2] == msgval
   3526			 && type == AHCMSG_EXT) {
   3527
   3528				if (full) {
   3529					if (ahc->msgout_index > end_index)
   3530						found = TRUE;
   3531				} else if (ahc->msgout_index > index)
   3532					found = TRUE;
   3533			}
   3534			index = end_index;
   3535		} else if (ahc->msgout_buf[index] >= SIMPLE_QUEUE_TAG
   3536			&& ahc->msgout_buf[index] <= IGNORE_WIDE_RESIDUE) {
   3537
   3538			/* Skip tag type and tag id or residue param*/
   3539			index += 2;
   3540		} else {
   3541			/* Single byte message */
   3542			if (type == AHCMSG_1B
   3543			 && ahc->msgout_buf[index] == msgval
   3544			 && ahc->msgout_index > index)
   3545				found = TRUE;
   3546			index++;
   3547		}
   3548
   3549		if (found)
   3550			break;
   3551	}
   3552	return (found);
   3553}
   3554
   3555/*
   3556 * Wait for a complete incoming message, parse it, and respond accordingly.
   3557 */
   3558static int
   3559ahc_parse_msg(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
   3560{
   3561	struct	ahc_initiator_tinfo *tinfo;
   3562	struct	ahc_tmode_tstate *tstate;
   3563	int	reject;
   3564	int	done;
   3565	int	response;
   3566	u_int	targ_scsirate;
   3567
   3568	done = MSGLOOP_IN_PROG;
   3569	response = FALSE;
   3570	reject = FALSE;
   3571	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
   3572				    devinfo->target, &tstate);
   3573	targ_scsirate = tinfo->scsirate;
   3574
   3575	/*
   3576	 * Parse as much of the message as is available,
   3577	 * rejecting it if we don't support it.  When
   3578	 * the entire message is available and has been
   3579	 * handled, return MSGLOOP_MSGCOMPLETE, indicating
   3580	 * that we have parsed an entire message.
   3581	 *
   3582	 * In the case of extended messages, we accept the length
   3583	 * byte outright and perform more checking once we know the
   3584	 * extended message type.
   3585	 */
   3586	switch (ahc->msgin_buf[0]) {
   3587	case DISCONNECT:
   3588	case SAVE_POINTERS:
   3589	case COMMAND_COMPLETE:
   3590	case RESTORE_POINTERS:
   3591	case IGNORE_WIDE_RESIDUE:
   3592		/*
   3593		 * End our message loop as these are messages
   3594		 * the sequencer handles on its own.
   3595		 */
   3596		done = MSGLOOP_TERMINATED;
   3597		break;
   3598	case MESSAGE_REJECT:
   3599		response = ahc_handle_msg_reject(ahc, devinfo);
   3600		fallthrough;
   3601	case NOP:
   3602		done = MSGLOOP_MSGCOMPLETE;
   3603		break;
   3604	case EXTENDED_MESSAGE:
   3605	{
   3606		/* Wait for enough of the message to begin validation */
   3607		if (ahc->msgin_index < 2)
   3608			break;
   3609		switch (ahc->msgin_buf[2]) {
   3610		case EXTENDED_SDTR:
   3611		{
   3612			const struct ahc_syncrate *syncrate;
   3613			u_int	 period;
   3614			u_int	 ppr_options;
   3615			u_int	 offset;
   3616			u_int	 saved_offset;
   3617
   3618			if (ahc->msgin_buf[1] != MSG_EXT_SDTR_LEN) {
   3619				reject = TRUE;
   3620				break;
   3621			}
   3622
   3623			/*
   3624			 * Wait until we have both args before validating
   3625			 * and acting on this message.
   3626			 *
   3627			 * Add one to MSG_EXT_SDTR_LEN to account for
   3628			 * the extended message preamble.
   3629			 */
   3630			if (ahc->msgin_index < (MSG_EXT_SDTR_LEN + 1))
   3631				break;
   3632
   3633			period = ahc->msgin_buf[3];
   3634			ppr_options = 0;
   3635			saved_offset = offset = ahc->msgin_buf[4];
   3636			syncrate = ahc_devlimited_syncrate(ahc, tinfo, &period,
   3637							   &ppr_options,
   3638							   devinfo->role);
   3639			ahc_validate_offset(ahc, tinfo, syncrate, &offset,
   3640					    targ_scsirate & WIDEXFER,
   3641					    devinfo->role);
   3642			if (bootverbose) {
   3643				printk("(%s:%c:%d:%d): Received "
   3644				       "SDTR period %x, offset %x\n\t"
   3645				       "Filtered to period %x, offset %x\n",
   3646				       ahc_name(ahc), devinfo->channel,
   3647				       devinfo->target, devinfo->lun,
   3648				       ahc->msgin_buf[3], saved_offset,
   3649				       period, offset);
   3650			}
   3651			ahc_set_syncrate(ahc, devinfo,
   3652					 syncrate, period,
   3653					 offset, ppr_options,
   3654					 AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
   3655					 /*paused*/TRUE);
   3656
   3657			/*
   3658			 * See if we initiated Sync Negotiation
   3659			 * and didn't have to fall down to async
   3660			 * transfers.
   3661			 */
   3662			if (ahc_sent_msg(ahc, AHCMSG_EXT, EXTENDED_SDTR, TRUE)) {
   3663				/* We started it */
   3664				if (saved_offset != offset) {
   3665					/* Went too low - force async */
   3666					reject = TRUE;
   3667				}
   3668			} else {
   3669				/*
   3670				 * Send our own SDTR in reply
   3671				 */
   3672				if (bootverbose
   3673				 && devinfo->role == ROLE_INITIATOR) {
   3674					printk("(%s:%c:%d:%d): Target "
   3675					       "Initiated SDTR\n",
   3676					       ahc_name(ahc), devinfo->channel,
   3677					       devinfo->target, devinfo->lun);
   3678				}
   3679				ahc->msgout_index = 0;
   3680				ahc->msgout_len = 0;
   3681				ahc_construct_sdtr(ahc, devinfo,
   3682						   period, offset);
   3683				ahc->msgout_index = 0;
   3684				response = TRUE;
   3685			}
   3686			done = MSGLOOP_MSGCOMPLETE;
   3687			break;
   3688		}
   3689		case EXTENDED_WDTR:
   3690		{
   3691			u_int bus_width;
   3692			u_int saved_width;
   3693			u_int sending_reply;
   3694
   3695			sending_reply = FALSE;
   3696			if (ahc->msgin_buf[1] != MSG_EXT_WDTR_LEN) {
   3697				reject = TRUE;
   3698				break;
   3699			}
   3700
   3701			/*
   3702			 * Wait until we have our arg before validating
   3703			 * and acting on this message.
   3704			 *
   3705			 * Add one to MSG_EXT_WDTR_LEN to account for
   3706			 * the extended message preamble.
   3707			 */
   3708			if (ahc->msgin_index < (MSG_EXT_WDTR_LEN + 1))
   3709				break;
   3710
   3711			bus_width = ahc->msgin_buf[3];
   3712			saved_width = bus_width;
   3713			ahc_validate_width(ahc, tinfo, &bus_width,
   3714					   devinfo->role);
   3715			if (bootverbose) {
   3716				printk("(%s:%c:%d:%d): Received WDTR "
   3717				       "%x filtered to %x\n",
   3718				       ahc_name(ahc), devinfo->channel,
   3719				       devinfo->target, devinfo->lun,
   3720				       saved_width, bus_width);
   3721			}
   3722
   3723			if (ahc_sent_msg(ahc, AHCMSG_EXT, EXTENDED_WDTR, TRUE)) {
   3724				/*
   3725				 * Don't send a WDTR back to the
   3726				 * target, since we asked first.
   3727				 * If the width went higher than our
   3728				 * request, reject it.
   3729				 */
   3730				if (saved_width > bus_width) {
   3731					reject = TRUE;
   3732					printk("(%s:%c:%d:%d): requested %dBit "
   3733					       "transfers.  Rejecting...\n",
   3734					       ahc_name(ahc), devinfo->channel,
   3735					       devinfo->target, devinfo->lun,
   3736					       8 * (0x01 << bus_width));
   3737					bus_width = 0;
   3738				}
   3739			} else {
   3740				/*
   3741				 * Send our own WDTR in reply
   3742				 */
   3743				if (bootverbose
   3744				 && devinfo->role == ROLE_INITIATOR) {
   3745					printk("(%s:%c:%d:%d): Target "
   3746					       "Initiated WDTR\n",
   3747					       ahc_name(ahc), devinfo->channel,
   3748					       devinfo->target, devinfo->lun);
   3749				}
   3750				ahc->msgout_index = 0;
   3751				ahc->msgout_len = 0;
   3752				ahc_construct_wdtr(ahc, devinfo, bus_width);
   3753				ahc->msgout_index = 0;
   3754				response = TRUE;
   3755				sending_reply = TRUE;
   3756			}
   3757			/*
   3758			 * After a wide message, we are async, but
   3759			 * some devices don't seem to honor this portion
   3760			 * of the spec.  Force a renegotiation of the
   3761			 * sync component of our transfer agreement even
   3762			 * if our goal is async.  By updating our width
   3763			 * after forcing the negotiation, we avoid
   3764			 * renegotiating for width.
   3765			 */
   3766			ahc_update_neg_request(ahc, devinfo, tstate,
   3767					       tinfo, AHC_NEG_ALWAYS);
   3768			ahc_set_width(ahc, devinfo, bus_width,
   3769				      AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
   3770				      /*paused*/TRUE);
   3771			if (sending_reply == FALSE && reject == FALSE) {
   3772
   3773				/*
   3774				 * We will always have an SDTR to send.
   3775				 */
   3776				ahc->msgout_index = 0;
   3777				ahc->msgout_len = 0;
   3778				ahc_build_transfer_msg(ahc, devinfo);
   3779				ahc->msgout_index = 0;
   3780				response = TRUE;
   3781			}
   3782			done = MSGLOOP_MSGCOMPLETE;
   3783			break;
   3784		}
   3785		case EXTENDED_PPR:
   3786		{
   3787			const struct ahc_syncrate *syncrate;
   3788			u_int	period;
   3789			u_int	offset;
   3790			u_int	bus_width;
   3791			u_int	ppr_options;
   3792			u_int	saved_width;
   3793			u_int	saved_offset;
   3794			u_int	saved_ppr_options;
   3795
   3796			if (ahc->msgin_buf[1] != MSG_EXT_PPR_LEN) {
   3797				reject = TRUE;
   3798				break;
   3799			}
   3800
   3801			/*
   3802			 * Wait until we have all args before validating
   3803			 * and acting on this message.
   3804			 *
   3805			 * Add one to MSG_EXT_PPR_LEN to account for
   3806			 * the extended message preamble.
   3807			 */
   3808			if (ahc->msgin_index < (MSG_EXT_PPR_LEN + 1))
   3809				break;
   3810
   3811			period = ahc->msgin_buf[3];
   3812			offset = ahc->msgin_buf[5];
   3813			bus_width = ahc->msgin_buf[6];
   3814			saved_width = bus_width;
   3815			ppr_options = ahc->msgin_buf[7];
   3816			/*
   3817			 * According to the spec, a DT only
   3818			 * period factor with no DT option
   3819			 * set implies async.
   3820			 */
   3821			if ((ppr_options & MSG_EXT_PPR_DT_REQ) == 0
   3822			 && period == 9)
   3823				offset = 0;
   3824			saved_ppr_options = ppr_options;
   3825			saved_offset = offset;
   3826
   3827			/*
   3828			 * Mask out any options we don't support
   3829			 * on any controller.  Transfer options are
   3830			 * only available if we are negotiating wide.
   3831			 */
   3832			ppr_options &= MSG_EXT_PPR_DT_REQ;
   3833			if (bus_width == 0)
   3834				ppr_options = 0;
   3835
   3836			ahc_validate_width(ahc, tinfo, &bus_width,
   3837					   devinfo->role);
   3838			syncrate = ahc_devlimited_syncrate(ahc, tinfo, &period,
   3839							   &ppr_options,
   3840							   devinfo->role);
   3841			ahc_validate_offset(ahc, tinfo, syncrate,
   3842					    &offset, bus_width,
   3843					    devinfo->role);
   3844
   3845			if (ahc_sent_msg(ahc, AHCMSG_EXT, EXTENDED_PPR, TRUE)) {
   3846				/*
   3847				 * If we are unable to do any of the
   3848				 * requested options (we went too low),
   3849				 * then we'll have to reject the message.
   3850				 */
   3851				if (saved_width > bus_width
   3852				 || saved_offset != offset
   3853				 || saved_ppr_options != ppr_options) {
   3854					reject = TRUE;
   3855					period = 0;
   3856					offset = 0;
   3857					bus_width = 0;
   3858					ppr_options = 0;
   3859					syncrate = NULL;
   3860				}
   3861			} else {
   3862				if (devinfo->role != ROLE_TARGET)
   3863					printk("(%s:%c:%d:%d): Target "
   3864					       "Initiated PPR\n",
   3865					       ahc_name(ahc), devinfo->channel,
   3866					       devinfo->target, devinfo->lun);
   3867				else
   3868					printk("(%s:%c:%d:%d): Initiator "
   3869					       "Initiated PPR\n",
   3870					       ahc_name(ahc), devinfo->channel,
   3871					       devinfo->target, devinfo->lun);
   3872				ahc->msgout_index = 0;
   3873				ahc->msgout_len = 0;
   3874				ahc_construct_ppr(ahc, devinfo, period, offset,
   3875						  bus_width, ppr_options);
   3876				ahc->msgout_index = 0;
   3877				response = TRUE;
   3878			}
   3879			if (bootverbose) {
   3880				printk("(%s:%c:%d:%d): Received PPR width %x, "
   3881				       "period %x, offset %x,options %x\n"
   3882				       "\tFiltered to width %x, period %x, "
   3883				       "offset %x, options %x\n",
   3884				       ahc_name(ahc), devinfo->channel,
   3885				       devinfo->target, devinfo->lun,
   3886				       saved_width, ahc->msgin_buf[3],
   3887				       saved_offset, saved_ppr_options,
   3888				       bus_width, period, offset, ppr_options);
   3889			}
   3890			ahc_set_width(ahc, devinfo, bus_width,
   3891				      AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
   3892				      /*paused*/TRUE);
   3893			ahc_set_syncrate(ahc, devinfo,
   3894					 syncrate, period,
   3895					 offset, ppr_options,
   3896					 AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
   3897					 /*paused*/TRUE);
   3898			done = MSGLOOP_MSGCOMPLETE;
   3899			break;
   3900		}
   3901		default:
   3902			/* Unknown extended message.  Reject it. */
   3903			reject = TRUE;
   3904			break;
   3905		}
   3906		break;
   3907	}
   3908#ifdef AHC_TARGET_MODE
   3909	case TARGET_RESET:
   3910		ahc_handle_devreset(ahc, devinfo,
   3911				    CAM_BDR_SENT,
   3912				    "Bus Device Reset Received",
   3913				    /*verbose_level*/0);
   3914		ahc_restart(ahc);
   3915		done = MSGLOOP_TERMINATED;
   3916		break;
   3917	case ABORT_TASK:
   3918	case ABORT_TASK_SET:
   3919	case CLEAR_QUEUE_TASK_SET:
   3920	{
   3921		int tag;
   3922
   3923		/* Target mode messages */
   3924		if (devinfo->role != ROLE_TARGET) {
   3925			reject = TRUE;
   3926			break;
   3927		}
   3928		tag = SCB_LIST_NULL;
   3929		if (ahc->msgin_buf[0] == ABORT_TASK)
   3930			tag = ahc_inb(ahc, INITIATOR_TAG);
   3931		ahc_abort_scbs(ahc, devinfo->target, devinfo->channel,
   3932			       devinfo->lun, tag, ROLE_TARGET,
   3933			       CAM_REQ_ABORTED);
   3934
   3935		tstate = ahc->enabled_targets[devinfo->our_scsiid];
   3936		if (tstate != NULL) {
   3937			struct ahc_tmode_lstate* lstate;
   3938
   3939			lstate = tstate->enabled_luns[devinfo->lun];
   3940			if (lstate != NULL) {
   3941				ahc_queue_lstate_event(ahc, lstate,
   3942						       devinfo->our_scsiid,
   3943						       ahc->msgin_buf[0],
   3944						       /*arg*/tag);
   3945				ahc_send_lstate_events(ahc, lstate);
   3946			}
   3947		}
   3948		ahc_restart(ahc);
   3949		done = MSGLOOP_TERMINATED;
   3950		break;
   3951	}
   3952#endif
   3953	case TERMINATE_IO_PROC:
   3954	default:
   3955		reject = TRUE;
   3956		break;
   3957	}
   3958
   3959	if (reject) {
   3960		/*
   3961		 * Setup to reject the message.
   3962		 */
   3963		ahc->msgout_index = 0;
   3964		ahc->msgout_len = 1;
   3965		ahc->msgout_buf[0] = MESSAGE_REJECT;
   3966		done = MSGLOOP_MSGCOMPLETE;
   3967		response = TRUE;
   3968	}
   3969
   3970	if (done != MSGLOOP_IN_PROG && !response)
   3971		/* Clear the outgoing message buffer */
   3972		ahc->msgout_len = 0;
   3973
   3974	return (done);
   3975}
   3976
   3977/*
   3978 * Process a message reject message.
   3979 */
   3980static int
   3981ahc_handle_msg_reject(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
   3982{
   3983	/*
   3984	 * What we care about here is if we had an
   3985	 * outstanding SDTR or WDTR message for this
   3986	 * target.  If we did, this is a signal that
   3987	 * the target is refusing negotiation.
   3988	 */
   3989	struct scb *scb;
   3990	struct ahc_initiator_tinfo *tinfo;
   3991	struct ahc_tmode_tstate *tstate;
   3992	u_int scb_index;
   3993	u_int last_msg;
   3994	int   response = 0;
   3995
   3996	scb_index = ahc_inb(ahc, SCB_TAG);
   3997	scb = ahc_lookup_scb(ahc, scb_index);
   3998	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel,
   3999				    devinfo->our_scsiid,
   4000				    devinfo->target, &tstate);
   4001	/* Might be necessary */
   4002	last_msg = ahc_inb(ahc, LAST_MSG);
   4003
   4004	if (ahc_sent_msg(ahc, AHCMSG_EXT, EXTENDED_PPR, /*full*/FALSE)) {
   4005		/*
   4006		 * Target does not support the PPR message.
   4007		 * Attempt to negotiate SPI-2 style.
   4008		 */
   4009		if (bootverbose) {
   4010			printk("(%s:%c:%d:%d): PPR Rejected. "
   4011			       "Trying WDTR/SDTR\n",
   4012			       ahc_name(ahc), devinfo->channel,
   4013			       devinfo->target, devinfo->lun);
   4014		}
   4015		tinfo->goal.ppr_options = 0;
   4016		tinfo->curr.transport_version = 2;
   4017		tinfo->goal.transport_version = 2;
   4018		ahc->msgout_index = 0;
   4019		ahc->msgout_len = 0;
   4020		ahc_build_transfer_msg(ahc, devinfo);
   4021		ahc->msgout_index = 0;
   4022		response = 1;
   4023	} else if (ahc_sent_msg(ahc, AHCMSG_EXT, EXTENDED_WDTR, /*full*/FALSE)) {
   4024
   4025		/* note 8bit xfers */
   4026		printk("(%s:%c:%d:%d): refuses WIDE negotiation.  Using "
   4027		       "8bit transfers\n", ahc_name(ahc),
   4028		       devinfo->channel, devinfo->target, devinfo->lun);
   4029		ahc_set_width(ahc, devinfo, MSG_EXT_WDTR_BUS_8_BIT,
   4030			      AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
   4031			      /*paused*/TRUE);
   4032		/*
   4033		 * No need to clear the sync rate.  If the target
   4034		 * did not accept the command, our syncrate is
   4035		 * unaffected.  If the target started the negotiation,
   4036		 * but rejected our response, we already cleared the
   4037		 * sync rate before sending our WDTR.
   4038		 */
   4039		if (tinfo->goal.offset != tinfo->curr.offset) {
   4040
   4041			/* Start the sync negotiation */
   4042			ahc->msgout_index = 0;
   4043			ahc->msgout_len = 0;
   4044			ahc_build_transfer_msg(ahc, devinfo);
   4045			ahc->msgout_index = 0;
   4046			response = 1;
   4047		}
   4048	} else if (ahc_sent_msg(ahc, AHCMSG_EXT, EXTENDED_SDTR, /*full*/FALSE)) {
   4049		/* note asynch xfers and clear flag */
   4050		ahc_set_syncrate(ahc, devinfo, /*syncrate*/NULL, /*period*/0,
   4051				 /*offset*/0, /*ppr_options*/0,
   4052				 AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
   4053				 /*paused*/TRUE);
   4054		printk("(%s:%c:%d:%d): refuses synchronous negotiation. "
   4055		       "Using asynchronous transfers\n",
   4056		       ahc_name(ahc), devinfo->channel,
   4057		       devinfo->target, devinfo->lun);
   4058	} else if ((scb->hscb->control & SIMPLE_QUEUE_TAG) != 0) {
   4059		int tag_type;
   4060		int mask;
   4061
   4062		tag_type = (scb->hscb->control & SIMPLE_QUEUE_TAG);
   4063
   4064		if (tag_type == SIMPLE_QUEUE_TAG) {
   4065			printk("(%s:%c:%d:%d): refuses tagged commands.  "
   4066			       "Performing non-tagged I/O\n", ahc_name(ahc),
   4067			       devinfo->channel, devinfo->target, devinfo->lun);
   4068			ahc_set_tags(ahc, scb->io_ctx, devinfo, AHC_QUEUE_NONE);
   4069			mask = ~0x23;
   4070		} else {
   4071			printk("(%s:%c:%d:%d): refuses %s tagged commands.  "
   4072			       "Performing simple queue tagged I/O only\n",
   4073			       ahc_name(ahc), devinfo->channel, devinfo->target,
   4074			       devinfo->lun, tag_type == ORDERED_QUEUE_TAG
   4075			       ? "ordered" : "head of queue");
   4076			ahc_set_tags(ahc, scb->io_ctx, devinfo, AHC_QUEUE_BASIC);
   4077			mask = ~0x03;
   4078		}
   4079
   4080		/*
   4081		 * Resend the identify for this CCB as the target
   4082		 * may believe that the selection is invalid otherwise.
   4083		 */
   4084		ahc_outb(ahc, SCB_CONTROL,
   4085			 ahc_inb(ahc, SCB_CONTROL) & mask);
   4086		scb->hscb->control &= mask;
   4087		ahc_set_transaction_tag(scb, /*enabled*/FALSE,
   4088					/*type*/SIMPLE_QUEUE_TAG);
   4089		ahc_outb(ahc, MSG_OUT, MSG_IDENTIFYFLAG);
   4090		ahc_assert_atn(ahc);
   4091
   4092		/*
   4093		 * This transaction is now at the head of
   4094		 * the untagged queue for this target.
   4095		 */
   4096		if ((ahc->flags & AHC_SCB_BTT) == 0) {
   4097			struct scb_tailq *untagged_q;
   4098
   4099			untagged_q =
   4100			    &(ahc->untagged_queues[devinfo->target_offset]);
   4101			TAILQ_INSERT_HEAD(untagged_q, scb, links.tqe);
   4102			scb->flags |= SCB_UNTAGGEDQ;
   4103		}
   4104		ahc_busy_tcl(ahc, BUILD_TCL(scb->hscb->scsiid, devinfo->lun),
   4105			     scb->hscb->tag);
   4106
   4107		/*
   4108		 * Requeue all tagged commands for this target
   4109		 * currently in our possession so they can be
   4110		 * converted to untagged commands.
   4111		 */
   4112		ahc_search_qinfifo(ahc, SCB_GET_TARGET(ahc, scb),
   4113				   SCB_GET_CHANNEL(ahc, scb),
   4114				   SCB_GET_LUN(scb), /*tag*/SCB_LIST_NULL,
   4115				   ROLE_INITIATOR, CAM_REQUEUE_REQ,
   4116				   SEARCH_COMPLETE);
   4117	} else {
   4118		/*
   4119		 * Otherwise, we ignore it.
   4120		 */
   4121		printk("%s:%c:%d: Message reject for %x -- ignored\n",
   4122		       ahc_name(ahc), devinfo->channel, devinfo->target,
   4123		       last_msg);
   4124	}
   4125	return (response);
   4126}
   4127
   4128/*
   4129 * Process an ingnore wide residue message.
   4130 */
   4131static void
   4132ahc_handle_ign_wide_residue(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
   4133{
   4134	u_int scb_index;
   4135	struct scb *scb;
   4136
   4137	scb_index = ahc_inb(ahc, SCB_TAG);
   4138	scb = ahc_lookup_scb(ahc, scb_index);
   4139	/*
   4140	 * XXX Actually check data direction in the sequencer?
   4141	 * Perhaps add datadir to some spare bits in the hscb?
   4142	 */
   4143	if ((ahc_inb(ahc, SEQ_FLAGS) & DPHASE) == 0
   4144	 || ahc_get_transfer_dir(scb) != CAM_DIR_IN) {
   4145		/*
   4146		 * Ignore the message if we haven't
   4147		 * seen an appropriate data phase yet.
   4148		 */
   4149	} else {
   4150		/*
   4151		 * If the residual occurred on the last
   4152		 * transfer and the transfer request was
   4153		 * expected to end on an odd count, do
   4154		 * nothing.  Otherwise, subtract a byte
   4155		 * and update the residual count accordingly.
   4156		 */
   4157		uint32_t sgptr;
   4158
   4159		sgptr = ahc_inb(ahc, SCB_RESIDUAL_SGPTR);
   4160		if ((sgptr & SG_LIST_NULL) != 0
   4161		 && (ahc_inb(ahc, SCB_LUN) & SCB_XFERLEN_ODD) != 0) {
   4162			/*
   4163			 * If the residual occurred on the last
   4164			 * transfer and the transfer request was
   4165			 * expected to end on an odd count, do
   4166			 * nothing.
   4167			 */
   4168		} else {
   4169			struct ahc_dma_seg *sg;
   4170			uint32_t data_cnt;
   4171			uint32_t data_addr;
   4172			uint32_t sglen;
   4173
   4174			/* Pull in all of the sgptr */
   4175			sgptr = ahc_inl(ahc, SCB_RESIDUAL_SGPTR);
   4176			data_cnt = ahc_inl(ahc, SCB_RESIDUAL_DATACNT);
   4177
   4178			if ((sgptr & SG_LIST_NULL) != 0) {
   4179				/*
   4180				 * The residual data count is not updated
   4181				 * for the command run to completion case.
   4182				 * Explicitly zero the count.
   4183				 */
   4184				data_cnt &= ~AHC_SG_LEN_MASK;
   4185			}
   4186
   4187			data_addr = ahc_inl(ahc, SHADDR);
   4188
   4189			data_cnt += 1;
   4190			data_addr -= 1;
   4191			sgptr &= SG_PTR_MASK;
   4192
   4193			sg = ahc_sg_bus_to_virt(scb, sgptr);
   4194
   4195			/*
   4196			 * The residual sg ptr points to the next S/G
   4197			 * to load so we must go back one.
   4198			 */
   4199			sg--;
   4200			sglen = ahc_le32toh(sg->len) & AHC_SG_LEN_MASK;
   4201			if (sg != scb->sg_list
   4202			 && sglen < (data_cnt & AHC_SG_LEN_MASK)) {
   4203
   4204				sg--;
   4205				sglen = ahc_le32toh(sg->len);
   4206				/*
   4207				 * Preserve High Address and SG_LIST bits
   4208				 * while setting the count to 1.
   4209				 */
   4210				data_cnt = 1 | (sglen & (~AHC_SG_LEN_MASK));
   4211				data_addr = ahc_le32toh(sg->addr)
   4212					  + (sglen & AHC_SG_LEN_MASK) - 1;
   4213
   4214				/*
   4215				 * Increment sg so it points to the
   4216				 * "next" sg.
   4217				 */
   4218				sg++;
   4219				sgptr = ahc_sg_virt_to_bus(scb, sg);
   4220			}
   4221			ahc_outl(ahc, SCB_RESIDUAL_SGPTR, sgptr);
   4222			ahc_outl(ahc, SCB_RESIDUAL_DATACNT, data_cnt);
   4223			/*
   4224			 * Toggle the "oddness" of the transfer length
   4225			 * to handle this mid-transfer ignore wide
   4226			 * residue.  This ensures that the oddness is
   4227			 * correct for subsequent data transfers.
   4228			 */
   4229			ahc_outb(ahc, SCB_LUN,
   4230				 ahc_inb(ahc, SCB_LUN) ^ SCB_XFERLEN_ODD);
   4231		}
   4232	}
   4233}
   4234
   4235
   4236/*
   4237 * Reinitialize the data pointers for the active transfer
   4238 * based on its current residual.
   4239 */
   4240static void
   4241ahc_reinitialize_dataptrs(struct ahc_softc *ahc)
   4242{
   4243	struct	 scb *scb;
   4244	struct	 ahc_dma_seg *sg;
   4245	u_int	 scb_index;
   4246	uint32_t sgptr;
   4247	uint32_t resid;
   4248	uint32_t dataptr;
   4249
   4250	scb_index = ahc_inb(ahc, SCB_TAG);
   4251	scb = ahc_lookup_scb(ahc, scb_index);
   4252	sgptr = (ahc_inb(ahc, SCB_RESIDUAL_SGPTR + 3) << 24)
   4253	      | (ahc_inb(ahc, SCB_RESIDUAL_SGPTR + 2) << 16)
   4254	      | (ahc_inb(ahc, SCB_RESIDUAL_SGPTR + 1) << 8)
   4255	      |	ahc_inb(ahc, SCB_RESIDUAL_SGPTR);
   4256
   4257	sgptr &= SG_PTR_MASK;
   4258	sg = ahc_sg_bus_to_virt(scb, sgptr);
   4259
   4260	/* The residual sg_ptr always points to the next sg */
   4261	sg--;
   4262
   4263	resid = (ahc_inb(ahc, SCB_RESIDUAL_DATACNT + 2) << 16)
   4264	      | (ahc_inb(ahc, SCB_RESIDUAL_DATACNT + 1) << 8)
   4265	      | ahc_inb(ahc, SCB_RESIDUAL_DATACNT);
   4266
   4267	dataptr = ahc_le32toh(sg->addr)
   4268		+ (ahc_le32toh(sg->len) & AHC_SG_LEN_MASK)
   4269		- resid;
   4270	if ((ahc->flags & AHC_39BIT_ADDRESSING) != 0) {
   4271		u_int dscommand1;
   4272
   4273		dscommand1 = ahc_inb(ahc, DSCOMMAND1);
   4274		ahc_outb(ahc, DSCOMMAND1, dscommand1 | HADDLDSEL0);
   4275		ahc_outb(ahc, HADDR,
   4276			 (ahc_le32toh(sg->len) >> 24) & SG_HIGH_ADDR_BITS);
   4277		ahc_outb(ahc, DSCOMMAND1, dscommand1);
   4278	}
   4279	ahc_outb(ahc, HADDR + 3, dataptr >> 24);
   4280	ahc_outb(ahc, HADDR + 2, dataptr >> 16);
   4281	ahc_outb(ahc, HADDR + 1, dataptr >> 8);
   4282	ahc_outb(ahc, HADDR, dataptr);
   4283	ahc_outb(ahc, HCNT + 2, resid >> 16);
   4284	ahc_outb(ahc, HCNT + 1, resid >> 8);
   4285	ahc_outb(ahc, HCNT, resid);
   4286	if ((ahc->features & AHC_ULTRA2) == 0) {
   4287		ahc_outb(ahc, STCNT + 2, resid >> 16);
   4288		ahc_outb(ahc, STCNT + 1, resid >> 8);
   4289		ahc_outb(ahc, STCNT, resid);
   4290	}
   4291}
   4292
   4293/*
   4294 * Handle the effects of issuing a bus device reset message.
   4295 */
   4296static void
   4297ahc_handle_devreset(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   4298		    cam_status status, char *message, int verbose_level)
   4299{
   4300#ifdef AHC_TARGET_MODE
   4301	struct ahc_tmode_tstate* tstate;
   4302	u_int lun;
   4303#endif
   4304	int found;
   4305
   4306	found = ahc_abort_scbs(ahc, devinfo->target, devinfo->channel,
   4307			       CAM_LUN_WILDCARD, SCB_LIST_NULL, devinfo->role,
   4308			       status);
   4309
   4310#ifdef AHC_TARGET_MODE
   4311	/*
   4312	 * Send an immediate notify ccb to all target mord peripheral
   4313	 * drivers affected by this action.
   4314	 */
   4315	tstate = ahc->enabled_targets[devinfo->our_scsiid];
   4316	if (tstate != NULL) {
   4317		for (lun = 0; lun < AHC_NUM_LUNS; lun++) {
   4318			struct ahc_tmode_lstate* lstate;
   4319
   4320			lstate = tstate->enabled_luns[lun];
   4321			if (lstate == NULL)
   4322				continue;
   4323
   4324			ahc_queue_lstate_event(ahc, lstate, devinfo->our_scsiid,
   4325					       TARGET_RESET, /*arg*/0);
   4326			ahc_send_lstate_events(ahc, lstate);
   4327		}
   4328	}
   4329#endif
   4330
   4331	/*
   4332	 * Go back to async/narrow transfers and renegotiate.
   4333	 */
   4334	ahc_set_width(ahc, devinfo, MSG_EXT_WDTR_BUS_8_BIT,
   4335		      AHC_TRANS_CUR, /*paused*/TRUE);
   4336	ahc_set_syncrate(ahc, devinfo, /*syncrate*/NULL,
   4337			 /*period*/0, /*offset*/0, /*ppr_options*/0,
   4338			 AHC_TRANS_CUR, /*paused*/TRUE);
   4339
   4340	if (status != CAM_SEL_TIMEOUT)
   4341		ahc_send_async(ahc, devinfo->channel, devinfo->target,
   4342			       CAM_LUN_WILDCARD, AC_SENT_BDR);
   4343
   4344	if (message != NULL
   4345	 && (verbose_level <= bootverbose))
   4346		printk("%s: %s on %c:%d. %d SCBs aborted\n", ahc_name(ahc),
   4347		       message, devinfo->channel, devinfo->target, found);
   4348}
   4349
   4350#ifdef AHC_TARGET_MODE
   4351static void
   4352ahc_setup_target_msgin(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
   4353		       struct scb *scb)
   4354{
   4355
   4356	/*
   4357	 * To facilitate adding multiple messages together,
   4358	 * each routine should increment the index and len
   4359	 * variables instead of setting them explicitly.
   4360	 */
   4361	ahc->msgout_index = 0;
   4362	ahc->msgout_len = 0;
   4363
   4364	if (scb != NULL && (scb->flags & SCB_AUTO_NEGOTIATE) != 0)
   4365		ahc_build_transfer_msg(ahc, devinfo);
   4366	else
   4367		panic("ahc_intr: AWAITING target message with no message");
   4368
   4369	ahc->msgout_index = 0;
   4370	ahc->msg_type = MSG_TYPE_TARGET_MSGIN;
   4371}
   4372#endif
   4373/**************************** Initialization **********************************/
   4374/*
   4375 * Allocate a controller structure for a new device
   4376 * and perform initial initializion.
   4377 */
   4378struct ahc_softc *
   4379ahc_alloc(void *platform_arg, char *name)
   4380{
   4381	struct  ahc_softc *ahc;
   4382	int	i;
   4383
   4384	ahc = kzalloc(sizeof(*ahc), GFP_ATOMIC);
   4385	if (!ahc) {
   4386		printk("aic7xxx: cannot malloc softc!\n");
   4387		kfree(name);
   4388		return NULL;
   4389	}
   4390
   4391	ahc->seep_config = kmalloc(sizeof(*ahc->seep_config), GFP_ATOMIC);
   4392	if (ahc->seep_config == NULL) {
   4393		kfree(ahc);
   4394		kfree(name);
   4395		return (NULL);
   4396	}
   4397	LIST_INIT(&ahc->pending_scbs);
   4398	/* We don't know our unit number until the OSM sets it */
   4399	ahc->name = name;
   4400	ahc->unit = -1;
   4401	ahc->description = NULL;
   4402	ahc->channel = 'A';
   4403	ahc->channel_b = 'B';
   4404	ahc->chip = AHC_NONE;
   4405	ahc->features = AHC_FENONE;
   4406	ahc->bugs = AHC_BUGNONE;
   4407	ahc->flags = AHC_FNONE;
   4408	/*
   4409	 * Default to all error reporting enabled with the
   4410	 * sequencer operating at its fastest speed.
   4411	 * The bus attach code may modify this.
   4412	 */
   4413	ahc->seqctl = FASTMODE;
   4414
   4415	for (i = 0; i < AHC_NUM_TARGETS; i++)
   4416		TAILQ_INIT(&ahc->untagged_queues[i]);
   4417	if (ahc_platform_alloc(ahc, platform_arg) != 0) {
   4418		ahc_free(ahc);
   4419		ahc = NULL;
   4420	}
   4421	return (ahc);
   4422}
   4423
   4424int
   4425ahc_softc_init(struct ahc_softc *ahc)
   4426{
   4427
   4428	/* The IRQMS bit is only valid on VL and EISA chips */
   4429	if ((ahc->chip & AHC_PCI) == 0)
   4430		ahc->unpause = ahc_inb(ahc, HCNTRL) & IRQMS;
   4431	else
   4432		ahc->unpause = 0;
   4433	ahc->pause = ahc->unpause | PAUSE;
   4434	/* XXX The shared scb data stuff should be deprecated */
   4435	if (ahc->scb_data == NULL) {
   4436		ahc->scb_data = kzalloc(sizeof(*ahc->scb_data), GFP_ATOMIC);
   4437		if (ahc->scb_data == NULL)
   4438			return (ENOMEM);
   4439	}
   4440
   4441	return (0);
   4442}
   4443
   4444void
   4445ahc_set_unit(struct ahc_softc *ahc, int unit)
   4446{
   4447	ahc->unit = unit;
   4448}
   4449
   4450void
   4451ahc_set_name(struct ahc_softc *ahc, char *name)
   4452{
   4453	kfree(ahc->name);
   4454	ahc->name = name;
   4455}
   4456
   4457void
   4458ahc_free(struct ahc_softc *ahc)
   4459{
   4460	int i;
   4461
   4462	switch (ahc->init_level) {
   4463	default:
   4464	case 5:
   4465		ahc_shutdown(ahc);
   4466		fallthrough;
   4467	case 4:
   4468		ahc_dmamap_unload(ahc, ahc->shared_data_dmat,
   4469				  ahc->shared_data_dmamap);
   4470		fallthrough;
   4471	case 3:
   4472		ahc_dmamem_free(ahc, ahc->shared_data_dmat, ahc->qoutfifo,
   4473				ahc->shared_data_dmamap);
   4474		ahc_dmamap_destroy(ahc, ahc->shared_data_dmat,
   4475				   ahc->shared_data_dmamap);
   4476		fallthrough;
   4477	case 2:
   4478		ahc_dma_tag_destroy(ahc, ahc->shared_data_dmat);
   4479		fallthrough;
   4480	case 1:
   4481		break;
   4482	case 0:
   4483		break;
   4484	}
   4485
   4486	ahc_platform_free(ahc);
   4487	ahc_fini_scbdata(ahc);
   4488	for (i = 0; i < AHC_NUM_TARGETS; i++) {
   4489		struct ahc_tmode_tstate *tstate;
   4490
   4491		tstate = ahc->enabled_targets[i];
   4492		if (tstate != NULL) {
   4493#ifdef AHC_TARGET_MODE
   4494			int j;
   4495
   4496			for (j = 0; j < AHC_NUM_LUNS; j++) {
   4497				struct ahc_tmode_lstate *lstate;
   4498
   4499				lstate = tstate->enabled_luns[j];
   4500				if (lstate != NULL) {
   4501					xpt_free_path(lstate->path);
   4502					kfree(lstate);
   4503				}
   4504			}
   4505#endif
   4506			kfree(tstate);
   4507		}
   4508	}
   4509#ifdef AHC_TARGET_MODE
   4510	if (ahc->black_hole != NULL) {
   4511		xpt_free_path(ahc->black_hole->path);
   4512		kfree(ahc->black_hole);
   4513	}
   4514#endif
   4515	kfree(ahc->name);
   4516	kfree(ahc->seep_config);
   4517	kfree(ahc);
   4518	return;
   4519}
   4520
   4521static void
   4522ahc_shutdown(void *arg)
   4523{
   4524	struct	ahc_softc *ahc;
   4525	int	i;
   4526
   4527	ahc = (struct ahc_softc *)arg;
   4528
   4529	/* This will reset most registers to 0, but not all */
   4530	ahc_reset(ahc, /*reinit*/FALSE);
   4531	ahc_outb(ahc, SCSISEQ, 0);
   4532	ahc_outb(ahc, SXFRCTL0, 0);
   4533	ahc_outb(ahc, DSPCISTATUS, 0);
   4534
   4535	for (i = TARG_SCSIRATE; i < SCSICONF; i++)
   4536		ahc_outb(ahc, i, 0);
   4537}
   4538
   4539/*
   4540 * Reset the controller and record some information about it
   4541 * that is only available just after a reset.  If "reinit" is
   4542 * non-zero, this reset occurred after initial configuration
   4543 * and the caller requests that the chip be fully reinitialized
   4544 * to a runable state.  Chip interrupts are *not* enabled after
   4545 * a reinitialization.  The caller must enable interrupts via
   4546 * ahc_intr_enable().
   4547 */
   4548int
   4549ahc_reset(struct ahc_softc *ahc, int reinit)
   4550{
   4551	u_int	sblkctl;
   4552	u_int	sxfrctl1_a, sxfrctl1_b;
   4553	int	error;
   4554	int	wait;
   4555
   4556	/*
   4557	 * Preserve the value of the SXFRCTL1 register for all channels.
   4558	 * It contains settings that affect termination and we don't want
   4559	 * to disturb the integrity of the bus.
   4560	 */
   4561	ahc_pause(ahc);
   4562	sxfrctl1_b = 0;
   4563	if ((ahc->chip & AHC_CHIPID_MASK) == AHC_AIC7770) {
   4564		u_int sblkctl;
   4565
   4566		/*
   4567		 * Save channel B's settings in case this chip
   4568		 * is setup for TWIN channel operation.
   4569		 */
   4570		sblkctl = ahc_inb(ahc, SBLKCTL);
   4571		ahc_outb(ahc, SBLKCTL, sblkctl | SELBUSB);
   4572		sxfrctl1_b = ahc_inb(ahc, SXFRCTL1);
   4573		ahc_outb(ahc, SBLKCTL, sblkctl & ~SELBUSB);
   4574	}
   4575	sxfrctl1_a = ahc_inb(ahc, SXFRCTL1);
   4576
   4577	ahc_outb(ahc, HCNTRL, CHIPRST | ahc->pause);
   4578
   4579	/*
   4580	 * Ensure that the reset has finished.  We delay 1000us
   4581	 * prior to reading the register to make sure the chip
   4582	 * has sufficiently completed its reset to handle register
   4583	 * accesses.
   4584	 */
   4585	wait = 1000;
   4586	do {
   4587		ahc_delay(1000);
   4588	} while (--wait && !(ahc_inb(ahc, HCNTRL) & CHIPRSTACK));
   4589
   4590	if (wait == 0) {
   4591		printk("%s: WARNING - Failed chip reset!  "
   4592		       "Trying to initialize anyway.\n", ahc_name(ahc));
   4593	}
   4594	ahc_outb(ahc, HCNTRL, ahc->pause);
   4595
   4596	/* Determine channel configuration */
   4597	sblkctl = ahc_inb(ahc, SBLKCTL) & (SELBUSB|SELWIDE);
   4598	/* No Twin Channel PCI cards */
   4599	if ((ahc->chip & AHC_PCI) != 0)
   4600		sblkctl &= ~SELBUSB;
   4601	switch (sblkctl) {
   4602	case 0:
   4603		/* Single Narrow Channel */
   4604		break;
   4605	case 2:
   4606		/* Wide Channel */
   4607		ahc->features |= AHC_WIDE;
   4608		break;
   4609	case 8:
   4610		/* Twin Channel */
   4611		ahc->features |= AHC_TWIN;
   4612		break;
   4613	default:
   4614		printk(" Unsupported adapter type.  Ignoring\n");
   4615		return(-1);
   4616	}
   4617
   4618	/*
   4619	 * Reload sxfrctl1.
   4620	 *
   4621	 * We must always initialize STPWEN to 1 before we
   4622	 * restore the saved values.  STPWEN is initialized
   4623	 * to a tri-state condition which can only be cleared
   4624	 * by turning it on.
   4625	 */
   4626	if ((ahc->features & AHC_TWIN) != 0) {
   4627		u_int sblkctl;
   4628
   4629		sblkctl = ahc_inb(ahc, SBLKCTL);
   4630		ahc_outb(ahc, SBLKCTL, sblkctl | SELBUSB);
   4631		ahc_outb(ahc, SXFRCTL1, sxfrctl1_b);
   4632		ahc_outb(ahc, SBLKCTL, sblkctl & ~SELBUSB);
   4633	}
   4634	ahc_outb(ahc, SXFRCTL1, sxfrctl1_a);
   4635
   4636	error = 0;
   4637	if (reinit != 0)
   4638		/*
   4639		 * If a recovery action has forced a chip reset,
   4640		 * re-initialize the chip to our liking.
   4641		 */
   4642		error = ahc->bus_chip_init(ahc);
   4643#ifdef AHC_DUMP_SEQ
   4644	else
   4645		ahc_dumpseq(ahc);
   4646#endif
   4647
   4648	return (error);
   4649}
   4650
   4651/*
   4652 * Determine the number of SCBs available on the controller
   4653 */
   4654int
   4655ahc_probe_scbs(struct ahc_softc *ahc) {
   4656	int i;
   4657
   4658	for (i = 0; i < AHC_SCB_MAX; i++) {
   4659
   4660		ahc_outb(ahc, SCBPTR, i);
   4661		ahc_outb(ahc, SCB_BASE, i);
   4662		if (ahc_inb(ahc, SCB_BASE) != i)
   4663			break;
   4664		ahc_outb(ahc, SCBPTR, 0);
   4665		if (ahc_inb(ahc, SCB_BASE) != 0)
   4666			break;
   4667	}
   4668	return (i);
   4669}
   4670
   4671static void
   4672ahc_dmamap_cb(void *arg, bus_dma_segment_t *segs, int nseg, int error) 
   4673{
   4674	dma_addr_t *baddr;
   4675
   4676	baddr = (dma_addr_t *)arg;
   4677	*baddr = segs->ds_addr;
   4678}
   4679
   4680static void
   4681ahc_build_free_scb_list(struct ahc_softc *ahc)
   4682{
   4683	int scbsize;
   4684	int i;
   4685
   4686	scbsize = 32;
   4687	if ((ahc->flags & AHC_LSCBS_ENABLED) != 0)
   4688		scbsize = 64;
   4689
   4690	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
   4691		int j;
   4692
   4693		ahc_outb(ahc, SCBPTR, i);
   4694
   4695		/*
   4696		 * Touch all SCB bytes to avoid parity errors
   4697		 * should one of our debugging routines read
   4698		 * an otherwise uninitiatlized byte.
   4699		 */
   4700		for (j = 0; j < scbsize; j++)
   4701			ahc_outb(ahc, SCB_BASE+j, 0xFF);
   4702
   4703		/* Clear the control byte. */
   4704		ahc_outb(ahc, SCB_CONTROL, 0);
   4705
   4706		/* Set the next pointer */
   4707		if ((ahc->flags & AHC_PAGESCBS) != 0)
   4708			ahc_outb(ahc, SCB_NEXT, i+1);
   4709		else
   4710			ahc_outb(ahc, SCB_NEXT, SCB_LIST_NULL);
   4711
   4712		/* Make the tag number, SCSIID, and lun invalid */
   4713		ahc_outb(ahc, SCB_TAG, SCB_LIST_NULL);
   4714		ahc_outb(ahc, SCB_SCSIID, 0xFF);
   4715		ahc_outb(ahc, SCB_LUN, 0xFF);
   4716	}
   4717
   4718	if ((ahc->flags & AHC_PAGESCBS) != 0) {
   4719		/* SCB 0 heads the free list. */
   4720		ahc_outb(ahc, FREE_SCBH, 0);
   4721	} else {
   4722		/* No free list. */
   4723		ahc_outb(ahc, FREE_SCBH, SCB_LIST_NULL);
   4724	}
   4725
   4726	/* Make sure that the last SCB terminates the free list */
   4727	ahc_outb(ahc, SCBPTR, i-1);
   4728	ahc_outb(ahc, SCB_NEXT, SCB_LIST_NULL);
   4729}
   4730
   4731static int
   4732ahc_init_scbdata(struct ahc_softc *ahc)
   4733{
   4734	struct scb_data *scb_data;
   4735
   4736	scb_data = ahc->scb_data;
   4737	SLIST_INIT(&scb_data->free_scbs);
   4738	SLIST_INIT(&scb_data->sg_maps);
   4739
   4740	/* Allocate SCB resources */
   4741	scb_data->scbarray = kcalloc(AHC_SCB_MAX_ALLOC, sizeof(struct scb),
   4742				     GFP_ATOMIC);
   4743	if (scb_data->scbarray == NULL)
   4744		return (ENOMEM);
   4745
   4746	/* Determine the number of hardware SCBs and initialize them */
   4747
   4748	scb_data->maxhscbs = ahc_probe_scbs(ahc);
   4749	if (ahc->scb_data->maxhscbs == 0) {
   4750		printk("%s: No SCB space found\n", ahc_name(ahc));
   4751		return (ENXIO);
   4752	}
   4753
   4754	/*
   4755	 * Create our DMA tags.  These tags define the kinds of device
   4756	 * accessible memory allocations and memory mappings we will
   4757	 * need to perform during normal operation.
   4758	 *
   4759	 * Unless we need to further restrict the allocation, we rely
   4760	 * on the restrictions of the parent dmat, hence the common
   4761	 * use of MAXADDR and MAXSIZE.
   4762	 */
   4763
   4764	/* DMA tag for our hardware scb structures */
   4765	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/1,
   4766			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
   4767			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
   4768			       /*highaddr*/BUS_SPACE_MAXADDR,
   4769			       /*filter*/NULL, /*filterarg*/NULL,
   4770			       AHC_SCB_MAX_ALLOC * sizeof(struct hardware_scb),
   4771			       /*nsegments*/1,
   4772			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
   4773			       /*flags*/0, &scb_data->hscb_dmat) != 0) {
   4774		goto error_exit;
   4775	}
   4776
   4777	scb_data->init_level++;
   4778
   4779	/* Allocation for our hscbs */
   4780	if (ahc_dmamem_alloc(ahc, scb_data->hscb_dmat,
   4781			     (void **)&scb_data->hscbs,
   4782			     BUS_DMA_NOWAIT, &scb_data->hscb_dmamap) != 0) {
   4783		goto error_exit;
   4784	}
   4785
   4786	scb_data->init_level++;
   4787
   4788	/* And permanently map them */
   4789	ahc_dmamap_load(ahc, scb_data->hscb_dmat, scb_data->hscb_dmamap,
   4790			scb_data->hscbs,
   4791			AHC_SCB_MAX_ALLOC * sizeof(struct hardware_scb),
   4792			ahc_dmamap_cb, &scb_data->hscb_busaddr, /*flags*/0);
   4793
   4794	scb_data->init_level++;
   4795
   4796	/* DMA tag for our sense buffers */
   4797	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/1,
   4798			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
   4799			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
   4800			       /*highaddr*/BUS_SPACE_MAXADDR,
   4801			       /*filter*/NULL, /*filterarg*/NULL,
   4802			       AHC_SCB_MAX_ALLOC * sizeof(struct scsi_sense_data),
   4803			       /*nsegments*/1,
   4804			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
   4805			       /*flags*/0, &scb_data->sense_dmat) != 0) {
   4806		goto error_exit;
   4807	}
   4808
   4809	scb_data->init_level++;
   4810
   4811	/* Allocate them */
   4812	if (ahc_dmamem_alloc(ahc, scb_data->sense_dmat,
   4813			     (void **)&scb_data->sense,
   4814			     BUS_DMA_NOWAIT, &scb_data->sense_dmamap) != 0) {
   4815		goto error_exit;
   4816	}
   4817
   4818	scb_data->init_level++;
   4819
   4820	/* And permanently map them */
   4821	ahc_dmamap_load(ahc, scb_data->sense_dmat, scb_data->sense_dmamap,
   4822			scb_data->sense,
   4823			AHC_SCB_MAX_ALLOC * sizeof(struct scsi_sense_data),
   4824			ahc_dmamap_cb, &scb_data->sense_busaddr, /*flags*/0);
   4825
   4826	scb_data->init_level++;
   4827
   4828	/* DMA tag for our S/G structures.  We allocate in page sized chunks */
   4829	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/8,
   4830			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
   4831			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
   4832			       /*highaddr*/BUS_SPACE_MAXADDR,
   4833			       /*filter*/NULL, /*filterarg*/NULL,
   4834			       PAGE_SIZE, /*nsegments*/1,
   4835			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
   4836			       /*flags*/0, &scb_data->sg_dmat) != 0) {
   4837		goto error_exit;
   4838	}
   4839
   4840	scb_data->init_level++;
   4841
   4842	/* Perform initial CCB allocation */
   4843	memset(scb_data->hscbs, 0,
   4844	       AHC_SCB_MAX_ALLOC * sizeof(struct hardware_scb));
   4845	ahc_alloc_scbs(ahc);
   4846
   4847	if (scb_data->numscbs == 0) {
   4848		printk("%s: ahc_init_scbdata - "
   4849		       "Unable to allocate initial scbs\n",
   4850		       ahc_name(ahc));
   4851		goto error_exit;
   4852	}
   4853
   4854	/*
   4855	 * Reserve the next queued SCB.
   4856	 */
   4857	ahc->next_queued_scb = ahc_get_scb(ahc);
   4858
   4859	/*
   4860	 * Note that we were successful
   4861	 */
   4862	return (0);
   4863
   4864error_exit:
   4865
   4866	return (ENOMEM);
   4867}
   4868
   4869static void
   4870ahc_fini_scbdata(struct ahc_softc *ahc)
   4871{
   4872	struct scb_data *scb_data;
   4873
   4874	scb_data = ahc->scb_data;
   4875	if (scb_data == NULL)
   4876		return;
   4877
   4878	switch (scb_data->init_level) {
   4879	default:
   4880	case 7:
   4881	{
   4882		struct sg_map_node *sg_map;
   4883
   4884		while ((sg_map = SLIST_FIRST(&scb_data->sg_maps))!= NULL) {
   4885			SLIST_REMOVE_HEAD(&scb_data->sg_maps, links);
   4886			ahc_dmamap_unload(ahc, scb_data->sg_dmat,
   4887					  sg_map->sg_dmamap);
   4888			ahc_dmamem_free(ahc, scb_data->sg_dmat,
   4889					sg_map->sg_vaddr,
   4890					sg_map->sg_dmamap);
   4891			kfree(sg_map);
   4892		}
   4893		ahc_dma_tag_destroy(ahc, scb_data->sg_dmat);
   4894	}
   4895		fallthrough;
   4896	case 6:
   4897		ahc_dmamap_unload(ahc, scb_data->sense_dmat,
   4898				  scb_data->sense_dmamap);
   4899		fallthrough;
   4900	case 5:
   4901		ahc_dmamem_free(ahc, scb_data->sense_dmat, scb_data->sense,
   4902				scb_data->sense_dmamap);
   4903		ahc_dmamap_destroy(ahc, scb_data->sense_dmat,
   4904				   scb_data->sense_dmamap);
   4905		fallthrough;
   4906	case 4:
   4907		ahc_dma_tag_destroy(ahc, scb_data->sense_dmat);
   4908		fallthrough;
   4909	case 3:
   4910		ahc_dmamap_unload(ahc, scb_data->hscb_dmat,
   4911				  scb_data->hscb_dmamap);
   4912		fallthrough;
   4913	case 2:
   4914		ahc_dmamem_free(ahc, scb_data->hscb_dmat, scb_data->hscbs,
   4915				scb_data->hscb_dmamap);
   4916		ahc_dmamap_destroy(ahc, scb_data->hscb_dmat,
   4917				   scb_data->hscb_dmamap);
   4918		fallthrough;
   4919	case 1:
   4920		ahc_dma_tag_destroy(ahc, scb_data->hscb_dmat);
   4921		break;
   4922	case 0:
   4923		break;
   4924	}
   4925	kfree(scb_data->scbarray);
   4926}
   4927
   4928static void
   4929ahc_alloc_scbs(struct ahc_softc *ahc)
   4930{
   4931	struct scb_data *scb_data;
   4932	struct scb *next_scb;
   4933	struct sg_map_node *sg_map;
   4934	dma_addr_t physaddr;
   4935	struct ahc_dma_seg *segs;
   4936	int newcount;
   4937	int i;
   4938
   4939	scb_data = ahc->scb_data;
   4940	if (scb_data->numscbs >= AHC_SCB_MAX_ALLOC)
   4941		/* Can't allocate any more */
   4942		return;
   4943
   4944	next_scb = &scb_data->scbarray[scb_data->numscbs];
   4945
   4946	sg_map = kmalloc(sizeof(*sg_map), GFP_ATOMIC);
   4947
   4948	if (sg_map == NULL)
   4949		return;
   4950
   4951	/* Allocate S/G space for the next batch of SCBS */
   4952	if (ahc_dmamem_alloc(ahc, scb_data->sg_dmat,
   4953			     (void **)&sg_map->sg_vaddr,
   4954			     BUS_DMA_NOWAIT, &sg_map->sg_dmamap) != 0) {
   4955		kfree(sg_map);
   4956		return;
   4957	}
   4958
   4959	SLIST_INSERT_HEAD(&scb_data->sg_maps, sg_map, links);
   4960
   4961	ahc_dmamap_load(ahc, scb_data->sg_dmat, sg_map->sg_dmamap,
   4962			sg_map->sg_vaddr, PAGE_SIZE, ahc_dmamap_cb,
   4963			&sg_map->sg_physaddr, /*flags*/0);
   4964
   4965	segs = sg_map->sg_vaddr;
   4966	physaddr = sg_map->sg_physaddr;
   4967
   4968	newcount = (PAGE_SIZE / (AHC_NSEG * sizeof(struct ahc_dma_seg)));
   4969	newcount = min(newcount, (AHC_SCB_MAX_ALLOC - scb_data->numscbs));
   4970	for (i = 0; i < newcount; i++) {
   4971		struct scb_platform_data *pdata;
   4972
   4973		pdata = kmalloc(sizeof(*pdata), GFP_ATOMIC);
   4974		if (pdata == NULL)
   4975			break;
   4976		next_scb->platform_data = pdata;
   4977		next_scb->sg_map = sg_map;
   4978		next_scb->sg_list = segs;
   4979		/*
   4980		 * The sequencer always starts with the second entry.
   4981		 * The first entry is embedded in the scb.
   4982		 */
   4983		next_scb->sg_list_phys = physaddr + sizeof(struct ahc_dma_seg);
   4984		next_scb->ahc_softc = ahc;
   4985		next_scb->flags = SCB_FREE;
   4986		next_scb->hscb = &scb_data->hscbs[scb_data->numscbs];
   4987		next_scb->hscb->tag = ahc->scb_data->numscbs;
   4988		SLIST_INSERT_HEAD(&ahc->scb_data->free_scbs,
   4989				  next_scb, links.sle);
   4990		segs += AHC_NSEG;
   4991		physaddr += (AHC_NSEG * sizeof(struct ahc_dma_seg));
   4992		next_scb++;
   4993		ahc->scb_data->numscbs++;
   4994	}
   4995}
   4996
   4997void
   4998ahc_controller_info(struct ahc_softc *ahc, char *buf)
   4999{
   5000	int len;
   5001
   5002	len = sprintf(buf, "%s: ", ahc_chip_names[ahc->chip & AHC_CHIPID_MASK]);
   5003	buf += len;
   5004	if ((ahc->features & AHC_TWIN) != 0)
   5005		len = sprintf(buf, "Twin Channel, A SCSI Id=%d, "
   5006			      "B SCSI Id=%d, primary %c, ",
   5007			      ahc->our_id, ahc->our_id_b,
   5008			      (ahc->flags & AHC_PRIMARY_CHANNEL) + 'A');
   5009	else {
   5010		const char *speed;
   5011		const char *type;
   5012
   5013		speed = "";
   5014		if ((ahc->features & AHC_ULTRA) != 0) {
   5015			speed = "Ultra ";
   5016		} else if ((ahc->features & AHC_DT) != 0) {
   5017			speed = "Ultra160 ";
   5018		} else if ((ahc->features & AHC_ULTRA2) != 0) {
   5019			speed = "Ultra2 ";
   5020		}
   5021		if ((ahc->features & AHC_WIDE) != 0) {
   5022			type = "Wide";
   5023		} else {
   5024			type = "Single";
   5025		}
   5026		len = sprintf(buf, "%s%s Channel %c, SCSI Id=%d, ",
   5027			      speed, type, ahc->channel, ahc->our_id);
   5028	}
   5029	buf += len;
   5030
   5031	if ((ahc->flags & AHC_PAGESCBS) != 0)
   5032		sprintf(buf, "%d/%d SCBs",
   5033			ahc->scb_data->maxhscbs, AHC_MAX_QUEUE);
   5034	else
   5035		sprintf(buf, "%d SCBs", ahc->scb_data->maxhscbs);
   5036}
   5037
   5038int
   5039ahc_chip_init(struct ahc_softc *ahc)
   5040{
   5041	int	 term;
   5042	int	 error;
   5043	u_int	 i;
   5044	u_int	 scsi_conf;
   5045	u_int	 scsiseq_template;
   5046	uint32_t physaddr;
   5047
   5048	ahc_outb(ahc, SEQ_FLAGS, 0);
   5049	ahc_outb(ahc, SEQ_FLAGS2, 0);
   5050
   5051	/* Set the SCSI Id, SXFRCTL0, SXFRCTL1, and SIMODE1, for both channels*/
   5052	if (ahc->features & AHC_TWIN) {
   5053
   5054		/*
   5055		 * Setup Channel B first.
   5056		 */
   5057		ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) | SELBUSB);
   5058		term = (ahc->flags & AHC_TERM_ENB_B) != 0 ? STPWEN : 0;
   5059		ahc_outb(ahc, SCSIID, ahc->our_id_b);
   5060		scsi_conf = ahc_inb(ahc, SCSICONF + 1);
   5061		ahc_outb(ahc, SXFRCTL1, (scsi_conf & (ENSPCHK|STIMESEL))
   5062					|term|ahc->seltime_b|ENSTIMER|ACTNEGEN);
   5063		if ((ahc->features & AHC_ULTRA2) != 0)
   5064			ahc_outb(ahc, SIMODE0, ahc_inb(ahc, SIMODE0)|ENIOERR);
   5065		ahc_outb(ahc, SIMODE1, ENSELTIMO|ENSCSIRST|ENSCSIPERR);
   5066		ahc_outb(ahc, SXFRCTL0, DFON|SPIOEN);
   5067
   5068		/* Select Channel A */
   5069		ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) & ~SELBUSB);
   5070	}
   5071	term = (ahc->flags & AHC_TERM_ENB_A) != 0 ? STPWEN : 0;
   5072	if ((ahc->features & AHC_ULTRA2) != 0)
   5073		ahc_outb(ahc, SCSIID_ULTRA2, ahc->our_id);
   5074	else
   5075		ahc_outb(ahc, SCSIID, ahc->our_id);
   5076	scsi_conf = ahc_inb(ahc, SCSICONF);
   5077	ahc_outb(ahc, SXFRCTL1, (scsi_conf & (ENSPCHK|STIMESEL))
   5078				|term|ahc->seltime
   5079				|ENSTIMER|ACTNEGEN);
   5080	if ((ahc->features & AHC_ULTRA2) != 0)
   5081		ahc_outb(ahc, SIMODE0, ahc_inb(ahc, SIMODE0)|ENIOERR);
   5082	ahc_outb(ahc, SIMODE1, ENSELTIMO|ENSCSIRST|ENSCSIPERR);
   5083	ahc_outb(ahc, SXFRCTL0, DFON|SPIOEN);
   5084
   5085	/* There are no untagged SCBs active yet. */
   5086	for (i = 0; i < 16; i++) {
   5087		ahc_unbusy_tcl(ahc, BUILD_TCL(i << 4, 0));
   5088		if ((ahc->flags & AHC_SCB_BTT) != 0) {
   5089			int lun;
   5090
   5091			/*
   5092			 * The SCB based BTT allows an entry per
   5093			 * target and lun pair.
   5094			 */
   5095			for (lun = 1; lun < AHC_NUM_LUNS; lun++)
   5096				ahc_unbusy_tcl(ahc, BUILD_TCL(i << 4, lun));
   5097		}
   5098	}
   5099
   5100	/* All of our queues are empty */
   5101	for (i = 0; i < 256; i++)
   5102		ahc->qoutfifo[i] = SCB_LIST_NULL;
   5103	ahc_sync_qoutfifo(ahc, BUS_DMASYNC_PREREAD);
   5104
   5105	for (i = 0; i < 256; i++)
   5106		ahc->qinfifo[i] = SCB_LIST_NULL;
   5107
   5108	if ((ahc->features & AHC_MULTI_TID) != 0) {
   5109		ahc_outb(ahc, TARGID, 0);
   5110		ahc_outb(ahc, TARGID + 1, 0);
   5111	}
   5112
   5113	/*
   5114	 * Tell the sequencer where it can find our arrays in memory.
   5115	 */
   5116	physaddr = ahc->scb_data->hscb_busaddr;
   5117	ahc_outb(ahc, HSCB_ADDR, physaddr & 0xFF);
   5118	ahc_outb(ahc, HSCB_ADDR + 1, (physaddr >> 8) & 0xFF);
   5119	ahc_outb(ahc, HSCB_ADDR + 2, (physaddr >> 16) & 0xFF);
   5120	ahc_outb(ahc, HSCB_ADDR + 3, (physaddr >> 24) & 0xFF);
   5121
   5122	physaddr = ahc->shared_data_busaddr;
   5123	ahc_outb(ahc, SHARED_DATA_ADDR, physaddr & 0xFF);
   5124	ahc_outb(ahc, SHARED_DATA_ADDR + 1, (physaddr >> 8) & 0xFF);
   5125	ahc_outb(ahc, SHARED_DATA_ADDR + 2, (physaddr >> 16) & 0xFF);
   5126	ahc_outb(ahc, SHARED_DATA_ADDR + 3, (physaddr >> 24) & 0xFF);
   5127
   5128	/*
   5129	 * Initialize the group code to command length table.
   5130	 * This overrides the values in TARG_SCSIRATE, so only
   5131	 * setup the table after we have processed that information.
   5132	 */
   5133	ahc_outb(ahc, CMDSIZE_TABLE, 5);
   5134	ahc_outb(ahc, CMDSIZE_TABLE + 1, 9);
   5135	ahc_outb(ahc, CMDSIZE_TABLE + 2, 9);
   5136	ahc_outb(ahc, CMDSIZE_TABLE + 3, 0);
   5137	ahc_outb(ahc, CMDSIZE_TABLE + 4, 15);
   5138	ahc_outb(ahc, CMDSIZE_TABLE + 5, 11);
   5139	ahc_outb(ahc, CMDSIZE_TABLE + 6, 0);
   5140	ahc_outb(ahc, CMDSIZE_TABLE + 7, 0);
   5141
   5142	if ((ahc->features & AHC_HS_MAILBOX) != 0)
   5143		ahc_outb(ahc, HS_MAILBOX, 0);
   5144
   5145	/* Tell the sequencer of our initial queue positions */
   5146	if ((ahc->features & AHC_TARGETMODE) != 0) {
   5147		ahc->tqinfifonext = 1;
   5148		ahc_outb(ahc, KERNEL_TQINPOS, ahc->tqinfifonext - 1);
   5149		ahc_outb(ahc, TQINPOS, ahc->tqinfifonext);
   5150	}
   5151	ahc->qinfifonext = 0;
   5152	ahc->qoutfifonext = 0;
   5153	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
   5154		ahc_outb(ahc, QOFF_CTLSTA, SCB_QSIZE_256);
   5155		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
   5156		ahc_outb(ahc, SNSCB_QOFF, ahc->qinfifonext);
   5157		ahc_outb(ahc, SDSCB_QOFF, 0);
   5158	} else {
   5159		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
   5160		ahc_outb(ahc, QINPOS, ahc->qinfifonext);
   5161		ahc_outb(ahc, QOUTPOS, ahc->qoutfifonext);
   5162	}
   5163
   5164	/* We don't have any waiting selections */
   5165	ahc_outb(ahc, WAITING_SCBH, SCB_LIST_NULL);
   5166
   5167	/* Our disconnection list is empty too */
   5168	ahc_outb(ahc, DISCONNECTED_SCBH, SCB_LIST_NULL);
   5169
   5170	/* Message out buffer starts empty */
   5171	ahc_outb(ahc, MSG_OUT, NOP);
   5172
   5173	/*
   5174	 * Setup the allowed SCSI Sequences based on operational mode.
   5175	 * If we are a target, we'll enable select in operations once
   5176	 * we've had a lun enabled.
   5177	 */
   5178	scsiseq_template = ENSELO|ENAUTOATNO|ENAUTOATNP;
   5179	if ((ahc->flags & AHC_INITIATORROLE) != 0)
   5180		scsiseq_template |= ENRSELI;
   5181	ahc_outb(ahc, SCSISEQ_TEMPLATE, scsiseq_template);
   5182
   5183	/* Initialize our list of free SCBs. */
   5184	ahc_build_free_scb_list(ahc);
   5185
   5186	/*
   5187	 * Tell the sequencer which SCB will be the next one it receives.
   5188	 */
   5189	ahc_outb(ahc, NEXT_QUEUED_SCB, ahc->next_queued_scb->hscb->tag);
   5190
   5191	/*
   5192	 * Load the Sequencer program and Enable the adapter
   5193	 * in "fast" mode.
   5194	 */
   5195	if (bootverbose)
   5196		printk("%s: Downloading Sequencer Program...",
   5197		       ahc_name(ahc));
   5198
   5199	error = ahc_loadseq(ahc);
   5200	if (error != 0)
   5201		return (error);
   5202
   5203	if ((ahc->features & AHC_ULTRA2) != 0) {
   5204		int wait;
   5205
   5206		/*
   5207		 * Wait for up to 500ms for our transceivers
   5208		 * to settle.  If the adapter does not have
   5209		 * a cable attached, the transceivers may
   5210		 * never settle, so don't complain if we
   5211		 * fail here.
   5212		 */
   5213		for (wait = 5000;
   5214		     (ahc_inb(ahc, SBLKCTL) & (ENAB40|ENAB20)) == 0 && wait;
   5215		     wait--)
   5216			ahc_delay(100);
   5217	}
   5218	ahc_restart(ahc);
   5219	return (0);
   5220}
   5221
   5222/*
   5223 * Start the board, ready for normal operation
   5224 */
   5225int
   5226ahc_init(struct ahc_softc *ahc)
   5227{
   5228	int	 max_targ;
   5229	u_int	 i;
   5230	u_int	 scsi_conf;
   5231	u_int	 ultraenb;
   5232	u_int	 discenable;
   5233	u_int	 tagenable;
   5234	size_t	 driver_data_size;
   5235
   5236#ifdef AHC_DEBUG
   5237	if ((ahc_debug & AHC_DEBUG_SEQUENCER) != 0)
   5238		ahc->flags |= AHC_SEQUENCER_DEBUG;
   5239#endif
   5240
   5241#ifdef AHC_PRINT_SRAM
   5242	printk("Scratch Ram:");
   5243	for (i = 0x20; i < 0x5f; i++) {
   5244		if (((i % 8) == 0) && (i != 0)) {
   5245			printk ("\n              ");
   5246		}
   5247		printk (" 0x%x", ahc_inb(ahc, i));
   5248	}
   5249	if ((ahc->features & AHC_MORE_SRAM) != 0) {
   5250		for (i = 0x70; i < 0x7f; i++) {
   5251			if (((i % 8) == 0) && (i != 0)) {
   5252				printk ("\n              ");
   5253			}
   5254			printk (" 0x%x", ahc_inb(ahc, i));
   5255		}
   5256	}
   5257	printk ("\n");
   5258	/*
   5259	 * Reading uninitialized scratch ram may
   5260	 * generate parity errors.
   5261	 */
   5262	ahc_outb(ahc, CLRINT, CLRPARERR);
   5263	ahc_outb(ahc, CLRINT, CLRBRKADRINT);
   5264#endif
   5265	max_targ = 15;
   5266
   5267	/*
   5268	 * Assume we have a board at this stage and it has been reset.
   5269	 */
   5270	if ((ahc->flags & AHC_USEDEFAULTS) != 0)
   5271		ahc->our_id = ahc->our_id_b = 7;
   5272
   5273	/*
   5274	 * Default to allowing initiator operations.
   5275	 */
   5276	ahc->flags |= AHC_INITIATORROLE;
   5277
   5278	/*
   5279	 * Only allow target mode features if this unit has them enabled.
   5280	 */
   5281	if ((AHC_TMODE_ENABLE & (0x1 << ahc->unit)) == 0)
   5282		ahc->features &= ~AHC_TARGETMODE;
   5283
   5284	ahc->init_level++;
   5285
   5286	/*
   5287	 * DMA tag for our command fifos and other data in system memory
   5288	 * the card's sequencer must be able to access.  For initiator
   5289	 * roles, we need to allocate space for the qinfifo and qoutfifo.
   5290	 * The qinfifo and qoutfifo are composed of 256 1 byte elements.
   5291	 * When providing for the target mode role, we must additionally
   5292	 * provide space for the incoming target command fifo and an extra
   5293	 * byte to deal with a dma bug in some chip versions.
   5294	 */
   5295	driver_data_size = 2 * 256 * sizeof(uint8_t);
   5296	if ((ahc->features & AHC_TARGETMODE) != 0)
   5297		driver_data_size += AHC_TMODE_CMDS * sizeof(struct target_cmd)
   5298				 + /*DMA WideOdd Bug Buffer*/1;
   5299	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/1,
   5300			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
   5301			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
   5302			       /*highaddr*/BUS_SPACE_MAXADDR,
   5303			       /*filter*/NULL, /*filterarg*/NULL,
   5304			       driver_data_size,
   5305			       /*nsegments*/1,
   5306			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
   5307			       /*flags*/0, &ahc->shared_data_dmat) != 0) {
   5308		return (ENOMEM);
   5309	}
   5310
   5311	ahc->init_level++;
   5312
   5313	/* Allocation of driver data */
   5314	if (ahc_dmamem_alloc(ahc, ahc->shared_data_dmat,
   5315			     (void **)&ahc->qoutfifo,
   5316			     BUS_DMA_NOWAIT, &ahc->shared_data_dmamap) != 0) {
   5317		return (ENOMEM);
   5318	}
   5319
   5320	ahc->init_level++;
   5321
   5322	/* And permanently map it in */
   5323	ahc_dmamap_load(ahc, ahc->shared_data_dmat, ahc->shared_data_dmamap,
   5324			ahc->qoutfifo, driver_data_size, ahc_dmamap_cb,
   5325			&ahc->shared_data_busaddr, /*flags*/0);
   5326
   5327	if ((ahc->features & AHC_TARGETMODE) != 0) {
   5328		ahc->targetcmds = (struct target_cmd *)ahc->qoutfifo;
   5329		ahc->qoutfifo = (uint8_t *)&ahc->targetcmds[AHC_TMODE_CMDS];
   5330		ahc->dma_bug_buf = ahc->shared_data_busaddr
   5331				 + driver_data_size - 1;
   5332		/* All target command blocks start out invalid. */
   5333		for (i = 0; i < AHC_TMODE_CMDS; i++)
   5334			ahc->targetcmds[i].cmd_valid = 0;
   5335		ahc_sync_tqinfifo(ahc, BUS_DMASYNC_PREREAD);
   5336		ahc->qoutfifo = (uint8_t *)&ahc->targetcmds[256];
   5337	}
   5338	ahc->qinfifo = &ahc->qoutfifo[256];
   5339
   5340	ahc->init_level++;
   5341
   5342	/* Allocate SCB data now that buffer_dmat is initialized */
   5343	if (ahc->scb_data->maxhscbs == 0)
   5344		if (ahc_init_scbdata(ahc) != 0)
   5345			return (ENOMEM);
   5346
   5347	/*
   5348	 * Allocate a tstate to house information for our
   5349	 * initiator presence on the bus as well as the user
   5350	 * data for any target mode initiator.
   5351	 */
   5352	if (ahc_alloc_tstate(ahc, ahc->our_id, 'A') == NULL) {
   5353		printk("%s: unable to allocate ahc_tmode_tstate.  "
   5354		       "Failing attach\n", ahc_name(ahc));
   5355		return (ENOMEM);
   5356	}
   5357
   5358	if ((ahc->features & AHC_TWIN) != 0) {
   5359		if (ahc_alloc_tstate(ahc, ahc->our_id_b, 'B') == NULL) {
   5360			printk("%s: unable to allocate ahc_tmode_tstate.  "
   5361			       "Failing attach\n", ahc_name(ahc));
   5362			return (ENOMEM);
   5363		}
   5364	}
   5365
   5366	if (ahc->scb_data->maxhscbs < AHC_SCB_MAX_ALLOC) {
   5367		ahc->flags |= AHC_PAGESCBS;
   5368	} else {
   5369		ahc->flags &= ~AHC_PAGESCBS;
   5370	}
   5371
   5372#ifdef AHC_DEBUG
   5373	if (ahc_debug & AHC_SHOW_MISC) {
   5374		printk("%s: hardware scb %u bytes; kernel scb %u bytes; "
   5375		       "ahc_dma %u bytes\n",
   5376			ahc_name(ahc),
   5377			(u_int)sizeof(struct hardware_scb),
   5378			(u_int)sizeof(struct scb),
   5379			(u_int)sizeof(struct ahc_dma_seg));
   5380	}
   5381#endif /* AHC_DEBUG */
   5382
   5383	/*
   5384	 * Look at the information that board initialization or
   5385	 * the board bios has left us.
   5386	 */
   5387	if (ahc->features & AHC_TWIN) {
   5388		scsi_conf = ahc_inb(ahc, SCSICONF + 1);
   5389		if ((scsi_conf & RESET_SCSI) != 0
   5390		 && (ahc->flags & AHC_INITIATORROLE) != 0)
   5391			ahc->flags |= AHC_RESET_BUS_B;
   5392	}
   5393
   5394	scsi_conf = ahc_inb(ahc, SCSICONF);
   5395	if ((scsi_conf & RESET_SCSI) != 0
   5396	 && (ahc->flags & AHC_INITIATORROLE) != 0)
   5397		ahc->flags |= AHC_RESET_BUS_A;
   5398
   5399	ultraenb = 0;
   5400	tagenable = ALL_TARGETS_MASK;
   5401
   5402	/* Grab the disconnection disable table and invert it for our needs */
   5403	if ((ahc->flags & AHC_USEDEFAULTS) != 0) {
   5404		printk("%s: Host Adapter Bios disabled.  Using default SCSI "
   5405			"device parameters\n", ahc_name(ahc));
   5406		ahc->flags |= AHC_EXTENDED_TRANS_A|AHC_EXTENDED_TRANS_B|
   5407			      AHC_TERM_ENB_A|AHC_TERM_ENB_B;
   5408		discenable = ALL_TARGETS_MASK;
   5409		if ((ahc->features & AHC_ULTRA) != 0)
   5410			ultraenb = ALL_TARGETS_MASK;
   5411	} else {
   5412		discenable = ~((ahc_inb(ahc, DISC_DSB + 1) << 8)
   5413			   | ahc_inb(ahc, DISC_DSB));
   5414		if ((ahc->features & (AHC_ULTRA|AHC_ULTRA2)) != 0)
   5415			ultraenb = (ahc_inb(ahc, ULTRA_ENB + 1) << 8)
   5416				      | ahc_inb(ahc, ULTRA_ENB);
   5417	}
   5418
   5419	if ((ahc->features & (AHC_WIDE|AHC_TWIN)) == 0)
   5420		max_targ = 7;
   5421
   5422	for (i = 0; i <= max_targ; i++) {
   5423		struct ahc_initiator_tinfo *tinfo;
   5424		struct ahc_tmode_tstate *tstate;
   5425		u_int our_id;
   5426		u_int target_id;
   5427		char channel;
   5428
   5429		channel = 'A';
   5430		our_id = ahc->our_id;
   5431		target_id = i;
   5432		if (i > 7 && (ahc->features & AHC_TWIN) != 0) {
   5433			channel = 'B';
   5434			our_id = ahc->our_id_b;
   5435			target_id = i % 8;
   5436		}
   5437		tinfo = ahc_fetch_transinfo(ahc, channel, our_id,
   5438					    target_id, &tstate);
   5439		/* Default to async narrow across the board */
   5440		memset(tinfo, 0, sizeof(*tinfo));
   5441		if (ahc->flags & AHC_USEDEFAULTS) {
   5442			if ((ahc->features & AHC_WIDE) != 0)
   5443				tinfo->user.width = MSG_EXT_WDTR_BUS_16_BIT;
   5444
   5445			/*
   5446			 * These will be truncated when we determine the
   5447			 * connection type we have with the target.
   5448			 */
   5449			tinfo->user.period = ahc_syncrates->period;
   5450			tinfo->user.offset = MAX_OFFSET;
   5451		} else {
   5452			u_int scsirate;
   5453			uint16_t mask;
   5454
   5455			/* Take the settings leftover in scratch RAM. */
   5456			scsirate = ahc_inb(ahc, TARG_SCSIRATE + i);
   5457			mask = (0x01 << i);
   5458			if ((ahc->features & AHC_ULTRA2) != 0) {
   5459				u_int offset;
   5460				u_int maxsync;
   5461
   5462				if ((scsirate & SOFS) == 0x0F) {
   5463					/*
   5464					 * Haven't negotiated yet,
   5465					 * so the format is different.
   5466					 */
   5467					scsirate = (scsirate & SXFR) >> 4
   5468						 | (ultraenb & mask)
   5469						  ? 0x08 : 0x0
   5470						 | (scsirate & WIDEXFER);
   5471					offset = MAX_OFFSET_ULTRA2;
   5472				} else
   5473					offset = ahc_inb(ahc, TARG_OFFSET + i);
   5474				if ((scsirate & ~WIDEXFER) == 0 && offset != 0)
   5475					/* Set to the lowest sync rate, 5MHz */
   5476					scsirate |= 0x1c;
   5477				maxsync = AHC_SYNCRATE_ULTRA2;
   5478				if ((ahc->features & AHC_DT) != 0)
   5479					maxsync = AHC_SYNCRATE_DT;
   5480				tinfo->user.period =
   5481				    ahc_find_period(ahc, scsirate, maxsync);
   5482				if (offset == 0)
   5483					tinfo->user.period = 0;
   5484				else
   5485					tinfo->user.offset = MAX_OFFSET;
   5486				if ((scsirate & SXFR_ULTRA2) <= 8/*10MHz*/
   5487				 && (ahc->features & AHC_DT) != 0)
   5488					tinfo->user.ppr_options =
   5489					    MSG_EXT_PPR_DT_REQ;
   5490			} else if ((scsirate & SOFS) != 0) {
   5491				if ((scsirate & SXFR) == 0x40
   5492				 && (ultraenb & mask) != 0) {
   5493					/* Treat 10MHz as a non-ultra speed */
   5494					scsirate &= ~SXFR;
   5495					ultraenb &= ~mask;
   5496				}
   5497				tinfo->user.period =
   5498				    ahc_find_period(ahc, scsirate,
   5499						    (ultraenb & mask)
   5500						   ? AHC_SYNCRATE_ULTRA
   5501						   : AHC_SYNCRATE_FAST);
   5502				if (tinfo->user.period != 0)
   5503					tinfo->user.offset = MAX_OFFSET;
   5504			}
   5505			if (tinfo->user.period == 0)
   5506				tinfo->user.offset = 0;
   5507			if ((scsirate & WIDEXFER) != 0
   5508			 && (ahc->features & AHC_WIDE) != 0)
   5509				tinfo->user.width = MSG_EXT_WDTR_BUS_16_BIT;
   5510			tinfo->user.protocol_version = 4;
   5511			if ((ahc->features & AHC_DT) != 0)
   5512				tinfo->user.transport_version = 3;
   5513			else
   5514				tinfo->user.transport_version = 2;
   5515			tinfo->goal.protocol_version = 2;
   5516			tinfo->goal.transport_version = 2;
   5517			tinfo->curr.protocol_version = 2;
   5518			tinfo->curr.transport_version = 2;
   5519		}
   5520		tstate->ultraenb = 0;
   5521	}
   5522	ahc->user_discenable = discenable;
   5523	ahc->user_tagenable = tagenable;
   5524
   5525	return (ahc->bus_chip_init(ahc));
   5526}
   5527
   5528void
   5529ahc_intr_enable(struct ahc_softc *ahc, int enable)
   5530{
   5531	u_int hcntrl;
   5532
   5533	hcntrl = ahc_inb(ahc, HCNTRL);
   5534	hcntrl &= ~INTEN;
   5535	ahc->pause &= ~INTEN;
   5536	ahc->unpause &= ~INTEN;
   5537	if (enable) {
   5538		hcntrl |= INTEN;
   5539		ahc->pause |= INTEN;
   5540		ahc->unpause |= INTEN;
   5541	}
   5542	ahc_outb(ahc, HCNTRL, hcntrl);
   5543}
   5544
   5545/*
   5546 * Ensure that the card is paused in a location
   5547 * outside of all critical sections and that all
   5548 * pending work is completed prior to returning.
   5549 * This routine should only be called from outside
   5550 * an interrupt context.
   5551 */
   5552void
   5553ahc_pause_and_flushwork(struct ahc_softc *ahc)
   5554{
   5555	int intstat;
   5556	int maxloops;
   5557	int paused;
   5558
   5559	maxloops = 1000;
   5560	ahc->flags |= AHC_ALL_INTERRUPTS;
   5561	paused = FALSE;
   5562	do {
   5563		if (paused) {
   5564			ahc_unpause(ahc);
   5565			/*
   5566			 * Give the sequencer some time to service
   5567			 * any active selections.
   5568			 */
   5569			ahc_delay(500);
   5570		}
   5571		ahc_intr(ahc);
   5572		ahc_pause(ahc);
   5573		paused = TRUE;
   5574		ahc_outb(ahc, SCSISEQ, ahc_inb(ahc, SCSISEQ) & ~ENSELO);
   5575		intstat = ahc_inb(ahc, INTSTAT);
   5576		if ((intstat & INT_PEND) == 0) {
   5577			ahc_clear_critical_section(ahc);
   5578			intstat = ahc_inb(ahc, INTSTAT);
   5579		}
   5580	} while (--maxloops
   5581	      && (intstat != 0xFF || (ahc->features & AHC_REMOVABLE) == 0)
   5582	      && ((intstat & INT_PEND) != 0
   5583	       || (ahc_inb(ahc, SSTAT0) & (SELDO|SELINGO)) != 0));
   5584	if (maxloops == 0) {
   5585		printk("Infinite interrupt loop, INTSTAT = %x",
   5586		       ahc_inb(ahc, INTSTAT));
   5587	}
   5588	ahc_platform_flushwork(ahc);
   5589	ahc->flags &= ~AHC_ALL_INTERRUPTS;
   5590}
   5591
   5592int __maybe_unused
   5593ahc_suspend(struct ahc_softc *ahc)
   5594{
   5595
   5596	ahc_pause_and_flushwork(ahc);
   5597
   5598	if (LIST_FIRST(&ahc->pending_scbs) != NULL) {
   5599		ahc_unpause(ahc);
   5600		return (EBUSY);
   5601	}
   5602
   5603#ifdef AHC_TARGET_MODE
   5604	/*
   5605	 * XXX What about ATIOs that have not yet been serviced?
   5606	 * Perhaps we should just refuse to be suspended if we
   5607	 * are acting in a target role.
   5608	 */
   5609	if (ahc->pending_device != NULL) {
   5610		ahc_unpause(ahc);
   5611		return (EBUSY);
   5612	}
   5613#endif
   5614	ahc_shutdown(ahc);
   5615	return (0);
   5616}
   5617
   5618int __maybe_unused
   5619ahc_resume(struct ahc_softc *ahc)
   5620{
   5621
   5622	ahc_reset(ahc, /*reinit*/TRUE);
   5623	ahc_intr_enable(ahc, TRUE);
   5624	ahc_restart(ahc);
   5625	return (0);
   5626}
   5627/************************** Busy Target Table *********************************/
   5628/*
   5629 * Return the untagged transaction id for a given target/channel lun.
   5630 * Optionally, clear the entry.
   5631 */
   5632static u_int
   5633ahc_index_busy_tcl(struct ahc_softc *ahc, u_int tcl)
   5634{
   5635	u_int scbid;
   5636	u_int target_offset;
   5637
   5638	if ((ahc->flags & AHC_SCB_BTT) != 0) {
   5639		u_int saved_scbptr;
   5640
   5641		saved_scbptr = ahc_inb(ahc, SCBPTR);
   5642		ahc_outb(ahc, SCBPTR, TCL_LUN(tcl));
   5643		scbid = ahc_inb(ahc, SCB_64_BTT + TCL_TARGET_OFFSET(tcl));
   5644		ahc_outb(ahc, SCBPTR, saved_scbptr);
   5645	} else {
   5646		target_offset = TCL_TARGET_OFFSET(tcl);
   5647		scbid = ahc_inb(ahc, BUSY_TARGETS + target_offset);
   5648	}
   5649
   5650	return (scbid);
   5651}
   5652
   5653static void
   5654ahc_unbusy_tcl(struct ahc_softc *ahc, u_int tcl)
   5655{
   5656	u_int target_offset;
   5657
   5658	if ((ahc->flags & AHC_SCB_BTT) != 0) {
   5659		u_int saved_scbptr;
   5660
   5661		saved_scbptr = ahc_inb(ahc, SCBPTR);
   5662		ahc_outb(ahc, SCBPTR, TCL_LUN(tcl));
   5663		ahc_outb(ahc, SCB_64_BTT+TCL_TARGET_OFFSET(tcl), SCB_LIST_NULL);
   5664		ahc_outb(ahc, SCBPTR, saved_scbptr);
   5665	} else {
   5666		target_offset = TCL_TARGET_OFFSET(tcl);
   5667		ahc_outb(ahc, BUSY_TARGETS + target_offset, SCB_LIST_NULL);
   5668	}
   5669}
   5670
   5671static void
   5672ahc_busy_tcl(struct ahc_softc *ahc, u_int tcl, u_int scbid)
   5673{
   5674	u_int target_offset;
   5675
   5676	if ((ahc->flags & AHC_SCB_BTT) != 0) {
   5677		u_int saved_scbptr;
   5678
   5679		saved_scbptr = ahc_inb(ahc, SCBPTR);
   5680		ahc_outb(ahc, SCBPTR, TCL_LUN(tcl));
   5681		ahc_outb(ahc, SCB_64_BTT + TCL_TARGET_OFFSET(tcl), scbid);
   5682		ahc_outb(ahc, SCBPTR, saved_scbptr);
   5683	} else {
   5684		target_offset = TCL_TARGET_OFFSET(tcl);
   5685		ahc_outb(ahc, BUSY_TARGETS + target_offset, scbid);
   5686	}
   5687}
   5688
   5689/************************** SCB and SCB queue management **********************/
   5690int
   5691ahc_match_scb(struct ahc_softc *ahc, struct scb *scb, int target,
   5692	      char channel, int lun, u_int tag, role_t role)
   5693{
   5694	int targ = SCB_GET_TARGET(ahc, scb);
   5695	char chan = SCB_GET_CHANNEL(ahc, scb);
   5696	int slun = SCB_GET_LUN(scb);
   5697	int match;
   5698
   5699	match = ((chan == channel) || (channel == ALL_CHANNELS));
   5700	if (match != 0)
   5701		match = ((targ == target) || (target == CAM_TARGET_WILDCARD));
   5702	if (match != 0)
   5703		match = ((lun == slun) || (lun == CAM_LUN_WILDCARD));
   5704	if (match != 0) {
   5705#ifdef AHC_TARGET_MODE
   5706		int group;
   5707
   5708		group = XPT_FC_GROUP(scb->io_ctx->ccb_h.func_code);
   5709		if (role == ROLE_INITIATOR) {
   5710			match = (group != XPT_FC_GROUP_TMODE)
   5711			      && ((tag == scb->hscb->tag)
   5712			       || (tag == SCB_LIST_NULL));
   5713		} else if (role == ROLE_TARGET) {
   5714			match = (group == XPT_FC_GROUP_TMODE)
   5715			      && ((tag == scb->io_ctx->csio.tag_id)
   5716			       || (tag == SCB_LIST_NULL));
   5717		}
   5718#else /* !AHC_TARGET_MODE */
   5719		match = ((tag == scb->hscb->tag) || (tag == SCB_LIST_NULL));
   5720#endif /* AHC_TARGET_MODE */
   5721	}
   5722
   5723	return match;
   5724}
   5725
   5726static void
   5727ahc_freeze_devq(struct ahc_softc *ahc, struct scb *scb)
   5728{
   5729	int	target;
   5730	char	channel;
   5731	int	lun;
   5732
   5733	target = SCB_GET_TARGET(ahc, scb);
   5734	lun = SCB_GET_LUN(scb);
   5735	channel = SCB_GET_CHANNEL(ahc, scb);
   5736
   5737	ahc_search_qinfifo(ahc, target, channel, lun,
   5738			   /*tag*/SCB_LIST_NULL, ROLE_UNKNOWN,
   5739			   CAM_REQUEUE_REQ, SEARCH_COMPLETE);
   5740
   5741	ahc_platform_freeze_devq(ahc, scb);
   5742}
   5743
   5744void
   5745ahc_qinfifo_requeue_tail(struct ahc_softc *ahc, struct scb *scb)
   5746{
   5747	struct scb *prev_scb;
   5748
   5749	prev_scb = NULL;
   5750	if (ahc_qinfifo_count(ahc) != 0) {
   5751		u_int prev_tag;
   5752		uint8_t prev_pos;
   5753
   5754		prev_pos = ahc->qinfifonext - 1;
   5755		prev_tag = ahc->qinfifo[prev_pos];
   5756		prev_scb = ahc_lookup_scb(ahc, prev_tag);
   5757	}
   5758	ahc_qinfifo_requeue(ahc, prev_scb, scb);
   5759	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
   5760		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
   5761	} else {
   5762		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
   5763	}
   5764}
   5765
   5766static void
   5767ahc_qinfifo_requeue(struct ahc_softc *ahc, struct scb *prev_scb,
   5768		    struct scb *scb)
   5769{
   5770	if (prev_scb == NULL) {
   5771		ahc_outb(ahc, NEXT_QUEUED_SCB, scb->hscb->tag);
   5772	} else {
   5773		prev_scb->hscb->next = scb->hscb->tag;
   5774		ahc_sync_scb(ahc, prev_scb,
   5775			     BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
   5776	}
   5777	ahc->qinfifo[ahc->qinfifonext++] = scb->hscb->tag;
   5778	scb->hscb->next = ahc->next_queued_scb->hscb->tag;
   5779	ahc_sync_scb(ahc, scb, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
   5780}
   5781
   5782static int
   5783ahc_qinfifo_count(struct ahc_softc *ahc)
   5784{
   5785	uint8_t qinpos;
   5786	uint8_t diff;
   5787
   5788	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
   5789		qinpos = ahc_inb(ahc, SNSCB_QOFF);
   5790		ahc_outb(ahc, SNSCB_QOFF, qinpos);
   5791	} else
   5792		qinpos = ahc_inb(ahc, QINPOS);
   5793	diff = ahc->qinfifonext - qinpos;
   5794	return (diff);
   5795}
   5796
   5797int
   5798ahc_search_qinfifo(struct ahc_softc *ahc, int target, char channel,
   5799		   int lun, u_int tag, role_t role, uint32_t status,
   5800		   ahc_search_action action)
   5801{
   5802	struct	scb *scb;
   5803	struct	scb *prev_scb;
   5804	uint8_t qinstart;
   5805	uint8_t qinpos;
   5806	uint8_t qintail;
   5807	uint8_t next;
   5808	uint8_t prev;
   5809	uint8_t curscbptr;
   5810	int	found;
   5811	int	have_qregs;
   5812
   5813	qintail = ahc->qinfifonext;
   5814	have_qregs = (ahc->features & AHC_QUEUE_REGS) != 0;
   5815	if (have_qregs) {
   5816		qinstart = ahc_inb(ahc, SNSCB_QOFF);
   5817		ahc_outb(ahc, SNSCB_QOFF, qinstart);
   5818	} else
   5819		qinstart = ahc_inb(ahc, QINPOS);
   5820	qinpos = qinstart;
   5821	found = 0;
   5822	prev_scb = NULL;
   5823
   5824	if (action == SEARCH_COMPLETE) {
   5825		/*
   5826		 * Don't attempt to run any queued untagged transactions
   5827		 * until we are done with the abort process.
   5828		 */
   5829		ahc_freeze_untagged_queues(ahc);
   5830	}
   5831
   5832	/*
   5833	 * Start with an empty queue.  Entries that are not chosen
   5834	 * for removal will be re-added to the queue as we go.
   5835	 */
   5836	ahc->qinfifonext = qinpos;
   5837	ahc_outb(ahc, NEXT_QUEUED_SCB, ahc->next_queued_scb->hscb->tag);
   5838
   5839	while (qinpos != qintail) {
   5840		scb = ahc_lookup_scb(ahc, ahc->qinfifo[qinpos]);
   5841		if (scb == NULL) {
   5842			printk("qinpos = %d, SCB index = %d\n",
   5843				qinpos, ahc->qinfifo[qinpos]);
   5844			panic("Loop 1\n");
   5845		}
   5846
   5847		if (ahc_match_scb(ahc, scb, target, channel, lun, tag, role)) {
   5848			/*
   5849			 * We found an scb that needs to be acted on.
   5850			 */
   5851			found++;
   5852			switch (action) {
   5853			case SEARCH_COMPLETE:
   5854			{
   5855				cam_status ostat;
   5856				cam_status cstat;
   5857
   5858				ostat = ahc_get_transaction_status(scb);
   5859				if (ostat == CAM_REQ_INPROG)
   5860					ahc_set_transaction_status(scb, status);
   5861				cstat = ahc_get_transaction_status(scb);
   5862				if (cstat != CAM_REQ_CMP)
   5863					ahc_freeze_scb(scb);
   5864				if ((scb->flags & SCB_ACTIVE) == 0)
   5865					printk("Inactive SCB in qinfifo\n");
   5866				ahc_done(ahc, scb);
   5867			}
   5868				fallthrough;
   5869			case SEARCH_REMOVE:
   5870				break;
   5871			case SEARCH_COUNT:
   5872				ahc_qinfifo_requeue(ahc, prev_scb, scb);
   5873				prev_scb = scb;
   5874				break;
   5875			}
   5876		} else {
   5877			ahc_qinfifo_requeue(ahc, prev_scb, scb);
   5878			prev_scb = scb;
   5879		}
   5880		qinpos++;
   5881	}
   5882
   5883	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
   5884		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
   5885	} else {
   5886		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
   5887	}
   5888
   5889	if (action != SEARCH_COUNT
   5890	 && (found != 0)
   5891	 && (qinstart != ahc->qinfifonext)) {
   5892		/*
   5893		 * The sequencer may be in the process of dmaing
   5894		 * down the SCB at the beginning of the queue.
   5895		 * This could be problematic if either the first,
   5896		 * or the second SCB is removed from the queue
   5897		 * (the first SCB includes a pointer to the "next"
   5898		 * SCB to dma). If we have removed any entries, swap
   5899		 * the first element in the queue with the next HSCB
   5900		 * so the sequencer will notice that NEXT_QUEUED_SCB
   5901		 * has changed during its dma attempt and will retry
   5902		 * the DMA.
   5903		 */
   5904		scb = ahc_lookup_scb(ahc, ahc->qinfifo[qinstart]);
   5905
   5906		if (scb == NULL) {
   5907			printk("found = %d, qinstart = %d, qinfifionext = %d\n",
   5908				found, qinstart, ahc->qinfifonext);
   5909			panic("First/Second Qinfifo fixup\n");
   5910		}
   5911		/*
   5912		 * ahc_swap_with_next_hscb forces our next pointer to
   5913		 * point to the reserved SCB for future commands.  Save
   5914		 * and restore our original next pointer to maintain
   5915		 * queue integrity.
   5916		 */
   5917		next = scb->hscb->next;
   5918		ahc->scb_data->scbindex[scb->hscb->tag] = NULL;
   5919		ahc_swap_with_next_hscb(ahc, scb);
   5920		scb->hscb->next = next;
   5921		ahc->qinfifo[qinstart] = scb->hscb->tag;
   5922
   5923		/* Tell the card about the new head of the qinfifo. */
   5924		ahc_outb(ahc, NEXT_QUEUED_SCB, scb->hscb->tag);
   5925
   5926		/* Fixup the tail "next" pointer. */
   5927		qintail = ahc->qinfifonext - 1;
   5928		scb = ahc_lookup_scb(ahc, ahc->qinfifo[qintail]);
   5929		scb->hscb->next = ahc->next_queued_scb->hscb->tag;
   5930	}
   5931
   5932	/*
   5933	 * Search waiting for selection list.
   5934	 */
   5935	curscbptr = ahc_inb(ahc, SCBPTR);
   5936	next = ahc_inb(ahc, WAITING_SCBH);  /* Start at head of list. */
   5937	prev = SCB_LIST_NULL;
   5938
   5939	while (next != SCB_LIST_NULL) {
   5940		uint8_t scb_index;
   5941
   5942		ahc_outb(ahc, SCBPTR, next);
   5943		scb_index = ahc_inb(ahc, SCB_TAG);
   5944		if (scb_index >= ahc->scb_data->numscbs) {
   5945			printk("Waiting List inconsistency. "
   5946			       "SCB index == %d, yet numscbs == %d.",
   5947			       scb_index, ahc->scb_data->numscbs);
   5948			ahc_dump_card_state(ahc);
   5949			panic("for safety");
   5950		}
   5951		scb = ahc_lookup_scb(ahc, scb_index);
   5952		if (scb == NULL) {
   5953			printk("scb_index = %d, next = %d\n",
   5954				scb_index, next);
   5955			panic("Waiting List traversal\n");
   5956		}
   5957		if (ahc_match_scb(ahc, scb, target, channel,
   5958				  lun, SCB_LIST_NULL, role)) {
   5959			/*
   5960			 * We found an scb that needs to be acted on.
   5961			 */
   5962			found++;
   5963			switch (action) {
   5964			case SEARCH_COMPLETE:
   5965			{
   5966				cam_status ostat;
   5967				cam_status cstat;
   5968
   5969				ostat = ahc_get_transaction_status(scb);
   5970				if (ostat == CAM_REQ_INPROG)
   5971					ahc_set_transaction_status(scb,
   5972								   status);
   5973				cstat = ahc_get_transaction_status(scb);
   5974				if (cstat != CAM_REQ_CMP)
   5975					ahc_freeze_scb(scb);
   5976				if ((scb->flags & SCB_ACTIVE) == 0)
   5977					printk("Inactive SCB in Waiting List\n");
   5978				ahc_done(ahc, scb);
   5979			}
   5980				fallthrough;
   5981			case SEARCH_REMOVE:
   5982				next = ahc_rem_wscb(ahc, next, prev);
   5983				break;
   5984			case SEARCH_COUNT:
   5985				prev = next;
   5986				next = ahc_inb(ahc, SCB_NEXT);
   5987				break;
   5988			}
   5989		} else {
   5990			prev = next;
   5991			next = ahc_inb(ahc, SCB_NEXT);
   5992		}
   5993	}
   5994	ahc_outb(ahc, SCBPTR, curscbptr);
   5995
   5996	found += ahc_search_untagged_queues(ahc, /*ahc_io_ctx_t*/NULL, target,
   5997					    channel, lun, status, action);
   5998
   5999	if (action == SEARCH_COMPLETE)
   6000		ahc_release_untagged_queues(ahc);
   6001	return (found);
   6002}
   6003
   6004int
   6005ahc_search_untagged_queues(struct ahc_softc *ahc, ahc_io_ctx_t ctx,
   6006			   int target, char channel, int lun, uint32_t status,
   6007			   ahc_search_action action)
   6008{
   6009	struct	scb *scb;
   6010	int	maxtarget;
   6011	int	found;
   6012	int	i;
   6013
   6014	if (action == SEARCH_COMPLETE) {
   6015		/*
   6016		 * Don't attempt to run any queued untagged transactions
   6017		 * until we are done with the abort process.
   6018		 */
   6019		ahc_freeze_untagged_queues(ahc);
   6020	}
   6021
   6022	found = 0;
   6023	i = 0;
   6024	if ((ahc->flags & AHC_SCB_BTT) == 0) {
   6025
   6026		maxtarget = 16;
   6027		if (target != CAM_TARGET_WILDCARD) {
   6028
   6029			i = target;
   6030			if (channel == 'B')
   6031				i += 8;
   6032			maxtarget = i + 1;
   6033		}
   6034	} else {
   6035		maxtarget = 0;
   6036	}
   6037
   6038	for (; i < maxtarget; i++) {
   6039		struct scb_tailq *untagged_q;
   6040		struct scb *next_scb;
   6041
   6042		untagged_q = &(ahc->untagged_queues[i]);
   6043		next_scb = TAILQ_FIRST(untagged_q);
   6044		while (next_scb != NULL) {
   6045
   6046			scb = next_scb;
   6047			next_scb = TAILQ_NEXT(scb, links.tqe);
   6048
   6049			/*
   6050			 * The head of the list may be the currently
   6051			 * active untagged command for a device.
   6052			 * We're only searching for commands that
   6053			 * have not been started.  A transaction
   6054			 * marked active but still in the qinfifo
   6055			 * is removed by the qinfifo scanning code
   6056			 * above.
   6057			 */
   6058			if ((scb->flags & SCB_ACTIVE) != 0)
   6059				continue;
   6060
   6061			if (ahc_match_scb(ahc, scb, target, channel, lun,
   6062					  SCB_LIST_NULL, ROLE_INITIATOR) == 0
   6063			 || (ctx != NULL && ctx != scb->io_ctx))
   6064				continue;
   6065
   6066			/*
   6067			 * We found an scb that needs to be acted on.
   6068			 */
   6069			found++;
   6070			switch (action) {
   6071			case SEARCH_COMPLETE:
   6072			{
   6073				cam_status ostat;
   6074				cam_status cstat;
   6075
   6076				ostat = ahc_get_transaction_status(scb);
   6077				if (ostat == CAM_REQ_INPROG)
   6078					ahc_set_transaction_status(scb, status);
   6079				cstat = ahc_get_transaction_status(scb);
   6080				if (cstat != CAM_REQ_CMP)
   6081					ahc_freeze_scb(scb);
   6082				if ((scb->flags & SCB_ACTIVE) == 0)
   6083					printk("Inactive SCB in untaggedQ\n");
   6084				ahc_done(ahc, scb);
   6085				break;
   6086			}
   6087			case SEARCH_REMOVE:
   6088				scb->flags &= ~SCB_UNTAGGEDQ;
   6089				TAILQ_REMOVE(untagged_q, scb, links.tqe);
   6090				break;
   6091			case SEARCH_COUNT:
   6092				break;
   6093			}
   6094		}
   6095	}
   6096
   6097	if (action == SEARCH_COMPLETE)
   6098		ahc_release_untagged_queues(ahc);
   6099	return (found);
   6100}
   6101
   6102int
   6103ahc_search_disc_list(struct ahc_softc *ahc, int target, char channel,
   6104		     int lun, u_int tag, int stop_on_first, int remove,
   6105		     int save_state)
   6106{
   6107	struct	scb *scbp;
   6108	u_int	next;
   6109	u_int	prev;
   6110	u_int	count;
   6111	u_int	active_scb;
   6112
   6113	count = 0;
   6114	next = ahc_inb(ahc, DISCONNECTED_SCBH);
   6115	prev = SCB_LIST_NULL;
   6116
   6117	if (save_state) {
   6118		/* restore this when we're done */
   6119		active_scb = ahc_inb(ahc, SCBPTR);
   6120	} else
   6121		/* Silence compiler */
   6122		active_scb = SCB_LIST_NULL;
   6123
   6124	while (next != SCB_LIST_NULL) {
   6125		u_int scb_index;
   6126
   6127		ahc_outb(ahc, SCBPTR, next);
   6128		scb_index = ahc_inb(ahc, SCB_TAG);
   6129		if (scb_index >= ahc->scb_data->numscbs) {
   6130			printk("Disconnected List inconsistency. "
   6131			       "SCB index == %d, yet numscbs == %d.",
   6132			       scb_index, ahc->scb_data->numscbs);
   6133			ahc_dump_card_state(ahc);
   6134			panic("for safety");
   6135		}
   6136
   6137		if (next == prev) {
   6138			panic("Disconnected List Loop. "
   6139			      "cur SCBPTR == %x, prev SCBPTR == %x.",
   6140			      next, prev);
   6141		}
   6142		scbp = ahc_lookup_scb(ahc, scb_index);
   6143		if (ahc_match_scb(ahc, scbp, target, channel, lun,
   6144				  tag, ROLE_INITIATOR)) {
   6145			count++;
   6146			if (remove) {
   6147				next =
   6148				    ahc_rem_scb_from_disc_list(ahc, prev, next);
   6149			} else {
   6150				prev = next;
   6151				next = ahc_inb(ahc, SCB_NEXT);
   6152			}
   6153			if (stop_on_first)
   6154				break;
   6155		} else {
   6156			prev = next;
   6157			next = ahc_inb(ahc, SCB_NEXT);
   6158		}
   6159	}
   6160	if (save_state)
   6161		ahc_outb(ahc, SCBPTR, active_scb);
   6162	return (count);
   6163}
   6164
   6165/*
   6166 * Remove an SCB from the on chip list of disconnected transactions.
   6167 * This is empty/unused if we are not performing SCB paging.
   6168 */
   6169static u_int
   6170ahc_rem_scb_from_disc_list(struct ahc_softc *ahc, u_int prev, u_int scbptr)
   6171{
   6172	u_int next;
   6173
   6174	ahc_outb(ahc, SCBPTR, scbptr);
   6175	next = ahc_inb(ahc, SCB_NEXT);
   6176
   6177	ahc_outb(ahc, SCB_CONTROL, 0);
   6178
   6179	ahc_add_curscb_to_free_list(ahc);
   6180
   6181	if (prev != SCB_LIST_NULL) {
   6182		ahc_outb(ahc, SCBPTR, prev);
   6183		ahc_outb(ahc, SCB_NEXT, next);
   6184	} else
   6185		ahc_outb(ahc, DISCONNECTED_SCBH, next);
   6186
   6187	return (next);
   6188}
   6189
   6190/*
   6191 * Add the SCB as selected by SCBPTR onto the on chip list of
   6192 * free hardware SCBs.  This list is empty/unused if we are not
   6193 * performing SCB paging.
   6194 */
   6195static void
   6196ahc_add_curscb_to_free_list(struct ahc_softc *ahc)
   6197{
   6198	/*
   6199	 * Invalidate the tag so that our abort
   6200	 * routines don't think it's active.
   6201	 */
   6202	ahc_outb(ahc, SCB_TAG, SCB_LIST_NULL);
   6203
   6204	if ((ahc->flags & AHC_PAGESCBS) != 0) {
   6205		ahc_outb(ahc, SCB_NEXT, ahc_inb(ahc, FREE_SCBH));
   6206		ahc_outb(ahc, FREE_SCBH, ahc_inb(ahc, SCBPTR));
   6207	}
   6208}
   6209
   6210/*
   6211 * Manipulate the waiting for selection list and return the
   6212 * scb that follows the one that we remove.
   6213 */
   6214static u_int
   6215ahc_rem_wscb(struct ahc_softc *ahc, u_int scbpos, u_int prev)
   6216{
   6217	u_int curscb, next;
   6218
   6219	/*
   6220	 * Select the SCB we want to abort and
   6221	 * pull the next pointer out of it.
   6222	 */
   6223	curscb = ahc_inb(ahc, SCBPTR);
   6224	ahc_outb(ahc, SCBPTR, scbpos);
   6225	next = ahc_inb(ahc, SCB_NEXT);
   6226
   6227	/* Clear the necessary fields */
   6228	ahc_outb(ahc, SCB_CONTROL, 0);
   6229
   6230	ahc_add_curscb_to_free_list(ahc);
   6231
   6232	/* update the waiting list */
   6233	if (prev == SCB_LIST_NULL) {
   6234		/* First in the list */
   6235		ahc_outb(ahc, WAITING_SCBH, next);
   6236
   6237		/*
   6238		 * Ensure we aren't attempting to perform
   6239		 * selection for this entry.
   6240		 */
   6241		ahc_outb(ahc, SCSISEQ, (ahc_inb(ahc, SCSISEQ) & ~ENSELO));
   6242	} else {
   6243		/*
   6244		 * Select the scb that pointed to us
   6245		 * and update its next pointer.
   6246		 */
   6247		ahc_outb(ahc, SCBPTR, prev);
   6248		ahc_outb(ahc, SCB_NEXT, next);
   6249	}
   6250
   6251	/*
   6252	 * Point us back at the original scb position.
   6253	 */
   6254	ahc_outb(ahc, SCBPTR, curscb);
   6255	return next;
   6256}
   6257
   6258/******************************** Error Handling ******************************/
   6259/*
   6260 * Abort all SCBs that match the given description (target/channel/lun/tag),
   6261 * setting their status to the passed in status if the status has not already
   6262 * been modified from CAM_REQ_INPROG.  This routine assumes that the sequencer
   6263 * is paused before it is called.
   6264 */
   6265static int
   6266ahc_abort_scbs(struct ahc_softc *ahc, int target, char channel,
   6267	       int lun, u_int tag, role_t role, uint32_t status)
   6268{
   6269	struct	scb *scbp;
   6270	struct	scb *scbp_next;
   6271	u_int	active_scb;
   6272	int	i, j;
   6273	int	maxtarget;
   6274	int	minlun;
   6275	int	maxlun;
   6276
   6277	int	found;
   6278
   6279	/*
   6280	 * Don't attempt to run any queued untagged transactions
   6281	 * until we are done with the abort process.
   6282	 */
   6283	ahc_freeze_untagged_queues(ahc);
   6284
   6285	/* restore this when we're done */
   6286	active_scb = ahc_inb(ahc, SCBPTR);
   6287
   6288	found = ahc_search_qinfifo(ahc, target, channel, lun, SCB_LIST_NULL,
   6289				   role, CAM_REQUEUE_REQ, SEARCH_COMPLETE);
   6290
   6291	/*
   6292	 * Clean out the busy target table for any untagged commands.
   6293	 */
   6294	i = 0;
   6295	maxtarget = 16;
   6296	if (target != CAM_TARGET_WILDCARD) {
   6297		i = target;
   6298		if (channel == 'B')
   6299			i += 8;
   6300		maxtarget = i + 1;
   6301	}
   6302
   6303	if (lun == CAM_LUN_WILDCARD) {
   6304
   6305		/*
   6306		 * Unless we are using an SCB based
   6307		 * busy targets table, there is only
   6308		 * one table entry for all luns of
   6309		 * a target.
   6310		 */
   6311		minlun = 0;
   6312		maxlun = 1;
   6313		if ((ahc->flags & AHC_SCB_BTT) != 0)
   6314			maxlun = AHC_NUM_LUNS;
   6315	} else {
   6316		minlun = lun;
   6317		maxlun = lun + 1;
   6318	}
   6319
   6320	if (role != ROLE_TARGET) {
   6321		for (;i < maxtarget; i++) {
   6322			for (j = minlun;j < maxlun; j++) {
   6323				u_int scbid;
   6324				u_int tcl;
   6325
   6326				tcl = BUILD_TCL(i << 4, j);
   6327				scbid = ahc_index_busy_tcl(ahc, tcl);
   6328				scbp = ahc_lookup_scb(ahc, scbid);
   6329				if (scbp == NULL
   6330				 || ahc_match_scb(ahc, scbp, target, channel,
   6331						  lun, tag, role) == 0)
   6332					continue;
   6333				ahc_unbusy_tcl(ahc, BUILD_TCL(i << 4, j));
   6334			}
   6335		}
   6336
   6337		/*
   6338		 * Go through the disconnected list and remove any entries we
   6339		 * have queued for completion, 0'ing their control byte too.
   6340		 * We save the active SCB and restore it ourselves, so there
   6341		 * is no reason for this search to restore it too.
   6342		 */
   6343		ahc_search_disc_list(ahc, target, channel, lun, tag,
   6344				     /*stop_on_first*/FALSE, /*remove*/TRUE,
   6345				     /*save_state*/FALSE);
   6346	}
   6347
   6348	/*
   6349	 * Go through the hardware SCB array looking for commands that
   6350	 * were active but not on any list.  In some cases, these remnants
   6351	 * might not still have mappings in the scbindex array (e.g. unexpected
   6352	 * bus free with the same scb queued for an abort).  Don't hold this
   6353	 * against them.
   6354	 */
   6355	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
   6356		u_int scbid;
   6357
   6358		ahc_outb(ahc, SCBPTR, i);
   6359		scbid = ahc_inb(ahc, SCB_TAG);
   6360		scbp = ahc_lookup_scb(ahc, scbid);
   6361		if ((scbp == NULL && scbid != SCB_LIST_NULL)
   6362		 || (scbp != NULL
   6363		  && ahc_match_scb(ahc, scbp, target, channel, lun, tag, role)))
   6364			ahc_add_curscb_to_free_list(ahc);
   6365	}
   6366
   6367	/*
   6368	 * Go through the pending CCB list and look for
   6369	 * commands for this target that are still active.
   6370	 * These are other tagged commands that were
   6371	 * disconnected when the reset occurred.
   6372	 */
   6373	scbp_next = LIST_FIRST(&ahc->pending_scbs);
   6374	while (scbp_next != NULL) {
   6375		scbp = scbp_next;
   6376		scbp_next = LIST_NEXT(scbp, pending_links);
   6377		if (ahc_match_scb(ahc, scbp, target, channel, lun, tag, role)) {
   6378			cam_status ostat;
   6379
   6380			ostat = ahc_get_transaction_status(scbp);
   6381			if (ostat == CAM_REQ_INPROG)
   6382				ahc_set_transaction_status(scbp, status);
   6383			if (ahc_get_transaction_status(scbp) != CAM_REQ_CMP)
   6384				ahc_freeze_scb(scbp);
   6385			if ((scbp->flags & SCB_ACTIVE) == 0)
   6386				printk("Inactive SCB on pending list\n");
   6387			ahc_done(ahc, scbp);
   6388			found++;
   6389		}
   6390	}
   6391	ahc_outb(ahc, SCBPTR, active_scb);
   6392	ahc_platform_abort_scbs(ahc, target, channel, lun, tag, role, status);
   6393	ahc_release_untagged_queues(ahc);
   6394	return found;
   6395}
   6396
   6397static void
   6398ahc_reset_current_bus(struct ahc_softc *ahc)
   6399{
   6400	uint8_t scsiseq;
   6401
   6402	ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) & ~ENSCSIRST);
   6403	scsiseq = ahc_inb(ahc, SCSISEQ);
   6404	ahc_outb(ahc, SCSISEQ, scsiseq | SCSIRSTO);
   6405	ahc_flush_device_writes(ahc);
   6406	ahc_delay(AHC_BUSRESET_DELAY);
   6407	/* Turn off the bus reset */
   6408	ahc_outb(ahc, SCSISEQ, scsiseq & ~SCSIRSTO);
   6409
   6410	ahc_clear_intstat(ahc);
   6411
   6412	/* Re-enable reset interrupts */
   6413	ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) | ENSCSIRST);
   6414}
   6415
   6416int
   6417ahc_reset_channel(struct ahc_softc *ahc, char channel, int initiate_reset)
   6418{
   6419	struct	ahc_devinfo devinfo;
   6420	u_int	initiator, target, max_scsiid;
   6421	u_int	sblkctl;
   6422	u_int	scsiseq;
   6423	u_int	simode1;
   6424	int	found;
   6425	int	restart_needed;
   6426	char	cur_channel;
   6427
   6428	ahc->pending_device = NULL;
   6429
   6430	ahc_compile_devinfo(&devinfo,
   6431			    CAM_TARGET_WILDCARD,
   6432			    CAM_TARGET_WILDCARD,
   6433			    CAM_LUN_WILDCARD,
   6434			    channel, ROLE_UNKNOWN);
   6435	ahc_pause(ahc);
   6436
   6437	/* Make sure the sequencer is in a safe location. */
   6438	ahc_clear_critical_section(ahc);
   6439
   6440	/*
   6441	 * Run our command complete fifos to ensure that we perform
   6442	 * completion processing on any commands that 'completed'
   6443	 * before the reset occurred.
   6444	 */
   6445	ahc_run_qoutfifo(ahc);
   6446#ifdef AHC_TARGET_MODE
   6447	/*
   6448	 * XXX - In Twin mode, the tqinfifo may have commands
   6449	 *	 for an unaffected channel in it.  However, if
   6450	 *	 we have run out of ATIO resources to drain that
   6451	 *	 queue, we may not get them all out here.  Further,
   6452	 *	 the blocked transactions for the reset channel
   6453	 *	 should just be killed off, irrespecitve of whether
   6454	 *	 we are blocked on ATIO resources.  Write a routine
   6455	 *	 to compact the tqinfifo appropriately.
   6456	 */
   6457	if ((ahc->flags & AHC_TARGETROLE) != 0) {
   6458		ahc_run_tqinfifo(ahc, /*paused*/TRUE);
   6459	}
   6460#endif
   6461
   6462	/*
   6463	 * Reset the bus if we are initiating this reset
   6464	 */
   6465	sblkctl = ahc_inb(ahc, SBLKCTL);
   6466	cur_channel = 'A';
   6467	if ((ahc->features & AHC_TWIN) != 0
   6468	 && ((sblkctl & SELBUSB) != 0))
   6469	    cur_channel = 'B';
   6470	scsiseq = ahc_inb(ahc, SCSISEQ_TEMPLATE);
   6471	if (cur_channel != channel) {
   6472		/* Case 1: Command for another bus is active
   6473		 * Stealthily reset the other bus without
   6474		 * upsetting the current bus.
   6475		 */
   6476		ahc_outb(ahc, SBLKCTL, sblkctl ^ SELBUSB);
   6477		simode1 = ahc_inb(ahc, SIMODE1) & ~(ENBUSFREE|ENSCSIRST);
   6478#ifdef AHC_TARGET_MODE
   6479		/*
   6480		 * Bus resets clear ENSELI, so we cannot
   6481		 * defer re-enabling bus reset interrupts
   6482		 * if we are in target mode.
   6483		 */
   6484		if ((ahc->flags & AHC_TARGETROLE) != 0)
   6485			simode1 |= ENSCSIRST;
   6486#endif
   6487		ahc_outb(ahc, SIMODE1, simode1);
   6488		if (initiate_reset)
   6489			ahc_reset_current_bus(ahc);
   6490		ahc_clear_intstat(ahc);
   6491		ahc_outb(ahc, SCSISEQ, scsiseq & (ENSELI|ENRSELI|ENAUTOATNP));
   6492		ahc_outb(ahc, SBLKCTL, sblkctl);
   6493		restart_needed = FALSE;
   6494	} else {
   6495		/* Case 2: A command from this bus is active or we're idle */
   6496		simode1 = ahc_inb(ahc, SIMODE1) & ~(ENBUSFREE|ENSCSIRST);
   6497#ifdef AHC_TARGET_MODE
   6498		/*
   6499		 * Bus resets clear ENSELI, so we cannot
   6500		 * defer re-enabling bus reset interrupts
   6501		 * if we are in target mode.
   6502		 */
   6503		if ((ahc->flags & AHC_TARGETROLE) != 0)
   6504			simode1 |= ENSCSIRST;
   6505#endif
   6506		ahc_outb(ahc, SIMODE1, simode1);
   6507		if (initiate_reset)
   6508			ahc_reset_current_bus(ahc);
   6509		ahc_clear_intstat(ahc);
   6510		ahc_outb(ahc, SCSISEQ, scsiseq & (ENSELI|ENRSELI|ENAUTOATNP));
   6511		restart_needed = TRUE;
   6512	}
   6513
   6514	/*
   6515	 * Clean up all the state information for the
   6516	 * pending transactions on this bus.
   6517	 */
   6518	found = ahc_abort_scbs(ahc, CAM_TARGET_WILDCARD, channel,
   6519			       CAM_LUN_WILDCARD, SCB_LIST_NULL,
   6520			       ROLE_UNKNOWN, CAM_SCSI_BUS_RESET);
   6521
   6522	max_scsiid = (ahc->features & AHC_WIDE) ? 15 : 7;
   6523
   6524#ifdef AHC_TARGET_MODE
   6525	/*
   6526	 * Send an immediate notify ccb to all target more peripheral
   6527	 * drivers affected by this action.
   6528	 */
   6529	for (target = 0; target <= max_scsiid; target++) {
   6530		struct ahc_tmode_tstate* tstate;
   6531		u_int lun;
   6532
   6533		tstate = ahc->enabled_targets[target];
   6534		if (tstate == NULL)
   6535			continue;
   6536		for (lun = 0; lun < AHC_NUM_LUNS; lun++) {
   6537			struct ahc_tmode_lstate* lstate;
   6538
   6539			lstate = tstate->enabled_luns[lun];
   6540			if (lstate == NULL)
   6541				continue;
   6542
   6543			ahc_queue_lstate_event(ahc, lstate, CAM_TARGET_WILDCARD,
   6544					       EVENT_TYPE_BUS_RESET, /*arg*/0);
   6545			ahc_send_lstate_events(ahc, lstate);
   6546		}
   6547	}
   6548#endif
   6549	/* Notify the XPT that a bus reset occurred */
   6550	ahc_send_async(ahc, devinfo.channel, CAM_TARGET_WILDCARD,
   6551		       CAM_LUN_WILDCARD, AC_BUS_RESET);
   6552
   6553	/*
   6554	 * Revert to async/narrow transfers until we renegotiate.
   6555	 */
   6556	for (target = 0; target <= max_scsiid; target++) {
   6557
   6558		if (ahc->enabled_targets[target] == NULL)
   6559			continue;
   6560		for (initiator = 0; initiator <= max_scsiid; initiator++) {
   6561			struct ahc_devinfo devinfo;
   6562
   6563			ahc_compile_devinfo(&devinfo, target, initiator,
   6564					    CAM_LUN_WILDCARD,
   6565					    channel, ROLE_UNKNOWN);
   6566			ahc_set_width(ahc, &devinfo, MSG_EXT_WDTR_BUS_8_BIT,
   6567				      AHC_TRANS_CUR, /*paused*/TRUE);
   6568			ahc_set_syncrate(ahc, &devinfo, /*syncrate*/NULL,
   6569					 /*period*/0, /*offset*/0,
   6570					 /*ppr_options*/0, AHC_TRANS_CUR,
   6571					 /*paused*/TRUE);
   6572		}
   6573	}
   6574
   6575	if (restart_needed)
   6576		ahc_restart(ahc);
   6577	else
   6578		ahc_unpause(ahc);
   6579	return found;
   6580}
   6581
   6582
   6583/***************************** Residual Processing ****************************/
   6584/*
   6585 * Calculate the residual for a just completed SCB.
   6586 */
   6587static void
   6588ahc_calc_residual(struct ahc_softc *ahc, struct scb *scb)
   6589{
   6590	struct hardware_scb *hscb;
   6591	struct status_pkt *spkt;
   6592	uint32_t sgptr;
   6593	uint32_t resid_sgptr;
   6594	uint32_t resid;
   6595
   6596	/*
   6597	 * 5 cases.
   6598	 * 1) No residual.
   6599	 *    SG_RESID_VALID clear in sgptr.
   6600	 * 2) Transferless command
   6601	 * 3) Never performed any transfers.
   6602	 *    sgptr has SG_FULL_RESID set.
   6603	 * 4) No residual but target did not
   6604	 *    save data pointers after the
   6605	 *    last transfer, so sgptr was
   6606	 *    never updated.
   6607	 * 5) We have a partial residual.
   6608	 *    Use residual_sgptr to determine
   6609	 *    where we are.
   6610	 */
   6611
   6612	hscb = scb->hscb;
   6613	sgptr = ahc_le32toh(hscb->sgptr);
   6614	if ((sgptr & SG_RESID_VALID) == 0)
   6615		/* Case 1 */
   6616		return;
   6617	sgptr &= ~SG_RESID_VALID;
   6618
   6619	if ((sgptr & SG_LIST_NULL) != 0)
   6620		/* Case 2 */
   6621		return;
   6622
   6623	spkt = &hscb->shared_data.status;
   6624	resid_sgptr = ahc_le32toh(spkt->residual_sg_ptr);
   6625	if ((sgptr & SG_FULL_RESID) != 0) {
   6626		/* Case 3 */
   6627		resid = ahc_get_transfer_length(scb);
   6628	} else if ((resid_sgptr & SG_LIST_NULL) != 0) {
   6629		/* Case 4 */
   6630		return;
   6631	} else if ((resid_sgptr & ~SG_PTR_MASK) != 0) {
   6632		panic("Bogus resid sgptr value 0x%x\n", resid_sgptr);
   6633	} else {
   6634		struct ahc_dma_seg *sg;
   6635
   6636		/*
   6637		 * Remainder of the SG where the transfer
   6638		 * stopped.
   6639		 */
   6640		resid = ahc_le32toh(spkt->residual_datacnt) & AHC_SG_LEN_MASK;
   6641		sg = ahc_sg_bus_to_virt(scb, resid_sgptr & SG_PTR_MASK);
   6642
   6643		/* The residual sg_ptr always points to the next sg */
   6644		sg--;
   6645
   6646		/*
   6647		 * Add up the contents of all residual
   6648		 * SG segments that are after the SG where
   6649		 * the transfer stopped.
   6650		 */
   6651		while ((ahc_le32toh(sg->len) & AHC_DMA_LAST_SEG) == 0) {
   6652			sg++;
   6653			resid += ahc_le32toh(sg->len) & AHC_SG_LEN_MASK;
   6654		}
   6655	}
   6656	if ((scb->flags & SCB_SENSE) == 0)
   6657		ahc_set_residual(scb, resid);
   6658	else
   6659		ahc_set_sense_residual(scb, resid);
   6660
   6661#ifdef AHC_DEBUG
   6662	if ((ahc_debug & AHC_SHOW_MISC) != 0) {
   6663		ahc_print_path(ahc, scb);
   6664		printk("Handled %sResidual of %d bytes\n",
   6665		       (scb->flags & SCB_SENSE) ? "Sense " : "", resid);
   6666	}
   6667#endif
   6668}
   6669
   6670/******************************* Target Mode **********************************/
   6671#ifdef AHC_TARGET_MODE
   6672/*
   6673 * Add a target mode event to this lun's queue
   6674 */
   6675static void
   6676ahc_queue_lstate_event(struct ahc_softc *ahc, struct ahc_tmode_lstate *lstate,
   6677		       u_int initiator_id, u_int event_type, u_int event_arg)
   6678{
   6679	struct ahc_tmode_event *event;
   6680	int pending;
   6681
   6682	xpt_freeze_devq(lstate->path, /*count*/1);
   6683	if (lstate->event_w_idx >= lstate->event_r_idx)
   6684		pending = lstate->event_w_idx - lstate->event_r_idx;
   6685	else
   6686		pending = AHC_TMODE_EVENT_BUFFER_SIZE + 1
   6687			- (lstate->event_r_idx - lstate->event_w_idx);
   6688
   6689	if (event_type == EVENT_TYPE_BUS_RESET
   6690	 || event_type == TARGET_RESET) {
   6691		/*
   6692		 * Any earlier events are irrelevant, so reset our buffer.
   6693		 * This has the effect of allowing us to deal with reset
   6694		 * floods (an external device holding down the reset line)
   6695		 * without losing the event that is really interesting.
   6696		 */
   6697		lstate->event_r_idx = 0;
   6698		lstate->event_w_idx = 0;
   6699		xpt_release_devq(lstate->path, pending, /*runqueue*/FALSE);
   6700	}
   6701
   6702	if (pending == AHC_TMODE_EVENT_BUFFER_SIZE) {
   6703		xpt_print_path(lstate->path);
   6704		printk("immediate event %x:%x lost\n",
   6705		       lstate->event_buffer[lstate->event_r_idx].event_type,
   6706		       lstate->event_buffer[lstate->event_r_idx].event_arg);
   6707		lstate->event_r_idx++;
   6708		if (lstate->event_r_idx == AHC_TMODE_EVENT_BUFFER_SIZE)
   6709			lstate->event_r_idx = 0;
   6710		xpt_release_devq(lstate->path, /*count*/1, /*runqueue*/FALSE);
   6711	}
   6712
   6713	event = &lstate->event_buffer[lstate->event_w_idx];
   6714	event->initiator_id = initiator_id;
   6715	event->event_type = event_type;
   6716	event->event_arg = event_arg;
   6717	lstate->event_w_idx++;
   6718	if (lstate->event_w_idx == AHC_TMODE_EVENT_BUFFER_SIZE)
   6719		lstate->event_w_idx = 0;
   6720}
   6721
   6722/*
   6723 * Send any target mode events queued up waiting
   6724 * for immediate notify resources.
   6725 */
   6726void
   6727ahc_send_lstate_events(struct ahc_softc *ahc, struct ahc_tmode_lstate *lstate)
   6728{
   6729	struct ccb_hdr *ccbh;
   6730	struct ccb_immed_notify *inot;
   6731
   6732	while (lstate->event_r_idx != lstate->event_w_idx
   6733	    && (ccbh = SLIST_FIRST(&lstate->immed_notifies)) != NULL) {
   6734		struct ahc_tmode_event *event;
   6735
   6736		event = &lstate->event_buffer[lstate->event_r_idx];
   6737		SLIST_REMOVE_HEAD(&lstate->immed_notifies, sim_links.sle);
   6738		inot = (struct ccb_immed_notify *)ccbh;
   6739		switch (event->event_type) {
   6740		case EVENT_TYPE_BUS_RESET:
   6741			ccbh->status = CAM_SCSI_BUS_RESET|CAM_DEV_QFRZN;
   6742			break;
   6743		default:
   6744			ccbh->status = CAM_MESSAGE_RECV|CAM_DEV_QFRZN;
   6745			inot->message_args[0] = event->event_type;
   6746			inot->message_args[1] = event->event_arg;
   6747			break;
   6748		}
   6749		inot->initiator_id = event->initiator_id;
   6750		inot->sense_len = 0;
   6751		xpt_done((union ccb *)inot);
   6752		lstate->event_r_idx++;
   6753		if (lstate->event_r_idx == AHC_TMODE_EVENT_BUFFER_SIZE)
   6754			lstate->event_r_idx = 0;
   6755	}
   6756}
   6757#endif
   6758
   6759/******************** Sequencer Program Patching/Download *********************/
   6760
   6761#ifdef AHC_DUMP_SEQ
   6762void
   6763ahc_dumpseq(struct ahc_softc* ahc)
   6764{
   6765	int i;
   6766
   6767	ahc_outb(ahc, SEQCTL, PERRORDIS|FAILDIS|FASTMODE|LOADRAM);
   6768	ahc_outb(ahc, SEQADDR0, 0);
   6769	ahc_outb(ahc, SEQADDR1, 0);
   6770	for (i = 0; i < ahc->instruction_ram_size; i++) {
   6771		uint8_t ins_bytes[4];
   6772
   6773		ahc_insb(ahc, SEQRAM, ins_bytes, 4);
   6774		printk("0x%08x\n", ins_bytes[0] << 24
   6775				 | ins_bytes[1] << 16
   6776				 | ins_bytes[2] << 8
   6777				 | ins_bytes[3]);
   6778	}
   6779}
   6780#endif
   6781
   6782static int
   6783ahc_loadseq(struct ahc_softc *ahc)
   6784{
   6785	struct	cs cs_table[NUM_CRITICAL_SECTIONS];
   6786	u_int	begin_set[NUM_CRITICAL_SECTIONS];
   6787	u_int	end_set[NUM_CRITICAL_SECTIONS];
   6788	const struct patch *cur_patch;
   6789	u_int	cs_count;
   6790	u_int	cur_cs;
   6791	u_int	i;
   6792	u_int	skip_addr;
   6793	u_int	sg_prefetch_cnt;
   6794	int	downloaded;
   6795	uint8_t	download_consts[7];
   6796
   6797	/*
   6798	 * Start out with 0 critical sections
   6799	 * that apply to this firmware load.
   6800	 */
   6801	cs_count = 0;
   6802	cur_cs = 0;
   6803	memset(begin_set, 0, sizeof(begin_set));
   6804	memset(end_set, 0, sizeof(end_set));
   6805
   6806	/* Setup downloadable constant table */
   6807	download_consts[QOUTFIFO_OFFSET] = 0;
   6808	if (ahc->targetcmds != NULL)
   6809		download_consts[QOUTFIFO_OFFSET] += 32;
   6810	download_consts[QINFIFO_OFFSET] = download_consts[QOUTFIFO_OFFSET] + 1;
   6811	download_consts[CACHESIZE_MASK] = ahc->pci_cachesize - 1;
   6812	download_consts[INVERTED_CACHESIZE_MASK] = ~(ahc->pci_cachesize - 1);
   6813	sg_prefetch_cnt = ahc->pci_cachesize;
   6814	if (sg_prefetch_cnt < (2 * sizeof(struct ahc_dma_seg)))
   6815		sg_prefetch_cnt = 2 * sizeof(struct ahc_dma_seg);
   6816	download_consts[SG_PREFETCH_CNT] = sg_prefetch_cnt;
   6817	download_consts[SG_PREFETCH_ALIGN_MASK] = ~(sg_prefetch_cnt - 1);
   6818	download_consts[SG_PREFETCH_ADDR_MASK] = (sg_prefetch_cnt - 1);
   6819
   6820	cur_patch = patches;
   6821	downloaded = 0;
   6822	skip_addr = 0;
   6823	ahc_outb(ahc, SEQCTL, PERRORDIS|FAILDIS|FASTMODE|LOADRAM);
   6824	ahc_outb(ahc, SEQADDR0, 0);
   6825	ahc_outb(ahc, SEQADDR1, 0);
   6826
   6827	for (i = 0; i < sizeof(seqprog)/4; i++) {
   6828		if (ahc_check_patch(ahc, &cur_patch, i, &skip_addr) == 0) {
   6829			/*
   6830			 * Don't download this instruction as it
   6831			 * is in a patch that was removed.
   6832			 */
   6833			continue;
   6834		}
   6835
   6836		if (downloaded == ahc->instruction_ram_size) {
   6837			/*
   6838			 * We're about to exceed the instruction
   6839			 * storage capacity for this chip.  Fail
   6840			 * the load.
   6841			 */
   6842			printk("\n%s: Program too large for instruction memory "
   6843			       "size of %d!\n", ahc_name(ahc),
   6844			       ahc->instruction_ram_size);
   6845			return (ENOMEM);
   6846		}
   6847
   6848		/*
   6849		 * Move through the CS table until we find a CS
   6850		 * that might apply to this instruction.
   6851		 */
   6852		for (; cur_cs < NUM_CRITICAL_SECTIONS; cur_cs++) {
   6853			if (critical_sections[cur_cs].end <= i) {
   6854				if (begin_set[cs_count] == TRUE
   6855				 && end_set[cs_count] == FALSE) {
   6856					cs_table[cs_count].end = downloaded;
   6857					end_set[cs_count] = TRUE;
   6858					cs_count++;
   6859				}
   6860				continue;
   6861			}
   6862			if (critical_sections[cur_cs].begin <= i
   6863			 && begin_set[cs_count] == FALSE) {
   6864				cs_table[cs_count].begin = downloaded;
   6865				begin_set[cs_count] = TRUE;
   6866			}
   6867			break;
   6868		}
   6869		ahc_download_instr(ahc, i, download_consts);
   6870		downloaded++;
   6871	}
   6872
   6873	ahc->num_critical_sections = cs_count;
   6874	if (cs_count != 0) {
   6875
   6876		cs_count *= sizeof(struct cs);
   6877		ahc->critical_sections = kmemdup(cs_table, cs_count, GFP_ATOMIC);
   6878		if (ahc->critical_sections == NULL)
   6879			panic("ahc_loadseq: Could not malloc");
   6880	}
   6881	ahc_outb(ahc, SEQCTL, PERRORDIS|FAILDIS|FASTMODE);
   6882
   6883	if (bootverbose) {
   6884		printk(" %d instructions downloaded\n", downloaded);
   6885		printk("%s: Features 0x%x, Bugs 0x%x, Flags 0x%x\n",
   6886		       ahc_name(ahc), ahc->features, ahc->bugs, ahc->flags);
   6887	}
   6888	return (0);
   6889}
   6890
   6891static int
   6892ahc_check_patch(struct ahc_softc *ahc, const struct patch **start_patch,
   6893		u_int start_instr, u_int *skip_addr)
   6894{
   6895	const struct patch *cur_patch;
   6896	const struct patch *last_patch;
   6897	u_int	num_patches;
   6898
   6899	num_patches = ARRAY_SIZE(patches);
   6900	last_patch = &patches[num_patches];
   6901	cur_patch = *start_patch;
   6902
   6903	while (cur_patch < last_patch && start_instr == cur_patch->begin) {
   6904
   6905		if (cur_patch->patch_func(ahc) == 0) {
   6906
   6907			/* Start rejecting code */
   6908			*skip_addr = start_instr + cur_patch->skip_instr;
   6909			cur_patch += cur_patch->skip_patch;
   6910		} else {
   6911			/* Accepted this patch.  Advance to the next
   6912			 * one and wait for our intruction pointer to
   6913			 * hit this point.
   6914			 */
   6915			cur_patch++;
   6916		}
   6917	}
   6918
   6919	*start_patch = cur_patch;
   6920	if (start_instr < *skip_addr)
   6921		/* Still skipping */
   6922		return (0);
   6923
   6924	return (1);
   6925}
   6926
   6927static void
   6928ahc_download_instr(struct ahc_softc *ahc, u_int instrptr, uint8_t *dconsts)
   6929{
   6930	union	ins_formats instr;
   6931	struct	ins_format1 *fmt1_ins;
   6932	struct	ins_format3 *fmt3_ins;
   6933	u_int	opcode;
   6934
   6935	/*
   6936	 * The firmware is always compiled into a little endian format.
   6937	 */
   6938	instr.integer = ahc_le32toh(*(uint32_t*)&seqprog[instrptr * 4]);
   6939
   6940	fmt1_ins = &instr.format1;
   6941	fmt3_ins = NULL;
   6942
   6943	/* Pull the opcode */
   6944	opcode = instr.format1.opcode;
   6945	switch (opcode) {
   6946	case AIC_OP_JMP:
   6947	case AIC_OP_JC:
   6948	case AIC_OP_JNC:
   6949	case AIC_OP_CALL:
   6950	case AIC_OP_JNE:
   6951	case AIC_OP_JNZ:
   6952	case AIC_OP_JE:
   6953	case AIC_OP_JZ:
   6954	{
   6955		const struct patch *cur_patch;
   6956		int address_offset;
   6957		u_int address;
   6958		u_int skip_addr;
   6959		u_int i;
   6960
   6961		fmt3_ins = &instr.format3;
   6962		address_offset = 0;
   6963		address = fmt3_ins->address;
   6964		cur_patch = patches;
   6965		skip_addr = 0;
   6966
   6967		for (i = 0; i < address;) {
   6968
   6969			ahc_check_patch(ahc, &cur_patch, i, &skip_addr);
   6970
   6971			if (skip_addr > i) {
   6972				int end_addr;
   6973
   6974				end_addr = min(address, skip_addr);
   6975				address_offset += end_addr - i;
   6976				i = skip_addr;
   6977			} else {
   6978				i++;
   6979			}
   6980		}
   6981		address -= address_offset;
   6982		fmt3_ins->address = address;
   6983	}
   6984		fallthrough;
   6985	case AIC_OP_OR:
   6986	case AIC_OP_AND:
   6987	case AIC_OP_XOR:
   6988	case AIC_OP_ADD:
   6989	case AIC_OP_ADC:
   6990	case AIC_OP_BMOV:
   6991		if (fmt1_ins->parity != 0) {
   6992			fmt1_ins->immediate = dconsts[fmt1_ins->immediate];
   6993		}
   6994		fmt1_ins->parity = 0;
   6995		if ((ahc->features & AHC_CMD_CHAN) == 0
   6996		 && opcode == AIC_OP_BMOV) {
   6997			/*
   6998			 * Block move was added at the same time
   6999			 * as the command channel.  Verify that
   7000			 * this is only a move of a single element
   7001			 * and convert the BMOV to a MOV
   7002			 * (AND with an immediate of FF).
   7003			 */
   7004			if (fmt1_ins->immediate != 1)
   7005				panic("%s: BMOV not supported\n",
   7006				      ahc_name(ahc));
   7007			fmt1_ins->opcode = AIC_OP_AND;
   7008			fmt1_ins->immediate = 0xff;
   7009		}
   7010		fallthrough;
   7011	case AIC_OP_ROL:
   7012		if ((ahc->features & AHC_ULTRA2) != 0) {
   7013			int i, count;
   7014
   7015			/* Calculate odd parity for the instruction */
   7016			for (i = 0, count = 0; i < 31; i++) {
   7017				uint32_t mask;
   7018
   7019				mask = 0x01 << i;
   7020				if ((instr.integer & mask) != 0)
   7021					count++;
   7022			}
   7023			if ((count & 0x01) == 0)
   7024				instr.format1.parity = 1;
   7025		} else {
   7026			/* Compress the instruction for older sequencers */
   7027			if (fmt3_ins != NULL) {
   7028				instr.integer =
   7029					fmt3_ins->immediate
   7030				      | (fmt3_ins->source << 8)
   7031				      | (fmt3_ins->address << 16)
   7032				      |	(fmt3_ins->opcode << 25);
   7033			} else {
   7034				instr.integer =
   7035					fmt1_ins->immediate
   7036				      | (fmt1_ins->source << 8)
   7037				      | (fmt1_ins->destination << 16)
   7038				      |	(fmt1_ins->ret << 24)
   7039				      |	(fmt1_ins->opcode << 25);
   7040			}
   7041		}
   7042		/* The sequencer is a little endian cpu */
   7043		instr.integer = ahc_htole32(instr.integer);
   7044		ahc_outsb(ahc, SEQRAM, instr.bytes, 4);
   7045		break;
   7046	default:
   7047		panic("Unknown opcode encountered in seq program");
   7048		break;
   7049	}
   7050}
   7051
   7052int
   7053ahc_print_register(const ahc_reg_parse_entry_t *table, u_int num_entries,
   7054		   const char *name, u_int address, u_int value,
   7055		   u_int *cur_column, u_int wrap_point)
   7056{
   7057	int	printed;
   7058	u_int	printed_mask;
   7059
   7060	if (cur_column != NULL && *cur_column >= wrap_point) {
   7061		printk("\n");
   7062		*cur_column = 0;
   7063	}
   7064	printed  = printk("%s[0x%x]", name, value);
   7065	if (table == NULL) {
   7066		printed += printk(" ");
   7067		*cur_column += printed;
   7068		return (printed);
   7069	}
   7070	printed_mask = 0;
   7071	while (printed_mask != 0xFF) {
   7072		int entry;
   7073
   7074		for (entry = 0; entry < num_entries; entry++) {
   7075			if (((value & table[entry].mask)
   7076			  != table[entry].value)
   7077			 || ((printed_mask & table[entry].mask)
   7078			  == table[entry].mask))
   7079				continue;
   7080
   7081			printed += printk("%s%s",
   7082					  printed_mask == 0 ? ":(" : "|",
   7083					  table[entry].name);
   7084			printed_mask |= table[entry].mask;
   7085			break;
   7086		}
   7087		if (entry >= num_entries)
   7088			break;
   7089	}
   7090	if (printed_mask != 0)
   7091		printed += printk(") ");
   7092	else
   7093		printed += printk(" ");
   7094	if (cur_column != NULL)
   7095		*cur_column += printed;
   7096	return (printed);
   7097}
   7098
   7099void
   7100ahc_dump_card_state(struct ahc_softc *ahc)
   7101{
   7102	struct	scb *scb;
   7103	struct	scb_tailq *untagged_q;
   7104	u_int	cur_col;
   7105	int	paused;
   7106	int	target;
   7107	int	maxtarget;
   7108	int	i;
   7109	uint8_t last_phase;
   7110	uint8_t qinpos;
   7111	uint8_t qintail;
   7112	uint8_t qoutpos;
   7113	uint8_t scb_index;
   7114	uint8_t saved_scbptr;
   7115
   7116	if (ahc_is_paused(ahc)) {
   7117		paused = 1;
   7118	} else {
   7119		paused = 0;
   7120		ahc_pause(ahc);
   7121	}
   7122
   7123	saved_scbptr = ahc_inb(ahc, SCBPTR);
   7124	last_phase = ahc_inb(ahc, LASTPHASE);
   7125	printk(">>>>>>>>>>>>>>>>>> Dump Card State Begins <<<<<<<<<<<<<<<<<\n"
   7126	       "%s: Dumping Card State %s, at SEQADDR 0x%x\n",
   7127	       ahc_name(ahc), ahc_lookup_phase_entry(last_phase)->phasemsg,
   7128	       ahc_inb(ahc, SEQADDR0) | (ahc_inb(ahc, SEQADDR1) << 8));
   7129	if (paused)
   7130		printk("Card was paused\n");
   7131	printk("ACCUM = 0x%x, SINDEX = 0x%x, DINDEX = 0x%x, ARG_2 = 0x%x\n",
   7132	       ahc_inb(ahc, ACCUM), ahc_inb(ahc, SINDEX), ahc_inb(ahc, DINDEX),
   7133	       ahc_inb(ahc, ARG_2));
   7134	printk("HCNT = 0x%x SCBPTR = 0x%x\n", ahc_inb(ahc, HCNT),
   7135	       ahc_inb(ahc, SCBPTR));
   7136	cur_col = 0;
   7137	if ((ahc->features & AHC_DT) != 0)
   7138		ahc_scsiphase_print(ahc_inb(ahc, SCSIPHASE), &cur_col, 50);
   7139	ahc_scsisigi_print(ahc_inb(ahc, SCSISIGI), &cur_col, 50);
   7140	ahc_error_print(ahc_inb(ahc, ERROR), &cur_col, 50);
   7141	ahc_scsibusl_print(ahc_inb(ahc, SCSIBUSL), &cur_col, 50);
   7142	ahc_lastphase_print(ahc_inb(ahc, LASTPHASE), &cur_col, 50);
   7143	ahc_scsiseq_print(ahc_inb(ahc, SCSISEQ), &cur_col, 50);
   7144	ahc_sblkctl_print(ahc_inb(ahc, SBLKCTL), &cur_col, 50);
   7145	ahc_scsirate_print(ahc_inb(ahc, SCSIRATE), &cur_col, 50);
   7146	ahc_seqctl_print(ahc_inb(ahc, SEQCTL), &cur_col, 50);
   7147	ahc_seq_flags_print(ahc_inb(ahc, SEQ_FLAGS), &cur_col, 50);
   7148	ahc_sstat0_print(ahc_inb(ahc, SSTAT0), &cur_col, 50);
   7149	ahc_sstat1_print(ahc_inb(ahc, SSTAT1), &cur_col, 50);
   7150	ahc_sstat2_print(ahc_inb(ahc, SSTAT2), &cur_col, 50);
   7151	ahc_sstat3_print(ahc_inb(ahc, SSTAT3), &cur_col, 50);
   7152	ahc_simode0_print(ahc_inb(ahc, SIMODE0), &cur_col, 50);
   7153	ahc_simode1_print(ahc_inb(ahc, SIMODE1), &cur_col, 50);
   7154	ahc_sxfrctl0_print(ahc_inb(ahc, SXFRCTL0), &cur_col, 50);
   7155	ahc_dfcntrl_print(ahc_inb(ahc, DFCNTRL), &cur_col, 50);
   7156	ahc_dfstatus_print(ahc_inb(ahc, DFSTATUS), &cur_col, 50);
   7157	if (cur_col != 0)
   7158		printk("\n");
   7159	printk("STACK:");
   7160	for (i = 0; i < STACK_SIZE; i++)
   7161		printk(" 0x%x", ahc_inb(ahc, STACK)|(ahc_inb(ahc, STACK) << 8));
   7162	printk("\nSCB count = %d\n", ahc->scb_data->numscbs);
   7163	printk("Kernel NEXTQSCB = %d\n", ahc->next_queued_scb->hscb->tag);
   7164	printk("Card NEXTQSCB = %d\n", ahc_inb(ahc, NEXT_QUEUED_SCB));
   7165	/* QINFIFO */
   7166	printk("QINFIFO entries: ");
   7167	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
   7168		qinpos = ahc_inb(ahc, SNSCB_QOFF);
   7169		ahc_outb(ahc, SNSCB_QOFF, qinpos);
   7170	} else
   7171		qinpos = ahc_inb(ahc, QINPOS);
   7172	qintail = ahc->qinfifonext;
   7173	while (qinpos != qintail) {
   7174		printk("%d ", ahc->qinfifo[qinpos]);
   7175		qinpos++;
   7176	}
   7177	printk("\n");
   7178
   7179	printk("Waiting Queue entries: ");
   7180	scb_index = ahc_inb(ahc, WAITING_SCBH);
   7181	i = 0;
   7182	while (scb_index != SCB_LIST_NULL && i++ < 256) {
   7183		ahc_outb(ahc, SCBPTR, scb_index);
   7184		printk("%d:%d ", scb_index, ahc_inb(ahc, SCB_TAG));
   7185		scb_index = ahc_inb(ahc, SCB_NEXT);
   7186	}
   7187	printk("\n");
   7188
   7189	printk("Disconnected Queue entries: ");
   7190	scb_index = ahc_inb(ahc, DISCONNECTED_SCBH);
   7191	i = 0;
   7192	while (scb_index != SCB_LIST_NULL && i++ < 256) {
   7193		ahc_outb(ahc, SCBPTR, scb_index);
   7194		printk("%d:%d ", scb_index, ahc_inb(ahc, SCB_TAG));
   7195		scb_index = ahc_inb(ahc, SCB_NEXT);
   7196	}
   7197	printk("\n");
   7198
   7199	ahc_sync_qoutfifo(ahc, BUS_DMASYNC_POSTREAD);
   7200	printk("QOUTFIFO entries: ");
   7201	qoutpos = ahc->qoutfifonext;
   7202	i = 0;
   7203	while (ahc->qoutfifo[qoutpos] != SCB_LIST_NULL && i++ < 256) {
   7204		printk("%d ", ahc->qoutfifo[qoutpos]);
   7205		qoutpos++;
   7206	}
   7207	printk("\n");
   7208
   7209	printk("Sequencer Free SCB List: ");
   7210	scb_index = ahc_inb(ahc, FREE_SCBH);
   7211	i = 0;
   7212	while (scb_index != SCB_LIST_NULL && i++ < 256) {
   7213		ahc_outb(ahc, SCBPTR, scb_index);
   7214		printk("%d ", scb_index);
   7215		scb_index = ahc_inb(ahc, SCB_NEXT);
   7216	}
   7217	printk("\n");
   7218
   7219	printk("Sequencer SCB Info: ");
   7220	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
   7221		ahc_outb(ahc, SCBPTR, i);
   7222		cur_col  = printk("\n%3d ", i);
   7223
   7224		ahc_scb_control_print(ahc_inb(ahc, SCB_CONTROL), &cur_col, 60);
   7225		ahc_scb_scsiid_print(ahc_inb(ahc, SCB_SCSIID), &cur_col, 60);
   7226		ahc_scb_lun_print(ahc_inb(ahc, SCB_LUN), &cur_col, 60);
   7227		ahc_scb_tag_print(ahc_inb(ahc, SCB_TAG), &cur_col, 60);
   7228	}
   7229	printk("\n");
   7230
   7231	printk("Pending list: ");
   7232	i = 0;
   7233	LIST_FOREACH(scb, &ahc->pending_scbs, pending_links) {
   7234		if (i++ > 256)
   7235			break;
   7236		cur_col  = printk("\n%3d ", scb->hscb->tag);
   7237		ahc_scb_control_print(scb->hscb->control, &cur_col, 60);
   7238		ahc_scb_scsiid_print(scb->hscb->scsiid, &cur_col, 60);
   7239		ahc_scb_lun_print(scb->hscb->lun, &cur_col, 60);
   7240		if ((ahc->flags & AHC_PAGESCBS) == 0) {
   7241			ahc_outb(ahc, SCBPTR, scb->hscb->tag);
   7242			printk("(");
   7243			ahc_scb_control_print(ahc_inb(ahc, SCB_CONTROL),
   7244					      &cur_col, 60);
   7245			ahc_scb_tag_print(ahc_inb(ahc, SCB_TAG), &cur_col, 60);
   7246			printk(")");
   7247		}
   7248	}
   7249	printk("\n");
   7250
   7251	printk("Kernel Free SCB list: ");
   7252	i = 0;
   7253	SLIST_FOREACH(scb, &ahc->scb_data->free_scbs, links.sle) {
   7254		if (i++ > 256)
   7255			break;
   7256		printk("%d ", scb->hscb->tag);
   7257	}
   7258	printk("\n");
   7259
   7260	maxtarget = (ahc->features & (AHC_WIDE|AHC_TWIN)) ? 15 : 7;
   7261	for (target = 0; target <= maxtarget; target++) {
   7262		untagged_q = &ahc->untagged_queues[target];
   7263		if (TAILQ_FIRST(untagged_q) == NULL)
   7264			continue;
   7265		printk("Untagged Q(%d): ", target);
   7266		i = 0;
   7267		TAILQ_FOREACH(scb, untagged_q, links.tqe) {
   7268			if (i++ > 256)
   7269				break;
   7270			printk("%d ", scb->hscb->tag);
   7271		}
   7272		printk("\n");
   7273	}
   7274
   7275	printk("\n<<<<<<<<<<<<<<<<< Dump Card State Ends >>>>>>>>>>>>>>>>>>\n");
   7276	ahc_outb(ahc, SCBPTR, saved_scbptr);
   7277	if (paused == 0)
   7278		ahc_unpause(ahc);
   7279}
   7280
   7281/************************* Target Mode ****************************************/
   7282#ifdef AHC_TARGET_MODE
   7283cam_status
   7284ahc_find_tmode_devs(struct ahc_softc *ahc, struct cam_sim *sim, union ccb *ccb,
   7285		    struct ahc_tmode_tstate **tstate,
   7286		    struct ahc_tmode_lstate **lstate,
   7287		    int notfound_failure)
   7288{
   7289
   7290	if ((ahc->features & AHC_TARGETMODE) == 0)
   7291		return (CAM_REQ_INVALID);
   7292
   7293	/*
   7294	 * Handle the 'black hole' device that sucks up
   7295	 * requests to unattached luns on enabled targets.
   7296	 */
   7297	if (ccb->ccb_h.target_id == CAM_TARGET_WILDCARD
   7298	 && ccb->ccb_h.target_lun == CAM_LUN_WILDCARD) {
   7299		*tstate = NULL;
   7300		*lstate = ahc->black_hole;
   7301	} else {
   7302		u_int max_id;
   7303
   7304		max_id = (ahc->features & AHC_WIDE) ? 16 : 8;
   7305		if (ccb->ccb_h.target_id >= max_id)
   7306			return (CAM_TID_INVALID);
   7307
   7308		if (ccb->ccb_h.target_lun >= AHC_NUM_LUNS)
   7309			return (CAM_LUN_INVALID);
   7310
   7311		*tstate = ahc->enabled_targets[ccb->ccb_h.target_id];
   7312		*lstate = NULL;
   7313		if (*tstate != NULL)
   7314			*lstate =
   7315			    (*tstate)->enabled_luns[ccb->ccb_h.target_lun];
   7316	}
   7317
   7318	if (notfound_failure != 0 && *lstate == NULL)
   7319		return (CAM_PATH_INVALID);
   7320
   7321	return (CAM_REQ_CMP);
   7322}
   7323
   7324void
   7325ahc_handle_en_lun(struct ahc_softc *ahc, struct cam_sim *sim, union ccb *ccb)
   7326{
   7327	struct	   ahc_tmode_tstate *tstate;
   7328	struct	   ahc_tmode_lstate *lstate;
   7329	struct	   ccb_en_lun *cel;
   7330	cam_status status;
   7331	u_long	   s;
   7332	u_int	   target;
   7333	u_int	   lun;
   7334	u_int	   target_mask;
   7335	u_int	   our_id;
   7336	int	   error;
   7337	char	   channel;
   7338
   7339	status = ahc_find_tmode_devs(ahc, sim, ccb, &tstate, &lstate,
   7340				     /*notfound_failure*/FALSE);
   7341
   7342	if (status != CAM_REQ_CMP) {
   7343		ccb->ccb_h.status = status;
   7344		return;
   7345	}
   7346
   7347	if (cam_sim_bus(sim) == 0)
   7348		our_id = ahc->our_id;
   7349	else
   7350		our_id = ahc->our_id_b;
   7351
   7352	if (ccb->ccb_h.target_id != our_id) {
   7353		/*
   7354		 * our_id represents our initiator ID, or
   7355		 * the ID of the first target to have an
   7356		 * enabled lun in target mode.  There are
   7357		 * two cases that may preclude enabling a
   7358		 * target id other than our_id.
   7359		 *
   7360		 *   o our_id is for an active initiator role.
   7361		 *     Since the hardware does not support
   7362		 *     reselections to the initiator role at
   7363		 *     anything other than our_id, and our_id
   7364		 *     is used by the hardware to indicate the
   7365		 *     ID to use for both select-out and
   7366		 *     reselect-out operations, the only target
   7367		 *     ID we can support in this mode is our_id.
   7368		 *
   7369		 *   o The MULTARGID feature is not available and
   7370		 *     a previous target mode ID has been enabled.
   7371		 */
   7372		if ((ahc->features & AHC_MULTIROLE) != 0) {
   7373
   7374			if ((ahc->features & AHC_MULTI_TID) != 0
   7375			 && (ahc->flags & AHC_INITIATORROLE) != 0) {
   7376				/*
   7377				 * Only allow additional targets if
   7378				 * the initiator role is disabled.
   7379				 * The hardware cannot handle a re-select-in
   7380				 * on the initiator id during a re-select-out
   7381				 * on a different target id.
   7382				 */
   7383				status = CAM_TID_INVALID;
   7384			} else if ((ahc->flags & AHC_INITIATORROLE) != 0
   7385				|| ahc->enabled_luns > 0) {
   7386				/*
   7387				 * Only allow our target id to change
   7388				 * if the initiator role is not configured
   7389				 * and there are no enabled luns which
   7390				 * are attached to the currently registered
   7391				 * scsi id.
   7392				 */
   7393				status = CAM_TID_INVALID;
   7394			}
   7395		} else if ((ahc->features & AHC_MULTI_TID) == 0
   7396			&& ahc->enabled_luns > 0) {
   7397
   7398			status = CAM_TID_INVALID;
   7399		}
   7400	}
   7401
   7402	if (status != CAM_REQ_CMP) {
   7403		ccb->ccb_h.status = status;
   7404		return;
   7405	}
   7406
   7407	/*
   7408	 * We now have an id that is valid.
   7409	 * If we aren't in target mode, switch modes.
   7410	 */
   7411	if ((ahc->flags & AHC_TARGETROLE) == 0
   7412	 && ccb->ccb_h.target_id != CAM_TARGET_WILDCARD) {
   7413		u_long	 s;
   7414		ahc_flag saved_flags;
   7415
   7416		printk("Configuring Target Mode\n");
   7417		ahc_lock(ahc, &s);
   7418		if (LIST_FIRST(&ahc->pending_scbs) != NULL) {
   7419			ccb->ccb_h.status = CAM_BUSY;
   7420			ahc_unlock(ahc, &s);
   7421			return;
   7422		}
   7423		saved_flags = ahc->flags;
   7424		ahc->flags |= AHC_TARGETROLE;
   7425		if ((ahc->features & AHC_MULTIROLE) == 0)
   7426			ahc->flags &= ~AHC_INITIATORROLE;
   7427		ahc_pause(ahc);
   7428		error = ahc_loadseq(ahc);
   7429		if (error != 0) {
   7430			/*
   7431			 * Restore original configuration and notify
   7432			 * the caller that we cannot support target mode.
   7433			 * Since the adapter started out in this
   7434			 * configuration, the firmware load will succeed,
   7435			 * so there is no point in checking ahc_loadseq's
   7436			 * return value.
   7437			 */
   7438			ahc->flags = saved_flags;
   7439			(void)ahc_loadseq(ahc);
   7440			ahc_restart(ahc);
   7441			ahc_unlock(ahc, &s);
   7442			ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
   7443			return;
   7444		}
   7445		ahc_restart(ahc);
   7446		ahc_unlock(ahc, &s);
   7447	}
   7448	cel = &ccb->cel;
   7449	target = ccb->ccb_h.target_id;
   7450	lun = ccb->ccb_h.target_lun;
   7451	channel = SIM_CHANNEL(ahc, sim);
   7452	target_mask = 0x01 << target;
   7453	if (channel == 'B')
   7454		target_mask <<= 8;
   7455
   7456	if (cel->enable != 0) {
   7457		u_int scsiseq;
   7458
   7459		/* Are we already enabled?? */
   7460		if (lstate != NULL) {
   7461			xpt_print_path(ccb->ccb_h.path);
   7462			printk("Lun already enabled\n");
   7463			ccb->ccb_h.status = CAM_LUN_ALRDY_ENA;
   7464			return;
   7465		}
   7466
   7467		if (cel->grp6_len != 0
   7468		 || cel->grp7_len != 0) {
   7469			/*
   7470			 * Don't (yet?) support vendor
   7471			 * specific commands.
   7472			 */
   7473			ccb->ccb_h.status = CAM_REQ_INVALID;
   7474			printk("Non-zero Group Codes\n");
   7475			return;
   7476		}
   7477
   7478		/*
   7479		 * Seems to be okay.
   7480		 * Setup our data structures.
   7481		 */
   7482		if (target != CAM_TARGET_WILDCARD && tstate == NULL) {
   7483			tstate = ahc_alloc_tstate(ahc, target, channel);
   7484			if (tstate == NULL) {
   7485				xpt_print_path(ccb->ccb_h.path);
   7486				printk("Couldn't allocate tstate\n");
   7487				ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
   7488				return;
   7489			}
   7490		}
   7491		lstate = kzalloc(sizeof(*lstate), GFP_ATOMIC);
   7492		if (lstate == NULL) {
   7493			xpt_print_path(ccb->ccb_h.path);
   7494			printk("Couldn't allocate lstate\n");
   7495			ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
   7496			return;
   7497		}
   7498		status = xpt_create_path(&lstate->path, /*periph*/NULL,
   7499					 xpt_path_path_id(ccb->ccb_h.path),
   7500					 xpt_path_target_id(ccb->ccb_h.path),
   7501					 xpt_path_lun_id(ccb->ccb_h.path));
   7502		if (status != CAM_REQ_CMP) {
   7503			kfree(lstate);
   7504			xpt_print_path(ccb->ccb_h.path);
   7505			printk("Couldn't allocate path\n");
   7506			ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
   7507			return;
   7508		}
   7509		SLIST_INIT(&lstate->accept_tios);
   7510		SLIST_INIT(&lstate->immed_notifies);
   7511		ahc_lock(ahc, &s);
   7512		ahc_pause(ahc);
   7513		if (target != CAM_TARGET_WILDCARD) {
   7514			tstate->enabled_luns[lun] = lstate;
   7515			ahc->enabled_luns++;
   7516
   7517			if ((ahc->features & AHC_MULTI_TID) != 0) {
   7518				u_int targid_mask;
   7519
   7520				targid_mask = ahc_inb(ahc, TARGID)
   7521					    | (ahc_inb(ahc, TARGID + 1) << 8);
   7522
   7523				targid_mask |= target_mask;
   7524				ahc_outb(ahc, TARGID, targid_mask);
   7525				ahc_outb(ahc, TARGID+1, (targid_mask >> 8));
   7526				ahc_update_scsiid(ahc, targid_mask);
   7527			} else {
   7528				u_int our_id;
   7529				char  channel;
   7530
   7531				channel = SIM_CHANNEL(ahc, sim);
   7532				our_id = SIM_SCSI_ID(ahc, sim);
   7533
   7534				/*
   7535				 * This can only happen if selections
   7536				 * are not enabled
   7537				 */
   7538				if (target != our_id) {
   7539					u_int sblkctl;
   7540					char  cur_channel;
   7541					int   swap;
   7542
   7543					sblkctl = ahc_inb(ahc, SBLKCTL);
   7544					cur_channel = (sblkctl & SELBUSB)
   7545						    ? 'B' : 'A';
   7546					if ((ahc->features & AHC_TWIN) == 0)
   7547						cur_channel = 'A';
   7548					swap = cur_channel != channel;
   7549					if (channel == 'A')
   7550						ahc->our_id = target;
   7551					else
   7552						ahc->our_id_b = target;
   7553
   7554					if (swap)
   7555						ahc_outb(ahc, SBLKCTL,
   7556							 sblkctl ^ SELBUSB);
   7557
   7558					ahc_outb(ahc, SCSIID, target);
   7559
   7560					if (swap)
   7561						ahc_outb(ahc, SBLKCTL, sblkctl);
   7562				}
   7563			}
   7564		} else
   7565			ahc->black_hole = lstate;
   7566		/* Allow select-in operations */
   7567		if (ahc->black_hole != NULL && ahc->enabled_luns > 0) {
   7568			scsiseq = ahc_inb(ahc, SCSISEQ_TEMPLATE);
   7569			scsiseq |= ENSELI;
   7570			ahc_outb(ahc, SCSISEQ_TEMPLATE, scsiseq);
   7571			scsiseq = ahc_inb(ahc, SCSISEQ);
   7572			scsiseq |= ENSELI;
   7573			ahc_outb(ahc, SCSISEQ, scsiseq);
   7574		}
   7575		ahc_unpause(ahc);
   7576		ahc_unlock(ahc, &s);
   7577		ccb->ccb_h.status = CAM_REQ_CMP;
   7578		xpt_print_path(ccb->ccb_h.path);
   7579		printk("Lun now enabled for target mode\n");
   7580	} else {
   7581		struct scb *scb;
   7582		int i, empty;
   7583
   7584		if (lstate == NULL) {
   7585			ccb->ccb_h.status = CAM_LUN_INVALID;
   7586			return;
   7587		}
   7588
   7589		ahc_lock(ahc, &s);
   7590
   7591		ccb->ccb_h.status = CAM_REQ_CMP;
   7592		LIST_FOREACH(scb, &ahc->pending_scbs, pending_links) {
   7593			struct ccb_hdr *ccbh;
   7594
   7595			ccbh = &scb->io_ctx->ccb_h;
   7596			if (ccbh->func_code == XPT_CONT_TARGET_IO
   7597			 && !xpt_path_comp(ccbh->path, ccb->ccb_h.path)){
   7598				printk("CTIO pending\n");
   7599				ccb->ccb_h.status = CAM_REQ_INVALID;
   7600				ahc_unlock(ahc, &s);
   7601				return;
   7602			}
   7603		}
   7604
   7605		if (SLIST_FIRST(&lstate->accept_tios) != NULL) {
   7606			printk("ATIOs pending\n");
   7607			ccb->ccb_h.status = CAM_REQ_INVALID;
   7608		}
   7609
   7610		if (SLIST_FIRST(&lstate->immed_notifies) != NULL) {
   7611			printk("INOTs pending\n");
   7612			ccb->ccb_h.status = CAM_REQ_INVALID;
   7613		}
   7614
   7615		if (ccb->ccb_h.status != CAM_REQ_CMP) {
   7616			ahc_unlock(ahc, &s);
   7617			return;
   7618		}
   7619
   7620		xpt_print_path(ccb->ccb_h.path);
   7621		printk("Target mode disabled\n");
   7622		xpt_free_path(lstate->path);
   7623		kfree(lstate);
   7624
   7625		ahc_pause(ahc);
   7626		/* Can we clean up the target too? */
   7627		if (target != CAM_TARGET_WILDCARD) {
   7628			tstate->enabled_luns[lun] = NULL;
   7629			ahc->enabled_luns--;
   7630			for (empty = 1, i = 0; i < 8; i++)
   7631				if (tstate->enabled_luns[i] != NULL) {
   7632					empty = 0;
   7633					break;
   7634				}
   7635
   7636			if (empty) {
   7637				ahc_free_tstate(ahc, target, channel,
   7638						/*force*/FALSE);
   7639				if (ahc->features & AHC_MULTI_TID) {
   7640					u_int targid_mask;
   7641
   7642					targid_mask = ahc_inb(ahc, TARGID)
   7643						    | (ahc_inb(ahc, TARGID + 1)
   7644						       << 8);
   7645
   7646					targid_mask &= ~target_mask;
   7647					ahc_outb(ahc, TARGID, targid_mask);
   7648					ahc_outb(ahc, TARGID+1,
   7649						 (targid_mask >> 8));
   7650					ahc_update_scsiid(ahc, targid_mask);
   7651				}
   7652			}
   7653		} else {
   7654
   7655			ahc->black_hole = NULL;
   7656
   7657			/*
   7658			 * We can't allow selections without
   7659			 * our black hole device.
   7660			 */
   7661			empty = TRUE;
   7662		}
   7663		if (ahc->enabled_luns == 0) {
   7664			/* Disallow select-in */
   7665			u_int scsiseq;
   7666
   7667			scsiseq = ahc_inb(ahc, SCSISEQ_TEMPLATE);
   7668			scsiseq &= ~ENSELI;
   7669			ahc_outb(ahc, SCSISEQ_TEMPLATE, scsiseq);
   7670			scsiseq = ahc_inb(ahc, SCSISEQ);
   7671			scsiseq &= ~ENSELI;
   7672			ahc_outb(ahc, SCSISEQ, scsiseq);
   7673
   7674			if ((ahc->features & AHC_MULTIROLE) == 0) {
   7675				printk("Configuring Initiator Mode\n");
   7676				ahc->flags &= ~AHC_TARGETROLE;
   7677				ahc->flags |= AHC_INITIATORROLE;
   7678				/*
   7679				 * Returning to a configuration that
   7680				 * fit previously will always succeed.
   7681				 */
   7682				(void)ahc_loadseq(ahc);
   7683				ahc_restart(ahc);
   7684				/*
   7685				 * Unpaused.  The extra unpause
   7686				 * that follows is harmless.
   7687				 */
   7688			}
   7689		}
   7690		ahc_unpause(ahc);
   7691		ahc_unlock(ahc, &s);
   7692	}
   7693}
   7694
   7695static void
   7696ahc_update_scsiid(struct ahc_softc *ahc, u_int targid_mask)
   7697{
   7698	u_int scsiid_mask;
   7699	u_int scsiid;
   7700
   7701	if ((ahc->features & AHC_MULTI_TID) == 0)
   7702		panic("ahc_update_scsiid called on non-multitid unit\n");
   7703
   7704	/*
   7705	 * Since we will rely on the TARGID mask
   7706	 * for selection enables, ensure that OID
   7707	 * in SCSIID is not set to some other ID
   7708	 * that we don't want to allow selections on.
   7709	 */
   7710	if ((ahc->features & AHC_ULTRA2) != 0)
   7711		scsiid = ahc_inb(ahc, SCSIID_ULTRA2);
   7712	else
   7713		scsiid = ahc_inb(ahc, SCSIID);
   7714	scsiid_mask = 0x1 << (scsiid & OID);
   7715	if ((targid_mask & scsiid_mask) == 0) {
   7716		u_int our_id;
   7717
   7718		/* ffs counts from 1 */
   7719		our_id = ffs(targid_mask);
   7720		if (our_id == 0)
   7721			our_id = ahc->our_id;
   7722		else
   7723			our_id--;
   7724		scsiid &= TID;
   7725		scsiid |= our_id;
   7726	}
   7727	if ((ahc->features & AHC_ULTRA2) != 0)
   7728		ahc_outb(ahc, SCSIID_ULTRA2, scsiid);
   7729	else
   7730		ahc_outb(ahc, SCSIID, scsiid);
   7731}
   7732
   7733static void
   7734ahc_run_tqinfifo(struct ahc_softc *ahc, int paused)
   7735{
   7736	struct target_cmd *cmd;
   7737
   7738	/*
   7739	 * If the card supports auto-access pause,
   7740	 * we can access the card directly regardless
   7741	 * of whether it is paused or not.
   7742	 */
   7743	if ((ahc->features & AHC_AUTOPAUSE) != 0)
   7744		paused = TRUE;
   7745
   7746	ahc_sync_tqinfifo(ahc, BUS_DMASYNC_POSTREAD);
   7747	while ((cmd = &ahc->targetcmds[ahc->tqinfifonext])->cmd_valid != 0) {
   7748
   7749		/*
   7750		 * Only advance through the queue if we
   7751		 * have the resources to process the command.
   7752		 */
   7753		if (ahc_handle_target_cmd(ahc, cmd) != 0)
   7754			break;
   7755
   7756		cmd->cmd_valid = 0;
   7757		ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
   7758				ahc->shared_data_dmamap,
   7759				ahc_targetcmd_offset(ahc, ahc->tqinfifonext),
   7760				sizeof(struct target_cmd),
   7761				BUS_DMASYNC_PREREAD);
   7762		ahc->tqinfifonext++;
   7763
   7764		/*
   7765		 * Lazily update our position in the target mode incoming
   7766		 * command queue as seen by the sequencer.
   7767		 */
   7768		if ((ahc->tqinfifonext & (HOST_TQINPOS - 1)) == 1) {
   7769			if ((ahc->features & AHC_HS_MAILBOX) != 0) {
   7770				u_int hs_mailbox;
   7771
   7772				hs_mailbox = ahc_inb(ahc, HS_MAILBOX);
   7773				hs_mailbox &= ~HOST_TQINPOS;
   7774				hs_mailbox |= ahc->tqinfifonext & HOST_TQINPOS;
   7775				ahc_outb(ahc, HS_MAILBOX, hs_mailbox);
   7776			} else {
   7777				if (!paused)
   7778					ahc_pause(ahc);
   7779				ahc_outb(ahc, KERNEL_TQINPOS,
   7780					 ahc->tqinfifonext & HOST_TQINPOS);
   7781				if (!paused)
   7782					ahc_unpause(ahc);
   7783			}
   7784		}
   7785	}
   7786}
   7787
   7788static int
   7789ahc_handle_target_cmd(struct ahc_softc *ahc, struct target_cmd *cmd)
   7790{
   7791	struct	  ahc_tmode_tstate *tstate;
   7792	struct	  ahc_tmode_lstate *lstate;
   7793	struct	  ccb_accept_tio *atio;
   7794	uint8_t *byte;
   7795	int	  initiator;
   7796	int	  target;
   7797	int	  lun;
   7798
   7799	initiator = SCSIID_TARGET(ahc, cmd->scsiid);
   7800	target = SCSIID_OUR_ID(cmd->scsiid);
   7801	lun    = (cmd->identify & MSG_IDENTIFY_LUNMASK);
   7802
   7803	byte = cmd->bytes;
   7804	tstate = ahc->enabled_targets[target];
   7805	lstate = NULL;
   7806	if (tstate != NULL)
   7807		lstate = tstate->enabled_luns[lun];
   7808
   7809	/*
   7810	 * Commands for disabled luns go to the black hole driver.
   7811	 */
   7812	if (lstate == NULL)
   7813		lstate = ahc->black_hole;
   7814
   7815	atio = (struct ccb_accept_tio*)SLIST_FIRST(&lstate->accept_tios);
   7816	if (atio == NULL) {
   7817		ahc->flags |= AHC_TQINFIFO_BLOCKED;
   7818		/*
   7819		 * Wait for more ATIOs from the peripheral driver for this lun.
   7820		 */
   7821		if (bootverbose)
   7822			printk("%s: ATIOs exhausted\n", ahc_name(ahc));
   7823		return (1);
   7824	} else
   7825		ahc->flags &= ~AHC_TQINFIFO_BLOCKED;
   7826#if 0
   7827	printk("Incoming command from %d for %d:%d%s\n",
   7828	       initiator, target, lun,
   7829	       lstate == ahc->black_hole ? "(Black Holed)" : "");
   7830#endif
   7831	SLIST_REMOVE_HEAD(&lstate->accept_tios, sim_links.sle);
   7832
   7833	if (lstate == ahc->black_hole) {
   7834		/* Fill in the wildcards */
   7835		atio->ccb_h.target_id = target;
   7836		atio->ccb_h.target_lun = lun;
   7837	}
   7838
   7839	/*
   7840	 * Package it up and send it off to
   7841	 * whomever has this lun enabled.
   7842	 */
   7843	atio->sense_len = 0;
   7844	atio->init_id = initiator;
   7845	if (byte[0] != 0xFF) {
   7846		/* Tag was included */
   7847		atio->tag_action = *byte++;
   7848		atio->tag_id = *byte++;
   7849		atio->ccb_h.flags = CAM_TAG_ACTION_VALID;
   7850	} else {
   7851		atio->ccb_h.flags = 0;
   7852	}
   7853	byte++;
   7854
   7855	/* Okay.  Now determine the cdb size based on the command code */
   7856	switch (*byte >> CMD_GROUP_CODE_SHIFT) {
   7857	case 0:
   7858		atio->cdb_len = 6;
   7859		break;
   7860	case 1:
   7861	case 2:
   7862		atio->cdb_len = 10;
   7863		break;
   7864	case 4:
   7865		atio->cdb_len = 16;
   7866		break;
   7867	case 5:
   7868		atio->cdb_len = 12;
   7869		break;
   7870	case 3:
   7871	default:
   7872		/* Only copy the opcode. */
   7873		atio->cdb_len = 1;
   7874		printk("Reserved or VU command code type encountered\n");
   7875		break;
   7876	}
   7877
   7878	memcpy(atio->cdb_io.cdb_bytes, byte, atio->cdb_len);
   7879
   7880	atio->ccb_h.status |= CAM_CDB_RECVD;
   7881
   7882	if ((cmd->identify & MSG_IDENTIFY_DISCFLAG) == 0) {
   7883		/*
   7884		 * We weren't allowed to disconnect.
   7885		 * We're hanging on the bus until a
   7886		 * continue target I/O comes in response
   7887		 * to this accept tio.
   7888		 */
   7889#if 0
   7890		printk("Received Immediate Command %d:%d:%d - %p\n",
   7891		       initiator, target, lun, ahc->pending_device);
   7892#endif
   7893		ahc->pending_device = lstate;
   7894		ahc_freeze_ccb((union ccb *)atio);
   7895		atio->ccb_h.flags |= CAM_DIS_DISCONNECT;
   7896	}
   7897	xpt_done((union ccb*)atio);
   7898	return (0);
   7899}
   7900
   7901#endif