cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

chip.c (461787B)


      1// SPDX-License-Identifier: GPL-2.0 or BSD-3-Clause
      2/*
      3 * Copyright(c) 2015 - 2020 Intel Corporation.
      4 * Copyright(c) 2021 Cornelis Networks.
      5 */
      6
      7/*
      8 * This file contains all of the code that is specific to the HFI chip
      9 */
     10
     11#include <linux/pci.h>
     12#include <linux/delay.h>
     13#include <linux/interrupt.h>
     14#include <linux/module.h>
     15
     16#include "hfi.h"
     17#include "trace.h"
     18#include "mad.h"
     19#include "pio.h"
     20#include "sdma.h"
     21#include "eprom.h"
     22#include "efivar.h"
     23#include "platform.h"
     24#include "aspm.h"
     25#include "affinity.h"
     26#include "debugfs.h"
     27#include "fault.h"
     28#include "netdev.h"
     29
     30uint num_vls = HFI1_MAX_VLS_SUPPORTED;
     31module_param(num_vls, uint, S_IRUGO);
     32MODULE_PARM_DESC(num_vls, "Set number of Virtual Lanes to use (1-8)");
     33
     34/*
     35 * Default time to aggregate two 10K packets from the idle state
     36 * (timer not running). The timer starts at the end of the first packet,
     37 * so only the time for one 10K packet and header plus a bit extra is needed.
     38 * 10 * 1024 + 64 header byte = 10304 byte
     39 * 10304 byte / 12.5 GB/s = 824.32ns
     40 */
     41uint rcv_intr_timeout = (824 + 16); /* 16 is for coalescing interrupt */
     42module_param(rcv_intr_timeout, uint, S_IRUGO);
     43MODULE_PARM_DESC(rcv_intr_timeout, "Receive interrupt mitigation timeout in ns");
     44
     45uint rcv_intr_count = 16; /* same as qib */
     46module_param(rcv_intr_count, uint, S_IRUGO);
     47MODULE_PARM_DESC(rcv_intr_count, "Receive interrupt mitigation count");
     48
     49ushort link_crc_mask = SUPPORTED_CRCS;
     50module_param(link_crc_mask, ushort, S_IRUGO);
     51MODULE_PARM_DESC(link_crc_mask, "CRCs to use on the link");
     52
     53uint loopback;
     54module_param_named(loopback, loopback, uint, S_IRUGO);
     55MODULE_PARM_DESC(loopback, "Put into loopback mode (1 = serdes, 3 = external cable");
     56
     57/* Other driver tunables */
     58uint rcv_intr_dynamic = 1; /* enable dynamic mode for rcv int mitigation*/
     59static ushort crc_14b_sideband = 1;
     60static uint use_flr = 1;
     61uint quick_linkup; /* skip LNI */
     62
     63struct flag_table {
     64	u64 flag;	/* the flag */
     65	char *str;	/* description string */
     66	u16 extra;	/* extra information */
     67	u16 unused0;
     68	u32 unused1;
     69};
     70
     71/* str must be a string constant */
     72#define FLAG_ENTRY(str, extra, flag) {flag, str, extra}
     73#define FLAG_ENTRY0(str, flag) {flag, str, 0}
     74
     75/* Send Error Consequences */
     76#define SEC_WRITE_DROPPED	0x1
     77#define SEC_PACKET_DROPPED	0x2
     78#define SEC_SC_HALTED		0x4	/* per-context only */
     79#define SEC_SPC_FREEZE		0x8	/* per-HFI only */
     80
     81#define DEFAULT_KRCVQS		  2
     82#define MIN_KERNEL_KCTXTS         2
     83#define FIRST_KERNEL_KCTXT        1
     84
     85/*
     86 * RSM instance allocation
     87 *   0 - User Fecn Handling
     88 *   1 - Vnic
     89 *   2 - AIP
     90 *   3 - Verbs
     91 */
     92#define RSM_INS_FECN              0
     93#define RSM_INS_VNIC              1
     94#define RSM_INS_AIP               2
     95#define RSM_INS_VERBS             3
     96
     97/* Bit offset into the GUID which carries HFI id information */
     98#define GUID_HFI_INDEX_SHIFT     39
     99
    100/* extract the emulation revision */
    101#define emulator_rev(dd) ((dd)->irev >> 8)
    102/* parallel and serial emulation versions are 3 and 4 respectively */
    103#define is_emulator_p(dd) ((((dd)->irev) & 0xf) == 3)
    104#define is_emulator_s(dd) ((((dd)->irev) & 0xf) == 4)
    105
    106/* RSM fields for Verbs */
    107/* packet type */
    108#define IB_PACKET_TYPE         2ull
    109#define QW_SHIFT               6ull
    110/* QPN[7..1] */
    111#define QPN_WIDTH              7ull
    112
    113/* LRH.BTH: QW 0, OFFSET 48 - for match */
    114#define LRH_BTH_QW             0ull
    115#define LRH_BTH_BIT_OFFSET     48ull
    116#define LRH_BTH_OFFSET(off)    ((LRH_BTH_QW << QW_SHIFT) | (off))
    117#define LRH_BTH_MATCH_OFFSET   LRH_BTH_OFFSET(LRH_BTH_BIT_OFFSET)
    118#define LRH_BTH_SELECT
    119#define LRH_BTH_MASK           3ull
    120#define LRH_BTH_VALUE          2ull
    121
    122/* LRH.SC[3..0] QW 0, OFFSET 56 - for match */
    123#define LRH_SC_QW              0ull
    124#define LRH_SC_BIT_OFFSET      56ull
    125#define LRH_SC_OFFSET(off)     ((LRH_SC_QW << QW_SHIFT) | (off))
    126#define LRH_SC_MATCH_OFFSET    LRH_SC_OFFSET(LRH_SC_BIT_OFFSET)
    127#define LRH_SC_MASK            128ull
    128#define LRH_SC_VALUE           0ull
    129
    130/* SC[n..0] QW 0, OFFSET 60 - for select */
    131#define LRH_SC_SELECT_OFFSET  ((LRH_SC_QW << QW_SHIFT) | (60ull))
    132
    133/* QPN[m+n:1] QW 1, OFFSET 1 */
    134#define QPN_SELECT_OFFSET      ((1ull << QW_SHIFT) | (1ull))
    135
    136/* RSM fields for AIP */
    137/* LRH.BTH above is reused for this rule */
    138
    139/* BTH.DESTQP: QW 1, OFFSET 16 for match */
    140#define BTH_DESTQP_QW           1ull
    141#define BTH_DESTQP_BIT_OFFSET   16ull
    142#define BTH_DESTQP_OFFSET(off) ((BTH_DESTQP_QW << QW_SHIFT) | (off))
    143#define BTH_DESTQP_MATCH_OFFSET BTH_DESTQP_OFFSET(BTH_DESTQP_BIT_OFFSET)
    144#define BTH_DESTQP_MASK         0xFFull
    145#define BTH_DESTQP_VALUE        0x81ull
    146
    147/* DETH.SQPN: QW 1 Offset 56 for select */
    148/* We use 8 most significant Soure QPN bits as entropy fpr AIP */
    149#define DETH_AIP_SQPN_QW 3ull
    150#define DETH_AIP_SQPN_BIT_OFFSET 56ull
    151#define DETH_AIP_SQPN_OFFSET(off) ((DETH_AIP_SQPN_QW << QW_SHIFT) | (off))
    152#define DETH_AIP_SQPN_SELECT_OFFSET \
    153	DETH_AIP_SQPN_OFFSET(DETH_AIP_SQPN_BIT_OFFSET)
    154
    155/* RSM fields for Vnic */
    156/* L2_TYPE: QW 0, OFFSET 61 - for match */
    157#define L2_TYPE_QW             0ull
    158#define L2_TYPE_BIT_OFFSET     61ull
    159#define L2_TYPE_OFFSET(off)    ((L2_TYPE_QW << QW_SHIFT) | (off))
    160#define L2_TYPE_MATCH_OFFSET   L2_TYPE_OFFSET(L2_TYPE_BIT_OFFSET)
    161#define L2_TYPE_MASK           3ull
    162#define L2_16B_VALUE           2ull
    163
    164/* L4_TYPE QW 1, OFFSET 0 - for match */
    165#define L4_TYPE_QW              1ull
    166#define L4_TYPE_BIT_OFFSET      0ull
    167#define L4_TYPE_OFFSET(off)     ((L4_TYPE_QW << QW_SHIFT) | (off))
    168#define L4_TYPE_MATCH_OFFSET    L4_TYPE_OFFSET(L4_TYPE_BIT_OFFSET)
    169#define L4_16B_TYPE_MASK        0xFFull
    170#define L4_16B_ETH_VALUE        0x78ull
    171
    172/* 16B VESWID - for select */
    173#define L4_16B_HDR_VESWID_OFFSET  ((2 << QW_SHIFT) | (16ull))
    174/* 16B ENTROPY - for select */
    175#define L2_16B_ENTROPY_OFFSET     ((1 << QW_SHIFT) | (32ull))
    176
    177/* defines to build power on SC2VL table */
    178#define SC2VL_VAL( \
    179	num, \
    180	sc0, sc0val, \
    181	sc1, sc1val, \
    182	sc2, sc2val, \
    183	sc3, sc3val, \
    184	sc4, sc4val, \
    185	sc5, sc5val, \
    186	sc6, sc6val, \
    187	sc7, sc7val) \
    188( \
    189	((u64)(sc0val) << SEND_SC2VLT##num##_SC##sc0##_SHIFT) | \
    190	((u64)(sc1val) << SEND_SC2VLT##num##_SC##sc1##_SHIFT) | \
    191	((u64)(sc2val) << SEND_SC2VLT##num##_SC##sc2##_SHIFT) | \
    192	((u64)(sc3val) << SEND_SC2VLT##num##_SC##sc3##_SHIFT) | \
    193	((u64)(sc4val) << SEND_SC2VLT##num##_SC##sc4##_SHIFT) | \
    194	((u64)(sc5val) << SEND_SC2VLT##num##_SC##sc5##_SHIFT) | \
    195	((u64)(sc6val) << SEND_SC2VLT##num##_SC##sc6##_SHIFT) | \
    196	((u64)(sc7val) << SEND_SC2VLT##num##_SC##sc7##_SHIFT)   \
    197)
    198
    199#define DC_SC_VL_VAL( \
    200	range, \
    201	e0, e0val, \
    202	e1, e1val, \
    203	e2, e2val, \
    204	e3, e3val, \
    205	e4, e4val, \
    206	e5, e5val, \
    207	e6, e6val, \
    208	e7, e7val, \
    209	e8, e8val, \
    210	e9, e9val, \
    211	e10, e10val, \
    212	e11, e11val, \
    213	e12, e12val, \
    214	e13, e13val, \
    215	e14, e14val, \
    216	e15, e15val) \
    217( \
    218	((u64)(e0val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e0##_SHIFT) | \
    219	((u64)(e1val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e1##_SHIFT) | \
    220	((u64)(e2val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e2##_SHIFT) | \
    221	((u64)(e3val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e3##_SHIFT) | \
    222	((u64)(e4val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e4##_SHIFT) | \
    223	((u64)(e5val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e5##_SHIFT) | \
    224	((u64)(e6val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e6##_SHIFT) | \
    225	((u64)(e7val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e7##_SHIFT) | \
    226	((u64)(e8val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e8##_SHIFT) | \
    227	((u64)(e9val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e9##_SHIFT) | \
    228	((u64)(e10val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e10##_SHIFT) | \
    229	((u64)(e11val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e11##_SHIFT) | \
    230	((u64)(e12val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e12##_SHIFT) | \
    231	((u64)(e13val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e13##_SHIFT) | \
    232	((u64)(e14val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e14##_SHIFT) | \
    233	((u64)(e15val) << DCC_CFG_SC_VL_TABLE_##range##_ENTRY##e15##_SHIFT) \
    234)
    235
    236/* all CceStatus sub-block freeze bits */
    237#define ALL_FROZE (CCE_STATUS_SDMA_FROZE_SMASK \
    238			| CCE_STATUS_RXE_FROZE_SMASK \
    239			| CCE_STATUS_TXE_FROZE_SMASK \
    240			| CCE_STATUS_TXE_PIO_FROZE_SMASK)
    241/* all CceStatus sub-block TXE pause bits */
    242#define ALL_TXE_PAUSE (CCE_STATUS_TXE_PIO_PAUSED_SMASK \
    243			| CCE_STATUS_TXE_PAUSED_SMASK \
    244			| CCE_STATUS_SDMA_PAUSED_SMASK)
    245/* all CceStatus sub-block RXE pause bits */
    246#define ALL_RXE_PAUSE CCE_STATUS_RXE_PAUSED_SMASK
    247
    248#define CNTR_MAX 0xFFFFFFFFFFFFFFFFULL
    249#define CNTR_32BIT_MAX 0x00000000FFFFFFFF
    250
    251/*
    252 * CCE Error flags.
    253 */
    254static struct flag_table cce_err_status_flags[] = {
    255/* 0*/	FLAG_ENTRY0("CceCsrParityErr",
    256		CCE_ERR_STATUS_CCE_CSR_PARITY_ERR_SMASK),
    257/* 1*/	FLAG_ENTRY0("CceCsrReadBadAddrErr",
    258		CCE_ERR_STATUS_CCE_CSR_READ_BAD_ADDR_ERR_SMASK),
    259/* 2*/	FLAG_ENTRY0("CceCsrWriteBadAddrErr",
    260		CCE_ERR_STATUS_CCE_CSR_WRITE_BAD_ADDR_ERR_SMASK),
    261/* 3*/	FLAG_ENTRY0("CceTrgtAsyncFifoParityErr",
    262		CCE_ERR_STATUS_CCE_TRGT_ASYNC_FIFO_PARITY_ERR_SMASK),
    263/* 4*/	FLAG_ENTRY0("CceTrgtAccessErr",
    264		CCE_ERR_STATUS_CCE_TRGT_ACCESS_ERR_SMASK),
    265/* 5*/	FLAG_ENTRY0("CceRspdDataParityErr",
    266		CCE_ERR_STATUS_CCE_RSPD_DATA_PARITY_ERR_SMASK),
    267/* 6*/	FLAG_ENTRY0("CceCli0AsyncFifoParityErr",
    268		CCE_ERR_STATUS_CCE_CLI0_ASYNC_FIFO_PARITY_ERR_SMASK),
    269/* 7*/	FLAG_ENTRY0("CceCsrCfgBusParityErr",
    270		CCE_ERR_STATUS_CCE_CSR_CFG_BUS_PARITY_ERR_SMASK),
    271/* 8*/	FLAG_ENTRY0("CceCli2AsyncFifoParityErr",
    272		CCE_ERR_STATUS_CCE_CLI2_ASYNC_FIFO_PARITY_ERR_SMASK),
    273/* 9*/	FLAG_ENTRY0("CceCli1AsyncFifoPioCrdtParityErr",
    274	    CCE_ERR_STATUS_CCE_CLI1_ASYNC_FIFO_PIO_CRDT_PARITY_ERR_SMASK),
    275/*10*/	FLAG_ENTRY0("CceCli1AsyncFifoPioCrdtParityErr",
    276	    CCE_ERR_STATUS_CCE_CLI1_ASYNC_FIFO_SDMA_HD_PARITY_ERR_SMASK),
    277/*11*/	FLAG_ENTRY0("CceCli1AsyncFifoRxdmaParityError",
    278	    CCE_ERR_STATUS_CCE_CLI1_ASYNC_FIFO_RXDMA_PARITY_ERROR_SMASK),
    279/*12*/	FLAG_ENTRY0("CceCli1AsyncFifoDbgParityError",
    280		CCE_ERR_STATUS_CCE_CLI1_ASYNC_FIFO_DBG_PARITY_ERROR_SMASK),
    281/*13*/	FLAG_ENTRY0("PcicRetryMemCorErr",
    282		CCE_ERR_STATUS_PCIC_RETRY_MEM_COR_ERR_SMASK),
    283/*14*/	FLAG_ENTRY0("PcicRetryMemCorErr",
    284		CCE_ERR_STATUS_PCIC_RETRY_SOT_MEM_COR_ERR_SMASK),
    285/*15*/	FLAG_ENTRY0("PcicPostHdQCorErr",
    286		CCE_ERR_STATUS_PCIC_POST_HD_QCOR_ERR_SMASK),
    287/*16*/	FLAG_ENTRY0("PcicPostHdQCorErr",
    288		CCE_ERR_STATUS_PCIC_POST_DAT_QCOR_ERR_SMASK),
    289/*17*/	FLAG_ENTRY0("PcicPostHdQCorErr",
    290		CCE_ERR_STATUS_PCIC_CPL_HD_QCOR_ERR_SMASK),
    291/*18*/	FLAG_ENTRY0("PcicCplDatQCorErr",
    292		CCE_ERR_STATUS_PCIC_CPL_DAT_QCOR_ERR_SMASK),
    293/*19*/	FLAG_ENTRY0("PcicNPostHQParityErr",
    294		CCE_ERR_STATUS_PCIC_NPOST_HQ_PARITY_ERR_SMASK),
    295/*20*/	FLAG_ENTRY0("PcicNPostDatQParityErr",
    296		CCE_ERR_STATUS_PCIC_NPOST_DAT_QPARITY_ERR_SMASK),
    297/*21*/	FLAG_ENTRY0("PcicRetryMemUncErr",
    298		CCE_ERR_STATUS_PCIC_RETRY_MEM_UNC_ERR_SMASK),
    299/*22*/	FLAG_ENTRY0("PcicRetrySotMemUncErr",
    300		CCE_ERR_STATUS_PCIC_RETRY_SOT_MEM_UNC_ERR_SMASK),
    301/*23*/	FLAG_ENTRY0("PcicPostHdQUncErr",
    302		CCE_ERR_STATUS_PCIC_POST_HD_QUNC_ERR_SMASK),
    303/*24*/	FLAG_ENTRY0("PcicPostDatQUncErr",
    304		CCE_ERR_STATUS_PCIC_POST_DAT_QUNC_ERR_SMASK),
    305/*25*/	FLAG_ENTRY0("PcicCplHdQUncErr",
    306		CCE_ERR_STATUS_PCIC_CPL_HD_QUNC_ERR_SMASK),
    307/*26*/	FLAG_ENTRY0("PcicCplDatQUncErr",
    308		CCE_ERR_STATUS_PCIC_CPL_DAT_QUNC_ERR_SMASK),
    309/*27*/	FLAG_ENTRY0("PcicTransmitFrontParityErr",
    310		CCE_ERR_STATUS_PCIC_TRANSMIT_FRONT_PARITY_ERR_SMASK),
    311/*28*/	FLAG_ENTRY0("PcicTransmitBackParityErr",
    312		CCE_ERR_STATUS_PCIC_TRANSMIT_BACK_PARITY_ERR_SMASK),
    313/*29*/	FLAG_ENTRY0("PcicReceiveParityErr",
    314		CCE_ERR_STATUS_PCIC_RECEIVE_PARITY_ERR_SMASK),
    315/*30*/	FLAG_ENTRY0("CceTrgtCplTimeoutErr",
    316		CCE_ERR_STATUS_CCE_TRGT_CPL_TIMEOUT_ERR_SMASK),
    317/*31*/	FLAG_ENTRY0("LATriggered",
    318		CCE_ERR_STATUS_LA_TRIGGERED_SMASK),
    319/*32*/	FLAG_ENTRY0("CceSegReadBadAddrErr",
    320		CCE_ERR_STATUS_CCE_SEG_READ_BAD_ADDR_ERR_SMASK),
    321/*33*/	FLAG_ENTRY0("CceSegWriteBadAddrErr",
    322		CCE_ERR_STATUS_CCE_SEG_WRITE_BAD_ADDR_ERR_SMASK),
    323/*34*/	FLAG_ENTRY0("CceRcplAsyncFifoParityErr",
    324		CCE_ERR_STATUS_CCE_RCPL_ASYNC_FIFO_PARITY_ERR_SMASK),
    325/*35*/	FLAG_ENTRY0("CceRxdmaConvFifoParityErr",
    326		CCE_ERR_STATUS_CCE_RXDMA_CONV_FIFO_PARITY_ERR_SMASK),
    327/*36*/	FLAG_ENTRY0("CceMsixTableCorErr",
    328		CCE_ERR_STATUS_CCE_MSIX_TABLE_COR_ERR_SMASK),
    329/*37*/	FLAG_ENTRY0("CceMsixTableUncErr",
    330		CCE_ERR_STATUS_CCE_MSIX_TABLE_UNC_ERR_SMASK),
    331/*38*/	FLAG_ENTRY0("CceIntMapCorErr",
    332		CCE_ERR_STATUS_CCE_INT_MAP_COR_ERR_SMASK),
    333/*39*/	FLAG_ENTRY0("CceIntMapUncErr",
    334		CCE_ERR_STATUS_CCE_INT_MAP_UNC_ERR_SMASK),
    335/*40*/	FLAG_ENTRY0("CceMsixCsrParityErr",
    336		CCE_ERR_STATUS_CCE_MSIX_CSR_PARITY_ERR_SMASK),
    337/*41-63 reserved*/
    338};
    339
    340/*
    341 * Misc Error flags
    342 */
    343#define MES(text) MISC_ERR_STATUS_MISC_##text##_ERR_SMASK
    344static struct flag_table misc_err_status_flags[] = {
    345/* 0*/	FLAG_ENTRY0("CSR_PARITY", MES(CSR_PARITY)),
    346/* 1*/	FLAG_ENTRY0("CSR_READ_BAD_ADDR", MES(CSR_READ_BAD_ADDR)),
    347/* 2*/	FLAG_ENTRY0("CSR_WRITE_BAD_ADDR", MES(CSR_WRITE_BAD_ADDR)),
    348/* 3*/	FLAG_ENTRY0("SBUS_WRITE_FAILED", MES(SBUS_WRITE_FAILED)),
    349/* 4*/	FLAG_ENTRY0("KEY_MISMATCH", MES(KEY_MISMATCH)),
    350/* 5*/	FLAG_ENTRY0("FW_AUTH_FAILED", MES(FW_AUTH_FAILED)),
    351/* 6*/	FLAG_ENTRY0("EFUSE_CSR_PARITY", MES(EFUSE_CSR_PARITY)),
    352/* 7*/	FLAG_ENTRY0("EFUSE_READ_BAD_ADDR", MES(EFUSE_READ_BAD_ADDR)),
    353/* 8*/	FLAG_ENTRY0("EFUSE_WRITE", MES(EFUSE_WRITE)),
    354/* 9*/	FLAG_ENTRY0("EFUSE_DONE_PARITY", MES(EFUSE_DONE_PARITY)),
    355/*10*/	FLAG_ENTRY0("INVALID_EEP_CMD", MES(INVALID_EEP_CMD)),
    356/*11*/	FLAG_ENTRY0("MBIST_FAIL", MES(MBIST_FAIL)),
    357/*12*/	FLAG_ENTRY0("PLL_LOCK_FAIL", MES(PLL_LOCK_FAIL))
    358};
    359
    360/*
    361 * TXE PIO Error flags and consequences
    362 */
    363static struct flag_table pio_err_status_flags[] = {
    364/* 0*/	FLAG_ENTRY("PioWriteBadCtxt",
    365	SEC_WRITE_DROPPED,
    366	SEND_PIO_ERR_STATUS_PIO_WRITE_BAD_CTXT_ERR_SMASK),
    367/* 1*/	FLAG_ENTRY("PioWriteAddrParity",
    368	SEC_SPC_FREEZE,
    369	SEND_PIO_ERR_STATUS_PIO_WRITE_ADDR_PARITY_ERR_SMASK),
    370/* 2*/	FLAG_ENTRY("PioCsrParity",
    371	SEC_SPC_FREEZE,
    372	SEND_PIO_ERR_STATUS_PIO_CSR_PARITY_ERR_SMASK),
    373/* 3*/	FLAG_ENTRY("PioSbMemFifo0",
    374	SEC_SPC_FREEZE,
    375	SEND_PIO_ERR_STATUS_PIO_SB_MEM_FIFO0_ERR_SMASK),
    376/* 4*/	FLAG_ENTRY("PioSbMemFifo1",
    377	SEC_SPC_FREEZE,
    378	SEND_PIO_ERR_STATUS_PIO_SB_MEM_FIFO1_ERR_SMASK),
    379/* 5*/	FLAG_ENTRY("PioPccFifoParity",
    380	SEC_SPC_FREEZE,
    381	SEND_PIO_ERR_STATUS_PIO_PCC_FIFO_PARITY_ERR_SMASK),
    382/* 6*/	FLAG_ENTRY("PioPecFifoParity",
    383	SEC_SPC_FREEZE,
    384	SEND_PIO_ERR_STATUS_PIO_PEC_FIFO_PARITY_ERR_SMASK),
    385/* 7*/	FLAG_ENTRY("PioSbrdctlCrrelParity",
    386	SEC_SPC_FREEZE,
    387	SEND_PIO_ERR_STATUS_PIO_SBRDCTL_CRREL_PARITY_ERR_SMASK),
    388/* 8*/	FLAG_ENTRY("PioSbrdctrlCrrelFifoParity",
    389	SEC_SPC_FREEZE,
    390	SEND_PIO_ERR_STATUS_PIO_SBRDCTRL_CRREL_FIFO_PARITY_ERR_SMASK),
    391/* 9*/	FLAG_ENTRY("PioPktEvictFifoParityErr",
    392	SEC_SPC_FREEZE,
    393	SEND_PIO_ERR_STATUS_PIO_PKT_EVICT_FIFO_PARITY_ERR_SMASK),
    394/*10*/	FLAG_ENTRY("PioSmPktResetParity",
    395	SEC_SPC_FREEZE,
    396	SEND_PIO_ERR_STATUS_PIO_SM_PKT_RESET_PARITY_ERR_SMASK),
    397/*11*/	FLAG_ENTRY("PioVlLenMemBank0Unc",
    398	SEC_SPC_FREEZE,
    399	SEND_PIO_ERR_STATUS_PIO_VL_LEN_MEM_BANK0_UNC_ERR_SMASK),
    400/*12*/	FLAG_ENTRY("PioVlLenMemBank1Unc",
    401	SEC_SPC_FREEZE,
    402	SEND_PIO_ERR_STATUS_PIO_VL_LEN_MEM_BANK1_UNC_ERR_SMASK),
    403/*13*/	FLAG_ENTRY("PioVlLenMemBank0Cor",
    404	0,
    405	SEND_PIO_ERR_STATUS_PIO_VL_LEN_MEM_BANK0_COR_ERR_SMASK),
    406/*14*/	FLAG_ENTRY("PioVlLenMemBank1Cor",
    407	0,
    408	SEND_PIO_ERR_STATUS_PIO_VL_LEN_MEM_BANK1_COR_ERR_SMASK),
    409/*15*/	FLAG_ENTRY("PioCreditRetFifoParity",
    410	SEC_SPC_FREEZE,
    411	SEND_PIO_ERR_STATUS_PIO_CREDIT_RET_FIFO_PARITY_ERR_SMASK),
    412/*16*/	FLAG_ENTRY("PioPpmcPblFifo",
    413	SEC_SPC_FREEZE,
    414	SEND_PIO_ERR_STATUS_PIO_PPMC_PBL_FIFO_ERR_SMASK),
    415/*17*/	FLAG_ENTRY("PioInitSmIn",
    416	0,
    417	SEND_PIO_ERR_STATUS_PIO_INIT_SM_IN_ERR_SMASK),
    418/*18*/	FLAG_ENTRY("PioPktEvictSmOrArbSm",
    419	SEC_SPC_FREEZE,
    420	SEND_PIO_ERR_STATUS_PIO_PKT_EVICT_SM_OR_ARB_SM_ERR_SMASK),
    421/*19*/	FLAG_ENTRY("PioHostAddrMemUnc",
    422	SEC_SPC_FREEZE,
    423	SEND_PIO_ERR_STATUS_PIO_HOST_ADDR_MEM_UNC_ERR_SMASK),
    424/*20*/	FLAG_ENTRY("PioHostAddrMemCor",
    425	0,
    426	SEND_PIO_ERR_STATUS_PIO_HOST_ADDR_MEM_COR_ERR_SMASK),
    427/*21*/	FLAG_ENTRY("PioWriteDataParity",
    428	SEC_SPC_FREEZE,
    429	SEND_PIO_ERR_STATUS_PIO_WRITE_DATA_PARITY_ERR_SMASK),
    430/*22*/	FLAG_ENTRY("PioStateMachine",
    431	SEC_SPC_FREEZE,
    432	SEND_PIO_ERR_STATUS_PIO_STATE_MACHINE_ERR_SMASK),
    433/*23*/	FLAG_ENTRY("PioWriteQwValidParity",
    434	SEC_WRITE_DROPPED | SEC_SPC_FREEZE,
    435	SEND_PIO_ERR_STATUS_PIO_WRITE_QW_VALID_PARITY_ERR_SMASK),
    436/*24*/	FLAG_ENTRY("PioBlockQwCountParity",
    437	SEC_WRITE_DROPPED | SEC_SPC_FREEZE,
    438	SEND_PIO_ERR_STATUS_PIO_BLOCK_QW_COUNT_PARITY_ERR_SMASK),
    439/*25*/	FLAG_ENTRY("PioVlfVlLenParity",
    440	SEC_SPC_FREEZE,
    441	SEND_PIO_ERR_STATUS_PIO_VLF_VL_LEN_PARITY_ERR_SMASK),
    442/*26*/	FLAG_ENTRY("PioVlfSopParity",
    443	SEC_SPC_FREEZE,
    444	SEND_PIO_ERR_STATUS_PIO_VLF_SOP_PARITY_ERR_SMASK),
    445/*27*/	FLAG_ENTRY("PioVlFifoParity",
    446	SEC_SPC_FREEZE,
    447	SEND_PIO_ERR_STATUS_PIO_VL_FIFO_PARITY_ERR_SMASK),
    448/*28*/	FLAG_ENTRY("PioPpmcBqcMemParity",
    449	SEC_SPC_FREEZE,
    450	SEND_PIO_ERR_STATUS_PIO_PPMC_BQC_MEM_PARITY_ERR_SMASK),
    451/*29*/	FLAG_ENTRY("PioPpmcSopLen",
    452	SEC_SPC_FREEZE,
    453	SEND_PIO_ERR_STATUS_PIO_PPMC_SOP_LEN_ERR_SMASK),
    454/*30-31 reserved*/
    455/*32*/	FLAG_ENTRY("PioCurrentFreeCntParity",
    456	SEC_SPC_FREEZE,
    457	SEND_PIO_ERR_STATUS_PIO_CURRENT_FREE_CNT_PARITY_ERR_SMASK),
    458/*33*/	FLAG_ENTRY("PioLastReturnedCntParity",
    459	SEC_SPC_FREEZE,
    460	SEND_PIO_ERR_STATUS_PIO_LAST_RETURNED_CNT_PARITY_ERR_SMASK),
    461/*34*/	FLAG_ENTRY("PioPccSopHeadParity",
    462	SEC_SPC_FREEZE,
    463	SEND_PIO_ERR_STATUS_PIO_PCC_SOP_HEAD_PARITY_ERR_SMASK),
    464/*35*/	FLAG_ENTRY("PioPecSopHeadParityErr",
    465	SEC_SPC_FREEZE,
    466	SEND_PIO_ERR_STATUS_PIO_PEC_SOP_HEAD_PARITY_ERR_SMASK),
    467/*36-63 reserved*/
    468};
    469
    470/* TXE PIO errors that cause an SPC freeze */
    471#define ALL_PIO_FREEZE_ERR \
    472	(SEND_PIO_ERR_STATUS_PIO_WRITE_ADDR_PARITY_ERR_SMASK \
    473	| SEND_PIO_ERR_STATUS_PIO_CSR_PARITY_ERR_SMASK \
    474	| SEND_PIO_ERR_STATUS_PIO_SB_MEM_FIFO0_ERR_SMASK \
    475	| SEND_PIO_ERR_STATUS_PIO_SB_MEM_FIFO1_ERR_SMASK \
    476	| SEND_PIO_ERR_STATUS_PIO_PCC_FIFO_PARITY_ERR_SMASK \
    477	| SEND_PIO_ERR_STATUS_PIO_PEC_FIFO_PARITY_ERR_SMASK \
    478	| SEND_PIO_ERR_STATUS_PIO_SBRDCTL_CRREL_PARITY_ERR_SMASK \
    479	| SEND_PIO_ERR_STATUS_PIO_SBRDCTRL_CRREL_FIFO_PARITY_ERR_SMASK \
    480	| SEND_PIO_ERR_STATUS_PIO_PKT_EVICT_FIFO_PARITY_ERR_SMASK \
    481	| SEND_PIO_ERR_STATUS_PIO_SM_PKT_RESET_PARITY_ERR_SMASK \
    482	| SEND_PIO_ERR_STATUS_PIO_VL_LEN_MEM_BANK0_UNC_ERR_SMASK \
    483	| SEND_PIO_ERR_STATUS_PIO_VL_LEN_MEM_BANK1_UNC_ERR_SMASK \
    484	| SEND_PIO_ERR_STATUS_PIO_CREDIT_RET_FIFO_PARITY_ERR_SMASK \
    485	| SEND_PIO_ERR_STATUS_PIO_PPMC_PBL_FIFO_ERR_SMASK \
    486	| SEND_PIO_ERR_STATUS_PIO_PKT_EVICT_SM_OR_ARB_SM_ERR_SMASK \
    487	| SEND_PIO_ERR_STATUS_PIO_HOST_ADDR_MEM_UNC_ERR_SMASK \
    488	| SEND_PIO_ERR_STATUS_PIO_WRITE_DATA_PARITY_ERR_SMASK \
    489	| SEND_PIO_ERR_STATUS_PIO_STATE_MACHINE_ERR_SMASK \
    490	| SEND_PIO_ERR_STATUS_PIO_WRITE_QW_VALID_PARITY_ERR_SMASK \
    491	| SEND_PIO_ERR_STATUS_PIO_BLOCK_QW_COUNT_PARITY_ERR_SMASK \
    492	| SEND_PIO_ERR_STATUS_PIO_VLF_VL_LEN_PARITY_ERR_SMASK \
    493	| SEND_PIO_ERR_STATUS_PIO_VLF_SOP_PARITY_ERR_SMASK \
    494	| SEND_PIO_ERR_STATUS_PIO_VL_FIFO_PARITY_ERR_SMASK \
    495	| SEND_PIO_ERR_STATUS_PIO_PPMC_BQC_MEM_PARITY_ERR_SMASK \
    496	| SEND_PIO_ERR_STATUS_PIO_PPMC_SOP_LEN_ERR_SMASK \
    497	| SEND_PIO_ERR_STATUS_PIO_CURRENT_FREE_CNT_PARITY_ERR_SMASK \
    498	| SEND_PIO_ERR_STATUS_PIO_LAST_RETURNED_CNT_PARITY_ERR_SMASK \
    499	| SEND_PIO_ERR_STATUS_PIO_PCC_SOP_HEAD_PARITY_ERR_SMASK \
    500	| SEND_PIO_ERR_STATUS_PIO_PEC_SOP_HEAD_PARITY_ERR_SMASK)
    501
    502/*
    503 * TXE SDMA Error flags
    504 */
    505static struct flag_table sdma_err_status_flags[] = {
    506/* 0*/	FLAG_ENTRY0("SDmaRpyTagErr",
    507		SEND_DMA_ERR_STATUS_SDMA_RPY_TAG_ERR_SMASK),
    508/* 1*/	FLAG_ENTRY0("SDmaCsrParityErr",
    509		SEND_DMA_ERR_STATUS_SDMA_CSR_PARITY_ERR_SMASK),
    510/* 2*/	FLAG_ENTRY0("SDmaPcieReqTrackingUncErr",
    511		SEND_DMA_ERR_STATUS_SDMA_PCIE_REQ_TRACKING_UNC_ERR_SMASK),
    512/* 3*/	FLAG_ENTRY0("SDmaPcieReqTrackingCorErr",
    513		SEND_DMA_ERR_STATUS_SDMA_PCIE_REQ_TRACKING_COR_ERR_SMASK),
    514/*04-63 reserved*/
    515};
    516
    517/* TXE SDMA errors that cause an SPC freeze */
    518#define ALL_SDMA_FREEZE_ERR  \
    519		(SEND_DMA_ERR_STATUS_SDMA_RPY_TAG_ERR_SMASK \
    520		| SEND_DMA_ERR_STATUS_SDMA_CSR_PARITY_ERR_SMASK \
    521		| SEND_DMA_ERR_STATUS_SDMA_PCIE_REQ_TRACKING_UNC_ERR_SMASK)
    522
    523/* SendEgressErrInfo bits that correspond to a PortXmitDiscard counter */
    524#define PORT_DISCARD_EGRESS_ERRS \
    525	(SEND_EGRESS_ERR_INFO_TOO_LONG_IB_PACKET_ERR_SMASK \
    526	| SEND_EGRESS_ERR_INFO_VL_MAPPING_ERR_SMASK \
    527	| SEND_EGRESS_ERR_INFO_VL_ERR_SMASK)
    528
    529/*
    530 * TXE Egress Error flags
    531 */
    532#define SEES(text) SEND_EGRESS_ERR_STATUS_##text##_ERR_SMASK
    533static struct flag_table egress_err_status_flags[] = {
    534/* 0*/	FLAG_ENTRY0("TxPktIntegrityMemCorErr", SEES(TX_PKT_INTEGRITY_MEM_COR)),
    535/* 1*/	FLAG_ENTRY0("TxPktIntegrityMemUncErr", SEES(TX_PKT_INTEGRITY_MEM_UNC)),
    536/* 2 reserved */
    537/* 3*/	FLAG_ENTRY0("TxEgressFifoUnderrunOrParityErr",
    538		SEES(TX_EGRESS_FIFO_UNDERRUN_OR_PARITY)),
    539/* 4*/	FLAG_ENTRY0("TxLinkdownErr", SEES(TX_LINKDOWN)),
    540/* 5*/	FLAG_ENTRY0("TxIncorrectLinkStateErr", SEES(TX_INCORRECT_LINK_STATE)),
    541/* 6 reserved */
    542/* 7*/	FLAG_ENTRY0("TxPioLaunchIntfParityErr",
    543		SEES(TX_PIO_LAUNCH_INTF_PARITY)),
    544/* 8*/	FLAG_ENTRY0("TxSdmaLaunchIntfParityErr",
    545		SEES(TX_SDMA_LAUNCH_INTF_PARITY)),
    546/* 9-10 reserved */
    547/*11*/	FLAG_ENTRY0("TxSbrdCtlStateMachineParityErr",
    548		SEES(TX_SBRD_CTL_STATE_MACHINE_PARITY)),
    549/*12*/	FLAG_ENTRY0("TxIllegalVLErr", SEES(TX_ILLEGAL_VL)),
    550/*13*/	FLAG_ENTRY0("TxLaunchCsrParityErr", SEES(TX_LAUNCH_CSR_PARITY)),
    551/*14*/	FLAG_ENTRY0("TxSbrdCtlCsrParityErr", SEES(TX_SBRD_CTL_CSR_PARITY)),
    552/*15*/	FLAG_ENTRY0("TxConfigParityErr", SEES(TX_CONFIG_PARITY)),
    553/*16*/	FLAG_ENTRY0("TxSdma0DisallowedPacketErr",
    554		SEES(TX_SDMA0_DISALLOWED_PACKET)),
    555/*17*/	FLAG_ENTRY0("TxSdma1DisallowedPacketErr",
    556		SEES(TX_SDMA1_DISALLOWED_PACKET)),
    557/*18*/	FLAG_ENTRY0("TxSdma2DisallowedPacketErr",
    558		SEES(TX_SDMA2_DISALLOWED_PACKET)),
    559/*19*/	FLAG_ENTRY0("TxSdma3DisallowedPacketErr",
    560		SEES(TX_SDMA3_DISALLOWED_PACKET)),
    561/*20*/	FLAG_ENTRY0("TxSdma4DisallowedPacketErr",
    562		SEES(TX_SDMA4_DISALLOWED_PACKET)),
    563/*21*/	FLAG_ENTRY0("TxSdma5DisallowedPacketErr",
    564		SEES(TX_SDMA5_DISALLOWED_PACKET)),
    565/*22*/	FLAG_ENTRY0("TxSdma6DisallowedPacketErr",
    566		SEES(TX_SDMA6_DISALLOWED_PACKET)),
    567/*23*/	FLAG_ENTRY0("TxSdma7DisallowedPacketErr",
    568		SEES(TX_SDMA7_DISALLOWED_PACKET)),
    569/*24*/	FLAG_ENTRY0("TxSdma8DisallowedPacketErr",
    570		SEES(TX_SDMA8_DISALLOWED_PACKET)),
    571/*25*/	FLAG_ENTRY0("TxSdma9DisallowedPacketErr",
    572		SEES(TX_SDMA9_DISALLOWED_PACKET)),
    573/*26*/	FLAG_ENTRY0("TxSdma10DisallowedPacketErr",
    574		SEES(TX_SDMA10_DISALLOWED_PACKET)),
    575/*27*/	FLAG_ENTRY0("TxSdma11DisallowedPacketErr",
    576		SEES(TX_SDMA11_DISALLOWED_PACKET)),
    577/*28*/	FLAG_ENTRY0("TxSdma12DisallowedPacketErr",
    578		SEES(TX_SDMA12_DISALLOWED_PACKET)),
    579/*29*/	FLAG_ENTRY0("TxSdma13DisallowedPacketErr",
    580		SEES(TX_SDMA13_DISALLOWED_PACKET)),
    581/*30*/	FLAG_ENTRY0("TxSdma14DisallowedPacketErr",
    582		SEES(TX_SDMA14_DISALLOWED_PACKET)),
    583/*31*/	FLAG_ENTRY0("TxSdma15DisallowedPacketErr",
    584		SEES(TX_SDMA15_DISALLOWED_PACKET)),
    585/*32*/	FLAG_ENTRY0("TxLaunchFifo0UncOrParityErr",
    586		SEES(TX_LAUNCH_FIFO0_UNC_OR_PARITY)),
    587/*33*/	FLAG_ENTRY0("TxLaunchFifo1UncOrParityErr",
    588		SEES(TX_LAUNCH_FIFO1_UNC_OR_PARITY)),
    589/*34*/	FLAG_ENTRY0("TxLaunchFifo2UncOrParityErr",
    590		SEES(TX_LAUNCH_FIFO2_UNC_OR_PARITY)),
    591/*35*/	FLAG_ENTRY0("TxLaunchFifo3UncOrParityErr",
    592		SEES(TX_LAUNCH_FIFO3_UNC_OR_PARITY)),
    593/*36*/	FLAG_ENTRY0("TxLaunchFifo4UncOrParityErr",
    594		SEES(TX_LAUNCH_FIFO4_UNC_OR_PARITY)),
    595/*37*/	FLAG_ENTRY0("TxLaunchFifo5UncOrParityErr",
    596		SEES(TX_LAUNCH_FIFO5_UNC_OR_PARITY)),
    597/*38*/	FLAG_ENTRY0("TxLaunchFifo6UncOrParityErr",
    598		SEES(TX_LAUNCH_FIFO6_UNC_OR_PARITY)),
    599/*39*/	FLAG_ENTRY0("TxLaunchFifo7UncOrParityErr",
    600		SEES(TX_LAUNCH_FIFO7_UNC_OR_PARITY)),
    601/*40*/	FLAG_ENTRY0("TxLaunchFifo8UncOrParityErr",
    602		SEES(TX_LAUNCH_FIFO8_UNC_OR_PARITY)),
    603/*41*/	FLAG_ENTRY0("TxCreditReturnParityErr", SEES(TX_CREDIT_RETURN_PARITY)),
    604/*42*/	FLAG_ENTRY0("TxSbHdrUncErr", SEES(TX_SB_HDR_UNC)),
    605/*43*/	FLAG_ENTRY0("TxReadSdmaMemoryUncErr", SEES(TX_READ_SDMA_MEMORY_UNC)),
    606/*44*/	FLAG_ENTRY0("TxReadPioMemoryUncErr", SEES(TX_READ_PIO_MEMORY_UNC)),
    607/*45*/	FLAG_ENTRY0("TxEgressFifoUncErr", SEES(TX_EGRESS_FIFO_UNC)),
    608/*46*/	FLAG_ENTRY0("TxHcrcInsertionErr", SEES(TX_HCRC_INSERTION)),
    609/*47*/	FLAG_ENTRY0("TxCreditReturnVLErr", SEES(TX_CREDIT_RETURN_VL)),
    610/*48*/	FLAG_ENTRY0("TxLaunchFifo0CorErr", SEES(TX_LAUNCH_FIFO0_COR)),
    611/*49*/	FLAG_ENTRY0("TxLaunchFifo1CorErr", SEES(TX_LAUNCH_FIFO1_COR)),
    612/*50*/	FLAG_ENTRY0("TxLaunchFifo2CorErr", SEES(TX_LAUNCH_FIFO2_COR)),
    613/*51*/	FLAG_ENTRY0("TxLaunchFifo3CorErr", SEES(TX_LAUNCH_FIFO3_COR)),
    614/*52*/	FLAG_ENTRY0("TxLaunchFifo4CorErr", SEES(TX_LAUNCH_FIFO4_COR)),
    615/*53*/	FLAG_ENTRY0("TxLaunchFifo5CorErr", SEES(TX_LAUNCH_FIFO5_COR)),
    616/*54*/	FLAG_ENTRY0("TxLaunchFifo6CorErr", SEES(TX_LAUNCH_FIFO6_COR)),
    617/*55*/	FLAG_ENTRY0("TxLaunchFifo7CorErr", SEES(TX_LAUNCH_FIFO7_COR)),
    618/*56*/	FLAG_ENTRY0("TxLaunchFifo8CorErr", SEES(TX_LAUNCH_FIFO8_COR)),
    619/*57*/	FLAG_ENTRY0("TxCreditOverrunErr", SEES(TX_CREDIT_OVERRUN)),
    620/*58*/	FLAG_ENTRY0("TxSbHdrCorErr", SEES(TX_SB_HDR_COR)),
    621/*59*/	FLAG_ENTRY0("TxReadSdmaMemoryCorErr", SEES(TX_READ_SDMA_MEMORY_COR)),
    622/*60*/	FLAG_ENTRY0("TxReadPioMemoryCorErr", SEES(TX_READ_PIO_MEMORY_COR)),
    623/*61*/	FLAG_ENTRY0("TxEgressFifoCorErr", SEES(TX_EGRESS_FIFO_COR)),
    624/*62*/	FLAG_ENTRY0("TxReadSdmaMemoryCsrUncErr",
    625		SEES(TX_READ_SDMA_MEMORY_CSR_UNC)),
    626/*63*/	FLAG_ENTRY0("TxReadPioMemoryCsrUncErr",
    627		SEES(TX_READ_PIO_MEMORY_CSR_UNC)),
    628};
    629
    630/*
    631 * TXE Egress Error Info flags
    632 */
    633#define SEEI(text) SEND_EGRESS_ERR_INFO_##text##_ERR_SMASK
    634static struct flag_table egress_err_info_flags[] = {
    635/* 0*/	FLAG_ENTRY0("Reserved", 0ull),
    636/* 1*/	FLAG_ENTRY0("VLErr", SEEI(VL)),
    637/* 2*/	FLAG_ENTRY0("JobKeyErr", SEEI(JOB_KEY)),
    638/* 3*/	FLAG_ENTRY0("JobKeyErr", SEEI(JOB_KEY)),
    639/* 4*/	FLAG_ENTRY0("PartitionKeyErr", SEEI(PARTITION_KEY)),
    640/* 5*/	FLAG_ENTRY0("SLIDErr", SEEI(SLID)),
    641/* 6*/	FLAG_ENTRY0("OpcodeErr", SEEI(OPCODE)),
    642/* 7*/	FLAG_ENTRY0("VLMappingErr", SEEI(VL_MAPPING)),
    643/* 8*/	FLAG_ENTRY0("RawErr", SEEI(RAW)),
    644/* 9*/	FLAG_ENTRY0("RawIPv6Err", SEEI(RAW_IPV6)),
    645/*10*/	FLAG_ENTRY0("GRHErr", SEEI(GRH)),
    646/*11*/	FLAG_ENTRY0("BypassErr", SEEI(BYPASS)),
    647/*12*/	FLAG_ENTRY0("KDETHPacketsErr", SEEI(KDETH_PACKETS)),
    648/*13*/	FLAG_ENTRY0("NonKDETHPacketsErr", SEEI(NON_KDETH_PACKETS)),
    649/*14*/	FLAG_ENTRY0("TooSmallIBPacketsErr", SEEI(TOO_SMALL_IB_PACKETS)),
    650/*15*/	FLAG_ENTRY0("TooSmallBypassPacketsErr", SEEI(TOO_SMALL_BYPASS_PACKETS)),
    651/*16*/	FLAG_ENTRY0("PbcTestErr", SEEI(PBC_TEST)),
    652/*17*/	FLAG_ENTRY0("BadPktLenErr", SEEI(BAD_PKT_LEN)),
    653/*18*/	FLAG_ENTRY0("TooLongIBPacketErr", SEEI(TOO_LONG_IB_PACKET)),
    654/*19*/	FLAG_ENTRY0("TooLongBypassPacketsErr", SEEI(TOO_LONG_BYPASS_PACKETS)),
    655/*20*/	FLAG_ENTRY0("PbcStaticRateControlErr", SEEI(PBC_STATIC_RATE_CONTROL)),
    656/*21*/	FLAG_ENTRY0("BypassBadPktLenErr", SEEI(BAD_PKT_LEN)),
    657};
    658
    659/* TXE Egress errors that cause an SPC freeze */
    660#define ALL_TXE_EGRESS_FREEZE_ERR \
    661	(SEES(TX_EGRESS_FIFO_UNDERRUN_OR_PARITY) \
    662	| SEES(TX_PIO_LAUNCH_INTF_PARITY) \
    663	| SEES(TX_SDMA_LAUNCH_INTF_PARITY) \
    664	| SEES(TX_SBRD_CTL_STATE_MACHINE_PARITY) \
    665	| SEES(TX_LAUNCH_CSR_PARITY) \
    666	| SEES(TX_SBRD_CTL_CSR_PARITY) \
    667	| SEES(TX_CONFIG_PARITY) \
    668	| SEES(TX_LAUNCH_FIFO0_UNC_OR_PARITY) \
    669	| SEES(TX_LAUNCH_FIFO1_UNC_OR_PARITY) \
    670	| SEES(TX_LAUNCH_FIFO2_UNC_OR_PARITY) \
    671	| SEES(TX_LAUNCH_FIFO3_UNC_OR_PARITY) \
    672	| SEES(TX_LAUNCH_FIFO4_UNC_OR_PARITY) \
    673	| SEES(TX_LAUNCH_FIFO5_UNC_OR_PARITY) \
    674	| SEES(TX_LAUNCH_FIFO6_UNC_OR_PARITY) \
    675	| SEES(TX_LAUNCH_FIFO7_UNC_OR_PARITY) \
    676	| SEES(TX_LAUNCH_FIFO8_UNC_OR_PARITY) \
    677	| SEES(TX_CREDIT_RETURN_PARITY))
    678
    679/*
    680 * TXE Send error flags
    681 */
    682#define SES(name) SEND_ERR_STATUS_SEND_##name##_ERR_SMASK
    683static struct flag_table send_err_status_flags[] = {
    684/* 0*/	FLAG_ENTRY0("SendCsrParityErr", SES(CSR_PARITY)),
    685/* 1*/	FLAG_ENTRY0("SendCsrReadBadAddrErr", SES(CSR_READ_BAD_ADDR)),
    686/* 2*/	FLAG_ENTRY0("SendCsrWriteBadAddrErr", SES(CSR_WRITE_BAD_ADDR))
    687};
    688
    689/*
    690 * TXE Send Context Error flags and consequences
    691 */
    692static struct flag_table sc_err_status_flags[] = {
    693/* 0*/	FLAG_ENTRY("InconsistentSop",
    694		SEC_PACKET_DROPPED | SEC_SC_HALTED,
    695		SEND_CTXT_ERR_STATUS_PIO_INCONSISTENT_SOP_ERR_SMASK),
    696/* 1*/	FLAG_ENTRY("DisallowedPacket",
    697		SEC_PACKET_DROPPED | SEC_SC_HALTED,
    698		SEND_CTXT_ERR_STATUS_PIO_DISALLOWED_PACKET_ERR_SMASK),
    699/* 2*/	FLAG_ENTRY("WriteCrossesBoundary",
    700		SEC_WRITE_DROPPED | SEC_SC_HALTED,
    701		SEND_CTXT_ERR_STATUS_PIO_WRITE_CROSSES_BOUNDARY_ERR_SMASK),
    702/* 3*/	FLAG_ENTRY("WriteOverflow",
    703		SEC_WRITE_DROPPED | SEC_SC_HALTED,
    704		SEND_CTXT_ERR_STATUS_PIO_WRITE_OVERFLOW_ERR_SMASK),
    705/* 4*/	FLAG_ENTRY("WriteOutOfBounds",
    706		SEC_WRITE_DROPPED | SEC_SC_HALTED,
    707		SEND_CTXT_ERR_STATUS_PIO_WRITE_OUT_OF_BOUNDS_ERR_SMASK),
    708/* 5-63 reserved*/
    709};
    710
    711/*
    712 * RXE Receive Error flags
    713 */
    714#define RXES(name) RCV_ERR_STATUS_RX_##name##_ERR_SMASK
    715static struct flag_table rxe_err_status_flags[] = {
    716/* 0*/	FLAG_ENTRY0("RxDmaCsrCorErr", RXES(DMA_CSR_COR)),
    717/* 1*/	FLAG_ENTRY0("RxDcIntfParityErr", RXES(DC_INTF_PARITY)),
    718/* 2*/	FLAG_ENTRY0("RxRcvHdrUncErr", RXES(RCV_HDR_UNC)),
    719/* 3*/	FLAG_ENTRY0("RxRcvHdrCorErr", RXES(RCV_HDR_COR)),
    720/* 4*/	FLAG_ENTRY0("RxRcvDataUncErr", RXES(RCV_DATA_UNC)),
    721/* 5*/	FLAG_ENTRY0("RxRcvDataCorErr", RXES(RCV_DATA_COR)),
    722/* 6*/	FLAG_ENTRY0("RxRcvQpMapTableUncErr", RXES(RCV_QP_MAP_TABLE_UNC)),
    723/* 7*/	FLAG_ENTRY0("RxRcvQpMapTableCorErr", RXES(RCV_QP_MAP_TABLE_COR)),
    724/* 8*/	FLAG_ENTRY0("RxRcvCsrParityErr", RXES(RCV_CSR_PARITY)),
    725/* 9*/	FLAG_ENTRY0("RxDcSopEopParityErr", RXES(DC_SOP_EOP_PARITY)),
    726/*10*/	FLAG_ENTRY0("RxDmaFlagUncErr", RXES(DMA_FLAG_UNC)),
    727/*11*/	FLAG_ENTRY0("RxDmaFlagCorErr", RXES(DMA_FLAG_COR)),
    728/*12*/	FLAG_ENTRY0("RxRcvFsmEncodingErr", RXES(RCV_FSM_ENCODING)),
    729/*13*/	FLAG_ENTRY0("RxRbufFreeListUncErr", RXES(RBUF_FREE_LIST_UNC)),
    730/*14*/	FLAG_ENTRY0("RxRbufFreeListCorErr", RXES(RBUF_FREE_LIST_COR)),
    731/*15*/	FLAG_ENTRY0("RxRbufLookupDesRegUncErr", RXES(RBUF_LOOKUP_DES_REG_UNC)),
    732/*16*/	FLAG_ENTRY0("RxRbufLookupDesRegUncCorErr",
    733		RXES(RBUF_LOOKUP_DES_REG_UNC_COR)),
    734/*17*/	FLAG_ENTRY0("RxRbufLookupDesUncErr", RXES(RBUF_LOOKUP_DES_UNC)),
    735/*18*/	FLAG_ENTRY0("RxRbufLookupDesCorErr", RXES(RBUF_LOOKUP_DES_COR)),
    736/*19*/	FLAG_ENTRY0("RxRbufBlockListReadUncErr",
    737		RXES(RBUF_BLOCK_LIST_READ_UNC)),
    738/*20*/	FLAG_ENTRY0("RxRbufBlockListReadCorErr",
    739		RXES(RBUF_BLOCK_LIST_READ_COR)),
    740/*21*/	FLAG_ENTRY0("RxRbufCsrQHeadBufNumParityErr",
    741		RXES(RBUF_CSR_QHEAD_BUF_NUM_PARITY)),
    742/*22*/	FLAG_ENTRY0("RxRbufCsrQEntCntParityErr",
    743		RXES(RBUF_CSR_QENT_CNT_PARITY)),
    744/*23*/	FLAG_ENTRY0("RxRbufCsrQNextBufParityErr",
    745		RXES(RBUF_CSR_QNEXT_BUF_PARITY)),
    746/*24*/	FLAG_ENTRY0("RxRbufCsrQVldBitParityErr",
    747		RXES(RBUF_CSR_QVLD_BIT_PARITY)),
    748/*25*/	FLAG_ENTRY0("RxRbufCsrQHdPtrParityErr", RXES(RBUF_CSR_QHD_PTR_PARITY)),
    749/*26*/	FLAG_ENTRY0("RxRbufCsrQTlPtrParityErr", RXES(RBUF_CSR_QTL_PTR_PARITY)),
    750/*27*/	FLAG_ENTRY0("RxRbufCsrQNumOfPktParityErr",
    751		RXES(RBUF_CSR_QNUM_OF_PKT_PARITY)),
    752/*28*/	FLAG_ENTRY0("RxRbufCsrQEOPDWParityErr", RXES(RBUF_CSR_QEOPDW_PARITY)),
    753/*29*/	FLAG_ENTRY0("RxRbufCtxIdParityErr", RXES(RBUF_CTX_ID_PARITY)),
    754/*30*/	FLAG_ENTRY0("RxRBufBadLookupErr", RXES(RBUF_BAD_LOOKUP)),
    755/*31*/	FLAG_ENTRY0("RxRbufFullErr", RXES(RBUF_FULL)),
    756/*32*/	FLAG_ENTRY0("RxRbufEmptyErr", RXES(RBUF_EMPTY)),
    757/*33*/	FLAG_ENTRY0("RxRbufFlRdAddrParityErr", RXES(RBUF_FL_RD_ADDR_PARITY)),
    758/*34*/	FLAG_ENTRY0("RxRbufFlWrAddrParityErr", RXES(RBUF_FL_WR_ADDR_PARITY)),
    759/*35*/	FLAG_ENTRY0("RxRbufFlInitdoneParityErr",
    760		RXES(RBUF_FL_INITDONE_PARITY)),
    761/*36*/	FLAG_ENTRY0("RxRbufFlInitWrAddrParityErr",
    762		RXES(RBUF_FL_INIT_WR_ADDR_PARITY)),
    763/*37*/	FLAG_ENTRY0("RxRbufNextFreeBufUncErr", RXES(RBUF_NEXT_FREE_BUF_UNC)),
    764/*38*/	FLAG_ENTRY0("RxRbufNextFreeBufCorErr", RXES(RBUF_NEXT_FREE_BUF_COR)),
    765/*39*/	FLAG_ENTRY0("RxLookupDesPart1UncErr", RXES(LOOKUP_DES_PART1_UNC)),
    766/*40*/	FLAG_ENTRY0("RxLookupDesPart1UncCorErr",
    767		RXES(LOOKUP_DES_PART1_UNC_COR)),
    768/*41*/	FLAG_ENTRY0("RxLookupDesPart2ParityErr",
    769		RXES(LOOKUP_DES_PART2_PARITY)),
    770/*42*/	FLAG_ENTRY0("RxLookupRcvArrayUncErr", RXES(LOOKUP_RCV_ARRAY_UNC)),
    771/*43*/	FLAG_ENTRY0("RxLookupRcvArrayCorErr", RXES(LOOKUP_RCV_ARRAY_COR)),
    772/*44*/	FLAG_ENTRY0("RxLookupCsrParityErr", RXES(LOOKUP_CSR_PARITY)),
    773/*45*/	FLAG_ENTRY0("RxHqIntrCsrParityErr", RXES(HQ_INTR_CSR_PARITY)),
    774/*46*/	FLAG_ENTRY0("RxHqIntrFsmErr", RXES(HQ_INTR_FSM)),
    775/*47*/	FLAG_ENTRY0("RxRbufDescPart1UncErr", RXES(RBUF_DESC_PART1_UNC)),
    776/*48*/	FLAG_ENTRY0("RxRbufDescPart1CorErr", RXES(RBUF_DESC_PART1_COR)),
    777/*49*/	FLAG_ENTRY0("RxRbufDescPart2UncErr", RXES(RBUF_DESC_PART2_UNC)),
    778/*50*/	FLAG_ENTRY0("RxRbufDescPart2CorErr", RXES(RBUF_DESC_PART2_COR)),
    779/*51*/	FLAG_ENTRY0("RxDmaHdrFifoRdUncErr", RXES(DMA_HDR_FIFO_RD_UNC)),
    780/*52*/	FLAG_ENTRY0("RxDmaHdrFifoRdCorErr", RXES(DMA_HDR_FIFO_RD_COR)),
    781/*53*/	FLAG_ENTRY0("RxDmaDataFifoRdUncErr", RXES(DMA_DATA_FIFO_RD_UNC)),
    782/*54*/	FLAG_ENTRY0("RxDmaDataFifoRdCorErr", RXES(DMA_DATA_FIFO_RD_COR)),
    783/*55*/	FLAG_ENTRY0("RxRbufDataUncErr", RXES(RBUF_DATA_UNC)),
    784/*56*/	FLAG_ENTRY0("RxRbufDataCorErr", RXES(RBUF_DATA_COR)),
    785/*57*/	FLAG_ENTRY0("RxDmaCsrParityErr", RXES(DMA_CSR_PARITY)),
    786/*58*/	FLAG_ENTRY0("RxDmaEqFsmEncodingErr", RXES(DMA_EQ_FSM_ENCODING)),
    787/*59*/	FLAG_ENTRY0("RxDmaDqFsmEncodingErr", RXES(DMA_DQ_FSM_ENCODING)),
    788/*60*/	FLAG_ENTRY0("RxDmaCsrUncErr", RXES(DMA_CSR_UNC)),
    789/*61*/	FLAG_ENTRY0("RxCsrReadBadAddrErr", RXES(CSR_READ_BAD_ADDR)),
    790/*62*/	FLAG_ENTRY0("RxCsrWriteBadAddrErr", RXES(CSR_WRITE_BAD_ADDR)),
    791/*63*/	FLAG_ENTRY0("RxCsrParityErr", RXES(CSR_PARITY))
    792};
    793
    794/* RXE errors that will trigger an SPC freeze */
    795#define ALL_RXE_FREEZE_ERR  \
    796	(RCV_ERR_STATUS_RX_RCV_QP_MAP_TABLE_UNC_ERR_SMASK \
    797	| RCV_ERR_STATUS_RX_RCV_CSR_PARITY_ERR_SMASK \
    798	| RCV_ERR_STATUS_RX_DMA_FLAG_UNC_ERR_SMASK \
    799	| RCV_ERR_STATUS_RX_RCV_FSM_ENCODING_ERR_SMASK \
    800	| RCV_ERR_STATUS_RX_RBUF_FREE_LIST_UNC_ERR_SMASK \
    801	| RCV_ERR_STATUS_RX_RBUF_LOOKUP_DES_REG_UNC_ERR_SMASK \
    802	| RCV_ERR_STATUS_RX_RBUF_LOOKUP_DES_REG_UNC_COR_ERR_SMASK \
    803	| RCV_ERR_STATUS_RX_RBUF_LOOKUP_DES_UNC_ERR_SMASK \
    804	| RCV_ERR_STATUS_RX_RBUF_BLOCK_LIST_READ_UNC_ERR_SMASK \
    805	| RCV_ERR_STATUS_RX_RBUF_CSR_QHEAD_BUF_NUM_PARITY_ERR_SMASK \
    806	| RCV_ERR_STATUS_RX_RBUF_CSR_QENT_CNT_PARITY_ERR_SMASK \
    807	| RCV_ERR_STATUS_RX_RBUF_CSR_QNEXT_BUF_PARITY_ERR_SMASK \
    808	| RCV_ERR_STATUS_RX_RBUF_CSR_QVLD_BIT_PARITY_ERR_SMASK \
    809	| RCV_ERR_STATUS_RX_RBUF_CSR_QHD_PTR_PARITY_ERR_SMASK \
    810	| RCV_ERR_STATUS_RX_RBUF_CSR_QTL_PTR_PARITY_ERR_SMASK \
    811	| RCV_ERR_STATUS_RX_RBUF_CSR_QNUM_OF_PKT_PARITY_ERR_SMASK \
    812	| RCV_ERR_STATUS_RX_RBUF_CSR_QEOPDW_PARITY_ERR_SMASK \
    813	| RCV_ERR_STATUS_RX_RBUF_CTX_ID_PARITY_ERR_SMASK \
    814	| RCV_ERR_STATUS_RX_RBUF_BAD_LOOKUP_ERR_SMASK \
    815	| RCV_ERR_STATUS_RX_RBUF_FULL_ERR_SMASK \
    816	| RCV_ERR_STATUS_RX_RBUF_EMPTY_ERR_SMASK \
    817	| RCV_ERR_STATUS_RX_RBUF_FL_RD_ADDR_PARITY_ERR_SMASK \
    818	| RCV_ERR_STATUS_RX_RBUF_FL_WR_ADDR_PARITY_ERR_SMASK \
    819	| RCV_ERR_STATUS_RX_RBUF_FL_INITDONE_PARITY_ERR_SMASK \
    820	| RCV_ERR_STATUS_RX_RBUF_FL_INIT_WR_ADDR_PARITY_ERR_SMASK \
    821	| RCV_ERR_STATUS_RX_RBUF_NEXT_FREE_BUF_UNC_ERR_SMASK \
    822	| RCV_ERR_STATUS_RX_LOOKUP_DES_PART1_UNC_ERR_SMASK \
    823	| RCV_ERR_STATUS_RX_LOOKUP_DES_PART1_UNC_COR_ERR_SMASK \
    824	| RCV_ERR_STATUS_RX_LOOKUP_DES_PART2_PARITY_ERR_SMASK \
    825	| RCV_ERR_STATUS_RX_LOOKUP_RCV_ARRAY_UNC_ERR_SMASK \
    826	| RCV_ERR_STATUS_RX_LOOKUP_CSR_PARITY_ERR_SMASK \
    827	| RCV_ERR_STATUS_RX_HQ_INTR_CSR_PARITY_ERR_SMASK \
    828	| RCV_ERR_STATUS_RX_HQ_INTR_FSM_ERR_SMASK \
    829	| RCV_ERR_STATUS_RX_RBUF_DESC_PART1_UNC_ERR_SMASK \
    830	| RCV_ERR_STATUS_RX_RBUF_DESC_PART1_COR_ERR_SMASK \
    831	| RCV_ERR_STATUS_RX_RBUF_DESC_PART2_UNC_ERR_SMASK \
    832	| RCV_ERR_STATUS_RX_DMA_HDR_FIFO_RD_UNC_ERR_SMASK \
    833	| RCV_ERR_STATUS_RX_DMA_DATA_FIFO_RD_UNC_ERR_SMASK \
    834	| RCV_ERR_STATUS_RX_RBUF_DATA_UNC_ERR_SMASK \
    835	| RCV_ERR_STATUS_RX_DMA_CSR_PARITY_ERR_SMASK \
    836	| RCV_ERR_STATUS_RX_DMA_EQ_FSM_ENCODING_ERR_SMASK \
    837	| RCV_ERR_STATUS_RX_DMA_DQ_FSM_ENCODING_ERR_SMASK \
    838	| RCV_ERR_STATUS_RX_DMA_CSR_UNC_ERR_SMASK \
    839	| RCV_ERR_STATUS_RX_CSR_PARITY_ERR_SMASK)
    840
    841#define RXE_FREEZE_ABORT_MASK \
    842	(RCV_ERR_STATUS_RX_DMA_CSR_UNC_ERR_SMASK | \
    843	RCV_ERR_STATUS_RX_DMA_HDR_FIFO_RD_UNC_ERR_SMASK | \
    844	RCV_ERR_STATUS_RX_DMA_DATA_FIFO_RD_UNC_ERR_SMASK)
    845
    846/*
    847 * DCC Error Flags
    848 */
    849#define DCCE(name) DCC_ERR_FLG_##name##_SMASK
    850static struct flag_table dcc_err_flags[] = {
    851	FLAG_ENTRY0("bad_l2_err", DCCE(BAD_L2_ERR)),
    852	FLAG_ENTRY0("bad_sc_err", DCCE(BAD_SC_ERR)),
    853	FLAG_ENTRY0("bad_mid_tail_err", DCCE(BAD_MID_TAIL_ERR)),
    854	FLAG_ENTRY0("bad_preemption_err", DCCE(BAD_PREEMPTION_ERR)),
    855	FLAG_ENTRY0("preemption_err", DCCE(PREEMPTION_ERR)),
    856	FLAG_ENTRY0("preemptionvl15_err", DCCE(PREEMPTIONVL15_ERR)),
    857	FLAG_ENTRY0("bad_vl_marker_err", DCCE(BAD_VL_MARKER_ERR)),
    858	FLAG_ENTRY0("bad_dlid_target_err", DCCE(BAD_DLID_TARGET_ERR)),
    859	FLAG_ENTRY0("bad_lver_err", DCCE(BAD_LVER_ERR)),
    860	FLAG_ENTRY0("uncorrectable_err", DCCE(UNCORRECTABLE_ERR)),
    861	FLAG_ENTRY0("bad_crdt_ack_err", DCCE(BAD_CRDT_ACK_ERR)),
    862	FLAG_ENTRY0("unsup_pkt_type", DCCE(UNSUP_PKT_TYPE)),
    863	FLAG_ENTRY0("bad_ctrl_flit_err", DCCE(BAD_CTRL_FLIT_ERR)),
    864	FLAG_ENTRY0("event_cntr_parity_err", DCCE(EVENT_CNTR_PARITY_ERR)),
    865	FLAG_ENTRY0("event_cntr_rollover_err", DCCE(EVENT_CNTR_ROLLOVER_ERR)),
    866	FLAG_ENTRY0("link_err", DCCE(LINK_ERR)),
    867	FLAG_ENTRY0("misc_cntr_rollover_err", DCCE(MISC_CNTR_ROLLOVER_ERR)),
    868	FLAG_ENTRY0("bad_ctrl_dist_err", DCCE(BAD_CTRL_DIST_ERR)),
    869	FLAG_ENTRY0("bad_tail_dist_err", DCCE(BAD_TAIL_DIST_ERR)),
    870	FLAG_ENTRY0("bad_head_dist_err", DCCE(BAD_HEAD_DIST_ERR)),
    871	FLAG_ENTRY0("nonvl15_state_err", DCCE(NONVL15_STATE_ERR)),
    872	FLAG_ENTRY0("vl15_multi_err", DCCE(VL15_MULTI_ERR)),
    873	FLAG_ENTRY0("bad_pkt_length_err", DCCE(BAD_PKT_LENGTH_ERR)),
    874	FLAG_ENTRY0("unsup_vl_err", DCCE(UNSUP_VL_ERR)),
    875	FLAG_ENTRY0("perm_nvl15_err", DCCE(PERM_NVL15_ERR)),
    876	FLAG_ENTRY0("slid_zero_err", DCCE(SLID_ZERO_ERR)),
    877	FLAG_ENTRY0("dlid_zero_err", DCCE(DLID_ZERO_ERR)),
    878	FLAG_ENTRY0("length_mtu_err", DCCE(LENGTH_MTU_ERR)),
    879	FLAG_ENTRY0("rx_early_drop_err", DCCE(RX_EARLY_DROP_ERR)),
    880	FLAG_ENTRY0("late_short_err", DCCE(LATE_SHORT_ERR)),
    881	FLAG_ENTRY0("late_long_err", DCCE(LATE_LONG_ERR)),
    882	FLAG_ENTRY0("late_ebp_err", DCCE(LATE_EBP_ERR)),
    883	FLAG_ENTRY0("fpe_tx_fifo_ovflw_err", DCCE(FPE_TX_FIFO_OVFLW_ERR)),
    884	FLAG_ENTRY0("fpe_tx_fifo_unflw_err", DCCE(FPE_TX_FIFO_UNFLW_ERR)),
    885	FLAG_ENTRY0("csr_access_blocked_host", DCCE(CSR_ACCESS_BLOCKED_HOST)),
    886	FLAG_ENTRY0("csr_access_blocked_uc", DCCE(CSR_ACCESS_BLOCKED_UC)),
    887	FLAG_ENTRY0("tx_ctrl_parity_err", DCCE(TX_CTRL_PARITY_ERR)),
    888	FLAG_ENTRY0("tx_ctrl_parity_mbe_err", DCCE(TX_CTRL_PARITY_MBE_ERR)),
    889	FLAG_ENTRY0("tx_sc_parity_err", DCCE(TX_SC_PARITY_ERR)),
    890	FLAG_ENTRY0("rx_ctrl_parity_mbe_err", DCCE(RX_CTRL_PARITY_MBE_ERR)),
    891	FLAG_ENTRY0("csr_parity_err", DCCE(CSR_PARITY_ERR)),
    892	FLAG_ENTRY0("csr_inval_addr", DCCE(CSR_INVAL_ADDR)),
    893	FLAG_ENTRY0("tx_byte_shft_parity_err", DCCE(TX_BYTE_SHFT_PARITY_ERR)),
    894	FLAG_ENTRY0("rx_byte_shft_parity_err", DCCE(RX_BYTE_SHFT_PARITY_ERR)),
    895	FLAG_ENTRY0("fmconfig_err", DCCE(FMCONFIG_ERR)),
    896	FLAG_ENTRY0("rcvport_err", DCCE(RCVPORT_ERR)),
    897};
    898
    899/*
    900 * LCB error flags
    901 */
    902#define LCBE(name) DC_LCB_ERR_FLG_##name##_SMASK
    903static struct flag_table lcb_err_flags[] = {
    904/* 0*/	FLAG_ENTRY0("CSR_PARITY_ERR", LCBE(CSR_PARITY_ERR)),
    905/* 1*/	FLAG_ENTRY0("INVALID_CSR_ADDR", LCBE(INVALID_CSR_ADDR)),
    906/* 2*/	FLAG_ENTRY0("RST_FOR_FAILED_DESKEW", LCBE(RST_FOR_FAILED_DESKEW)),
    907/* 3*/	FLAG_ENTRY0("ALL_LNS_FAILED_REINIT_TEST",
    908		LCBE(ALL_LNS_FAILED_REINIT_TEST)),
    909/* 4*/	FLAG_ENTRY0("LOST_REINIT_STALL_OR_TOS", LCBE(LOST_REINIT_STALL_OR_TOS)),
    910/* 5*/	FLAG_ENTRY0("TX_LESS_THAN_FOUR_LNS", LCBE(TX_LESS_THAN_FOUR_LNS)),
    911/* 6*/	FLAG_ENTRY0("RX_LESS_THAN_FOUR_LNS", LCBE(RX_LESS_THAN_FOUR_LNS)),
    912/* 7*/	FLAG_ENTRY0("SEQ_CRC_ERR", LCBE(SEQ_CRC_ERR)),
    913/* 8*/	FLAG_ENTRY0("REINIT_FROM_PEER", LCBE(REINIT_FROM_PEER)),
    914/* 9*/	FLAG_ENTRY0("REINIT_FOR_LN_DEGRADE", LCBE(REINIT_FOR_LN_DEGRADE)),
    915/*10*/	FLAG_ENTRY0("CRC_ERR_CNT_HIT_LIMIT", LCBE(CRC_ERR_CNT_HIT_LIMIT)),
    916/*11*/	FLAG_ENTRY0("RCLK_STOPPED", LCBE(RCLK_STOPPED)),
    917/*12*/	FLAG_ENTRY0("UNEXPECTED_REPLAY_MARKER", LCBE(UNEXPECTED_REPLAY_MARKER)),
    918/*13*/	FLAG_ENTRY0("UNEXPECTED_ROUND_TRIP_MARKER",
    919		LCBE(UNEXPECTED_ROUND_TRIP_MARKER)),
    920/*14*/	FLAG_ENTRY0("ILLEGAL_NULL_LTP", LCBE(ILLEGAL_NULL_LTP)),
    921/*15*/	FLAG_ENTRY0("ILLEGAL_FLIT_ENCODING", LCBE(ILLEGAL_FLIT_ENCODING)),
    922/*16*/	FLAG_ENTRY0("FLIT_INPUT_BUF_OFLW", LCBE(FLIT_INPUT_BUF_OFLW)),
    923/*17*/	FLAG_ENTRY0("VL_ACK_INPUT_BUF_OFLW", LCBE(VL_ACK_INPUT_BUF_OFLW)),
    924/*18*/	FLAG_ENTRY0("VL_ACK_INPUT_PARITY_ERR", LCBE(VL_ACK_INPUT_PARITY_ERR)),
    925/*19*/	FLAG_ENTRY0("VL_ACK_INPUT_WRONG_CRC_MODE",
    926		LCBE(VL_ACK_INPUT_WRONG_CRC_MODE)),
    927/*20*/	FLAG_ENTRY0("FLIT_INPUT_BUF_MBE", LCBE(FLIT_INPUT_BUF_MBE)),
    928/*21*/	FLAG_ENTRY0("FLIT_INPUT_BUF_SBE", LCBE(FLIT_INPUT_BUF_SBE)),
    929/*22*/	FLAG_ENTRY0("REPLAY_BUF_MBE", LCBE(REPLAY_BUF_MBE)),
    930/*23*/	FLAG_ENTRY0("REPLAY_BUF_SBE", LCBE(REPLAY_BUF_SBE)),
    931/*24*/	FLAG_ENTRY0("CREDIT_RETURN_FLIT_MBE", LCBE(CREDIT_RETURN_FLIT_MBE)),
    932/*25*/	FLAG_ENTRY0("RST_FOR_LINK_TIMEOUT", LCBE(RST_FOR_LINK_TIMEOUT)),
    933/*26*/	FLAG_ENTRY0("RST_FOR_INCOMPLT_RND_TRIP",
    934		LCBE(RST_FOR_INCOMPLT_RND_TRIP)),
    935/*27*/	FLAG_ENTRY0("HOLD_REINIT", LCBE(HOLD_REINIT)),
    936/*28*/	FLAG_ENTRY0("NEG_EDGE_LINK_TRANSFER_ACTIVE",
    937		LCBE(NEG_EDGE_LINK_TRANSFER_ACTIVE)),
    938/*29*/	FLAG_ENTRY0("REDUNDANT_FLIT_PARITY_ERR",
    939		LCBE(REDUNDANT_FLIT_PARITY_ERR))
    940};
    941
    942/*
    943 * DC8051 Error Flags
    944 */
    945#define D8E(name) DC_DC8051_ERR_FLG_##name##_SMASK
    946static struct flag_table dc8051_err_flags[] = {
    947	FLAG_ENTRY0("SET_BY_8051", D8E(SET_BY_8051)),
    948	FLAG_ENTRY0("LOST_8051_HEART_BEAT", D8E(LOST_8051_HEART_BEAT)),
    949	FLAG_ENTRY0("CRAM_MBE", D8E(CRAM_MBE)),
    950	FLAG_ENTRY0("CRAM_SBE", D8E(CRAM_SBE)),
    951	FLAG_ENTRY0("DRAM_MBE", D8E(DRAM_MBE)),
    952	FLAG_ENTRY0("DRAM_SBE", D8E(DRAM_SBE)),
    953	FLAG_ENTRY0("IRAM_MBE", D8E(IRAM_MBE)),
    954	FLAG_ENTRY0("IRAM_SBE", D8E(IRAM_SBE)),
    955	FLAG_ENTRY0("UNMATCHED_SECURE_MSG_ACROSS_BCC_LANES",
    956		    D8E(UNMATCHED_SECURE_MSG_ACROSS_BCC_LANES)),
    957	FLAG_ENTRY0("INVALID_CSR_ADDR", D8E(INVALID_CSR_ADDR)),
    958};
    959
    960/*
    961 * DC8051 Information Error flags
    962 *
    963 * Flags in DC8051_DBG_ERR_INFO_SET_BY_8051.ERROR field.
    964 */
    965static struct flag_table dc8051_info_err_flags[] = {
    966	FLAG_ENTRY0("Spico ROM check failed",  SPICO_ROM_FAILED),
    967	FLAG_ENTRY0("Unknown frame received",  UNKNOWN_FRAME),
    968	FLAG_ENTRY0("Target BER not met",      TARGET_BER_NOT_MET),
    969	FLAG_ENTRY0("Serdes internal loopback failure",
    970		    FAILED_SERDES_INTERNAL_LOOPBACK),
    971	FLAG_ENTRY0("Failed SerDes init",      FAILED_SERDES_INIT),
    972	FLAG_ENTRY0("Failed LNI(Polling)",     FAILED_LNI_POLLING),
    973	FLAG_ENTRY0("Failed LNI(Debounce)",    FAILED_LNI_DEBOUNCE),
    974	FLAG_ENTRY0("Failed LNI(EstbComm)",    FAILED_LNI_ESTBCOMM),
    975	FLAG_ENTRY0("Failed LNI(OptEq)",       FAILED_LNI_OPTEQ),
    976	FLAG_ENTRY0("Failed LNI(VerifyCap_1)", FAILED_LNI_VERIFY_CAP1),
    977	FLAG_ENTRY0("Failed LNI(VerifyCap_2)", FAILED_LNI_VERIFY_CAP2),
    978	FLAG_ENTRY0("Failed LNI(ConfigLT)",    FAILED_LNI_CONFIGLT),
    979	FLAG_ENTRY0("Host Handshake Timeout",  HOST_HANDSHAKE_TIMEOUT),
    980	FLAG_ENTRY0("External Device Request Timeout",
    981		    EXTERNAL_DEVICE_REQ_TIMEOUT),
    982};
    983
    984/*
    985 * DC8051 Information Host Information flags
    986 *
    987 * Flags in DC8051_DBG_ERR_INFO_SET_BY_8051.HOST_MSG field.
    988 */
    989static struct flag_table dc8051_info_host_msg_flags[] = {
    990	FLAG_ENTRY0("Host request done", 0x0001),
    991	FLAG_ENTRY0("BC PWR_MGM message", 0x0002),
    992	FLAG_ENTRY0("BC SMA message", 0x0004),
    993	FLAG_ENTRY0("BC Unknown message (BCC)", 0x0008),
    994	FLAG_ENTRY0("BC Unknown message (LCB)", 0x0010),
    995	FLAG_ENTRY0("External device config request", 0x0020),
    996	FLAG_ENTRY0("VerifyCap all frames received", 0x0040),
    997	FLAG_ENTRY0("LinkUp achieved", 0x0080),
    998	FLAG_ENTRY0("Link going down", 0x0100),
    999	FLAG_ENTRY0("Link width downgraded", 0x0200),
   1000};
   1001
   1002static u32 encoded_size(u32 size);
   1003static u32 chip_to_opa_lstate(struct hfi1_devdata *dd, u32 chip_lstate);
   1004static int set_physical_link_state(struct hfi1_devdata *dd, u64 state);
   1005static void read_vc_remote_phy(struct hfi1_devdata *dd, u8 *power_management,
   1006			       u8 *continuous);
   1007static void read_vc_remote_fabric(struct hfi1_devdata *dd, u8 *vau, u8 *z,
   1008				  u8 *vcu, u16 *vl15buf, u8 *crc_sizes);
   1009static void read_vc_remote_link_width(struct hfi1_devdata *dd,
   1010				      u8 *remote_tx_rate, u16 *link_widths);
   1011static void read_vc_local_link_mode(struct hfi1_devdata *dd, u8 *misc_bits,
   1012				    u8 *flag_bits, u16 *link_widths);
   1013static void read_remote_device_id(struct hfi1_devdata *dd, u16 *device_id,
   1014				  u8 *device_rev);
   1015static void read_local_lni(struct hfi1_devdata *dd, u8 *enable_lane_rx);
   1016static int read_tx_settings(struct hfi1_devdata *dd, u8 *enable_lane_tx,
   1017			    u8 *tx_polarity_inversion,
   1018			    u8 *rx_polarity_inversion, u8 *max_rate);
   1019static void handle_sdma_eng_err(struct hfi1_devdata *dd,
   1020				unsigned int context, u64 err_status);
   1021static void handle_qsfp_int(struct hfi1_devdata *dd, u32 source, u64 reg);
   1022static void handle_dcc_err(struct hfi1_devdata *dd,
   1023			   unsigned int context, u64 err_status);
   1024static void handle_lcb_err(struct hfi1_devdata *dd,
   1025			   unsigned int context, u64 err_status);
   1026static void handle_8051_interrupt(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1027static void handle_cce_err(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1028static void handle_rxe_err(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1029static void handle_misc_err(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1030static void handle_pio_err(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1031static void handle_sdma_err(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1032static void handle_egress_err(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1033static void handle_txe_err(struct hfi1_devdata *dd, u32 unused, u64 reg);
   1034static void set_partition_keys(struct hfi1_pportdata *ppd);
   1035static const char *link_state_name(u32 state);
   1036static const char *link_state_reason_name(struct hfi1_pportdata *ppd,
   1037					  u32 state);
   1038static int do_8051_command(struct hfi1_devdata *dd, u32 type, u64 in_data,
   1039			   u64 *out_data);
   1040static int read_idle_sma(struct hfi1_devdata *dd, u64 *data);
   1041static int thermal_init(struct hfi1_devdata *dd);
   1042
   1043static void update_statusp(struct hfi1_pportdata *ppd, u32 state);
   1044static int wait_phys_link_offline_substates(struct hfi1_pportdata *ppd,
   1045					    int msecs);
   1046static int wait_logical_linkstate(struct hfi1_pportdata *ppd, u32 state,
   1047				  int msecs);
   1048static void log_state_transition(struct hfi1_pportdata *ppd, u32 state);
   1049static void log_physical_state(struct hfi1_pportdata *ppd, u32 state);
   1050static int wait_physical_linkstate(struct hfi1_pportdata *ppd, u32 state,
   1051				   int msecs);
   1052static int wait_phys_link_out_of_offline(struct hfi1_pportdata *ppd,
   1053					 int msecs);
   1054static void read_planned_down_reason_code(struct hfi1_devdata *dd, u8 *pdrrc);
   1055static void read_link_down_reason(struct hfi1_devdata *dd, u8 *ldr);
   1056static void handle_temp_err(struct hfi1_devdata *dd);
   1057static void dc_shutdown(struct hfi1_devdata *dd);
   1058static void dc_start(struct hfi1_devdata *dd);
   1059static int qos_rmt_entries(struct hfi1_devdata *dd, unsigned int *mp,
   1060			   unsigned int *np);
   1061static void clear_full_mgmt_pkey(struct hfi1_pportdata *ppd);
   1062static int wait_link_transfer_active(struct hfi1_devdata *dd, int wait_ms);
   1063static void clear_rsm_rule(struct hfi1_devdata *dd, u8 rule_index);
   1064static void update_xmit_counters(struct hfi1_pportdata *ppd, u16 link_width);
   1065
   1066/*
   1067 * Error interrupt table entry.  This is used as input to the interrupt
   1068 * "clear down" routine used for all second tier error interrupt register.
   1069 * Second tier interrupt registers have a single bit representing them
   1070 * in the top-level CceIntStatus.
   1071 */
   1072struct err_reg_info {
   1073	u32 status;		/* status CSR offset */
   1074	u32 clear;		/* clear CSR offset */
   1075	u32 mask;		/* mask CSR offset */
   1076	void (*handler)(struct hfi1_devdata *dd, u32 source, u64 reg);
   1077	const char *desc;
   1078};
   1079
   1080#define NUM_MISC_ERRS (IS_GENERAL_ERR_END + 1 - IS_GENERAL_ERR_START)
   1081#define NUM_DC_ERRS (IS_DC_END + 1 - IS_DC_START)
   1082#define NUM_VARIOUS (IS_VARIOUS_END + 1 - IS_VARIOUS_START)
   1083
   1084/*
   1085 * Helpers for building HFI and DC error interrupt table entries.  Different
   1086 * helpers are needed because of inconsistent register names.
   1087 */
   1088#define EE(reg, handler, desc) \
   1089	{ reg##_STATUS, reg##_CLEAR, reg##_MASK, \
   1090		handler, desc }
   1091#define DC_EE1(reg, handler, desc) \
   1092	{ reg##_FLG, reg##_FLG_CLR, reg##_FLG_EN, handler, desc }
   1093#define DC_EE2(reg, handler, desc) \
   1094	{ reg##_FLG, reg##_CLR, reg##_EN, handler, desc }
   1095
   1096/*
   1097 * Table of the "misc" grouping of error interrupts.  Each entry refers to
   1098 * another register containing more information.
   1099 */
   1100static const struct err_reg_info misc_errs[NUM_MISC_ERRS] = {
   1101/* 0*/	EE(CCE_ERR,		handle_cce_err,    "CceErr"),
   1102/* 1*/	EE(RCV_ERR,		handle_rxe_err,    "RxeErr"),
   1103/* 2*/	EE(MISC_ERR,	handle_misc_err,   "MiscErr"),
   1104/* 3*/	{ 0, 0, 0, NULL }, /* reserved */
   1105/* 4*/	EE(SEND_PIO_ERR,    handle_pio_err,    "PioErr"),
   1106/* 5*/	EE(SEND_DMA_ERR,    handle_sdma_err,   "SDmaErr"),
   1107/* 6*/	EE(SEND_EGRESS_ERR, handle_egress_err, "EgressErr"),
   1108/* 7*/	EE(SEND_ERR,	handle_txe_err,    "TxeErr")
   1109	/* the rest are reserved */
   1110};
   1111
   1112/*
   1113 * Index into the Various section of the interrupt sources
   1114 * corresponding to the Critical Temperature interrupt.
   1115 */
   1116#define TCRIT_INT_SOURCE 4
   1117
   1118/*
   1119 * SDMA error interrupt entry - refers to another register containing more
   1120 * information.
   1121 */
   1122static const struct err_reg_info sdma_eng_err =
   1123	EE(SEND_DMA_ENG_ERR, handle_sdma_eng_err, "SDmaEngErr");
   1124
   1125static const struct err_reg_info various_err[NUM_VARIOUS] = {
   1126/* 0*/	{ 0, 0, 0, NULL }, /* PbcInt */
   1127/* 1*/	{ 0, 0, 0, NULL }, /* GpioAssertInt */
   1128/* 2*/	EE(ASIC_QSFP1,	handle_qsfp_int,	"QSFP1"),
   1129/* 3*/	EE(ASIC_QSFP2,	handle_qsfp_int,	"QSFP2"),
   1130/* 4*/	{ 0, 0, 0, NULL }, /* TCritInt */
   1131	/* rest are reserved */
   1132};
   1133
   1134/*
   1135 * The DC encoding of mtu_cap for 10K MTU in the DCC_CFG_PORT_CONFIG
   1136 * register can not be derived from the MTU value because 10K is not
   1137 * a power of 2. Therefore, we need a constant. Everything else can
   1138 * be calculated.
   1139 */
   1140#define DCC_CFG_PORT_MTU_CAP_10240 7
   1141
   1142/*
   1143 * Table of the DC grouping of error interrupts.  Each entry refers to
   1144 * another register containing more information.
   1145 */
   1146static const struct err_reg_info dc_errs[NUM_DC_ERRS] = {
   1147/* 0*/	DC_EE1(DCC_ERR,		handle_dcc_err,	       "DCC Err"),
   1148/* 1*/	DC_EE2(DC_LCB_ERR,	handle_lcb_err,	       "LCB Err"),
   1149/* 2*/	DC_EE2(DC_DC8051_ERR,	handle_8051_interrupt, "DC8051 Interrupt"),
   1150/* 3*/	/* dc_lbm_int - special, see is_dc_int() */
   1151	/* the rest are reserved */
   1152};
   1153
   1154struct cntr_entry {
   1155	/*
   1156	 * counter name
   1157	 */
   1158	char *name;
   1159
   1160	/*
   1161	 * csr to read for name (if applicable)
   1162	 */
   1163	u64 csr;
   1164
   1165	/*
   1166	 * offset into dd or ppd to store the counter's value
   1167	 */
   1168	int offset;
   1169
   1170	/*
   1171	 * flags
   1172	 */
   1173	u8 flags;
   1174
   1175	/*
   1176	 * accessor for stat element, context either dd or ppd
   1177	 */
   1178	u64 (*rw_cntr)(const struct cntr_entry *, void *context, int vl,
   1179		       int mode, u64 data);
   1180};
   1181
   1182#define C_RCV_HDR_OVF_FIRST C_RCV_HDR_OVF_0
   1183#define C_RCV_HDR_OVF_LAST C_RCV_HDR_OVF_159
   1184
   1185#define CNTR_ELEM(name, csr, offset, flags, accessor) \
   1186{ \
   1187	name, \
   1188	csr, \
   1189	offset, \
   1190	flags, \
   1191	accessor \
   1192}
   1193
   1194/* 32bit RXE */
   1195#define RXE32_PORT_CNTR_ELEM(name, counter, flags) \
   1196CNTR_ELEM(#name, \
   1197	  (counter * 8 + RCV_COUNTER_ARRAY32), \
   1198	  0, flags | CNTR_32BIT, \
   1199	  port_access_u32_csr)
   1200
   1201#define RXE32_DEV_CNTR_ELEM(name, counter, flags) \
   1202CNTR_ELEM(#name, \
   1203	  (counter * 8 + RCV_COUNTER_ARRAY32), \
   1204	  0, flags | CNTR_32BIT, \
   1205	  dev_access_u32_csr)
   1206
   1207/* 64bit RXE */
   1208#define RXE64_PORT_CNTR_ELEM(name, counter, flags) \
   1209CNTR_ELEM(#name, \
   1210	  (counter * 8 + RCV_COUNTER_ARRAY64), \
   1211	  0, flags, \
   1212	  port_access_u64_csr)
   1213
   1214#define RXE64_DEV_CNTR_ELEM(name, counter, flags) \
   1215CNTR_ELEM(#name, \
   1216	  (counter * 8 + RCV_COUNTER_ARRAY64), \
   1217	  0, flags, \
   1218	  dev_access_u64_csr)
   1219
   1220#define OVR_LBL(ctx) C_RCV_HDR_OVF_ ## ctx
   1221#define OVR_ELM(ctx) \
   1222CNTR_ELEM("RcvHdrOvr" #ctx, \
   1223	  (RCV_HDR_OVFL_CNT + ctx * 0x100), \
   1224	  0, CNTR_NORMAL, port_access_u64_csr)
   1225
   1226/* 32bit TXE */
   1227#define TXE32_PORT_CNTR_ELEM(name, counter, flags) \
   1228CNTR_ELEM(#name, \
   1229	  (counter * 8 + SEND_COUNTER_ARRAY32), \
   1230	  0, flags | CNTR_32BIT, \
   1231	  port_access_u32_csr)
   1232
   1233/* 64bit TXE */
   1234#define TXE64_PORT_CNTR_ELEM(name, counter, flags) \
   1235CNTR_ELEM(#name, \
   1236	  (counter * 8 + SEND_COUNTER_ARRAY64), \
   1237	  0, flags, \
   1238	  port_access_u64_csr)
   1239
   1240# define TX64_DEV_CNTR_ELEM(name, counter, flags) \
   1241CNTR_ELEM(#name,\
   1242	  counter * 8 + SEND_COUNTER_ARRAY64, \
   1243	  0, \
   1244	  flags, \
   1245	  dev_access_u64_csr)
   1246
   1247/* CCE */
   1248#define CCE_PERF_DEV_CNTR_ELEM(name, counter, flags) \
   1249CNTR_ELEM(#name, \
   1250	  (counter * 8 + CCE_COUNTER_ARRAY32), \
   1251	  0, flags | CNTR_32BIT, \
   1252	  dev_access_u32_csr)
   1253
   1254#define CCE_INT_DEV_CNTR_ELEM(name, counter, flags) \
   1255CNTR_ELEM(#name, \
   1256	  (counter * 8 + CCE_INT_COUNTER_ARRAY32), \
   1257	  0, flags | CNTR_32BIT, \
   1258	  dev_access_u32_csr)
   1259
   1260/* DC */
   1261#define DC_PERF_CNTR(name, counter, flags) \
   1262CNTR_ELEM(#name, \
   1263	  counter, \
   1264	  0, \
   1265	  flags, \
   1266	  dev_access_u64_csr)
   1267
   1268#define DC_PERF_CNTR_LCB(name, counter, flags) \
   1269CNTR_ELEM(#name, \
   1270	  counter, \
   1271	  0, \
   1272	  flags, \
   1273	  dc_access_lcb_cntr)
   1274
   1275/* ibp counters */
   1276#define SW_IBP_CNTR(name, cntr) \
   1277CNTR_ELEM(#name, \
   1278	  0, \
   1279	  0, \
   1280	  CNTR_SYNTH, \
   1281	  access_ibp_##cntr)
   1282
   1283/**
   1284 * hfi1_addr_from_offset - return addr for readq/writeq
   1285 * @dd: the dd device
   1286 * @offset: the offset of the CSR within bar0
   1287 *
   1288 * This routine selects the appropriate base address
   1289 * based on the indicated offset.
   1290 */
   1291static inline void __iomem *hfi1_addr_from_offset(
   1292	const struct hfi1_devdata *dd,
   1293	u32 offset)
   1294{
   1295	if (offset >= dd->base2_start)
   1296		return dd->kregbase2 + (offset - dd->base2_start);
   1297	return dd->kregbase1 + offset;
   1298}
   1299
   1300/**
   1301 * read_csr - read CSR at the indicated offset
   1302 * @dd: the dd device
   1303 * @offset: the offset of the CSR within bar0
   1304 *
   1305 * Return: the value read or all FF's if there
   1306 * is no mapping
   1307 */
   1308u64 read_csr(const struct hfi1_devdata *dd, u32 offset)
   1309{
   1310	if (dd->flags & HFI1_PRESENT)
   1311		return readq(hfi1_addr_from_offset(dd, offset));
   1312	return -1;
   1313}
   1314
   1315/**
   1316 * write_csr - write CSR at the indicated offset
   1317 * @dd: the dd device
   1318 * @offset: the offset of the CSR within bar0
   1319 * @value: value to write
   1320 */
   1321void write_csr(const struct hfi1_devdata *dd, u32 offset, u64 value)
   1322{
   1323	if (dd->flags & HFI1_PRESENT) {
   1324		void __iomem *base = hfi1_addr_from_offset(dd, offset);
   1325
   1326		/* avoid write to RcvArray */
   1327		if (WARN_ON(offset >= RCV_ARRAY && offset < dd->base2_start))
   1328			return;
   1329		writeq(value, base);
   1330	}
   1331}
   1332
   1333/**
   1334 * get_csr_addr - return te iomem address for offset
   1335 * @dd: the dd device
   1336 * @offset: the offset of the CSR within bar0
   1337 *
   1338 * Return: The iomem address to use in subsequent
   1339 * writeq/readq operations.
   1340 */
   1341void __iomem *get_csr_addr(
   1342	const struct hfi1_devdata *dd,
   1343	u32 offset)
   1344{
   1345	if (dd->flags & HFI1_PRESENT)
   1346		return hfi1_addr_from_offset(dd, offset);
   1347	return NULL;
   1348}
   1349
   1350static inline u64 read_write_csr(const struct hfi1_devdata *dd, u32 csr,
   1351				 int mode, u64 value)
   1352{
   1353	u64 ret;
   1354
   1355	if (mode == CNTR_MODE_R) {
   1356		ret = read_csr(dd, csr);
   1357	} else if (mode == CNTR_MODE_W) {
   1358		write_csr(dd, csr, value);
   1359		ret = value;
   1360	} else {
   1361		dd_dev_err(dd, "Invalid cntr register access mode");
   1362		return 0;
   1363	}
   1364
   1365	hfi1_cdbg(CNTR, "csr 0x%x val 0x%llx mode %d", csr, ret, mode);
   1366	return ret;
   1367}
   1368
   1369/* Dev Access */
   1370static u64 dev_access_u32_csr(const struct cntr_entry *entry,
   1371			      void *context, int vl, int mode, u64 data)
   1372{
   1373	struct hfi1_devdata *dd = context;
   1374	u64 csr = entry->csr;
   1375
   1376	if (entry->flags & CNTR_SDMA) {
   1377		if (vl == CNTR_INVALID_VL)
   1378			return 0;
   1379		csr += 0x100 * vl;
   1380	} else {
   1381		if (vl != CNTR_INVALID_VL)
   1382			return 0;
   1383	}
   1384	return read_write_csr(dd, csr, mode, data);
   1385}
   1386
   1387static u64 access_sde_err_cnt(const struct cntr_entry *entry,
   1388			      void *context, int idx, int mode, u64 data)
   1389{
   1390	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1391
   1392	if (dd->per_sdma && idx < dd->num_sdma)
   1393		return dd->per_sdma[idx].err_cnt;
   1394	return 0;
   1395}
   1396
   1397static u64 access_sde_int_cnt(const struct cntr_entry *entry,
   1398			      void *context, int idx, int mode, u64 data)
   1399{
   1400	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1401
   1402	if (dd->per_sdma && idx < dd->num_sdma)
   1403		return dd->per_sdma[idx].sdma_int_cnt;
   1404	return 0;
   1405}
   1406
   1407static u64 access_sde_idle_int_cnt(const struct cntr_entry *entry,
   1408				   void *context, int idx, int mode, u64 data)
   1409{
   1410	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1411
   1412	if (dd->per_sdma && idx < dd->num_sdma)
   1413		return dd->per_sdma[idx].idle_int_cnt;
   1414	return 0;
   1415}
   1416
   1417static u64 access_sde_progress_int_cnt(const struct cntr_entry *entry,
   1418				       void *context, int idx, int mode,
   1419				       u64 data)
   1420{
   1421	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1422
   1423	if (dd->per_sdma && idx < dd->num_sdma)
   1424		return dd->per_sdma[idx].progress_int_cnt;
   1425	return 0;
   1426}
   1427
   1428static u64 dev_access_u64_csr(const struct cntr_entry *entry, void *context,
   1429			      int vl, int mode, u64 data)
   1430{
   1431	struct hfi1_devdata *dd = context;
   1432
   1433	u64 val = 0;
   1434	u64 csr = entry->csr;
   1435
   1436	if (entry->flags & CNTR_VL) {
   1437		if (vl == CNTR_INVALID_VL)
   1438			return 0;
   1439		csr += 8 * vl;
   1440	} else {
   1441		if (vl != CNTR_INVALID_VL)
   1442			return 0;
   1443	}
   1444
   1445	val = read_write_csr(dd, csr, mode, data);
   1446	return val;
   1447}
   1448
   1449static u64 dc_access_lcb_cntr(const struct cntr_entry *entry, void *context,
   1450			      int vl, int mode, u64 data)
   1451{
   1452	struct hfi1_devdata *dd = context;
   1453	u32 csr = entry->csr;
   1454	int ret = 0;
   1455
   1456	if (vl != CNTR_INVALID_VL)
   1457		return 0;
   1458	if (mode == CNTR_MODE_R)
   1459		ret = read_lcb_csr(dd, csr, &data);
   1460	else if (mode == CNTR_MODE_W)
   1461		ret = write_lcb_csr(dd, csr, data);
   1462
   1463	if (ret) {
   1464		dd_dev_err(dd, "Could not acquire LCB for counter 0x%x", csr);
   1465		return 0;
   1466	}
   1467
   1468	hfi1_cdbg(CNTR, "csr 0x%x val 0x%llx mode %d", csr, data, mode);
   1469	return data;
   1470}
   1471
   1472/* Port Access */
   1473static u64 port_access_u32_csr(const struct cntr_entry *entry, void *context,
   1474			       int vl, int mode, u64 data)
   1475{
   1476	struct hfi1_pportdata *ppd = context;
   1477
   1478	if (vl != CNTR_INVALID_VL)
   1479		return 0;
   1480	return read_write_csr(ppd->dd, entry->csr, mode, data);
   1481}
   1482
   1483static u64 port_access_u64_csr(const struct cntr_entry *entry,
   1484			       void *context, int vl, int mode, u64 data)
   1485{
   1486	struct hfi1_pportdata *ppd = context;
   1487	u64 val;
   1488	u64 csr = entry->csr;
   1489
   1490	if (entry->flags & CNTR_VL) {
   1491		if (vl == CNTR_INVALID_VL)
   1492			return 0;
   1493		csr += 8 * vl;
   1494	} else {
   1495		if (vl != CNTR_INVALID_VL)
   1496			return 0;
   1497	}
   1498	val = read_write_csr(ppd->dd, csr, mode, data);
   1499	return val;
   1500}
   1501
   1502/* Software defined */
   1503static inline u64 read_write_sw(struct hfi1_devdata *dd, u64 *cntr, int mode,
   1504				u64 data)
   1505{
   1506	u64 ret;
   1507
   1508	if (mode == CNTR_MODE_R) {
   1509		ret = *cntr;
   1510	} else if (mode == CNTR_MODE_W) {
   1511		*cntr = data;
   1512		ret = data;
   1513	} else {
   1514		dd_dev_err(dd, "Invalid cntr sw access mode");
   1515		return 0;
   1516	}
   1517
   1518	hfi1_cdbg(CNTR, "val 0x%llx mode %d", ret, mode);
   1519
   1520	return ret;
   1521}
   1522
   1523static u64 access_sw_link_dn_cnt(const struct cntr_entry *entry, void *context,
   1524				 int vl, int mode, u64 data)
   1525{
   1526	struct hfi1_pportdata *ppd = context;
   1527
   1528	if (vl != CNTR_INVALID_VL)
   1529		return 0;
   1530	return read_write_sw(ppd->dd, &ppd->link_downed, mode, data);
   1531}
   1532
   1533static u64 access_sw_link_up_cnt(const struct cntr_entry *entry, void *context,
   1534				 int vl, int mode, u64 data)
   1535{
   1536	struct hfi1_pportdata *ppd = context;
   1537
   1538	if (vl != CNTR_INVALID_VL)
   1539		return 0;
   1540	return read_write_sw(ppd->dd, &ppd->link_up, mode, data);
   1541}
   1542
   1543static u64 access_sw_unknown_frame_cnt(const struct cntr_entry *entry,
   1544				       void *context, int vl, int mode,
   1545				       u64 data)
   1546{
   1547	struct hfi1_pportdata *ppd = (struct hfi1_pportdata *)context;
   1548
   1549	if (vl != CNTR_INVALID_VL)
   1550		return 0;
   1551	return read_write_sw(ppd->dd, &ppd->unknown_frame_count, mode, data);
   1552}
   1553
   1554static u64 access_sw_xmit_discards(const struct cntr_entry *entry,
   1555				   void *context, int vl, int mode, u64 data)
   1556{
   1557	struct hfi1_pportdata *ppd = (struct hfi1_pportdata *)context;
   1558	u64 zero = 0;
   1559	u64 *counter;
   1560
   1561	if (vl == CNTR_INVALID_VL)
   1562		counter = &ppd->port_xmit_discards;
   1563	else if (vl >= 0 && vl < C_VL_COUNT)
   1564		counter = &ppd->port_xmit_discards_vl[vl];
   1565	else
   1566		counter = &zero;
   1567
   1568	return read_write_sw(ppd->dd, counter, mode, data);
   1569}
   1570
   1571static u64 access_xmit_constraint_errs(const struct cntr_entry *entry,
   1572				       void *context, int vl, int mode,
   1573				       u64 data)
   1574{
   1575	struct hfi1_pportdata *ppd = context;
   1576
   1577	if (vl != CNTR_INVALID_VL)
   1578		return 0;
   1579
   1580	return read_write_sw(ppd->dd, &ppd->port_xmit_constraint_errors,
   1581			     mode, data);
   1582}
   1583
   1584static u64 access_rcv_constraint_errs(const struct cntr_entry *entry,
   1585				      void *context, int vl, int mode, u64 data)
   1586{
   1587	struct hfi1_pportdata *ppd = context;
   1588
   1589	if (vl != CNTR_INVALID_VL)
   1590		return 0;
   1591
   1592	return read_write_sw(ppd->dd, &ppd->port_rcv_constraint_errors,
   1593			     mode, data);
   1594}
   1595
   1596u64 get_all_cpu_total(u64 __percpu *cntr)
   1597{
   1598	int cpu;
   1599	u64 counter = 0;
   1600
   1601	for_each_possible_cpu(cpu)
   1602		counter += *per_cpu_ptr(cntr, cpu);
   1603	return counter;
   1604}
   1605
   1606static u64 read_write_cpu(struct hfi1_devdata *dd, u64 *z_val,
   1607			  u64 __percpu *cntr,
   1608			  int vl, int mode, u64 data)
   1609{
   1610	u64 ret = 0;
   1611
   1612	if (vl != CNTR_INVALID_VL)
   1613		return 0;
   1614
   1615	if (mode == CNTR_MODE_R) {
   1616		ret = get_all_cpu_total(cntr) - *z_val;
   1617	} else if (mode == CNTR_MODE_W) {
   1618		/* A write can only zero the counter */
   1619		if (data == 0)
   1620			*z_val = get_all_cpu_total(cntr);
   1621		else
   1622			dd_dev_err(dd, "Per CPU cntrs can only be zeroed");
   1623	} else {
   1624		dd_dev_err(dd, "Invalid cntr sw cpu access mode");
   1625		return 0;
   1626	}
   1627
   1628	return ret;
   1629}
   1630
   1631static u64 access_sw_cpu_intr(const struct cntr_entry *entry,
   1632			      void *context, int vl, int mode, u64 data)
   1633{
   1634	struct hfi1_devdata *dd = context;
   1635
   1636	return read_write_cpu(dd, &dd->z_int_counter, dd->int_counter, vl,
   1637			      mode, data);
   1638}
   1639
   1640static u64 access_sw_cpu_rcv_limit(const struct cntr_entry *entry,
   1641				   void *context, int vl, int mode, u64 data)
   1642{
   1643	struct hfi1_devdata *dd = context;
   1644
   1645	return read_write_cpu(dd, &dd->z_rcv_limit, dd->rcv_limit, vl,
   1646			      mode, data);
   1647}
   1648
   1649static u64 access_sw_pio_wait(const struct cntr_entry *entry,
   1650			      void *context, int vl, int mode, u64 data)
   1651{
   1652	struct hfi1_devdata *dd = context;
   1653
   1654	return dd->verbs_dev.n_piowait;
   1655}
   1656
   1657static u64 access_sw_pio_drain(const struct cntr_entry *entry,
   1658			       void *context, int vl, int mode, u64 data)
   1659{
   1660	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1661
   1662	return dd->verbs_dev.n_piodrain;
   1663}
   1664
   1665static u64 access_sw_ctx0_seq_drop(const struct cntr_entry *entry,
   1666				   void *context, int vl, int mode, u64 data)
   1667{
   1668	struct hfi1_devdata *dd = context;
   1669
   1670	return dd->ctx0_seq_drop;
   1671}
   1672
   1673static u64 access_sw_vtx_wait(const struct cntr_entry *entry,
   1674			      void *context, int vl, int mode, u64 data)
   1675{
   1676	struct hfi1_devdata *dd = context;
   1677
   1678	return dd->verbs_dev.n_txwait;
   1679}
   1680
   1681static u64 access_sw_kmem_wait(const struct cntr_entry *entry,
   1682			       void *context, int vl, int mode, u64 data)
   1683{
   1684	struct hfi1_devdata *dd = context;
   1685
   1686	return dd->verbs_dev.n_kmem_wait;
   1687}
   1688
   1689static u64 access_sw_send_schedule(const struct cntr_entry *entry,
   1690				   void *context, int vl, int mode, u64 data)
   1691{
   1692	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1693
   1694	return read_write_cpu(dd, &dd->z_send_schedule, dd->send_schedule, vl,
   1695			      mode, data);
   1696}
   1697
   1698/* Software counters for the error status bits within MISC_ERR_STATUS */
   1699static u64 access_misc_pll_lock_fail_err_cnt(const struct cntr_entry *entry,
   1700					     void *context, int vl, int mode,
   1701					     u64 data)
   1702{
   1703	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1704
   1705	return dd->misc_err_status_cnt[12];
   1706}
   1707
   1708static u64 access_misc_mbist_fail_err_cnt(const struct cntr_entry *entry,
   1709					  void *context, int vl, int mode,
   1710					  u64 data)
   1711{
   1712	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1713
   1714	return dd->misc_err_status_cnt[11];
   1715}
   1716
   1717static u64 access_misc_invalid_eep_cmd_err_cnt(const struct cntr_entry *entry,
   1718					       void *context, int vl, int mode,
   1719					       u64 data)
   1720{
   1721	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1722
   1723	return dd->misc_err_status_cnt[10];
   1724}
   1725
   1726static u64 access_misc_efuse_done_parity_err_cnt(const struct cntr_entry *entry,
   1727						 void *context, int vl,
   1728						 int mode, u64 data)
   1729{
   1730	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1731
   1732	return dd->misc_err_status_cnt[9];
   1733}
   1734
   1735static u64 access_misc_efuse_write_err_cnt(const struct cntr_entry *entry,
   1736					   void *context, int vl, int mode,
   1737					   u64 data)
   1738{
   1739	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1740
   1741	return dd->misc_err_status_cnt[8];
   1742}
   1743
   1744static u64 access_misc_efuse_read_bad_addr_err_cnt(
   1745				const struct cntr_entry *entry,
   1746				void *context, int vl, int mode, u64 data)
   1747{
   1748	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1749
   1750	return dd->misc_err_status_cnt[7];
   1751}
   1752
   1753static u64 access_misc_efuse_csr_parity_err_cnt(const struct cntr_entry *entry,
   1754						void *context, int vl,
   1755						int mode, u64 data)
   1756{
   1757	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1758
   1759	return dd->misc_err_status_cnt[6];
   1760}
   1761
   1762static u64 access_misc_fw_auth_failed_err_cnt(const struct cntr_entry *entry,
   1763					      void *context, int vl, int mode,
   1764					      u64 data)
   1765{
   1766	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1767
   1768	return dd->misc_err_status_cnt[5];
   1769}
   1770
   1771static u64 access_misc_key_mismatch_err_cnt(const struct cntr_entry *entry,
   1772					    void *context, int vl, int mode,
   1773					    u64 data)
   1774{
   1775	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1776
   1777	return dd->misc_err_status_cnt[4];
   1778}
   1779
   1780static u64 access_misc_sbus_write_failed_err_cnt(const struct cntr_entry *entry,
   1781						 void *context, int vl,
   1782						 int mode, u64 data)
   1783{
   1784	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1785
   1786	return dd->misc_err_status_cnt[3];
   1787}
   1788
   1789static u64 access_misc_csr_write_bad_addr_err_cnt(
   1790				const struct cntr_entry *entry,
   1791				void *context, int vl, int mode, u64 data)
   1792{
   1793	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1794
   1795	return dd->misc_err_status_cnt[2];
   1796}
   1797
   1798static u64 access_misc_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
   1799						 void *context, int vl,
   1800						 int mode, u64 data)
   1801{
   1802	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1803
   1804	return dd->misc_err_status_cnt[1];
   1805}
   1806
   1807static u64 access_misc_csr_parity_err_cnt(const struct cntr_entry *entry,
   1808					  void *context, int vl, int mode,
   1809					  u64 data)
   1810{
   1811	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1812
   1813	return dd->misc_err_status_cnt[0];
   1814}
   1815
   1816/*
   1817 * Software counter for the aggregate of
   1818 * individual CceErrStatus counters
   1819 */
   1820static u64 access_sw_cce_err_status_aggregated_cnt(
   1821				const struct cntr_entry *entry,
   1822				void *context, int vl, int mode, u64 data)
   1823{
   1824	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1825
   1826	return dd->sw_cce_err_status_aggregate;
   1827}
   1828
   1829/*
   1830 * Software counters corresponding to each of the
   1831 * error status bits within CceErrStatus
   1832 */
   1833static u64 access_cce_msix_csr_parity_err_cnt(const struct cntr_entry *entry,
   1834					      void *context, int vl, int mode,
   1835					      u64 data)
   1836{
   1837	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1838
   1839	return dd->cce_err_status_cnt[40];
   1840}
   1841
   1842static u64 access_cce_int_map_unc_err_cnt(const struct cntr_entry *entry,
   1843					  void *context, int vl, int mode,
   1844					  u64 data)
   1845{
   1846	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1847
   1848	return dd->cce_err_status_cnt[39];
   1849}
   1850
   1851static u64 access_cce_int_map_cor_err_cnt(const struct cntr_entry *entry,
   1852					  void *context, int vl, int mode,
   1853					  u64 data)
   1854{
   1855	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1856
   1857	return dd->cce_err_status_cnt[38];
   1858}
   1859
   1860static u64 access_cce_msix_table_unc_err_cnt(const struct cntr_entry *entry,
   1861					     void *context, int vl, int mode,
   1862					     u64 data)
   1863{
   1864	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1865
   1866	return dd->cce_err_status_cnt[37];
   1867}
   1868
   1869static u64 access_cce_msix_table_cor_err_cnt(const struct cntr_entry *entry,
   1870					     void *context, int vl, int mode,
   1871					     u64 data)
   1872{
   1873	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1874
   1875	return dd->cce_err_status_cnt[36];
   1876}
   1877
   1878static u64 access_cce_rxdma_conv_fifo_parity_err_cnt(
   1879				const struct cntr_entry *entry,
   1880				void *context, int vl, int mode, u64 data)
   1881{
   1882	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1883
   1884	return dd->cce_err_status_cnt[35];
   1885}
   1886
   1887static u64 access_cce_rcpl_async_fifo_parity_err_cnt(
   1888				const struct cntr_entry *entry,
   1889				void *context, int vl, int mode, u64 data)
   1890{
   1891	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1892
   1893	return dd->cce_err_status_cnt[34];
   1894}
   1895
   1896static u64 access_cce_seg_write_bad_addr_err_cnt(const struct cntr_entry *entry,
   1897						 void *context, int vl,
   1898						 int mode, u64 data)
   1899{
   1900	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1901
   1902	return dd->cce_err_status_cnt[33];
   1903}
   1904
   1905static u64 access_cce_seg_read_bad_addr_err_cnt(const struct cntr_entry *entry,
   1906						void *context, int vl, int mode,
   1907						u64 data)
   1908{
   1909	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1910
   1911	return dd->cce_err_status_cnt[32];
   1912}
   1913
   1914static u64 access_la_triggered_cnt(const struct cntr_entry *entry,
   1915				   void *context, int vl, int mode, u64 data)
   1916{
   1917	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1918
   1919	return dd->cce_err_status_cnt[31];
   1920}
   1921
   1922static u64 access_cce_trgt_cpl_timeout_err_cnt(const struct cntr_entry *entry,
   1923					       void *context, int vl, int mode,
   1924					       u64 data)
   1925{
   1926	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1927
   1928	return dd->cce_err_status_cnt[30];
   1929}
   1930
   1931static u64 access_pcic_receive_parity_err_cnt(const struct cntr_entry *entry,
   1932					      void *context, int vl, int mode,
   1933					      u64 data)
   1934{
   1935	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1936
   1937	return dd->cce_err_status_cnt[29];
   1938}
   1939
   1940static u64 access_pcic_transmit_back_parity_err_cnt(
   1941				const struct cntr_entry *entry,
   1942				void *context, int vl, int mode, u64 data)
   1943{
   1944	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1945
   1946	return dd->cce_err_status_cnt[28];
   1947}
   1948
   1949static u64 access_pcic_transmit_front_parity_err_cnt(
   1950				const struct cntr_entry *entry,
   1951				void *context, int vl, int mode, u64 data)
   1952{
   1953	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1954
   1955	return dd->cce_err_status_cnt[27];
   1956}
   1957
   1958static u64 access_pcic_cpl_dat_q_unc_err_cnt(const struct cntr_entry *entry,
   1959					     void *context, int vl, int mode,
   1960					     u64 data)
   1961{
   1962	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1963
   1964	return dd->cce_err_status_cnt[26];
   1965}
   1966
   1967static u64 access_pcic_cpl_hd_q_unc_err_cnt(const struct cntr_entry *entry,
   1968					    void *context, int vl, int mode,
   1969					    u64 data)
   1970{
   1971	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1972
   1973	return dd->cce_err_status_cnt[25];
   1974}
   1975
   1976static u64 access_pcic_post_dat_q_unc_err_cnt(const struct cntr_entry *entry,
   1977					      void *context, int vl, int mode,
   1978					      u64 data)
   1979{
   1980	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1981
   1982	return dd->cce_err_status_cnt[24];
   1983}
   1984
   1985static u64 access_pcic_post_hd_q_unc_err_cnt(const struct cntr_entry *entry,
   1986					     void *context, int vl, int mode,
   1987					     u64 data)
   1988{
   1989	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1990
   1991	return dd->cce_err_status_cnt[23];
   1992}
   1993
   1994static u64 access_pcic_retry_sot_mem_unc_err_cnt(const struct cntr_entry *entry,
   1995						 void *context, int vl,
   1996						 int mode, u64 data)
   1997{
   1998	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   1999
   2000	return dd->cce_err_status_cnt[22];
   2001}
   2002
   2003static u64 access_pcic_retry_mem_unc_err(const struct cntr_entry *entry,
   2004					 void *context, int vl, int mode,
   2005					 u64 data)
   2006{
   2007	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2008
   2009	return dd->cce_err_status_cnt[21];
   2010}
   2011
   2012static u64 access_pcic_n_post_dat_q_parity_err_cnt(
   2013				const struct cntr_entry *entry,
   2014				void *context, int vl, int mode, u64 data)
   2015{
   2016	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2017
   2018	return dd->cce_err_status_cnt[20];
   2019}
   2020
   2021static u64 access_pcic_n_post_h_q_parity_err_cnt(const struct cntr_entry *entry,
   2022						 void *context, int vl,
   2023						 int mode, u64 data)
   2024{
   2025	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2026
   2027	return dd->cce_err_status_cnt[19];
   2028}
   2029
   2030static u64 access_pcic_cpl_dat_q_cor_err_cnt(const struct cntr_entry *entry,
   2031					     void *context, int vl, int mode,
   2032					     u64 data)
   2033{
   2034	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2035
   2036	return dd->cce_err_status_cnt[18];
   2037}
   2038
   2039static u64 access_pcic_cpl_hd_q_cor_err_cnt(const struct cntr_entry *entry,
   2040					    void *context, int vl, int mode,
   2041					    u64 data)
   2042{
   2043	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2044
   2045	return dd->cce_err_status_cnt[17];
   2046}
   2047
   2048static u64 access_pcic_post_dat_q_cor_err_cnt(const struct cntr_entry *entry,
   2049					      void *context, int vl, int mode,
   2050					      u64 data)
   2051{
   2052	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2053
   2054	return dd->cce_err_status_cnt[16];
   2055}
   2056
   2057static u64 access_pcic_post_hd_q_cor_err_cnt(const struct cntr_entry *entry,
   2058					     void *context, int vl, int mode,
   2059					     u64 data)
   2060{
   2061	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2062
   2063	return dd->cce_err_status_cnt[15];
   2064}
   2065
   2066static u64 access_pcic_retry_sot_mem_cor_err_cnt(const struct cntr_entry *entry,
   2067						 void *context, int vl,
   2068						 int mode, u64 data)
   2069{
   2070	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2071
   2072	return dd->cce_err_status_cnt[14];
   2073}
   2074
   2075static u64 access_pcic_retry_mem_cor_err_cnt(const struct cntr_entry *entry,
   2076					     void *context, int vl, int mode,
   2077					     u64 data)
   2078{
   2079	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2080
   2081	return dd->cce_err_status_cnt[13];
   2082}
   2083
   2084static u64 access_cce_cli1_async_fifo_dbg_parity_err_cnt(
   2085				const struct cntr_entry *entry,
   2086				void *context, int vl, int mode, u64 data)
   2087{
   2088	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2089
   2090	return dd->cce_err_status_cnt[12];
   2091}
   2092
   2093static u64 access_cce_cli1_async_fifo_rxdma_parity_err_cnt(
   2094				const struct cntr_entry *entry,
   2095				void *context, int vl, int mode, u64 data)
   2096{
   2097	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2098
   2099	return dd->cce_err_status_cnt[11];
   2100}
   2101
   2102static u64 access_cce_cli1_async_fifo_sdma_hd_parity_err_cnt(
   2103				const struct cntr_entry *entry,
   2104				void *context, int vl, int mode, u64 data)
   2105{
   2106	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2107
   2108	return dd->cce_err_status_cnt[10];
   2109}
   2110
   2111static u64 access_cce_cl1_async_fifo_pio_crdt_parity_err_cnt(
   2112				const struct cntr_entry *entry,
   2113				void *context, int vl, int mode, u64 data)
   2114{
   2115	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2116
   2117	return dd->cce_err_status_cnt[9];
   2118}
   2119
   2120static u64 access_cce_cli2_async_fifo_parity_err_cnt(
   2121				const struct cntr_entry *entry,
   2122				void *context, int vl, int mode, u64 data)
   2123{
   2124	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2125
   2126	return dd->cce_err_status_cnt[8];
   2127}
   2128
   2129static u64 access_cce_csr_cfg_bus_parity_err_cnt(const struct cntr_entry *entry,
   2130						 void *context, int vl,
   2131						 int mode, u64 data)
   2132{
   2133	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2134
   2135	return dd->cce_err_status_cnt[7];
   2136}
   2137
   2138static u64 access_cce_cli0_async_fifo_parity_err_cnt(
   2139				const struct cntr_entry *entry,
   2140				void *context, int vl, int mode, u64 data)
   2141{
   2142	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2143
   2144	return dd->cce_err_status_cnt[6];
   2145}
   2146
   2147static u64 access_cce_rspd_data_parity_err_cnt(const struct cntr_entry *entry,
   2148					       void *context, int vl, int mode,
   2149					       u64 data)
   2150{
   2151	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2152
   2153	return dd->cce_err_status_cnt[5];
   2154}
   2155
   2156static u64 access_cce_trgt_access_err_cnt(const struct cntr_entry *entry,
   2157					  void *context, int vl, int mode,
   2158					  u64 data)
   2159{
   2160	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2161
   2162	return dd->cce_err_status_cnt[4];
   2163}
   2164
   2165static u64 access_cce_trgt_async_fifo_parity_err_cnt(
   2166				const struct cntr_entry *entry,
   2167				void *context, int vl, int mode, u64 data)
   2168{
   2169	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2170
   2171	return dd->cce_err_status_cnt[3];
   2172}
   2173
   2174static u64 access_cce_csr_write_bad_addr_err_cnt(const struct cntr_entry *entry,
   2175						 void *context, int vl,
   2176						 int mode, u64 data)
   2177{
   2178	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2179
   2180	return dd->cce_err_status_cnt[2];
   2181}
   2182
   2183static u64 access_cce_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
   2184						void *context, int vl,
   2185						int mode, u64 data)
   2186{
   2187	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2188
   2189	return dd->cce_err_status_cnt[1];
   2190}
   2191
   2192static u64 access_ccs_csr_parity_err_cnt(const struct cntr_entry *entry,
   2193					 void *context, int vl, int mode,
   2194					 u64 data)
   2195{
   2196	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2197
   2198	return dd->cce_err_status_cnt[0];
   2199}
   2200
   2201/*
   2202 * Software counters corresponding to each of the
   2203 * error status bits within RcvErrStatus
   2204 */
   2205static u64 access_rx_csr_parity_err_cnt(const struct cntr_entry *entry,
   2206					void *context, int vl, int mode,
   2207					u64 data)
   2208{
   2209	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2210
   2211	return dd->rcv_err_status_cnt[63];
   2212}
   2213
   2214static u64 access_rx_csr_write_bad_addr_err_cnt(const struct cntr_entry *entry,
   2215						void *context, int vl,
   2216						int mode, u64 data)
   2217{
   2218	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2219
   2220	return dd->rcv_err_status_cnt[62];
   2221}
   2222
   2223static u64 access_rx_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
   2224					       void *context, int vl, int mode,
   2225					       u64 data)
   2226{
   2227	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2228
   2229	return dd->rcv_err_status_cnt[61];
   2230}
   2231
   2232static u64 access_rx_dma_csr_unc_err_cnt(const struct cntr_entry *entry,
   2233					 void *context, int vl, int mode,
   2234					 u64 data)
   2235{
   2236	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2237
   2238	return dd->rcv_err_status_cnt[60];
   2239}
   2240
   2241static u64 access_rx_dma_dq_fsm_encoding_err_cnt(const struct cntr_entry *entry,
   2242						 void *context, int vl,
   2243						 int mode, u64 data)
   2244{
   2245	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2246
   2247	return dd->rcv_err_status_cnt[59];
   2248}
   2249
   2250static u64 access_rx_dma_eq_fsm_encoding_err_cnt(const struct cntr_entry *entry,
   2251						 void *context, int vl,
   2252						 int mode, u64 data)
   2253{
   2254	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2255
   2256	return dd->rcv_err_status_cnt[58];
   2257}
   2258
   2259static u64 access_rx_dma_csr_parity_err_cnt(const struct cntr_entry *entry,
   2260					    void *context, int vl, int mode,
   2261					    u64 data)
   2262{
   2263	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2264
   2265	return dd->rcv_err_status_cnt[57];
   2266}
   2267
   2268static u64 access_rx_rbuf_data_cor_err_cnt(const struct cntr_entry *entry,
   2269					   void *context, int vl, int mode,
   2270					   u64 data)
   2271{
   2272	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2273
   2274	return dd->rcv_err_status_cnt[56];
   2275}
   2276
   2277static u64 access_rx_rbuf_data_unc_err_cnt(const struct cntr_entry *entry,
   2278					   void *context, int vl, int mode,
   2279					   u64 data)
   2280{
   2281	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2282
   2283	return dd->rcv_err_status_cnt[55];
   2284}
   2285
   2286static u64 access_rx_dma_data_fifo_rd_cor_err_cnt(
   2287				const struct cntr_entry *entry,
   2288				void *context, int vl, int mode, u64 data)
   2289{
   2290	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2291
   2292	return dd->rcv_err_status_cnt[54];
   2293}
   2294
   2295static u64 access_rx_dma_data_fifo_rd_unc_err_cnt(
   2296				const struct cntr_entry *entry,
   2297				void *context, int vl, int mode, u64 data)
   2298{
   2299	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2300
   2301	return dd->rcv_err_status_cnt[53];
   2302}
   2303
   2304static u64 access_rx_dma_hdr_fifo_rd_cor_err_cnt(const struct cntr_entry *entry,
   2305						 void *context, int vl,
   2306						 int mode, u64 data)
   2307{
   2308	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2309
   2310	return dd->rcv_err_status_cnt[52];
   2311}
   2312
   2313static u64 access_rx_dma_hdr_fifo_rd_unc_err_cnt(const struct cntr_entry *entry,
   2314						 void *context, int vl,
   2315						 int mode, u64 data)
   2316{
   2317	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2318
   2319	return dd->rcv_err_status_cnt[51];
   2320}
   2321
   2322static u64 access_rx_rbuf_desc_part2_cor_err_cnt(const struct cntr_entry *entry,
   2323						 void *context, int vl,
   2324						 int mode, u64 data)
   2325{
   2326	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2327
   2328	return dd->rcv_err_status_cnt[50];
   2329}
   2330
   2331static u64 access_rx_rbuf_desc_part2_unc_err_cnt(const struct cntr_entry *entry,
   2332						 void *context, int vl,
   2333						 int mode, u64 data)
   2334{
   2335	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2336
   2337	return dd->rcv_err_status_cnt[49];
   2338}
   2339
   2340static u64 access_rx_rbuf_desc_part1_cor_err_cnt(const struct cntr_entry *entry,
   2341						 void *context, int vl,
   2342						 int mode, u64 data)
   2343{
   2344	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2345
   2346	return dd->rcv_err_status_cnt[48];
   2347}
   2348
   2349static u64 access_rx_rbuf_desc_part1_unc_err_cnt(const struct cntr_entry *entry,
   2350						 void *context, int vl,
   2351						 int mode, u64 data)
   2352{
   2353	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2354
   2355	return dd->rcv_err_status_cnt[47];
   2356}
   2357
   2358static u64 access_rx_hq_intr_fsm_err_cnt(const struct cntr_entry *entry,
   2359					 void *context, int vl, int mode,
   2360					 u64 data)
   2361{
   2362	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2363
   2364	return dd->rcv_err_status_cnt[46];
   2365}
   2366
   2367static u64 access_rx_hq_intr_csr_parity_err_cnt(
   2368				const struct cntr_entry *entry,
   2369				void *context, int vl, int mode, u64 data)
   2370{
   2371	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2372
   2373	return dd->rcv_err_status_cnt[45];
   2374}
   2375
   2376static u64 access_rx_lookup_csr_parity_err_cnt(
   2377				const struct cntr_entry *entry,
   2378				void *context, int vl, int mode, u64 data)
   2379{
   2380	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2381
   2382	return dd->rcv_err_status_cnt[44];
   2383}
   2384
   2385static u64 access_rx_lookup_rcv_array_cor_err_cnt(
   2386				const struct cntr_entry *entry,
   2387				void *context, int vl, int mode, u64 data)
   2388{
   2389	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2390
   2391	return dd->rcv_err_status_cnt[43];
   2392}
   2393
   2394static u64 access_rx_lookup_rcv_array_unc_err_cnt(
   2395				const struct cntr_entry *entry,
   2396				void *context, int vl, int mode, u64 data)
   2397{
   2398	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2399
   2400	return dd->rcv_err_status_cnt[42];
   2401}
   2402
   2403static u64 access_rx_lookup_des_part2_parity_err_cnt(
   2404				const struct cntr_entry *entry,
   2405				void *context, int vl, int mode, u64 data)
   2406{
   2407	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2408
   2409	return dd->rcv_err_status_cnt[41];
   2410}
   2411
   2412static u64 access_rx_lookup_des_part1_unc_cor_err_cnt(
   2413				const struct cntr_entry *entry,
   2414				void *context, int vl, int mode, u64 data)
   2415{
   2416	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2417
   2418	return dd->rcv_err_status_cnt[40];
   2419}
   2420
   2421static u64 access_rx_lookup_des_part1_unc_err_cnt(
   2422				const struct cntr_entry *entry,
   2423				void *context, int vl, int mode, u64 data)
   2424{
   2425	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2426
   2427	return dd->rcv_err_status_cnt[39];
   2428}
   2429
   2430static u64 access_rx_rbuf_next_free_buf_cor_err_cnt(
   2431				const struct cntr_entry *entry,
   2432				void *context, int vl, int mode, u64 data)
   2433{
   2434	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2435
   2436	return dd->rcv_err_status_cnt[38];
   2437}
   2438
   2439static u64 access_rx_rbuf_next_free_buf_unc_err_cnt(
   2440				const struct cntr_entry *entry,
   2441				void *context, int vl, int mode, u64 data)
   2442{
   2443	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2444
   2445	return dd->rcv_err_status_cnt[37];
   2446}
   2447
   2448static u64 access_rbuf_fl_init_wr_addr_parity_err_cnt(
   2449				const struct cntr_entry *entry,
   2450				void *context, int vl, int mode, u64 data)
   2451{
   2452	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2453
   2454	return dd->rcv_err_status_cnt[36];
   2455}
   2456
   2457static u64 access_rx_rbuf_fl_initdone_parity_err_cnt(
   2458				const struct cntr_entry *entry,
   2459				void *context, int vl, int mode, u64 data)
   2460{
   2461	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2462
   2463	return dd->rcv_err_status_cnt[35];
   2464}
   2465
   2466static u64 access_rx_rbuf_fl_write_addr_parity_err_cnt(
   2467				const struct cntr_entry *entry,
   2468				void *context, int vl, int mode, u64 data)
   2469{
   2470	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2471
   2472	return dd->rcv_err_status_cnt[34];
   2473}
   2474
   2475static u64 access_rx_rbuf_fl_rd_addr_parity_err_cnt(
   2476				const struct cntr_entry *entry,
   2477				void *context, int vl, int mode, u64 data)
   2478{
   2479	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2480
   2481	return dd->rcv_err_status_cnt[33];
   2482}
   2483
   2484static u64 access_rx_rbuf_empty_err_cnt(const struct cntr_entry *entry,
   2485					void *context, int vl, int mode,
   2486					u64 data)
   2487{
   2488	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2489
   2490	return dd->rcv_err_status_cnt[32];
   2491}
   2492
   2493static u64 access_rx_rbuf_full_err_cnt(const struct cntr_entry *entry,
   2494				       void *context, int vl, int mode,
   2495				       u64 data)
   2496{
   2497	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2498
   2499	return dd->rcv_err_status_cnt[31];
   2500}
   2501
   2502static u64 access_rbuf_bad_lookup_err_cnt(const struct cntr_entry *entry,
   2503					  void *context, int vl, int mode,
   2504					  u64 data)
   2505{
   2506	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2507
   2508	return dd->rcv_err_status_cnt[30];
   2509}
   2510
   2511static u64 access_rbuf_ctx_id_parity_err_cnt(const struct cntr_entry *entry,
   2512					     void *context, int vl, int mode,
   2513					     u64 data)
   2514{
   2515	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2516
   2517	return dd->rcv_err_status_cnt[29];
   2518}
   2519
   2520static u64 access_rbuf_csr_qeopdw_parity_err_cnt(const struct cntr_entry *entry,
   2521						 void *context, int vl,
   2522						 int mode, u64 data)
   2523{
   2524	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2525
   2526	return dd->rcv_err_status_cnt[28];
   2527}
   2528
   2529static u64 access_rx_rbuf_csr_q_num_of_pkt_parity_err_cnt(
   2530				const struct cntr_entry *entry,
   2531				void *context, int vl, int mode, u64 data)
   2532{
   2533	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2534
   2535	return dd->rcv_err_status_cnt[27];
   2536}
   2537
   2538static u64 access_rx_rbuf_csr_q_t1_ptr_parity_err_cnt(
   2539				const struct cntr_entry *entry,
   2540				void *context, int vl, int mode, u64 data)
   2541{
   2542	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2543
   2544	return dd->rcv_err_status_cnt[26];
   2545}
   2546
   2547static u64 access_rx_rbuf_csr_q_hd_ptr_parity_err_cnt(
   2548				const struct cntr_entry *entry,
   2549				void *context, int vl, int mode, u64 data)
   2550{
   2551	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2552
   2553	return dd->rcv_err_status_cnt[25];
   2554}
   2555
   2556static u64 access_rx_rbuf_csr_q_vld_bit_parity_err_cnt(
   2557				const struct cntr_entry *entry,
   2558				void *context, int vl, int mode, u64 data)
   2559{
   2560	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2561
   2562	return dd->rcv_err_status_cnt[24];
   2563}
   2564
   2565static u64 access_rx_rbuf_csr_q_next_buf_parity_err_cnt(
   2566				const struct cntr_entry *entry,
   2567				void *context, int vl, int mode, u64 data)
   2568{
   2569	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2570
   2571	return dd->rcv_err_status_cnt[23];
   2572}
   2573
   2574static u64 access_rx_rbuf_csr_q_ent_cnt_parity_err_cnt(
   2575				const struct cntr_entry *entry,
   2576				void *context, int vl, int mode, u64 data)
   2577{
   2578	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2579
   2580	return dd->rcv_err_status_cnt[22];
   2581}
   2582
   2583static u64 access_rx_rbuf_csr_q_head_buf_num_parity_err_cnt(
   2584				const struct cntr_entry *entry,
   2585				void *context, int vl, int mode, u64 data)
   2586{
   2587	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2588
   2589	return dd->rcv_err_status_cnt[21];
   2590}
   2591
   2592static u64 access_rx_rbuf_block_list_read_cor_err_cnt(
   2593				const struct cntr_entry *entry,
   2594				void *context, int vl, int mode, u64 data)
   2595{
   2596	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2597
   2598	return dd->rcv_err_status_cnt[20];
   2599}
   2600
   2601static u64 access_rx_rbuf_block_list_read_unc_err_cnt(
   2602				const struct cntr_entry *entry,
   2603				void *context, int vl, int mode, u64 data)
   2604{
   2605	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2606
   2607	return dd->rcv_err_status_cnt[19];
   2608}
   2609
   2610static u64 access_rx_rbuf_lookup_des_cor_err_cnt(const struct cntr_entry *entry,
   2611						 void *context, int vl,
   2612						 int mode, u64 data)
   2613{
   2614	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2615
   2616	return dd->rcv_err_status_cnt[18];
   2617}
   2618
   2619static u64 access_rx_rbuf_lookup_des_unc_err_cnt(const struct cntr_entry *entry,
   2620						 void *context, int vl,
   2621						 int mode, u64 data)
   2622{
   2623	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2624
   2625	return dd->rcv_err_status_cnt[17];
   2626}
   2627
   2628static u64 access_rx_rbuf_lookup_des_reg_unc_cor_err_cnt(
   2629				const struct cntr_entry *entry,
   2630				void *context, int vl, int mode, u64 data)
   2631{
   2632	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2633
   2634	return dd->rcv_err_status_cnt[16];
   2635}
   2636
   2637static u64 access_rx_rbuf_lookup_des_reg_unc_err_cnt(
   2638				const struct cntr_entry *entry,
   2639				void *context, int vl, int mode, u64 data)
   2640{
   2641	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2642
   2643	return dd->rcv_err_status_cnt[15];
   2644}
   2645
   2646static u64 access_rx_rbuf_free_list_cor_err_cnt(const struct cntr_entry *entry,
   2647						void *context, int vl,
   2648						int mode, u64 data)
   2649{
   2650	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2651
   2652	return dd->rcv_err_status_cnt[14];
   2653}
   2654
   2655static u64 access_rx_rbuf_free_list_unc_err_cnt(const struct cntr_entry *entry,
   2656						void *context, int vl,
   2657						int mode, u64 data)
   2658{
   2659	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2660
   2661	return dd->rcv_err_status_cnt[13];
   2662}
   2663
   2664static u64 access_rx_rcv_fsm_encoding_err_cnt(const struct cntr_entry *entry,
   2665					      void *context, int vl, int mode,
   2666					      u64 data)
   2667{
   2668	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2669
   2670	return dd->rcv_err_status_cnt[12];
   2671}
   2672
   2673static u64 access_rx_dma_flag_cor_err_cnt(const struct cntr_entry *entry,
   2674					  void *context, int vl, int mode,
   2675					  u64 data)
   2676{
   2677	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2678
   2679	return dd->rcv_err_status_cnt[11];
   2680}
   2681
   2682static u64 access_rx_dma_flag_unc_err_cnt(const struct cntr_entry *entry,
   2683					  void *context, int vl, int mode,
   2684					  u64 data)
   2685{
   2686	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2687
   2688	return dd->rcv_err_status_cnt[10];
   2689}
   2690
   2691static u64 access_rx_dc_sop_eop_parity_err_cnt(const struct cntr_entry *entry,
   2692					       void *context, int vl, int mode,
   2693					       u64 data)
   2694{
   2695	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2696
   2697	return dd->rcv_err_status_cnt[9];
   2698}
   2699
   2700static u64 access_rx_rcv_csr_parity_err_cnt(const struct cntr_entry *entry,
   2701					    void *context, int vl, int mode,
   2702					    u64 data)
   2703{
   2704	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2705
   2706	return dd->rcv_err_status_cnt[8];
   2707}
   2708
   2709static u64 access_rx_rcv_qp_map_table_cor_err_cnt(
   2710				const struct cntr_entry *entry,
   2711				void *context, int vl, int mode, u64 data)
   2712{
   2713	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2714
   2715	return dd->rcv_err_status_cnt[7];
   2716}
   2717
   2718static u64 access_rx_rcv_qp_map_table_unc_err_cnt(
   2719				const struct cntr_entry *entry,
   2720				void *context, int vl, int mode, u64 data)
   2721{
   2722	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2723
   2724	return dd->rcv_err_status_cnt[6];
   2725}
   2726
   2727static u64 access_rx_rcv_data_cor_err_cnt(const struct cntr_entry *entry,
   2728					  void *context, int vl, int mode,
   2729					  u64 data)
   2730{
   2731	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2732
   2733	return dd->rcv_err_status_cnt[5];
   2734}
   2735
   2736static u64 access_rx_rcv_data_unc_err_cnt(const struct cntr_entry *entry,
   2737					  void *context, int vl, int mode,
   2738					  u64 data)
   2739{
   2740	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2741
   2742	return dd->rcv_err_status_cnt[4];
   2743}
   2744
   2745static u64 access_rx_rcv_hdr_cor_err_cnt(const struct cntr_entry *entry,
   2746					 void *context, int vl, int mode,
   2747					 u64 data)
   2748{
   2749	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2750
   2751	return dd->rcv_err_status_cnt[3];
   2752}
   2753
   2754static u64 access_rx_rcv_hdr_unc_err_cnt(const struct cntr_entry *entry,
   2755					 void *context, int vl, int mode,
   2756					 u64 data)
   2757{
   2758	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2759
   2760	return dd->rcv_err_status_cnt[2];
   2761}
   2762
   2763static u64 access_rx_dc_intf_parity_err_cnt(const struct cntr_entry *entry,
   2764					    void *context, int vl, int mode,
   2765					    u64 data)
   2766{
   2767	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2768
   2769	return dd->rcv_err_status_cnt[1];
   2770}
   2771
   2772static u64 access_rx_dma_csr_cor_err_cnt(const struct cntr_entry *entry,
   2773					 void *context, int vl, int mode,
   2774					 u64 data)
   2775{
   2776	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2777
   2778	return dd->rcv_err_status_cnt[0];
   2779}
   2780
   2781/*
   2782 * Software counters corresponding to each of the
   2783 * error status bits within SendPioErrStatus
   2784 */
   2785static u64 access_pio_pec_sop_head_parity_err_cnt(
   2786				const struct cntr_entry *entry,
   2787				void *context, int vl, int mode, u64 data)
   2788{
   2789	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2790
   2791	return dd->send_pio_err_status_cnt[35];
   2792}
   2793
   2794static u64 access_pio_pcc_sop_head_parity_err_cnt(
   2795				const struct cntr_entry *entry,
   2796				void *context, int vl, int mode, u64 data)
   2797{
   2798	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2799
   2800	return dd->send_pio_err_status_cnt[34];
   2801}
   2802
   2803static u64 access_pio_last_returned_cnt_parity_err_cnt(
   2804				const struct cntr_entry *entry,
   2805				void *context, int vl, int mode, u64 data)
   2806{
   2807	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2808
   2809	return dd->send_pio_err_status_cnt[33];
   2810}
   2811
   2812static u64 access_pio_current_free_cnt_parity_err_cnt(
   2813				const struct cntr_entry *entry,
   2814				void *context, int vl, int mode, u64 data)
   2815{
   2816	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2817
   2818	return dd->send_pio_err_status_cnt[32];
   2819}
   2820
   2821static u64 access_pio_reserved_31_err_cnt(const struct cntr_entry *entry,
   2822					  void *context, int vl, int mode,
   2823					  u64 data)
   2824{
   2825	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2826
   2827	return dd->send_pio_err_status_cnt[31];
   2828}
   2829
   2830static u64 access_pio_reserved_30_err_cnt(const struct cntr_entry *entry,
   2831					  void *context, int vl, int mode,
   2832					  u64 data)
   2833{
   2834	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2835
   2836	return dd->send_pio_err_status_cnt[30];
   2837}
   2838
   2839static u64 access_pio_ppmc_sop_len_err_cnt(const struct cntr_entry *entry,
   2840					   void *context, int vl, int mode,
   2841					   u64 data)
   2842{
   2843	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2844
   2845	return dd->send_pio_err_status_cnt[29];
   2846}
   2847
   2848static u64 access_pio_ppmc_bqc_mem_parity_err_cnt(
   2849				const struct cntr_entry *entry,
   2850				void *context, int vl, int mode, u64 data)
   2851{
   2852	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2853
   2854	return dd->send_pio_err_status_cnt[28];
   2855}
   2856
   2857static u64 access_pio_vl_fifo_parity_err_cnt(const struct cntr_entry *entry,
   2858					     void *context, int vl, int mode,
   2859					     u64 data)
   2860{
   2861	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2862
   2863	return dd->send_pio_err_status_cnt[27];
   2864}
   2865
   2866static u64 access_pio_vlf_sop_parity_err_cnt(const struct cntr_entry *entry,
   2867					     void *context, int vl, int mode,
   2868					     u64 data)
   2869{
   2870	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2871
   2872	return dd->send_pio_err_status_cnt[26];
   2873}
   2874
   2875static u64 access_pio_vlf_v1_len_parity_err_cnt(const struct cntr_entry *entry,
   2876						void *context, int vl,
   2877						int mode, u64 data)
   2878{
   2879	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2880
   2881	return dd->send_pio_err_status_cnt[25];
   2882}
   2883
   2884static u64 access_pio_block_qw_count_parity_err_cnt(
   2885				const struct cntr_entry *entry,
   2886				void *context, int vl, int mode, u64 data)
   2887{
   2888	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2889
   2890	return dd->send_pio_err_status_cnt[24];
   2891}
   2892
   2893static u64 access_pio_write_qw_valid_parity_err_cnt(
   2894				const struct cntr_entry *entry,
   2895				void *context, int vl, int mode, u64 data)
   2896{
   2897	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2898
   2899	return dd->send_pio_err_status_cnt[23];
   2900}
   2901
   2902static u64 access_pio_state_machine_err_cnt(const struct cntr_entry *entry,
   2903					    void *context, int vl, int mode,
   2904					    u64 data)
   2905{
   2906	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2907
   2908	return dd->send_pio_err_status_cnt[22];
   2909}
   2910
   2911static u64 access_pio_write_data_parity_err_cnt(const struct cntr_entry *entry,
   2912						void *context, int vl,
   2913						int mode, u64 data)
   2914{
   2915	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2916
   2917	return dd->send_pio_err_status_cnt[21];
   2918}
   2919
   2920static u64 access_pio_host_addr_mem_cor_err_cnt(const struct cntr_entry *entry,
   2921						void *context, int vl,
   2922						int mode, u64 data)
   2923{
   2924	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2925
   2926	return dd->send_pio_err_status_cnt[20];
   2927}
   2928
   2929static u64 access_pio_host_addr_mem_unc_err_cnt(const struct cntr_entry *entry,
   2930						void *context, int vl,
   2931						int mode, u64 data)
   2932{
   2933	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2934
   2935	return dd->send_pio_err_status_cnt[19];
   2936}
   2937
   2938static u64 access_pio_pkt_evict_sm_or_arb_sm_err_cnt(
   2939				const struct cntr_entry *entry,
   2940				void *context, int vl, int mode, u64 data)
   2941{
   2942	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2943
   2944	return dd->send_pio_err_status_cnt[18];
   2945}
   2946
   2947static u64 access_pio_init_sm_in_err_cnt(const struct cntr_entry *entry,
   2948					 void *context, int vl, int mode,
   2949					 u64 data)
   2950{
   2951	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2952
   2953	return dd->send_pio_err_status_cnt[17];
   2954}
   2955
   2956static u64 access_pio_ppmc_pbl_fifo_err_cnt(const struct cntr_entry *entry,
   2957					    void *context, int vl, int mode,
   2958					    u64 data)
   2959{
   2960	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2961
   2962	return dd->send_pio_err_status_cnt[16];
   2963}
   2964
   2965static u64 access_pio_credit_ret_fifo_parity_err_cnt(
   2966				const struct cntr_entry *entry,
   2967				void *context, int vl, int mode, u64 data)
   2968{
   2969	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2970
   2971	return dd->send_pio_err_status_cnt[15];
   2972}
   2973
   2974static u64 access_pio_v1_len_mem_bank1_cor_err_cnt(
   2975				const struct cntr_entry *entry,
   2976				void *context, int vl, int mode, u64 data)
   2977{
   2978	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2979
   2980	return dd->send_pio_err_status_cnt[14];
   2981}
   2982
   2983static u64 access_pio_v1_len_mem_bank0_cor_err_cnt(
   2984				const struct cntr_entry *entry,
   2985				void *context, int vl, int mode, u64 data)
   2986{
   2987	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2988
   2989	return dd->send_pio_err_status_cnt[13];
   2990}
   2991
   2992static u64 access_pio_v1_len_mem_bank1_unc_err_cnt(
   2993				const struct cntr_entry *entry,
   2994				void *context, int vl, int mode, u64 data)
   2995{
   2996	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   2997
   2998	return dd->send_pio_err_status_cnt[12];
   2999}
   3000
   3001static u64 access_pio_v1_len_mem_bank0_unc_err_cnt(
   3002				const struct cntr_entry *entry,
   3003				void *context, int vl, int mode, u64 data)
   3004{
   3005	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3006
   3007	return dd->send_pio_err_status_cnt[11];
   3008}
   3009
   3010static u64 access_pio_sm_pkt_reset_parity_err_cnt(
   3011				const struct cntr_entry *entry,
   3012				void *context, int vl, int mode, u64 data)
   3013{
   3014	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3015
   3016	return dd->send_pio_err_status_cnt[10];
   3017}
   3018
   3019static u64 access_pio_pkt_evict_fifo_parity_err_cnt(
   3020				const struct cntr_entry *entry,
   3021				void *context, int vl, int mode, u64 data)
   3022{
   3023	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3024
   3025	return dd->send_pio_err_status_cnt[9];
   3026}
   3027
   3028static u64 access_pio_sbrdctrl_crrel_fifo_parity_err_cnt(
   3029				const struct cntr_entry *entry,
   3030				void *context, int vl, int mode, u64 data)
   3031{
   3032	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3033
   3034	return dd->send_pio_err_status_cnt[8];
   3035}
   3036
   3037static u64 access_pio_sbrdctl_crrel_parity_err_cnt(
   3038				const struct cntr_entry *entry,
   3039				void *context, int vl, int mode, u64 data)
   3040{
   3041	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3042
   3043	return dd->send_pio_err_status_cnt[7];
   3044}
   3045
   3046static u64 access_pio_pec_fifo_parity_err_cnt(const struct cntr_entry *entry,
   3047					      void *context, int vl, int mode,
   3048					      u64 data)
   3049{
   3050	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3051
   3052	return dd->send_pio_err_status_cnt[6];
   3053}
   3054
   3055static u64 access_pio_pcc_fifo_parity_err_cnt(const struct cntr_entry *entry,
   3056					      void *context, int vl, int mode,
   3057					      u64 data)
   3058{
   3059	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3060
   3061	return dd->send_pio_err_status_cnt[5];
   3062}
   3063
   3064static u64 access_pio_sb_mem_fifo1_err_cnt(const struct cntr_entry *entry,
   3065					   void *context, int vl, int mode,
   3066					   u64 data)
   3067{
   3068	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3069
   3070	return dd->send_pio_err_status_cnt[4];
   3071}
   3072
   3073static u64 access_pio_sb_mem_fifo0_err_cnt(const struct cntr_entry *entry,
   3074					   void *context, int vl, int mode,
   3075					   u64 data)
   3076{
   3077	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3078
   3079	return dd->send_pio_err_status_cnt[3];
   3080}
   3081
   3082static u64 access_pio_csr_parity_err_cnt(const struct cntr_entry *entry,
   3083					 void *context, int vl, int mode,
   3084					 u64 data)
   3085{
   3086	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3087
   3088	return dd->send_pio_err_status_cnt[2];
   3089}
   3090
   3091static u64 access_pio_write_addr_parity_err_cnt(const struct cntr_entry *entry,
   3092						void *context, int vl,
   3093						int mode, u64 data)
   3094{
   3095	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3096
   3097	return dd->send_pio_err_status_cnt[1];
   3098}
   3099
   3100static u64 access_pio_write_bad_ctxt_err_cnt(const struct cntr_entry *entry,
   3101					     void *context, int vl, int mode,
   3102					     u64 data)
   3103{
   3104	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3105
   3106	return dd->send_pio_err_status_cnt[0];
   3107}
   3108
   3109/*
   3110 * Software counters corresponding to each of the
   3111 * error status bits within SendDmaErrStatus
   3112 */
   3113static u64 access_sdma_pcie_req_tracking_cor_err_cnt(
   3114				const struct cntr_entry *entry,
   3115				void *context, int vl, int mode, u64 data)
   3116{
   3117	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3118
   3119	return dd->send_dma_err_status_cnt[3];
   3120}
   3121
   3122static u64 access_sdma_pcie_req_tracking_unc_err_cnt(
   3123				const struct cntr_entry *entry,
   3124				void *context, int vl, int mode, u64 data)
   3125{
   3126	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3127
   3128	return dd->send_dma_err_status_cnt[2];
   3129}
   3130
   3131static u64 access_sdma_csr_parity_err_cnt(const struct cntr_entry *entry,
   3132					  void *context, int vl, int mode,
   3133					  u64 data)
   3134{
   3135	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3136
   3137	return dd->send_dma_err_status_cnt[1];
   3138}
   3139
   3140static u64 access_sdma_rpy_tag_err_cnt(const struct cntr_entry *entry,
   3141				       void *context, int vl, int mode,
   3142				       u64 data)
   3143{
   3144	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3145
   3146	return dd->send_dma_err_status_cnt[0];
   3147}
   3148
   3149/*
   3150 * Software counters corresponding to each of the
   3151 * error status bits within SendEgressErrStatus
   3152 */
   3153static u64 access_tx_read_pio_memory_csr_unc_err_cnt(
   3154				const struct cntr_entry *entry,
   3155				void *context, int vl, int mode, u64 data)
   3156{
   3157	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3158
   3159	return dd->send_egress_err_status_cnt[63];
   3160}
   3161
   3162static u64 access_tx_read_sdma_memory_csr_err_cnt(
   3163				const struct cntr_entry *entry,
   3164				void *context, int vl, int mode, u64 data)
   3165{
   3166	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3167
   3168	return dd->send_egress_err_status_cnt[62];
   3169}
   3170
   3171static u64 access_tx_egress_fifo_cor_err_cnt(const struct cntr_entry *entry,
   3172					     void *context, int vl, int mode,
   3173					     u64 data)
   3174{
   3175	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3176
   3177	return dd->send_egress_err_status_cnt[61];
   3178}
   3179
   3180static u64 access_tx_read_pio_memory_cor_err_cnt(const struct cntr_entry *entry,
   3181						 void *context, int vl,
   3182						 int mode, u64 data)
   3183{
   3184	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3185
   3186	return dd->send_egress_err_status_cnt[60];
   3187}
   3188
   3189static u64 access_tx_read_sdma_memory_cor_err_cnt(
   3190				const struct cntr_entry *entry,
   3191				void *context, int vl, int mode, u64 data)
   3192{
   3193	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3194
   3195	return dd->send_egress_err_status_cnt[59];
   3196}
   3197
   3198static u64 access_tx_sb_hdr_cor_err_cnt(const struct cntr_entry *entry,
   3199					void *context, int vl, int mode,
   3200					u64 data)
   3201{
   3202	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3203
   3204	return dd->send_egress_err_status_cnt[58];
   3205}
   3206
   3207static u64 access_tx_credit_overrun_err_cnt(const struct cntr_entry *entry,
   3208					    void *context, int vl, int mode,
   3209					    u64 data)
   3210{
   3211	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3212
   3213	return dd->send_egress_err_status_cnt[57];
   3214}
   3215
   3216static u64 access_tx_launch_fifo8_cor_err_cnt(const struct cntr_entry *entry,
   3217					      void *context, int vl, int mode,
   3218					      u64 data)
   3219{
   3220	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3221
   3222	return dd->send_egress_err_status_cnt[56];
   3223}
   3224
   3225static u64 access_tx_launch_fifo7_cor_err_cnt(const struct cntr_entry *entry,
   3226					      void *context, int vl, int mode,
   3227					      u64 data)
   3228{
   3229	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3230
   3231	return dd->send_egress_err_status_cnt[55];
   3232}
   3233
   3234static u64 access_tx_launch_fifo6_cor_err_cnt(const struct cntr_entry *entry,
   3235					      void *context, int vl, int mode,
   3236					      u64 data)
   3237{
   3238	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3239
   3240	return dd->send_egress_err_status_cnt[54];
   3241}
   3242
   3243static u64 access_tx_launch_fifo5_cor_err_cnt(const struct cntr_entry *entry,
   3244					      void *context, int vl, int mode,
   3245					      u64 data)
   3246{
   3247	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3248
   3249	return dd->send_egress_err_status_cnt[53];
   3250}
   3251
   3252static u64 access_tx_launch_fifo4_cor_err_cnt(const struct cntr_entry *entry,
   3253					      void *context, int vl, int mode,
   3254					      u64 data)
   3255{
   3256	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3257
   3258	return dd->send_egress_err_status_cnt[52];
   3259}
   3260
   3261static u64 access_tx_launch_fifo3_cor_err_cnt(const struct cntr_entry *entry,
   3262					      void *context, int vl, int mode,
   3263					      u64 data)
   3264{
   3265	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3266
   3267	return dd->send_egress_err_status_cnt[51];
   3268}
   3269
   3270static u64 access_tx_launch_fifo2_cor_err_cnt(const struct cntr_entry *entry,
   3271					      void *context, int vl, int mode,
   3272					      u64 data)
   3273{
   3274	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3275
   3276	return dd->send_egress_err_status_cnt[50];
   3277}
   3278
   3279static u64 access_tx_launch_fifo1_cor_err_cnt(const struct cntr_entry *entry,
   3280					      void *context, int vl, int mode,
   3281					      u64 data)
   3282{
   3283	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3284
   3285	return dd->send_egress_err_status_cnt[49];
   3286}
   3287
   3288static u64 access_tx_launch_fifo0_cor_err_cnt(const struct cntr_entry *entry,
   3289					      void *context, int vl, int mode,
   3290					      u64 data)
   3291{
   3292	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3293
   3294	return dd->send_egress_err_status_cnt[48];
   3295}
   3296
   3297static u64 access_tx_credit_return_vl_err_cnt(const struct cntr_entry *entry,
   3298					      void *context, int vl, int mode,
   3299					      u64 data)
   3300{
   3301	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3302
   3303	return dd->send_egress_err_status_cnt[47];
   3304}
   3305
   3306static u64 access_tx_hcrc_insertion_err_cnt(const struct cntr_entry *entry,
   3307					    void *context, int vl, int mode,
   3308					    u64 data)
   3309{
   3310	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3311
   3312	return dd->send_egress_err_status_cnt[46];
   3313}
   3314
   3315static u64 access_tx_egress_fifo_unc_err_cnt(const struct cntr_entry *entry,
   3316					     void *context, int vl, int mode,
   3317					     u64 data)
   3318{
   3319	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3320
   3321	return dd->send_egress_err_status_cnt[45];
   3322}
   3323
   3324static u64 access_tx_read_pio_memory_unc_err_cnt(const struct cntr_entry *entry,
   3325						 void *context, int vl,
   3326						 int mode, u64 data)
   3327{
   3328	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3329
   3330	return dd->send_egress_err_status_cnt[44];
   3331}
   3332
   3333static u64 access_tx_read_sdma_memory_unc_err_cnt(
   3334				const struct cntr_entry *entry,
   3335				void *context, int vl, int mode, u64 data)
   3336{
   3337	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3338
   3339	return dd->send_egress_err_status_cnt[43];
   3340}
   3341
   3342static u64 access_tx_sb_hdr_unc_err_cnt(const struct cntr_entry *entry,
   3343					void *context, int vl, int mode,
   3344					u64 data)
   3345{
   3346	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3347
   3348	return dd->send_egress_err_status_cnt[42];
   3349}
   3350
   3351static u64 access_tx_credit_return_partiy_err_cnt(
   3352				const struct cntr_entry *entry,
   3353				void *context, int vl, int mode, u64 data)
   3354{
   3355	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3356
   3357	return dd->send_egress_err_status_cnt[41];
   3358}
   3359
   3360static u64 access_tx_launch_fifo8_unc_or_parity_err_cnt(
   3361				const struct cntr_entry *entry,
   3362				void *context, int vl, int mode, u64 data)
   3363{
   3364	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3365
   3366	return dd->send_egress_err_status_cnt[40];
   3367}
   3368
   3369static u64 access_tx_launch_fifo7_unc_or_parity_err_cnt(
   3370				const struct cntr_entry *entry,
   3371				void *context, int vl, int mode, u64 data)
   3372{
   3373	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3374
   3375	return dd->send_egress_err_status_cnt[39];
   3376}
   3377
   3378static u64 access_tx_launch_fifo6_unc_or_parity_err_cnt(
   3379				const struct cntr_entry *entry,
   3380				void *context, int vl, int mode, u64 data)
   3381{
   3382	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3383
   3384	return dd->send_egress_err_status_cnt[38];
   3385}
   3386
   3387static u64 access_tx_launch_fifo5_unc_or_parity_err_cnt(
   3388				const struct cntr_entry *entry,
   3389				void *context, int vl, int mode, u64 data)
   3390{
   3391	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3392
   3393	return dd->send_egress_err_status_cnt[37];
   3394}
   3395
   3396static u64 access_tx_launch_fifo4_unc_or_parity_err_cnt(
   3397				const struct cntr_entry *entry,
   3398				void *context, int vl, int mode, u64 data)
   3399{
   3400	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3401
   3402	return dd->send_egress_err_status_cnt[36];
   3403}
   3404
   3405static u64 access_tx_launch_fifo3_unc_or_parity_err_cnt(
   3406				const struct cntr_entry *entry,
   3407				void *context, int vl, int mode, u64 data)
   3408{
   3409	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3410
   3411	return dd->send_egress_err_status_cnt[35];
   3412}
   3413
   3414static u64 access_tx_launch_fifo2_unc_or_parity_err_cnt(
   3415				const struct cntr_entry *entry,
   3416				void *context, int vl, int mode, u64 data)
   3417{
   3418	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3419
   3420	return dd->send_egress_err_status_cnt[34];
   3421}
   3422
   3423static u64 access_tx_launch_fifo1_unc_or_parity_err_cnt(
   3424				const struct cntr_entry *entry,
   3425				void *context, int vl, int mode, u64 data)
   3426{
   3427	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3428
   3429	return dd->send_egress_err_status_cnt[33];
   3430}
   3431
   3432static u64 access_tx_launch_fifo0_unc_or_parity_err_cnt(
   3433				const struct cntr_entry *entry,
   3434				void *context, int vl, int mode, u64 data)
   3435{
   3436	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3437
   3438	return dd->send_egress_err_status_cnt[32];
   3439}
   3440
   3441static u64 access_tx_sdma15_disallowed_packet_err_cnt(
   3442				const struct cntr_entry *entry,
   3443				void *context, int vl, int mode, u64 data)
   3444{
   3445	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3446
   3447	return dd->send_egress_err_status_cnt[31];
   3448}
   3449
   3450static u64 access_tx_sdma14_disallowed_packet_err_cnt(
   3451				const struct cntr_entry *entry,
   3452				void *context, int vl, int mode, u64 data)
   3453{
   3454	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3455
   3456	return dd->send_egress_err_status_cnt[30];
   3457}
   3458
   3459static u64 access_tx_sdma13_disallowed_packet_err_cnt(
   3460				const struct cntr_entry *entry,
   3461				void *context, int vl, int mode, u64 data)
   3462{
   3463	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3464
   3465	return dd->send_egress_err_status_cnt[29];
   3466}
   3467
   3468static u64 access_tx_sdma12_disallowed_packet_err_cnt(
   3469				const struct cntr_entry *entry,
   3470				void *context, int vl, int mode, u64 data)
   3471{
   3472	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3473
   3474	return dd->send_egress_err_status_cnt[28];
   3475}
   3476
   3477static u64 access_tx_sdma11_disallowed_packet_err_cnt(
   3478				const struct cntr_entry *entry,
   3479				void *context, int vl, int mode, u64 data)
   3480{
   3481	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3482
   3483	return dd->send_egress_err_status_cnt[27];
   3484}
   3485
   3486static u64 access_tx_sdma10_disallowed_packet_err_cnt(
   3487				const struct cntr_entry *entry,
   3488				void *context, int vl, int mode, u64 data)
   3489{
   3490	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3491
   3492	return dd->send_egress_err_status_cnt[26];
   3493}
   3494
   3495static u64 access_tx_sdma9_disallowed_packet_err_cnt(
   3496				const struct cntr_entry *entry,
   3497				void *context, int vl, int mode, u64 data)
   3498{
   3499	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3500
   3501	return dd->send_egress_err_status_cnt[25];
   3502}
   3503
   3504static u64 access_tx_sdma8_disallowed_packet_err_cnt(
   3505				const struct cntr_entry *entry,
   3506				void *context, int vl, int mode, u64 data)
   3507{
   3508	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3509
   3510	return dd->send_egress_err_status_cnt[24];
   3511}
   3512
   3513static u64 access_tx_sdma7_disallowed_packet_err_cnt(
   3514				const struct cntr_entry *entry,
   3515				void *context, int vl, int mode, u64 data)
   3516{
   3517	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3518
   3519	return dd->send_egress_err_status_cnt[23];
   3520}
   3521
   3522static u64 access_tx_sdma6_disallowed_packet_err_cnt(
   3523				const struct cntr_entry *entry,
   3524				void *context, int vl, int mode, u64 data)
   3525{
   3526	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3527
   3528	return dd->send_egress_err_status_cnt[22];
   3529}
   3530
   3531static u64 access_tx_sdma5_disallowed_packet_err_cnt(
   3532				const struct cntr_entry *entry,
   3533				void *context, int vl, int mode, u64 data)
   3534{
   3535	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3536
   3537	return dd->send_egress_err_status_cnt[21];
   3538}
   3539
   3540static u64 access_tx_sdma4_disallowed_packet_err_cnt(
   3541				const struct cntr_entry *entry,
   3542				void *context, int vl, int mode, u64 data)
   3543{
   3544	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3545
   3546	return dd->send_egress_err_status_cnt[20];
   3547}
   3548
   3549static u64 access_tx_sdma3_disallowed_packet_err_cnt(
   3550				const struct cntr_entry *entry,
   3551				void *context, int vl, int mode, u64 data)
   3552{
   3553	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3554
   3555	return dd->send_egress_err_status_cnt[19];
   3556}
   3557
   3558static u64 access_tx_sdma2_disallowed_packet_err_cnt(
   3559				const struct cntr_entry *entry,
   3560				void *context, int vl, int mode, u64 data)
   3561{
   3562	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3563
   3564	return dd->send_egress_err_status_cnt[18];
   3565}
   3566
   3567static u64 access_tx_sdma1_disallowed_packet_err_cnt(
   3568				const struct cntr_entry *entry,
   3569				void *context, int vl, int mode, u64 data)
   3570{
   3571	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3572
   3573	return dd->send_egress_err_status_cnt[17];
   3574}
   3575
   3576static u64 access_tx_sdma0_disallowed_packet_err_cnt(
   3577				const struct cntr_entry *entry,
   3578				void *context, int vl, int mode, u64 data)
   3579{
   3580	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3581
   3582	return dd->send_egress_err_status_cnt[16];
   3583}
   3584
   3585static u64 access_tx_config_parity_err_cnt(const struct cntr_entry *entry,
   3586					   void *context, int vl, int mode,
   3587					   u64 data)
   3588{
   3589	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3590
   3591	return dd->send_egress_err_status_cnt[15];
   3592}
   3593
   3594static u64 access_tx_sbrd_ctl_csr_parity_err_cnt(const struct cntr_entry *entry,
   3595						 void *context, int vl,
   3596						 int mode, u64 data)
   3597{
   3598	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3599
   3600	return dd->send_egress_err_status_cnt[14];
   3601}
   3602
   3603static u64 access_tx_launch_csr_parity_err_cnt(const struct cntr_entry *entry,
   3604					       void *context, int vl, int mode,
   3605					       u64 data)
   3606{
   3607	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3608
   3609	return dd->send_egress_err_status_cnt[13];
   3610}
   3611
   3612static u64 access_tx_illegal_vl_err_cnt(const struct cntr_entry *entry,
   3613					void *context, int vl, int mode,
   3614					u64 data)
   3615{
   3616	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3617
   3618	return dd->send_egress_err_status_cnt[12];
   3619}
   3620
   3621static u64 access_tx_sbrd_ctl_state_machine_parity_err_cnt(
   3622				const struct cntr_entry *entry,
   3623				void *context, int vl, int mode, u64 data)
   3624{
   3625	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3626
   3627	return dd->send_egress_err_status_cnt[11];
   3628}
   3629
   3630static u64 access_egress_reserved_10_err_cnt(const struct cntr_entry *entry,
   3631					     void *context, int vl, int mode,
   3632					     u64 data)
   3633{
   3634	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3635
   3636	return dd->send_egress_err_status_cnt[10];
   3637}
   3638
   3639static u64 access_egress_reserved_9_err_cnt(const struct cntr_entry *entry,
   3640					    void *context, int vl, int mode,
   3641					    u64 data)
   3642{
   3643	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3644
   3645	return dd->send_egress_err_status_cnt[9];
   3646}
   3647
   3648static u64 access_tx_sdma_launch_intf_parity_err_cnt(
   3649				const struct cntr_entry *entry,
   3650				void *context, int vl, int mode, u64 data)
   3651{
   3652	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3653
   3654	return dd->send_egress_err_status_cnt[8];
   3655}
   3656
   3657static u64 access_tx_pio_launch_intf_parity_err_cnt(
   3658				const struct cntr_entry *entry,
   3659				void *context, int vl, int mode, u64 data)
   3660{
   3661	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3662
   3663	return dd->send_egress_err_status_cnt[7];
   3664}
   3665
   3666static u64 access_egress_reserved_6_err_cnt(const struct cntr_entry *entry,
   3667					    void *context, int vl, int mode,
   3668					    u64 data)
   3669{
   3670	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3671
   3672	return dd->send_egress_err_status_cnt[6];
   3673}
   3674
   3675static u64 access_tx_incorrect_link_state_err_cnt(
   3676				const struct cntr_entry *entry,
   3677				void *context, int vl, int mode, u64 data)
   3678{
   3679	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3680
   3681	return dd->send_egress_err_status_cnt[5];
   3682}
   3683
   3684static u64 access_tx_linkdown_err_cnt(const struct cntr_entry *entry,
   3685				      void *context, int vl, int mode,
   3686				      u64 data)
   3687{
   3688	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3689
   3690	return dd->send_egress_err_status_cnt[4];
   3691}
   3692
   3693static u64 access_tx_egress_fifi_underrun_or_parity_err_cnt(
   3694				const struct cntr_entry *entry,
   3695				void *context, int vl, int mode, u64 data)
   3696{
   3697	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3698
   3699	return dd->send_egress_err_status_cnt[3];
   3700}
   3701
   3702static u64 access_egress_reserved_2_err_cnt(const struct cntr_entry *entry,
   3703					    void *context, int vl, int mode,
   3704					    u64 data)
   3705{
   3706	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3707
   3708	return dd->send_egress_err_status_cnt[2];
   3709}
   3710
   3711static u64 access_tx_pkt_integrity_mem_unc_err_cnt(
   3712				const struct cntr_entry *entry,
   3713				void *context, int vl, int mode, u64 data)
   3714{
   3715	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3716
   3717	return dd->send_egress_err_status_cnt[1];
   3718}
   3719
   3720static u64 access_tx_pkt_integrity_mem_cor_err_cnt(
   3721				const struct cntr_entry *entry,
   3722				void *context, int vl, int mode, u64 data)
   3723{
   3724	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3725
   3726	return dd->send_egress_err_status_cnt[0];
   3727}
   3728
   3729/*
   3730 * Software counters corresponding to each of the
   3731 * error status bits within SendErrStatus
   3732 */
   3733static u64 access_send_csr_write_bad_addr_err_cnt(
   3734				const struct cntr_entry *entry,
   3735				void *context, int vl, int mode, u64 data)
   3736{
   3737	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3738
   3739	return dd->send_err_status_cnt[2];
   3740}
   3741
   3742static u64 access_send_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry,
   3743						 void *context, int vl,
   3744						 int mode, u64 data)
   3745{
   3746	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3747
   3748	return dd->send_err_status_cnt[1];
   3749}
   3750
   3751static u64 access_send_csr_parity_cnt(const struct cntr_entry *entry,
   3752				      void *context, int vl, int mode,
   3753				      u64 data)
   3754{
   3755	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3756
   3757	return dd->send_err_status_cnt[0];
   3758}
   3759
   3760/*
   3761 * Software counters corresponding to each of the
   3762 * error status bits within SendCtxtErrStatus
   3763 */
   3764static u64 access_pio_write_out_of_bounds_err_cnt(
   3765				const struct cntr_entry *entry,
   3766				void *context, int vl, int mode, u64 data)
   3767{
   3768	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3769
   3770	return dd->sw_ctxt_err_status_cnt[4];
   3771}
   3772
   3773static u64 access_pio_write_overflow_err_cnt(const struct cntr_entry *entry,
   3774					     void *context, int vl, int mode,
   3775					     u64 data)
   3776{
   3777	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3778
   3779	return dd->sw_ctxt_err_status_cnt[3];
   3780}
   3781
   3782static u64 access_pio_write_crosses_boundary_err_cnt(
   3783				const struct cntr_entry *entry,
   3784				void *context, int vl, int mode, u64 data)
   3785{
   3786	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3787
   3788	return dd->sw_ctxt_err_status_cnt[2];
   3789}
   3790
   3791static u64 access_pio_disallowed_packet_err_cnt(const struct cntr_entry *entry,
   3792						void *context, int vl,
   3793						int mode, u64 data)
   3794{
   3795	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3796
   3797	return dd->sw_ctxt_err_status_cnt[1];
   3798}
   3799
   3800static u64 access_pio_inconsistent_sop_err_cnt(const struct cntr_entry *entry,
   3801					       void *context, int vl, int mode,
   3802					       u64 data)
   3803{
   3804	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3805
   3806	return dd->sw_ctxt_err_status_cnt[0];
   3807}
   3808
   3809/*
   3810 * Software counters corresponding to each of the
   3811 * error status bits within SendDmaEngErrStatus
   3812 */
   3813static u64 access_sdma_header_request_fifo_cor_err_cnt(
   3814				const struct cntr_entry *entry,
   3815				void *context, int vl, int mode, u64 data)
   3816{
   3817	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3818
   3819	return dd->sw_send_dma_eng_err_status_cnt[23];
   3820}
   3821
   3822static u64 access_sdma_header_storage_cor_err_cnt(
   3823				const struct cntr_entry *entry,
   3824				void *context, int vl, int mode, u64 data)
   3825{
   3826	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3827
   3828	return dd->sw_send_dma_eng_err_status_cnt[22];
   3829}
   3830
   3831static u64 access_sdma_packet_tracking_cor_err_cnt(
   3832				const struct cntr_entry *entry,
   3833				void *context, int vl, int mode, u64 data)
   3834{
   3835	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3836
   3837	return dd->sw_send_dma_eng_err_status_cnt[21];
   3838}
   3839
   3840static u64 access_sdma_assembly_cor_err_cnt(const struct cntr_entry *entry,
   3841					    void *context, int vl, int mode,
   3842					    u64 data)
   3843{
   3844	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3845
   3846	return dd->sw_send_dma_eng_err_status_cnt[20];
   3847}
   3848
   3849static u64 access_sdma_desc_table_cor_err_cnt(const struct cntr_entry *entry,
   3850					      void *context, int vl, int mode,
   3851					      u64 data)
   3852{
   3853	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3854
   3855	return dd->sw_send_dma_eng_err_status_cnt[19];
   3856}
   3857
   3858static u64 access_sdma_header_request_fifo_unc_err_cnt(
   3859				const struct cntr_entry *entry,
   3860				void *context, int vl, int mode, u64 data)
   3861{
   3862	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3863
   3864	return dd->sw_send_dma_eng_err_status_cnt[18];
   3865}
   3866
   3867static u64 access_sdma_header_storage_unc_err_cnt(
   3868				const struct cntr_entry *entry,
   3869				void *context, int vl, int mode, u64 data)
   3870{
   3871	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3872
   3873	return dd->sw_send_dma_eng_err_status_cnt[17];
   3874}
   3875
   3876static u64 access_sdma_packet_tracking_unc_err_cnt(
   3877				const struct cntr_entry *entry,
   3878				void *context, int vl, int mode, u64 data)
   3879{
   3880	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3881
   3882	return dd->sw_send_dma_eng_err_status_cnt[16];
   3883}
   3884
   3885static u64 access_sdma_assembly_unc_err_cnt(const struct cntr_entry *entry,
   3886					    void *context, int vl, int mode,
   3887					    u64 data)
   3888{
   3889	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3890
   3891	return dd->sw_send_dma_eng_err_status_cnt[15];
   3892}
   3893
   3894static u64 access_sdma_desc_table_unc_err_cnt(const struct cntr_entry *entry,
   3895					      void *context, int vl, int mode,
   3896					      u64 data)
   3897{
   3898	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3899
   3900	return dd->sw_send_dma_eng_err_status_cnt[14];
   3901}
   3902
   3903static u64 access_sdma_timeout_err_cnt(const struct cntr_entry *entry,
   3904				       void *context, int vl, int mode,
   3905				       u64 data)
   3906{
   3907	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3908
   3909	return dd->sw_send_dma_eng_err_status_cnt[13];
   3910}
   3911
   3912static u64 access_sdma_header_length_err_cnt(const struct cntr_entry *entry,
   3913					     void *context, int vl, int mode,
   3914					     u64 data)
   3915{
   3916	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3917
   3918	return dd->sw_send_dma_eng_err_status_cnt[12];
   3919}
   3920
   3921static u64 access_sdma_header_address_err_cnt(const struct cntr_entry *entry,
   3922					      void *context, int vl, int mode,
   3923					      u64 data)
   3924{
   3925	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3926
   3927	return dd->sw_send_dma_eng_err_status_cnt[11];
   3928}
   3929
   3930static u64 access_sdma_header_select_err_cnt(const struct cntr_entry *entry,
   3931					     void *context, int vl, int mode,
   3932					     u64 data)
   3933{
   3934	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3935
   3936	return dd->sw_send_dma_eng_err_status_cnt[10];
   3937}
   3938
   3939static u64 access_sdma_reserved_9_err_cnt(const struct cntr_entry *entry,
   3940					  void *context, int vl, int mode,
   3941					  u64 data)
   3942{
   3943	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3944
   3945	return dd->sw_send_dma_eng_err_status_cnt[9];
   3946}
   3947
   3948static u64 access_sdma_packet_desc_overflow_err_cnt(
   3949				const struct cntr_entry *entry,
   3950				void *context, int vl, int mode, u64 data)
   3951{
   3952	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3953
   3954	return dd->sw_send_dma_eng_err_status_cnt[8];
   3955}
   3956
   3957static u64 access_sdma_length_mismatch_err_cnt(const struct cntr_entry *entry,
   3958					       void *context, int vl,
   3959					       int mode, u64 data)
   3960{
   3961	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3962
   3963	return dd->sw_send_dma_eng_err_status_cnt[7];
   3964}
   3965
   3966static u64 access_sdma_halt_err_cnt(const struct cntr_entry *entry,
   3967				    void *context, int vl, int mode, u64 data)
   3968{
   3969	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3970
   3971	return dd->sw_send_dma_eng_err_status_cnt[6];
   3972}
   3973
   3974static u64 access_sdma_mem_read_err_cnt(const struct cntr_entry *entry,
   3975					void *context, int vl, int mode,
   3976					u64 data)
   3977{
   3978	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3979
   3980	return dd->sw_send_dma_eng_err_status_cnt[5];
   3981}
   3982
   3983static u64 access_sdma_first_desc_err_cnt(const struct cntr_entry *entry,
   3984					  void *context, int vl, int mode,
   3985					  u64 data)
   3986{
   3987	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3988
   3989	return dd->sw_send_dma_eng_err_status_cnt[4];
   3990}
   3991
   3992static u64 access_sdma_tail_out_of_bounds_err_cnt(
   3993				const struct cntr_entry *entry,
   3994				void *context, int vl, int mode, u64 data)
   3995{
   3996	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   3997
   3998	return dd->sw_send_dma_eng_err_status_cnt[3];
   3999}
   4000
   4001static u64 access_sdma_too_long_err_cnt(const struct cntr_entry *entry,
   4002					void *context, int vl, int mode,
   4003					u64 data)
   4004{
   4005	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   4006
   4007	return dd->sw_send_dma_eng_err_status_cnt[2];
   4008}
   4009
   4010static u64 access_sdma_gen_mismatch_err_cnt(const struct cntr_entry *entry,
   4011					    void *context, int vl, int mode,
   4012					    u64 data)
   4013{
   4014	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   4015
   4016	return dd->sw_send_dma_eng_err_status_cnt[1];
   4017}
   4018
   4019static u64 access_sdma_wrong_dw_err_cnt(const struct cntr_entry *entry,
   4020					void *context, int vl, int mode,
   4021					u64 data)
   4022{
   4023	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   4024
   4025	return dd->sw_send_dma_eng_err_status_cnt[0];
   4026}
   4027
   4028static u64 access_dc_rcv_err_cnt(const struct cntr_entry *entry,
   4029				 void *context, int vl, int mode,
   4030				 u64 data)
   4031{
   4032	struct hfi1_devdata *dd = (struct hfi1_devdata *)context;
   4033
   4034	u64 val = 0;
   4035	u64 csr = entry->csr;
   4036
   4037	val = read_write_csr(dd, csr, mode, data);
   4038	if (mode == CNTR_MODE_R) {
   4039		val = val > CNTR_MAX - dd->sw_rcv_bypass_packet_errors ?
   4040			CNTR_MAX : val + dd->sw_rcv_bypass_packet_errors;
   4041	} else if (mode == CNTR_MODE_W) {
   4042		dd->sw_rcv_bypass_packet_errors = 0;
   4043	} else {
   4044		dd_dev_err(dd, "Invalid cntr register access mode");
   4045		return 0;
   4046	}
   4047	return val;
   4048}
   4049
   4050#define def_access_sw_cpu(cntr) \
   4051static u64 access_sw_cpu_##cntr(const struct cntr_entry *entry,		      \
   4052			      void *context, int vl, int mode, u64 data)      \
   4053{									      \
   4054	struct hfi1_pportdata *ppd = (struct hfi1_pportdata *)context;	      \
   4055	return read_write_cpu(ppd->dd, &ppd->ibport_data.rvp.z_ ##cntr,	      \
   4056			      ppd->ibport_data.rvp.cntr, vl,		      \
   4057			      mode, data);				      \
   4058}
   4059
   4060def_access_sw_cpu(rc_acks);
   4061def_access_sw_cpu(rc_qacks);
   4062def_access_sw_cpu(rc_delayed_comp);
   4063
   4064#define def_access_ibp_counter(cntr) \
   4065static u64 access_ibp_##cntr(const struct cntr_entry *entry,		      \
   4066				void *context, int vl, int mode, u64 data)    \
   4067{									      \
   4068	struct hfi1_pportdata *ppd = (struct hfi1_pportdata *)context;	      \
   4069									      \
   4070	if (vl != CNTR_INVALID_VL)					      \
   4071		return 0;						      \
   4072									      \
   4073	return read_write_sw(ppd->dd, &ppd->ibport_data.rvp.n_ ##cntr,	      \
   4074			     mode, data);				      \
   4075}
   4076
   4077def_access_ibp_counter(loop_pkts);
   4078def_access_ibp_counter(rc_resends);
   4079def_access_ibp_counter(rnr_naks);
   4080def_access_ibp_counter(other_naks);
   4081def_access_ibp_counter(rc_timeouts);
   4082def_access_ibp_counter(pkt_drops);
   4083def_access_ibp_counter(dmawait);
   4084def_access_ibp_counter(rc_seqnak);
   4085def_access_ibp_counter(rc_dupreq);
   4086def_access_ibp_counter(rdma_seq);
   4087def_access_ibp_counter(unaligned);
   4088def_access_ibp_counter(seq_naks);
   4089def_access_ibp_counter(rc_crwaits);
   4090
   4091static struct cntr_entry dev_cntrs[DEV_CNTR_LAST] = {
   4092[C_RCV_OVF] = RXE32_DEV_CNTR_ELEM(RcvOverflow, RCV_BUF_OVFL_CNT, CNTR_SYNTH),
   4093[C_RX_LEN_ERR] = RXE32_DEV_CNTR_ELEM(RxLenErr, RCV_LENGTH_ERR_CNT, CNTR_SYNTH),
   4094[C_RX_SHORT_ERR] = RXE32_DEV_CNTR_ELEM(RxShrErr, RCV_SHORT_ERR_CNT, CNTR_SYNTH),
   4095[C_RX_ICRC_ERR] = RXE32_DEV_CNTR_ELEM(RxICrcErr, RCV_ICRC_ERR_CNT, CNTR_SYNTH),
   4096[C_RX_EBP] = RXE32_DEV_CNTR_ELEM(RxEbpCnt, RCV_EBP_CNT, CNTR_SYNTH),
   4097[C_RX_TID_FULL] = RXE32_DEV_CNTR_ELEM(RxTIDFullEr, RCV_TID_FULL_ERR_CNT,
   4098			CNTR_NORMAL),
   4099[C_RX_TID_INVALID] = RXE32_DEV_CNTR_ELEM(RxTIDInvalid, RCV_TID_VALID_ERR_CNT,
   4100			CNTR_NORMAL),
   4101[C_RX_TID_FLGMS] = RXE32_DEV_CNTR_ELEM(RxTidFLGMs,
   4102			RCV_TID_FLOW_GEN_MISMATCH_CNT,
   4103			CNTR_NORMAL),
   4104[C_RX_CTX_EGRS] = RXE32_DEV_CNTR_ELEM(RxCtxEgrS, RCV_CONTEXT_EGR_STALL,
   4105			CNTR_NORMAL),
   4106[C_RCV_TID_FLSMS] = RXE32_DEV_CNTR_ELEM(RxTidFLSMs,
   4107			RCV_TID_FLOW_SEQ_MISMATCH_CNT, CNTR_NORMAL),
   4108[C_CCE_PCI_CR_ST] = CCE_PERF_DEV_CNTR_ELEM(CcePciCrSt,
   4109			CCE_PCIE_POSTED_CRDT_STALL_CNT, CNTR_NORMAL),
   4110[C_CCE_PCI_TR_ST] = CCE_PERF_DEV_CNTR_ELEM(CcePciTrSt, CCE_PCIE_TRGT_STALL_CNT,
   4111			CNTR_NORMAL),
   4112[C_CCE_PIO_WR_ST] = CCE_PERF_DEV_CNTR_ELEM(CcePioWrSt, CCE_PIO_WR_STALL_CNT,
   4113			CNTR_NORMAL),
   4114[C_CCE_ERR_INT] = CCE_INT_DEV_CNTR_ELEM(CceErrInt, CCE_ERR_INT_CNT,
   4115			CNTR_NORMAL),
   4116[C_CCE_SDMA_INT] = CCE_INT_DEV_CNTR_ELEM(CceSdmaInt, CCE_SDMA_INT_CNT,
   4117			CNTR_NORMAL),
   4118[C_CCE_MISC_INT] = CCE_INT_DEV_CNTR_ELEM(CceMiscInt, CCE_MISC_INT_CNT,
   4119			CNTR_NORMAL),
   4120[C_CCE_RCV_AV_INT] = CCE_INT_DEV_CNTR_ELEM(CceRcvAvInt, CCE_RCV_AVAIL_INT_CNT,
   4121			CNTR_NORMAL),
   4122[C_CCE_RCV_URG_INT] = CCE_INT_DEV_CNTR_ELEM(CceRcvUrgInt,
   4123			CCE_RCV_URGENT_INT_CNT,	CNTR_NORMAL),
   4124[C_CCE_SEND_CR_INT] = CCE_INT_DEV_CNTR_ELEM(CceSndCrInt,
   4125			CCE_SEND_CREDIT_INT_CNT, CNTR_NORMAL),
   4126[C_DC_UNC_ERR] = DC_PERF_CNTR(DcUnctblErr, DCC_ERR_UNCORRECTABLE_CNT,
   4127			      CNTR_SYNTH),
   4128[C_DC_RCV_ERR] = CNTR_ELEM("DcRecvErr", DCC_ERR_PORTRCV_ERR_CNT, 0, CNTR_SYNTH,
   4129			    access_dc_rcv_err_cnt),
   4130[C_DC_FM_CFG_ERR] = DC_PERF_CNTR(DcFmCfgErr, DCC_ERR_FMCONFIG_ERR_CNT,
   4131				 CNTR_SYNTH),
   4132[C_DC_RMT_PHY_ERR] = DC_PERF_CNTR(DcRmtPhyErr, DCC_ERR_RCVREMOTE_PHY_ERR_CNT,
   4133				  CNTR_SYNTH),
   4134[C_DC_DROPPED_PKT] = DC_PERF_CNTR(DcDroppedPkt, DCC_ERR_DROPPED_PKT_CNT,
   4135				  CNTR_SYNTH),
   4136[C_DC_MC_XMIT_PKTS] = DC_PERF_CNTR(DcMcXmitPkts,
   4137				   DCC_PRF_PORT_XMIT_MULTICAST_CNT, CNTR_SYNTH),
   4138[C_DC_MC_RCV_PKTS] = DC_PERF_CNTR(DcMcRcvPkts,
   4139				  DCC_PRF_PORT_RCV_MULTICAST_PKT_CNT,
   4140				  CNTR_SYNTH),
   4141[C_DC_XMIT_CERR] = DC_PERF_CNTR(DcXmitCorr,
   4142				DCC_PRF_PORT_XMIT_CORRECTABLE_CNT, CNTR_SYNTH),
   4143[C_DC_RCV_CERR] = DC_PERF_CNTR(DcRcvCorrCnt, DCC_PRF_PORT_RCV_CORRECTABLE_CNT,
   4144			       CNTR_SYNTH),
   4145[C_DC_RCV_FCC] = DC_PERF_CNTR(DcRxFCntl, DCC_PRF_RX_FLOW_CRTL_CNT,
   4146			      CNTR_SYNTH),
   4147[C_DC_XMIT_FCC] = DC_PERF_CNTR(DcXmitFCntl, DCC_PRF_TX_FLOW_CRTL_CNT,
   4148			       CNTR_SYNTH),
   4149[C_DC_XMIT_FLITS] = DC_PERF_CNTR(DcXmitFlits, DCC_PRF_PORT_XMIT_DATA_CNT,
   4150				 CNTR_SYNTH),
   4151[C_DC_RCV_FLITS] = DC_PERF_CNTR(DcRcvFlits, DCC_PRF_PORT_RCV_DATA_CNT,
   4152				CNTR_SYNTH),
   4153[C_DC_XMIT_PKTS] = DC_PERF_CNTR(DcXmitPkts, DCC_PRF_PORT_XMIT_PKTS_CNT,
   4154				CNTR_SYNTH),
   4155[C_DC_RCV_PKTS] = DC_PERF_CNTR(DcRcvPkts, DCC_PRF_PORT_RCV_PKTS_CNT,
   4156			       CNTR_SYNTH),
   4157[C_DC_RX_FLIT_VL] = DC_PERF_CNTR(DcRxFlitVl, DCC_PRF_PORT_VL_RCV_DATA_CNT,
   4158				 CNTR_SYNTH | CNTR_VL),
   4159[C_DC_RX_PKT_VL] = DC_PERF_CNTR(DcRxPktVl, DCC_PRF_PORT_VL_RCV_PKTS_CNT,
   4160				CNTR_SYNTH | CNTR_VL),
   4161[C_DC_RCV_FCN] = DC_PERF_CNTR(DcRcvFcn, DCC_PRF_PORT_RCV_FECN_CNT, CNTR_SYNTH),
   4162[C_DC_RCV_FCN_VL] = DC_PERF_CNTR(DcRcvFcnVl, DCC_PRF_PORT_VL_RCV_FECN_CNT,
   4163				 CNTR_SYNTH | CNTR_VL),
   4164[C_DC_RCV_BCN] = DC_PERF_CNTR(DcRcvBcn, DCC_PRF_PORT_RCV_BECN_CNT, CNTR_SYNTH),
   4165[C_DC_RCV_BCN_VL] = DC_PERF_CNTR(DcRcvBcnVl, DCC_PRF_PORT_VL_RCV_BECN_CNT,
   4166				 CNTR_SYNTH | CNTR_VL),
   4167[C_DC_RCV_BBL] = DC_PERF_CNTR(DcRcvBbl, DCC_PRF_PORT_RCV_BUBBLE_CNT,
   4168			      CNTR_SYNTH),
   4169[C_DC_RCV_BBL_VL] = DC_PERF_CNTR(DcRcvBblVl, DCC_PRF_PORT_VL_RCV_BUBBLE_CNT,
   4170				 CNTR_SYNTH | CNTR_VL),
   4171[C_DC_MARK_FECN] = DC_PERF_CNTR(DcMarkFcn, DCC_PRF_PORT_MARK_FECN_CNT,
   4172				CNTR_SYNTH),
   4173[C_DC_MARK_FECN_VL] = DC_PERF_CNTR(DcMarkFcnVl, DCC_PRF_PORT_VL_MARK_FECN_CNT,
   4174				   CNTR_SYNTH | CNTR_VL),
   4175[C_DC_TOTAL_CRC] =
   4176	DC_PERF_CNTR_LCB(DcTotCrc, DC_LCB_ERR_INFO_TOTAL_CRC_ERR,
   4177			 CNTR_SYNTH),
   4178[C_DC_CRC_LN0] = DC_PERF_CNTR_LCB(DcCrcLn0, DC_LCB_ERR_INFO_CRC_ERR_LN0,
   4179				  CNTR_SYNTH),
   4180[C_DC_CRC_LN1] = DC_PERF_CNTR_LCB(DcCrcLn1, DC_LCB_ERR_INFO_CRC_ERR_LN1,
   4181				  CNTR_SYNTH),
   4182[C_DC_CRC_LN2] = DC_PERF_CNTR_LCB(DcCrcLn2, DC_LCB_ERR_INFO_CRC_ERR_LN2,
   4183				  CNTR_SYNTH),
   4184[C_DC_CRC_LN3] = DC_PERF_CNTR_LCB(DcCrcLn3, DC_LCB_ERR_INFO_CRC_ERR_LN3,
   4185				  CNTR_SYNTH),
   4186[C_DC_CRC_MULT_LN] =
   4187	DC_PERF_CNTR_LCB(DcMultLn, DC_LCB_ERR_INFO_CRC_ERR_MULTI_LN,
   4188			 CNTR_SYNTH),
   4189[C_DC_TX_REPLAY] = DC_PERF_CNTR_LCB(DcTxReplay, DC_LCB_ERR_INFO_TX_REPLAY_CNT,
   4190				    CNTR_SYNTH),
   4191[C_DC_RX_REPLAY] = DC_PERF_CNTR_LCB(DcRxReplay, DC_LCB_ERR_INFO_RX_REPLAY_CNT,
   4192				    CNTR_SYNTH),
   4193[C_DC_SEQ_CRC_CNT] =
   4194	DC_PERF_CNTR_LCB(DcLinkSeqCrc, DC_LCB_ERR_INFO_SEQ_CRC_CNT,
   4195			 CNTR_SYNTH),
   4196[C_DC_ESC0_ONLY_CNT] =
   4197	DC_PERF_CNTR_LCB(DcEsc0, DC_LCB_ERR_INFO_ESCAPE_0_ONLY_CNT,
   4198			 CNTR_SYNTH),
   4199[C_DC_ESC0_PLUS1_CNT] =
   4200	DC_PERF_CNTR_LCB(DcEsc1, DC_LCB_ERR_INFO_ESCAPE_0_PLUS1_CNT,
   4201			 CNTR_SYNTH),
   4202[C_DC_ESC0_PLUS2_CNT] =
   4203	DC_PERF_CNTR_LCB(DcEsc0Plus2, DC_LCB_ERR_INFO_ESCAPE_0_PLUS2_CNT,
   4204			 CNTR_SYNTH),
   4205[C_DC_REINIT_FROM_PEER_CNT] =
   4206	DC_PERF_CNTR_LCB(DcReinitPeer, DC_LCB_ERR_INFO_REINIT_FROM_PEER_CNT,
   4207			 CNTR_SYNTH),
   4208[C_DC_SBE_CNT] = DC_PERF_CNTR_LCB(DcSbe, DC_LCB_ERR_INFO_SBE_CNT,
   4209				  CNTR_SYNTH),
   4210[C_DC_MISC_FLG_CNT] =
   4211	DC_PERF_CNTR_LCB(DcMiscFlg, DC_LCB_ERR_INFO_MISC_FLG_CNT,
   4212			 CNTR_SYNTH),
   4213[C_DC_PRF_GOOD_LTP_CNT] =
   4214	DC_PERF_CNTR_LCB(DcGoodLTP, DC_LCB_PRF_GOOD_LTP_CNT, CNTR_SYNTH),
   4215[C_DC_PRF_ACCEPTED_LTP_CNT] =
   4216	DC_PERF_CNTR_LCB(DcAccLTP, DC_LCB_PRF_ACCEPTED_LTP_CNT,
   4217			 CNTR_SYNTH),
   4218[C_DC_PRF_RX_FLIT_CNT] =
   4219	DC_PERF_CNTR_LCB(DcPrfRxFlit, DC_LCB_PRF_RX_FLIT_CNT, CNTR_SYNTH),
   4220[C_DC_PRF_TX_FLIT_CNT] =
   4221	DC_PERF_CNTR_LCB(DcPrfTxFlit, DC_LCB_PRF_TX_FLIT_CNT, CNTR_SYNTH),
   4222[C_DC_PRF_CLK_CNTR] =
   4223	DC_PERF_CNTR_LCB(DcPrfClk, DC_LCB_PRF_CLK_CNTR, CNTR_SYNTH),
   4224[C_DC_PG_DBG_FLIT_CRDTS_CNT] =
   4225	DC_PERF_CNTR_LCB(DcFltCrdts, DC_LCB_PG_DBG_FLIT_CRDTS_CNT, CNTR_SYNTH),
   4226[C_DC_PG_STS_PAUSE_COMPLETE_CNT] =
   4227	DC_PERF_CNTR_LCB(DcPauseComp, DC_LCB_PG_STS_PAUSE_COMPLETE_CNT,
   4228			 CNTR_SYNTH),
   4229[C_DC_PG_STS_TX_SBE_CNT] =
   4230	DC_PERF_CNTR_LCB(DcStsTxSbe, DC_LCB_PG_STS_TX_SBE_CNT, CNTR_SYNTH),
   4231[C_DC_PG_STS_TX_MBE_CNT] =
   4232	DC_PERF_CNTR_LCB(DcStsTxMbe, DC_LCB_PG_STS_TX_MBE_CNT,
   4233			 CNTR_SYNTH),
   4234[C_SW_CPU_INTR] = CNTR_ELEM("Intr", 0, 0, CNTR_NORMAL,
   4235			    access_sw_cpu_intr),
   4236[C_SW_CPU_RCV_LIM] = CNTR_ELEM("RcvLimit", 0, 0, CNTR_NORMAL,
   4237			    access_sw_cpu_rcv_limit),
   4238[C_SW_CTX0_SEQ_DROP] = CNTR_ELEM("SeqDrop0", 0, 0, CNTR_NORMAL,
   4239			    access_sw_ctx0_seq_drop),
   4240[C_SW_VTX_WAIT] = CNTR_ELEM("vTxWait", 0, 0, CNTR_NORMAL,
   4241			    access_sw_vtx_wait),
   4242[C_SW_PIO_WAIT] = CNTR_ELEM("PioWait", 0, 0, CNTR_NORMAL,
   4243			    access_sw_pio_wait),
   4244[C_SW_PIO_DRAIN] = CNTR_ELEM("PioDrain", 0, 0, CNTR_NORMAL,
   4245			    access_sw_pio_drain),
   4246[C_SW_KMEM_WAIT] = CNTR_ELEM("KmemWait", 0, 0, CNTR_NORMAL,
   4247			    access_sw_kmem_wait),
   4248[C_SW_TID_WAIT] = CNTR_ELEM("TidWait", 0, 0, CNTR_NORMAL,
   4249			    hfi1_access_sw_tid_wait),
   4250[C_SW_SEND_SCHED] = CNTR_ELEM("SendSched", 0, 0, CNTR_NORMAL,
   4251			    access_sw_send_schedule),
   4252[C_SDMA_DESC_FETCHED_CNT] = CNTR_ELEM("SDEDscFdCn",
   4253				      SEND_DMA_DESC_FETCHED_CNT, 0,
   4254				      CNTR_NORMAL | CNTR_32BIT | CNTR_SDMA,
   4255				      dev_access_u32_csr),
   4256[C_SDMA_INT_CNT] = CNTR_ELEM("SDMAInt", 0, 0,
   4257			     CNTR_NORMAL | CNTR_32BIT | CNTR_SDMA,
   4258			     access_sde_int_cnt),
   4259[C_SDMA_ERR_CNT] = CNTR_ELEM("SDMAErrCt", 0, 0,
   4260			     CNTR_NORMAL | CNTR_32BIT | CNTR_SDMA,
   4261			     access_sde_err_cnt),
   4262[C_SDMA_IDLE_INT_CNT] = CNTR_ELEM("SDMAIdInt", 0, 0,
   4263				  CNTR_NORMAL | CNTR_32BIT | CNTR_SDMA,
   4264				  access_sde_idle_int_cnt),
   4265[C_SDMA_PROGRESS_INT_CNT] = CNTR_ELEM("SDMAPrIntCn", 0, 0,
   4266				      CNTR_NORMAL | CNTR_32BIT | CNTR_SDMA,
   4267				      access_sde_progress_int_cnt),
   4268/* MISC_ERR_STATUS */
   4269[C_MISC_PLL_LOCK_FAIL_ERR] = CNTR_ELEM("MISC_PLL_LOCK_FAIL_ERR", 0, 0,
   4270				CNTR_NORMAL,
   4271				access_misc_pll_lock_fail_err_cnt),
   4272[C_MISC_MBIST_FAIL_ERR] = CNTR_ELEM("MISC_MBIST_FAIL_ERR", 0, 0,
   4273				CNTR_NORMAL,
   4274				access_misc_mbist_fail_err_cnt),
   4275[C_MISC_INVALID_EEP_CMD_ERR] = CNTR_ELEM("MISC_INVALID_EEP_CMD_ERR", 0, 0,
   4276				CNTR_NORMAL,
   4277				access_misc_invalid_eep_cmd_err_cnt),
   4278[C_MISC_EFUSE_DONE_PARITY_ERR] = CNTR_ELEM("MISC_EFUSE_DONE_PARITY_ERR", 0, 0,
   4279				CNTR_NORMAL,
   4280				access_misc_efuse_done_parity_err_cnt),
   4281[C_MISC_EFUSE_WRITE_ERR] = CNTR_ELEM("MISC_EFUSE_WRITE_ERR", 0, 0,
   4282				CNTR_NORMAL,
   4283				access_misc_efuse_write_err_cnt),
   4284[C_MISC_EFUSE_READ_BAD_ADDR_ERR] = CNTR_ELEM("MISC_EFUSE_READ_BAD_ADDR_ERR", 0,
   4285				0, CNTR_NORMAL,
   4286				access_misc_efuse_read_bad_addr_err_cnt),
   4287[C_MISC_EFUSE_CSR_PARITY_ERR] = CNTR_ELEM("MISC_EFUSE_CSR_PARITY_ERR", 0, 0,
   4288				CNTR_NORMAL,
   4289				access_misc_efuse_csr_parity_err_cnt),
   4290[C_MISC_FW_AUTH_FAILED_ERR] = CNTR_ELEM("MISC_FW_AUTH_FAILED_ERR", 0, 0,
   4291				CNTR_NORMAL,
   4292				access_misc_fw_auth_failed_err_cnt),
   4293[C_MISC_KEY_MISMATCH_ERR] = CNTR_ELEM("MISC_KEY_MISMATCH_ERR", 0, 0,
   4294				CNTR_NORMAL,
   4295				access_misc_key_mismatch_err_cnt),
   4296[C_MISC_SBUS_WRITE_FAILED_ERR] = CNTR_ELEM("MISC_SBUS_WRITE_FAILED_ERR", 0, 0,
   4297				CNTR_NORMAL,
   4298				access_misc_sbus_write_failed_err_cnt),
   4299[C_MISC_CSR_WRITE_BAD_ADDR_ERR] = CNTR_ELEM("MISC_CSR_WRITE_BAD_ADDR_ERR", 0, 0,
   4300				CNTR_NORMAL,
   4301				access_misc_csr_write_bad_addr_err_cnt),
   4302[C_MISC_CSR_READ_BAD_ADDR_ERR] = CNTR_ELEM("MISC_CSR_READ_BAD_ADDR_ERR", 0, 0,
   4303				CNTR_NORMAL,
   4304				access_misc_csr_read_bad_addr_err_cnt),
   4305[C_MISC_CSR_PARITY_ERR] = CNTR_ELEM("MISC_CSR_PARITY_ERR", 0, 0,
   4306				CNTR_NORMAL,
   4307				access_misc_csr_parity_err_cnt),
   4308/* CceErrStatus */
   4309[C_CCE_ERR_STATUS_AGGREGATED_CNT] = CNTR_ELEM("CceErrStatusAggregatedCnt", 0, 0,
   4310				CNTR_NORMAL,
   4311				access_sw_cce_err_status_aggregated_cnt),
   4312[C_CCE_MSIX_CSR_PARITY_ERR] = CNTR_ELEM("CceMsixCsrParityErr", 0, 0,
   4313				CNTR_NORMAL,
   4314				access_cce_msix_csr_parity_err_cnt),
   4315[C_CCE_INT_MAP_UNC_ERR] = CNTR_ELEM("CceIntMapUncErr", 0, 0,
   4316				CNTR_NORMAL,
   4317				access_cce_int_map_unc_err_cnt),
   4318[C_CCE_INT_MAP_COR_ERR] = CNTR_ELEM("CceIntMapCorErr", 0, 0,
   4319				CNTR_NORMAL,
   4320				access_cce_int_map_cor_err_cnt),
   4321[C_CCE_MSIX_TABLE_UNC_ERR] = CNTR_ELEM("CceMsixTableUncErr", 0, 0,
   4322				CNTR_NORMAL,
   4323				access_cce_msix_table_unc_err_cnt),
   4324[C_CCE_MSIX_TABLE_COR_ERR] = CNTR_ELEM("CceMsixTableCorErr", 0, 0,
   4325				CNTR_NORMAL,
   4326				access_cce_msix_table_cor_err_cnt),
   4327[C_CCE_RXDMA_CONV_FIFO_PARITY_ERR] = CNTR_ELEM("CceRxdmaConvFifoParityErr", 0,
   4328				0, CNTR_NORMAL,
   4329				access_cce_rxdma_conv_fifo_parity_err_cnt),
   4330[C_CCE_RCPL_ASYNC_FIFO_PARITY_ERR] = CNTR_ELEM("CceRcplAsyncFifoParityErr", 0,
   4331				0, CNTR_NORMAL,
   4332				access_cce_rcpl_async_fifo_parity_err_cnt),
   4333[C_CCE_SEG_WRITE_BAD_ADDR_ERR] = CNTR_ELEM("CceSegWriteBadAddrErr", 0, 0,
   4334				CNTR_NORMAL,
   4335				access_cce_seg_write_bad_addr_err_cnt),
   4336[C_CCE_SEG_READ_BAD_ADDR_ERR] = CNTR_ELEM("CceSegReadBadAddrErr", 0, 0,
   4337				CNTR_NORMAL,
   4338				access_cce_seg_read_bad_addr_err_cnt),
   4339[C_LA_TRIGGERED] = CNTR_ELEM("Cce LATriggered", 0, 0,
   4340				CNTR_NORMAL,
   4341				access_la_triggered_cnt),
   4342[C_CCE_TRGT_CPL_TIMEOUT_ERR] = CNTR_ELEM("CceTrgtCplTimeoutErr", 0, 0,
   4343				CNTR_NORMAL,
   4344				access_cce_trgt_cpl_timeout_err_cnt),
   4345[C_PCIC_RECEIVE_PARITY_ERR] = CNTR_ELEM("PcicReceiveParityErr", 0, 0,
   4346				CNTR_NORMAL,
   4347				access_pcic_receive_parity_err_cnt),
   4348[C_PCIC_TRANSMIT_BACK_PARITY_ERR] = CNTR_ELEM("PcicTransmitBackParityErr", 0, 0,
   4349				CNTR_NORMAL,
   4350				access_pcic_transmit_back_parity_err_cnt),
   4351[C_PCIC_TRANSMIT_FRONT_PARITY_ERR] = CNTR_ELEM("PcicTransmitFrontParityErr", 0,
   4352				0, CNTR_NORMAL,
   4353				access_pcic_transmit_front_parity_err_cnt),
   4354[C_PCIC_CPL_DAT_Q_UNC_ERR] = CNTR_ELEM("PcicCplDatQUncErr", 0, 0,
   4355				CNTR_NORMAL,
   4356				access_pcic_cpl_dat_q_unc_err_cnt),
   4357[C_PCIC_CPL_HD_Q_UNC_ERR] = CNTR_ELEM("PcicCplHdQUncErr", 0, 0,
   4358				CNTR_NORMAL,
   4359				access_pcic_cpl_hd_q_unc_err_cnt),
   4360[C_PCIC_POST_DAT_Q_UNC_ERR] = CNTR_ELEM("PcicPostDatQUncErr", 0, 0,
   4361				CNTR_NORMAL,
   4362				access_pcic_post_dat_q_unc_err_cnt),
   4363[C_PCIC_POST_HD_Q_UNC_ERR] = CNTR_ELEM("PcicPostHdQUncErr", 0, 0,
   4364				CNTR_NORMAL,
   4365				access_pcic_post_hd_q_unc_err_cnt),
   4366[C_PCIC_RETRY_SOT_MEM_UNC_ERR] = CNTR_ELEM("PcicRetrySotMemUncErr", 0, 0,
   4367				CNTR_NORMAL,
   4368				access_pcic_retry_sot_mem_unc_err_cnt),
   4369[C_PCIC_RETRY_MEM_UNC_ERR] = CNTR_ELEM("PcicRetryMemUncErr", 0, 0,
   4370				CNTR_NORMAL,
   4371				access_pcic_retry_mem_unc_err),
   4372[C_PCIC_N_POST_DAT_Q_PARITY_ERR] = CNTR_ELEM("PcicNPostDatQParityErr", 0, 0,
   4373				CNTR_NORMAL,
   4374				access_pcic_n_post_dat_q_parity_err_cnt),
   4375[C_PCIC_N_POST_H_Q_PARITY_ERR] = CNTR_ELEM("PcicNPostHQParityErr", 0, 0,
   4376				CNTR_NORMAL,
   4377				access_pcic_n_post_h_q_parity_err_cnt),
   4378[C_PCIC_CPL_DAT_Q_COR_ERR] = CNTR_ELEM("PcicCplDatQCorErr", 0, 0,
   4379				CNTR_NORMAL,
   4380				access_pcic_cpl_dat_q_cor_err_cnt),
   4381[C_PCIC_CPL_HD_Q_COR_ERR] = CNTR_ELEM("PcicCplHdQCorErr", 0, 0,
   4382				CNTR_NORMAL,
   4383				access_pcic_cpl_hd_q_cor_err_cnt),
   4384[C_PCIC_POST_DAT_Q_COR_ERR] = CNTR_ELEM("PcicPostDatQCorErr", 0, 0,
   4385				CNTR_NORMAL,
   4386				access_pcic_post_dat_q_cor_err_cnt),
   4387[C_PCIC_POST_HD_Q_COR_ERR] = CNTR_ELEM("PcicPostHdQCorErr", 0, 0,
   4388				CNTR_NORMAL,
   4389				access_pcic_post_hd_q_cor_err_cnt),
   4390[C_PCIC_RETRY_SOT_MEM_COR_ERR] = CNTR_ELEM("PcicRetrySotMemCorErr", 0, 0,
   4391				CNTR_NORMAL,
   4392				access_pcic_retry_sot_mem_cor_err_cnt),
   4393[C_PCIC_RETRY_MEM_COR_ERR] = CNTR_ELEM("PcicRetryMemCorErr", 0, 0,
   4394				CNTR_NORMAL,
   4395				access_pcic_retry_mem_cor_err_cnt),
   4396[C_CCE_CLI1_ASYNC_FIFO_DBG_PARITY_ERR] = CNTR_ELEM(
   4397				"CceCli1AsyncFifoDbgParityError", 0, 0,
   4398				CNTR_NORMAL,
   4399				access_cce_cli1_async_fifo_dbg_parity_err_cnt),
   4400[C_CCE_CLI1_ASYNC_FIFO_RXDMA_PARITY_ERR] = CNTR_ELEM(
   4401				"CceCli1AsyncFifoRxdmaParityError", 0, 0,
   4402				CNTR_NORMAL,
   4403				access_cce_cli1_async_fifo_rxdma_parity_err_cnt
   4404				),
   4405[C_CCE_CLI1_ASYNC_FIFO_SDMA_HD_PARITY_ERR] = CNTR_ELEM(
   4406			"CceCli1AsyncFifoSdmaHdParityErr", 0, 0,
   4407			CNTR_NORMAL,
   4408			access_cce_cli1_async_fifo_sdma_hd_parity_err_cnt),
   4409[C_CCE_CLI1_ASYNC_FIFO_PIO_CRDT_PARITY_ERR] = CNTR_ELEM(
   4410			"CceCli1AsyncFifoPioCrdtParityErr", 0, 0,
   4411			CNTR_NORMAL,
   4412			access_cce_cl1_async_fifo_pio_crdt_parity_err_cnt),
   4413[C_CCE_CLI2_ASYNC_FIFO_PARITY_ERR] = CNTR_ELEM("CceCli2AsyncFifoParityErr", 0,
   4414			0, CNTR_NORMAL,
   4415			access_cce_cli2_async_fifo_parity_err_cnt),
   4416[C_CCE_CSR_CFG_BUS_PARITY_ERR] = CNTR_ELEM("CceCsrCfgBusParityErr", 0, 0,
   4417			CNTR_NORMAL,
   4418			access_cce_csr_cfg_bus_parity_err_cnt),
   4419[C_CCE_CLI0_ASYNC_FIFO_PARTIY_ERR] = CNTR_ELEM("CceCli0AsyncFifoParityErr", 0,
   4420			0, CNTR_NORMAL,
   4421			access_cce_cli0_async_fifo_parity_err_cnt),
   4422[C_CCE_RSPD_DATA_PARITY_ERR] = CNTR_ELEM("CceRspdDataParityErr", 0, 0,
   4423			CNTR_NORMAL,
   4424			access_cce_rspd_data_parity_err_cnt),
   4425[C_CCE_TRGT_ACCESS_ERR] = CNTR_ELEM("CceTrgtAccessErr", 0, 0,
   4426			CNTR_NORMAL,
   4427			access_cce_trgt_access_err_cnt),
   4428[C_CCE_TRGT_ASYNC_FIFO_PARITY_ERR] = CNTR_ELEM("CceTrgtAsyncFifoParityErr", 0,
   4429			0, CNTR_NORMAL,
   4430			access_cce_trgt_async_fifo_parity_err_cnt),
   4431[C_CCE_CSR_WRITE_BAD_ADDR_ERR] = CNTR_ELEM("CceCsrWriteBadAddrErr", 0, 0,
   4432			CNTR_NORMAL,
   4433			access_cce_csr_write_bad_addr_err_cnt),
   4434[C_CCE_CSR_READ_BAD_ADDR_ERR] = CNTR_ELEM("CceCsrReadBadAddrErr", 0, 0,
   4435			CNTR_NORMAL,
   4436			access_cce_csr_read_bad_addr_err_cnt),
   4437[C_CCE_CSR_PARITY_ERR] = CNTR_ELEM("CceCsrParityErr", 0, 0,
   4438			CNTR_NORMAL,
   4439			access_ccs_csr_parity_err_cnt),
   4440
   4441/* RcvErrStatus */
   4442[C_RX_CSR_PARITY_ERR] = CNTR_ELEM("RxCsrParityErr", 0, 0,
   4443			CNTR_NORMAL,
   4444			access_rx_csr_parity_err_cnt),
   4445[C_RX_CSR_WRITE_BAD_ADDR_ERR] = CNTR_ELEM("RxCsrWriteBadAddrErr", 0, 0,
   4446			CNTR_NORMAL,
   4447			access_rx_csr_write_bad_addr_err_cnt),
   4448[C_RX_CSR_READ_BAD_ADDR_ERR] = CNTR_ELEM("RxCsrReadBadAddrErr", 0, 0,
   4449			CNTR_NORMAL,
   4450			access_rx_csr_read_bad_addr_err_cnt),
   4451[C_RX_DMA_CSR_UNC_ERR] = CNTR_ELEM("RxDmaCsrUncErr", 0, 0,
   4452			CNTR_NORMAL,
   4453			access_rx_dma_csr_unc_err_cnt),
   4454[C_RX_DMA_DQ_FSM_ENCODING_ERR] = CNTR_ELEM("RxDmaDqFsmEncodingErr", 0, 0,
   4455			CNTR_NORMAL,
   4456			access_rx_dma_dq_fsm_encoding_err_cnt),
   4457[C_RX_DMA_EQ_FSM_ENCODING_ERR] = CNTR_ELEM("RxDmaEqFsmEncodingErr", 0, 0,
   4458			CNTR_NORMAL,
   4459			access_rx_dma_eq_fsm_encoding_err_cnt),
   4460[C_RX_DMA_CSR_PARITY_ERR] = CNTR_ELEM("RxDmaCsrParityErr", 0, 0,
   4461			CNTR_NORMAL,
   4462			access_rx_dma_csr_parity_err_cnt),
   4463[C_RX_RBUF_DATA_COR_ERR] = CNTR_ELEM("RxRbufDataCorErr", 0, 0,
   4464			CNTR_NORMAL,
   4465			access_rx_rbuf_data_cor_err_cnt),
   4466[C_RX_RBUF_DATA_UNC_ERR] = CNTR_ELEM("RxRbufDataUncErr", 0, 0,
   4467			CNTR_NORMAL,
   4468			access_rx_rbuf_data_unc_err_cnt),
   4469[C_RX_DMA_DATA_FIFO_RD_COR_ERR] = CNTR_ELEM("RxDmaDataFifoRdCorErr", 0, 0,
   4470			CNTR_NORMAL,
   4471			access_rx_dma_data_fifo_rd_cor_err_cnt),
   4472[C_RX_DMA_DATA_FIFO_RD_UNC_ERR] = CNTR_ELEM("RxDmaDataFifoRdUncErr", 0, 0,
   4473			CNTR_NORMAL,
   4474			access_rx_dma_data_fifo_rd_unc_err_cnt),
   4475[C_RX_DMA_HDR_FIFO_RD_COR_ERR] = CNTR_ELEM("RxDmaHdrFifoRdCorErr", 0, 0,
   4476			CNTR_NORMAL,
   4477			access_rx_dma_hdr_fifo_rd_cor_err_cnt),
   4478[C_RX_DMA_HDR_FIFO_RD_UNC_ERR] = CNTR_ELEM("RxDmaHdrFifoRdUncErr", 0, 0,
   4479			CNTR_NORMAL,
   4480			access_rx_dma_hdr_fifo_rd_unc_err_cnt),
   4481[C_RX_RBUF_DESC_PART2_COR_ERR] = CNTR_ELEM("RxRbufDescPart2CorErr", 0, 0,
   4482			CNTR_NORMAL,
   4483			access_rx_rbuf_desc_part2_cor_err_cnt),
   4484[C_RX_RBUF_DESC_PART2_UNC_ERR] = CNTR_ELEM("RxRbufDescPart2UncErr", 0, 0,
   4485			CNTR_NORMAL,
   4486			access_rx_rbuf_desc_part2_unc_err_cnt),
   4487[C_RX_RBUF_DESC_PART1_COR_ERR] = CNTR_ELEM("RxRbufDescPart1CorErr", 0, 0,
   4488			CNTR_NORMAL,
   4489			access_rx_rbuf_desc_part1_cor_err_cnt),
   4490[C_RX_RBUF_DESC_PART1_UNC_ERR] = CNTR_ELEM("RxRbufDescPart1UncErr", 0, 0,
   4491			CNTR_NORMAL,
   4492			access_rx_rbuf_desc_part1_unc_err_cnt),
   4493[C_RX_HQ_INTR_FSM_ERR] = CNTR_ELEM("RxHqIntrFsmErr", 0, 0,
   4494			CNTR_NORMAL,
   4495			access_rx_hq_intr_fsm_err_cnt),
   4496[C_RX_HQ_INTR_CSR_PARITY_ERR] = CNTR_ELEM("RxHqIntrCsrParityErr", 0, 0,
   4497			CNTR_NORMAL,
   4498			access_rx_hq_intr_csr_parity_err_cnt),
   4499[C_RX_LOOKUP_CSR_PARITY_ERR] = CNTR_ELEM("RxLookupCsrParityErr", 0, 0,
   4500			CNTR_NORMAL,
   4501			access_rx_lookup_csr_parity_err_cnt),
   4502[C_RX_LOOKUP_RCV_ARRAY_COR_ERR] = CNTR_ELEM("RxLookupRcvArrayCorErr", 0, 0,
   4503			CNTR_NORMAL,
   4504			access_rx_lookup_rcv_array_cor_err_cnt),
   4505[C_RX_LOOKUP_RCV_ARRAY_UNC_ERR] = CNTR_ELEM("RxLookupRcvArrayUncErr", 0, 0,
   4506			CNTR_NORMAL,
   4507			access_rx_lookup_rcv_array_unc_err_cnt),
   4508[C_RX_LOOKUP_DES_PART2_PARITY_ERR] = CNTR_ELEM("RxLookupDesPart2ParityErr", 0,
   4509			0, CNTR_NORMAL,
   4510			access_rx_lookup_des_part2_parity_err_cnt),
   4511[C_RX_LOOKUP_DES_PART1_UNC_COR_ERR] = CNTR_ELEM("RxLookupDesPart1UncCorErr", 0,
   4512			0, CNTR_NORMAL,
   4513			access_rx_lookup_des_part1_unc_cor_err_cnt),
   4514[C_RX_LOOKUP_DES_PART1_UNC_ERR] = CNTR_ELEM("RxLookupDesPart1UncErr", 0, 0,
   4515			CNTR_NORMAL,
   4516			access_rx_lookup_des_part1_unc_err_cnt),
   4517[C_RX_RBUF_NEXT_FREE_BUF_COR_ERR] = CNTR_ELEM("RxRbufNextFreeBufCorErr", 0, 0,
   4518			CNTR_NORMAL,
   4519			access_rx_rbuf_next_free_buf_cor_err_cnt),
   4520[C_RX_RBUF_NEXT_FREE_BUF_UNC_ERR] = CNTR_ELEM("RxRbufNextFreeBufUncErr", 0, 0,
   4521			CNTR_NORMAL,
   4522			access_rx_rbuf_next_free_buf_unc_err_cnt),
   4523[C_RX_RBUF_FL_INIT_WR_ADDR_PARITY_ERR] = CNTR_ELEM(
   4524			"RxRbufFlInitWrAddrParityErr", 0, 0,
   4525			CNTR_NORMAL,
   4526			access_rbuf_fl_init_wr_addr_parity_err_cnt),
   4527[C_RX_RBUF_FL_INITDONE_PARITY_ERR] = CNTR_ELEM("RxRbufFlInitdoneParityErr", 0,
   4528			0, CNTR_NORMAL,
   4529			access_rx_rbuf_fl_initdone_parity_err_cnt),
   4530[C_RX_RBUF_FL_WRITE_ADDR_PARITY_ERR] = CNTR_ELEM("RxRbufFlWrAddrParityErr", 0,
   4531			0, CNTR_NORMAL,
   4532			access_rx_rbuf_fl_write_addr_parity_err_cnt),
   4533[C_RX_RBUF_FL_RD_ADDR_PARITY_ERR] = CNTR_ELEM("RxRbufFlRdAddrParityErr", 0, 0,
   4534			CNTR_NORMAL,
   4535			access_rx_rbuf_fl_rd_addr_parity_err_cnt),
   4536[C_RX_RBUF_EMPTY_ERR] = CNTR_ELEM("RxRbufEmptyErr", 0, 0,
   4537			CNTR_NORMAL,
   4538			access_rx_rbuf_empty_err_cnt),
   4539[C_RX_RBUF_FULL_ERR] = CNTR_ELEM("RxRbufFullErr", 0, 0,
   4540			CNTR_NORMAL,
   4541			access_rx_rbuf_full_err_cnt),
   4542[C_RX_RBUF_BAD_LOOKUP_ERR] = CNTR_ELEM("RxRBufBadLookupErr", 0, 0,
   4543			CNTR_NORMAL,
   4544			access_rbuf_bad_lookup_err_cnt),
   4545[C_RX_RBUF_CTX_ID_PARITY_ERR] = CNTR_ELEM("RxRbufCtxIdParityErr", 0, 0,
   4546			CNTR_NORMAL,
   4547			access_rbuf_ctx_id_parity_err_cnt),
   4548[C_RX_RBUF_CSR_QEOPDW_PARITY_ERR] = CNTR_ELEM("RxRbufCsrQEOPDWParityErr", 0, 0,
   4549			CNTR_NORMAL,
   4550			access_rbuf_csr_qeopdw_parity_err_cnt),
   4551[C_RX_RBUF_CSR_Q_NUM_OF_PKT_PARITY_ERR] = CNTR_ELEM(
   4552			"RxRbufCsrQNumOfPktParityErr", 0, 0,
   4553			CNTR_NORMAL,
   4554			access_rx_rbuf_csr_q_num_of_pkt_parity_err_cnt),
   4555[C_RX_RBUF_CSR_Q_T1_PTR_PARITY_ERR] = CNTR_ELEM(
   4556			"RxRbufCsrQTlPtrParityErr", 0, 0,
   4557			CNTR_NORMAL,
   4558			access_rx_rbuf_csr_q_t1_ptr_parity_err_cnt),
   4559[C_RX_RBUF_CSR_Q_HD_PTR_PARITY_ERR] = CNTR_ELEM("RxRbufCsrQHdPtrParityErr", 0,
   4560			0, CNTR_NORMAL,
   4561			access_rx_rbuf_csr_q_hd_ptr_parity_err_cnt),
   4562[C_RX_RBUF_CSR_Q_VLD_BIT_PARITY_ERR] = CNTR_ELEM("RxRbufCsrQVldBitParityErr", 0,
   4563			0, CNTR_NORMAL,
   4564			access_rx_rbuf_csr_q_vld_bit_parity_err_cnt),
   4565[C_RX_RBUF_CSR_Q_NEXT_BUF_PARITY_ERR] = CNTR_ELEM("RxRbufCsrQNextBufParityErr",
   4566			0, 0, CNTR_NORMAL,
   4567			access_rx_rbuf_csr_q_next_buf_parity_err_cnt),
   4568[C_RX_RBUF_CSR_Q_ENT_CNT_PARITY_ERR] = CNTR_ELEM("RxRbufCsrQEntCntParityErr", 0,
   4569			0, CNTR_NORMAL,
   4570			access_rx_rbuf_csr_q_ent_cnt_parity_err_cnt),
   4571[C_RX_RBUF_CSR_Q_HEAD_BUF_NUM_PARITY_ERR] = CNTR_ELEM(
   4572			"RxRbufCsrQHeadBufNumParityErr", 0, 0,
   4573			CNTR_NORMAL,
   4574			access_rx_rbuf_csr_q_head_buf_num_parity_err_cnt),
   4575[C_RX_RBUF_BLOCK_LIST_READ_COR_ERR] = CNTR_ELEM("RxRbufBlockListReadCorErr", 0,
   4576			0, CNTR_NORMAL,
   4577			access_rx_rbuf_block_list_read_cor_err_cnt),
   4578[C_RX_RBUF_BLOCK_LIST_READ_UNC_ERR] = CNTR_ELEM("RxRbufBlockListReadUncErr", 0,
   4579			0, CNTR_NORMAL,
   4580			access_rx_rbuf_block_list_read_unc_err_cnt),
   4581[C_RX_RBUF_LOOKUP_DES_COR_ERR] = CNTR_ELEM("RxRbufLookupDesCorErr", 0, 0,
   4582			CNTR_NORMAL,
   4583			access_rx_rbuf_lookup_des_cor_err_cnt),
   4584[C_RX_RBUF_LOOKUP_DES_UNC_ERR] = CNTR_ELEM("RxRbufLookupDesUncErr", 0, 0,
   4585			CNTR_NORMAL,
   4586			access_rx_rbuf_lookup_des_unc_err_cnt),
   4587[C_RX_RBUF_LOOKUP_DES_REG_UNC_COR_ERR] = CNTR_ELEM(
   4588			"RxRbufLookupDesRegUncCorErr", 0, 0,
   4589			CNTR_NORMAL,
   4590			access_rx_rbuf_lookup_des_reg_unc_cor_err_cnt),
   4591[C_RX_RBUF_LOOKUP_DES_REG_UNC_ERR] = CNTR_ELEM("RxRbufLookupDesRegUncErr", 0, 0,
   4592			CNTR_NORMAL,
   4593			access_rx_rbuf_lookup_des_reg_unc_err_cnt),
   4594[C_RX_RBUF_FREE_LIST_COR_ERR] = CNTR_ELEM("RxRbufFreeListCorErr", 0, 0,
   4595			CNTR_NORMAL,
   4596			access_rx_rbuf_free_list_cor_err_cnt),
   4597[C_RX_RBUF_FREE_LIST_UNC_ERR] = CNTR_ELEM("RxRbufFreeListUncErr", 0, 0,
   4598			CNTR_NORMAL,
   4599			access_rx_rbuf_free_list_unc_err_cnt),
   4600[C_RX_RCV_FSM_ENCODING_ERR] = CNTR_ELEM("RxRcvFsmEncodingErr", 0, 0,
   4601			CNTR_NORMAL,
   4602			access_rx_rcv_fsm_encoding_err_cnt),
   4603[C_RX_DMA_FLAG_COR_ERR] = CNTR_ELEM("RxDmaFlagCorErr", 0, 0,
   4604			CNTR_NORMAL,
   4605			access_rx_dma_flag_cor_err_cnt),
   4606[C_RX_DMA_FLAG_UNC_ERR] = CNTR_ELEM("RxDmaFlagUncErr", 0, 0,
   4607			CNTR_NORMAL,
   4608			access_rx_dma_flag_unc_err_cnt),
   4609[C_RX_DC_SOP_EOP_PARITY_ERR] = CNTR_ELEM("RxDcSopEopParityErr", 0, 0,
   4610			CNTR_NORMAL,
   4611			access_rx_dc_sop_eop_parity_err_cnt),
   4612[C_RX_RCV_CSR_PARITY_ERR] = CNTR_ELEM("RxRcvCsrParityErr", 0, 0,
   4613			CNTR_NORMAL,
   4614			access_rx_rcv_csr_parity_err_cnt),
   4615[C_RX_RCV_QP_MAP_TABLE_COR_ERR] = CNTR_ELEM("RxRcvQpMapTableCorErr", 0, 0,
   4616			CNTR_NORMAL,
   4617			access_rx_rcv_qp_map_table_cor_err_cnt),
   4618[C_RX_RCV_QP_MAP_TABLE_UNC_ERR] = CNTR_ELEM("RxRcvQpMapTableUncErr", 0, 0,
   4619			CNTR_NORMAL,
   4620			access_rx_rcv_qp_map_table_unc_err_cnt),
   4621[C_RX_RCV_DATA_COR_ERR] = CNTR_ELEM("RxRcvDataCorErr", 0, 0,
   4622			CNTR_NORMAL,
   4623			access_rx_rcv_data_cor_err_cnt),
   4624[C_RX_RCV_DATA_UNC_ERR] = CNTR_ELEM("RxRcvDataUncErr", 0, 0,
   4625			CNTR_NORMAL,
   4626			access_rx_rcv_data_unc_err_cnt),
   4627[C_RX_RCV_HDR_COR_ERR] = CNTR_ELEM("RxRcvHdrCorErr", 0, 0,
   4628			CNTR_NORMAL,
   4629			access_rx_rcv_hdr_cor_err_cnt),
   4630[C_RX_RCV_HDR_UNC_ERR] = CNTR_ELEM("RxRcvHdrUncErr", 0, 0,
   4631			CNTR_NORMAL,
   4632			access_rx_rcv_hdr_unc_err_cnt),
   4633[C_RX_DC_INTF_PARITY_ERR] = CNTR_ELEM("RxDcIntfParityErr", 0, 0,
   4634			CNTR_NORMAL,
   4635			access_rx_dc_intf_parity_err_cnt),
   4636[C_RX_DMA_CSR_COR_ERR] = CNTR_ELEM("RxDmaCsrCorErr", 0, 0,
   4637			CNTR_NORMAL,
   4638			access_rx_dma_csr_cor_err_cnt),
   4639/* SendPioErrStatus */
   4640[C_PIO_PEC_SOP_HEAD_PARITY_ERR] = CNTR_ELEM("PioPecSopHeadParityErr", 0, 0,
   4641			CNTR_NORMAL,
   4642			access_pio_pec_sop_head_parity_err_cnt),
   4643[C_PIO_PCC_SOP_HEAD_PARITY_ERR] = CNTR_ELEM("PioPccSopHeadParityErr", 0, 0,
   4644			CNTR_NORMAL,
   4645			access_pio_pcc_sop_head_parity_err_cnt),
   4646[C_PIO_LAST_RETURNED_CNT_PARITY_ERR] = CNTR_ELEM("PioLastReturnedCntParityErr",
   4647			0, 0, CNTR_NORMAL,
   4648			access_pio_last_returned_cnt_parity_err_cnt),
   4649[C_PIO_CURRENT_FREE_CNT_PARITY_ERR] = CNTR_ELEM("PioCurrentFreeCntParityErr", 0,
   4650			0, CNTR_NORMAL,
   4651			access_pio_current_free_cnt_parity_err_cnt),
   4652[C_PIO_RSVD_31_ERR] = CNTR_ELEM("Pio Reserved 31", 0, 0,
   4653			CNTR_NORMAL,
   4654			access_pio_reserved_31_err_cnt),
   4655[C_PIO_RSVD_30_ERR] = CNTR_ELEM("Pio Reserved 30", 0, 0,
   4656			CNTR_NORMAL,
   4657			access_pio_reserved_30_err_cnt),
   4658[C_PIO_PPMC_SOP_LEN_ERR] = CNTR_ELEM("PioPpmcSopLenErr", 0, 0,
   4659			CNTR_NORMAL,
   4660			access_pio_ppmc_sop_len_err_cnt),
   4661[C_PIO_PPMC_BQC_MEM_PARITY_ERR] = CNTR_ELEM("PioPpmcBqcMemParityErr", 0, 0,
   4662			CNTR_NORMAL,
   4663			access_pio_ppmc_bqc_mem_parity_err_cnt),
   4664[C_PIO_VL_FIFO_PARITY_ERR] = CNTR_ELEM("PioVlFifoParityErr", 0, 0,
   4665			CNTR_NORMAL,
   4666			access_pio_vl_fifo_parity_err_cnt),
   4667[C_PIO_VLF_SOP_PARITY_ERR] = CNTR_ELEM("PioVlfSopParityErr", 0, 0,
   4668			CNTR_NORMAL,
   4669			access_pio_vlf_sop_parity_err_cnt),
   4670[C_PIO_VLF_V1_LEN_PARITY_ERR] = CNTR_ELEM("PioVlfVlLenParityErr", 0, 0,
   4671			CNTR_NORMAL,
   4672			access_pio_vlf_v1_len_parity_err_cnt),
   4673[C_PIO_BLOCK_QW_COUNT_PARITY_ERR] = CNTR_ELEM("PioBlockQwCountParityErr", 0, 0,
   4674			CNTR_NORMAL,
   4675			access_pio_block_qw_count_parity_err_cnt),
   4676[C_PIO_WRITE_QW_VALID_PARITY_ERR] = CNTR_ELEM("PioWriteQwValidParityErr", 0, 0,
   4677			CNTR_NORMAL,
   4678			access_pio_write_qw_valid_parity_err_cnt),
   4679[C_PIO_STATE_MACHINE_ERR] = CNTR_ELEM("PioStateMachineErr", 0, 0,
   4680			CNTR_NORMAL,
   4681			access_pio_state_machine_err_cnt),
   4682[C_PIO_WRITE_DATA_PARITY_ERR] = CNTR_ELEM("PioWriteDataParityErr", 0, 0,
   4683			CNTR_NORMAL,
   4684			access_pio_write_data_parity_err_cnt),
   4685[C_PIO_HOST_ADDR_MEM_COR_ERR] = CNTR_ELEM("PioHostAddrMemCorErr", 0, 0,
   4686			CNTR_NORMAL,
   4687			access_pio_host_addr_mem_cor_err_cnt),
   4688[C_PIO_HOST_ADDR_MEM_UNC_ERR] = CNTR_ELEM("PioHostAddrMemUncErr", 0, 0,
   4689			CNTR_NORMAL,
   4690			access_pio_host_addr_mem_unc_err_cnt),
   4691[C_PIO_PKT_EVICT_SM_OR_ARM_SM_ERR] = CNTR_ELEM("PioPktEvictSmOrArbSmErr", 0, 0,
   4692			CNTR_NORMAL,
   4693			access_pio_pkt_evict_sm_or_arb_sm_err_cnt),
   4694[C_PIO_INIT_SM_IN_ERR] = CNTR_ELEM("PioInitSmInErr", 0, 0,
   4695			CNTR_NORMAL,
   4696			access_pio_init_sm_in_err_cnt),
   4697[C_PIO_PPMC_PBL_FIFO_ERR] = CNTR_ELEM("PioPpmcPblFifoErr", 0, 0,
   4698			CNTR_NORMAL,
   4699			access_pio_ppmc_pbl_fifo_err_cnt),
   4700[C_PIO_CREDIT_RET_FIFO_PARITY_ERR] = CNTR_ELEM("PioCreditRetFifoParityErr", 0,
   4701			0, CNTR_NORMAL,
   4702			access_pio_credit_ret_fifo_parity_err_cnt),
   4703[C_PIO_V1_LEN_MEM_BANK1_COR_ERR] = CNTR_ELEM("PioVlLenMemBank1CorErr", 0, 0,
   4704			CNTR_NORMAL,
   4705			access_pio_v1_len_mem_bank1_cor_err_cnt),
   4706[C_PIO_V1_LEN_MEM_BANK0_COR_ERR] = CNTR_ELEM("PioVlLenMemBank0CorErr", 0, 0,
   4707			CNTR_NORMAL,
   4708			access_pio_v1_len_mem_bank0_cor_err_cnt),
   4709[C_PIO_V1_LEN_MEM_BANK1_UNC_ERR] = CNTR_ELEM("PioVlLenMemBank1UncErr", 0, 0,
   4710			CNTR_NORMAL,
   4711			access_pio_v1_len_mem_bank1_unc_err_cnt),
   4712[C_PIO_V1_LEN_MEM_BANK0_UNC_ERR] = CNTR_ELEM("PioVlLenMemBank0UncErr", 0, 0,
   4713			CNTR_NORMAL,
   4714			access_pio_v1_len_mem_bank0_unc_err_cnt),
   4715[C_PIO_SM_PKT_RESET_PARITY_ERR] = CNTR_ELEM("PioSmPktResetParityErr", 0, 0,
   4716			CNTR_NORMAL,
   4717			access_pio_sm_pkt_reset_parity_err_cnt),
   4718[C_PIO_PKT_EVICT_FIFO_PARITY_ERR] = CNTR_ELEM("PioPktEvictFifoParityErr", 0, 0,
   4719			CNTR_NORMAL,
   4720			access_pio_pkt_evict_fifo_parity_err_cnt),
   4721[C_PIO_SBRDCTRL_CRREL_FIFO_PARITY_ERR] = CNTR_ELEM(
   4722			"PioSbrdctrlCrrelFifoParityErr", 0, 0,
   4723			CNTR_NORMAL,
   4724			access_pio_sbrdctrl_crrel_fifo_parity_err_cnt),
   4725[C_PIO_SBRDCTL_CRREL_PARITY_ERR] = CNTR_ELEM("PioSbrdctlCrrelParityErr", 0, 0,
   4726			CNTR_NORMAL,
   4727			access_pio_sbrdctl_crrel_parity_err_cnt),
   4728[C_PIO_PEC_FIFO_PARITY_ERR] = CNTR_ELEM("PioPecFifoParityErr", 0, 0,
   4729			CNTR_NORMAL,
   4730			access_pio_pec_fifo_parity_err_cnt),
   4731[C_PIO_PCC_FIFO_PARITY_ERR] = CNTR_ELEM("PioPccFifoParityErr", 0, 0,
   4732			CNTR_NORMAL,
   4733			access_pio_pcc_fifo_parity_err_cnt),
   4734[C_PIO_SB_MEM_FIFO1_ERR] = CNTR_ELEM("PioSbMemFifo1Err", 0, 0,
   4735			CNTR_NORMAL,
   4736			access_pio_sb_mem_fifo1_err_cnt),
   4737[C_PIO_SB_MEM_FIFO0_ERR] = CNTR_ELEM("PioSbMemFifo0Err", 0, 0,
   4738			CNTR_NORMAL,
   4739			access_pio_sb_mem_fifo0_err_cnt),
   4740[C_PIO_CSR_PARITY_ERR] = CNTR_ELEM("PioCsrParityErr", 0, 0,
   4741			CNTR_NORMAL,
   4742			access_pio_csr_parity_err_cnt),
   4743[C_PIO_WRITE_ADDR_PARITY_ERR] = CNTR_ELEM("PioWriteAddrParityErr", 0, 0,
   4744			CNTR_NORMAL,
   4745			access_pio_write_addr_parity_err_cnt),
   4746[C_PIO_WRITE_BAD_CTXT_ERR] = CNTR_ELEM("PioWriteBadCtxtErr", 0, 0,
   4747			CNTR_NORMAL,
   4748			access_pio_write_bad_ctxt_err_cnt),
   4749/* SendDmaErrStatus */
   4750[C_SDMA_PCIE_REQ_TRACKING_COR_ERR] = CNTR_ELEM("SDmaPcieReqTrackingCorErr", 0,
   4751			0, CNTR_NORMAL,
   4752			access_sdma_pcie_req_tracking_cor_err_cnt),
   4753[C_SDMA_PCIE_REQ_TRACKING_UNC_ERR] = CNTR_ELEM("SDmaPcieReqTrackingUncErr", 0,
   4754			0, CNTR_NORMAL,
   4755			access_sdma_pcie_req_tracking_unc_err_cnt),
   4756[C_SDMA_CSR_PARITY_ERR] = CNTR_ELEM("SDmaCsrParityErr", 0, 0,
   4757			CNTR_NORMAL,
   4758			access_sdma_csr_parity_err_cnt),
   4759[C_SDMA_RPY_TAG_ERR] = CNTR_ELEM("SDmaRpyTagErr", 0, 0,
   4760			CNTR_NORMAL,
   4761			access_sdma_rpy_tag_err_cnt),
   4762/* SendEgressErrStatus */
   4763[C_TX_READ_PIO_MEMORY_CSR_UNC_ERR] = CNTR_ELEM("TxReadPioMemoryCsrUncErr", 0, 0,
   4764			CNTR_NORMAL,
   4765			access_tx_read_pio_memory_csr_unc_err_cnt),
   4766[C_TX_READ_SDMA_MEMORY_CSR_UNC_ERR] = CNTR_ELEM("TxReadSdmaMemoryCsrUncErr", 0,
   4767			0, CNTR_NORMAL,
   4768			access_tx_read_sdma_memory_csr_err_cnt),
   4769[C_TX_EGRESS_FIFO_COR_ERR] = CNTR_ELEM("TxEgressFifoCorErr", 0, 0,
   4770			CNTR_NORMAL,
   4771			access_tx_egress_fifo_cor_err_cnt),
   4772[C_TX_READ_PIO_MEMORY_COR_ERR] = CNTR_ELEM("TxReadPioMemoryCorErr", 0, 0,
   4773			CNTR_NORMAL,
   4774			access_tx_read_pio_memory_cor_err_cnt),
   4775[C_TX_READ_SDMA_MEMORY_COR_ERR] = CNTR_ELEM("TxReadSdmaMemoryCorErr", 0, 0,
   4776			CNTR_NORMAL,
   4777			access_tx_read_sdma_memory_cor_err_cnt),
   4778[C_TX_SB_HDR_COR_ERR] = CNTR_ELEM("TxSbHdrCorErr", 0, 0,
   4779			CNTR_NORMAL,
   4780			access_tx_sb_hdr_cor_err_cnt),
   4781[C_TX_CREDIT_OVERRUN_ERR] = CNTR_ELEM("TxCreditOverrunErr", 0, 0,
   4782			CNTR_NORMAL,
   4783			access_tx_credit_overrun_err_cnt),
   4784[C_TX_LAUNCH_FIFO8_COR_ERR] = CNTR_ELEM("TxLaunchFifo8CorErr", 0, 0,
   4785			CNTR_NORMAL,
   4786			access_tx_launch_fifo8_cor_err_cnt),
   4787[C_TX_LAUNCH_FIFO7_COR_ERR] = CNTR_ELEM("TxLaunchFifo7CorErr", 0, 0,
   4788			CNTR_NORMAL,
   4789			access_tx_launch_fifo7_cor_err_cnt),
   4790[C_TX_LAUNCH_FIFO6_COR_ERR] = CNTR_ELEM("TxLaunchFifo6CorErr", 0, 0,
   4791			CNTR_NORMAL,
   4792			access_tx_launch_fifo6_cor_err_cnt),
   4793[C_TX_LAUNCH_FIFO5_COR_ERR] = CNTR_ELEM("TxLaunchFifo5CorErr", 0, 0,
   4794			CNTR_NORMAL,
   4795			access_tx_launch_fifo5_cor_err_cnt),
   4796[C_TX_LAUNCH_FIFO4_COR_ERR] = CNTR_ELEM("TxLaunchFifo4CorErr", 0, 0,
   4797			CNTR_NORMAL,
   4798			access_tx_launch_fifo4_cor_err_cnt),
   4799[C_TX_LAUNCH_FIFO3_COR_ERR] = CNTR_ELEM("TxLaunchFifo3CorErr", 0, 0,
   4800			CNTR_NORMAL,
   4801			access_tx_launch_fifo3_cor_err_cnt),
   4802[C_TX_LAUNCH_FIFO2_COR_ERR] = CNTR_ELEM("TxLaunchFifo2CorErr", 0, 0,
   4803			CNTR_NORMAL,
   4804			access_tx_launch_fifo2_cor_err_cnt),
   4805[C_TX_LAUNCH_FIFO1_COR_ERR] = CNTR_ELEM("TxLaunchFifo1CorErr", 0, 0,
   4806			CNTR_NORMAL,
   4807			access_tx_launch_fifo1_cor_err_cnt),
   4808[C_TX_LAUNCH_FIFO0_COR_ERR] = CNTR_ELEM("TxLaunchFifo0CorErr", 0, 0,
   4809			CNTR_NORMAL,
   4810			access_tx_launch_fifo0_cor_err_cnt),
   4811[C_TX_CREDIT_RETURN_VL_ERR] = CNTR_ELEM("TxCreditReturnVLErr", 0, 0,
   4812			CNTR_NORMAL,
   4813			access_tx_credit_return_vl_err_cnt),
   4814[C_TX_HCRC_INSERTION_ERR] = CNTR_ELEM("TxHcrcInsertionErr", 0, 0,
   4815			CNTR_NORMAL,
   4816			access_tx_hcrc_insertion_err_cnt),
   4817[C_TX_EGRESS_FIFI_UNC_ERR] = CNTR_ELEM("TxEgressFifoUncErr", 0, 0,
   4818			CNTR_NORMAL,
   4819			access_tx_egress_fifo_unc_err_cnt),
   4820[C_TX_READ_PIO_MEMORY_UNC_ERR] = CNTR_ELEM("TxReadPioMemoryUncErr", 0, 0,
   4821			CNTR_NORMAL,
   4822			access_tx_read_pio_memory_unc_err_cnt),
   4823[C_TX_READ_SDMA_MEMORY_UNC_ERR] = CNTR_ELEM("TxReadSdmaMemoryUncErr", 0, 0,
   4824			CNTR_NORMAL,
   4825			access_tx_read_sdma_memory_unc_err_cnt),
   4826[C_TX_SB_HDR_UNC_ERR] = CNTR_ELEM("TxSbHdrUncErr", 0, 0,
   4827			CNTR_NORMAL,
   4828			access_tx_sb_hdr_unc_err_cnt),
   4829[C_TX_CREDIT_RETURN_PARITY_ERR] = CNTR_ELEM("TxCreditReturnParityErr", 0, 0,
   4830			CNTR_NORMAL,
   4831			access_tx_credit_return_partiy_err_cnt),
   4832[C_TX_LAUNCH_FIFO8_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo8UncOrParityErr",
   4833			0, 0, CNTR_NORMAL,
   4834			access_tx_launch_fifo8_unc_or_parity_err_cnt),
   4835[C_TX_LAUNCH_FIFO7_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo7UncOrParityErr",
   4836			0, 0, CNTR_NORMAL,
   4837			access_tx_launch_fifo7_unc_or_parity_err_cnt),
   4838[C_TX_LAUNCH_FIFO6_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo6UncOrParityErr",
   4839			0, 0, CNTR_NORMAL,
   4840			access_tx_launch_fifo6_unc_or_parity_err_cnt),
   4841[C_TX_LAUNCH_FIFO5_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo5UncOrParityErr",
   4842			0, 0, CNTR_NORMAL,
   4843			access_tx_launch_fifo5_unc_or_parity_err_cnt),
   4844[C_TX_LAUNCH_FIFO4_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo4UncOrParityErr",
   4845			0, 0, CNTR_NORMAL,
   4846			access_tx_launch_fifo4_unc_or_parity_err_cnt),
   4847[C_TX_LAUNCH_FIFO3_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo3UncOrParityErr",
   4848			0, 0, CNTR_NORMAL,
   4849			access_tx_launch_fifo3_unc_or_parity_err_cnt),
   4850[C_TX_LAUNCH_FIFO2_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo2UncOrParityErr",
   4851			0, 0, CNTR_NORMAL,
   4852			access_tx_launch_fifo2_unc_or_parity_err_cnt),
   4853[C_TX_LAUNCH_FIFO1_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo1UncOrParityErr",
   4854			0, 0, CNTR_NORMAL,
   4855			access_tx_launch_fifo1_unc_or_parity_err_cnt),
   4856[C_TX_LAUNCH_FIFO0_UNC_OR_PARITY_ERR] = CNTR_ELEM("TxLaunchFifo0UncOrParityErr",
   4857			0, 0, CNTR_NORMAL,
   4858			access_tx_launch_fifo0_unc_or_parity_err_cnt),
   4859[C_TX_SDMA15_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma15DisallowedPacketErr",
   4860			0, 0, CNTR_NORMAL,
   4861			access_tx_sdma15_disallowed_packet_err_cnt),
   4862[C_TX_SDMA14_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma14DisallowedPacketErr",
   4863			0, 0, CNTR_NORMAL,
   4864			access_tx_sdma14_disallowed_packet_err_cnt),
   4865[C_TX_SDMA13_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma13DisallowedPacketErr",
   4866			0, 0, CNTR_NORMAL,
   4867			access_tx_sdma13_disallowed_packet_err_cnt),
   4868[C_TX_SDMA12_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma12DisallowedPacketErr",
   4869			0, 0, CNTR_NORMAL,
   4870			access_tx_sdma12_disallowed_packet_err_cnt),
   4871[C_TX_SDMA11_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma11DisallowedPacketErr",
   4872			0, 0, CNTR_NORMAL,
   4873			access_tx_sdma11_disallowed_packet_err_cnt),
   4874[C_TX_SDMA10_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma10DisallowedPacketErr",
   4875			0, 0, CNTR_NORMAL,
   4876			access_tx_sdma10_disallowed_packet_err_cnt),
   4877[C_TX_SDMA9_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma9DisallowedPacketErr",
   4878			0, 0, CNTR_NORMAL,
   4879			access_tx_sdma9_disallowed_packet_err_cnt),
   4880[C_TX_SDMA8_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma8DisallowedPacketErr",
   4881			0, 0, CNTR_NORMAL,
   4882			access_tx_sdma8_disallowed_packet_err_cnt),
   4883[C_TX_SDMA7_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma7DisallowedPacketErr",
   4884			0, 0, CNTR_NORMAL,
   4885			access_tx_sdma7_disallowed_packet_err_cnt),
   4886[C_TX_SDMA6_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma6DisallowedPacketErr",
   4887			0, 0, CNTR_NORMAL,
   4888			access_tx_sdma6_disallowed_packet_err_cnt),
   4889[C_TX_SDMA5_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma5DisallowedPacketErr",
   4890			0, 0, CNTR_NORMAL,
   4891			access_tx_sdma5_disallowed_packet_err_cnt),
   4892[C_TX_SDMA4_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma4DisallowedPacketErr",
   4893			0, 0, CNTR_NORMAL,
   4894			access_tx_sdma4_disallowed_packet_err_cnt),
   4895[C_TX_SDMA3_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma3DisallowedPacketErr",
   4896			0, 0, CNTR_NORMAL,
   4897			access_tx_sdma3_disallowed_packet_err_cnt),
   4898[C_TX_SDMA2_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma2DisallowedPacketErr",
   4899			0, 0, CNTR_NORMAL,
   4900			access_tx_sdma2_disallowed_packet_err_cnt),
   4901[C_TX_SDMA1_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma1DisallowedPacketErr",
   4902			0, 0, CNTR_NORMAL,
   4903			access_tx_sdma1_disallowed_packet_err_cnt),
   4904[C_TX_SDMA0_DISALLOWED_PACKET_ERR] = CNTR_ELEM("TxSdma0DisallowedPacketErr",
   4905			0, 0, CNTR_NORMAL,
   4906			access_tx_sdma0_disallowed_packet_err_cnt),
   4907[C_TX_CONFIG_PARITY_ERR] = CNTR_ELEM("TxConfigParityErr", 0, 0,
   4908			CNTR_NORMAL,
   4909			access_tx_config_parity_err_cnt),
   4910[C_TX_SBRD_CTL_CSR_PARITY_ERR] = CNTR_ELEM("TxSbrdCtlCsrParityErr", 0, 0,
   4911			CNTR_NORMAL,
   4912			access_tx_sbrd_ctl_csr_parity_err_cnt),
   4913[C_TX_LAUNCH_CSR_PARITY_ERR] = CNTR_ELEM("TxLaunchCsrParityErr", 0, 0,
   4914			CNTR_NORMAL,
   4915			access_tx_launch_csr_parity_err_cnt),
   4916[C_TX_ILLEGAL_CL_ERR] = CNTR_ELEM("TxIllegalVLErr", 0, 0,
   4917			CNTR_NORMAL,
   4918			access_tx_illegal_vl_err_cnt),
   4919[C_TX_SBRD_CTL_STATE_MACHINE_PARITY_ERR] = CNTR_ELEM(
   4920			"TxSbrdCtlStateMachineParityErr", 0, 0,
   4921			CNTR_NORMAL,
   4922			access_tx_sbrd_ctl_state_machine_parity_err_cnt),
   4923[C_TX_RESERVED_10] = CNTR_ELEM("Tx Egress Reserved 10", 0, 0,
   4924			CNTR_NORMAL,
   4925			access_egress_reserved_10_err_cnt),
   4926[C_TX_RESERVED_9] = CNTR_ELEM("Tx Egress Reserved 9", 0, 0,
   4927			CNTR_NORMAL,
   4928			access_egress_reserved_9_err_cnt),
   4929[C_TX_SDMA_LAUNCH_INTF_PARITY_ERR] = CNTR_ELEM("TxSdmaLaunchIntfParityErr",
   4930			0, 0, CNTR_NORMAL,
   4931			access_tx_sdma_launch_intf_parity_err_cnt),
   4932[C_TX_PIO_LAUNCH_INTF_PARITY_ERR] = CNTR_ELEM("TxPioLaunchIntfParityErr", 0, 0,
   4933			CNTR_NORMAL,
   4934			access_tx_pio_launch_intf_parity_err_cnt),
   4935[C_TX_RESERVED_6] = CNTR_ELEM("Tx Egress Reserved 6", 0, 0,
   4936			CNTR_NORMAL,
   4937			access_egress_reserved_6_err_cnt),
   4938[C_TX_INCORRECT_LINK_STATE_ERR] = CNTR_ELEM("TxIncorrectLinkStateErr", 0, 0,
   4939			CNTR_NORMAL,
   4940			access_tx_incorrect_link_state_err_cnt),
   4941[C_TX_LINK_DOWN_ERR] = CNTR_ELEM("TxLinkdownErr", 0, 0,
   4942			CNTR_NORMAL,
   4943			access_tx_linkdown_err_cnt),
   4944[C_TX_EGRESS_FIFO_UNDERRUN_OR_PARITY_ERR] = CNTR_ELEM(
   4945			"EgressFifoUnderrunOrParityErr", 0, 0,
   4946			CNTR_NORMAL,
   4947			access_tx_egress_fifi_underrun_or_parity_err_cnt),
   4948[C_TX_RESERVED_2] = CNTR_ELEM("Tx Egress Reserved 2", 0, 0,
   4949			CNTR_NORMAL,
   4950			access_egress_reserved_2_err_cnt),
   4951[C_TX_PKT_INTEGRITY_MEM_UNC_ERR] = CNTR_ELEM("TxPktIntegrityMemUncErr", 0, 0,
   4952			CNTR_NORMAL,
   4953			access_tx_pkt_integrity_mem_unc_err_cnt),
   4954[C_TX_PKT_INTEGRITY_MEM_COR_ERR] = CNTR_ELEM("TxPktIntegrityMemCorErr", 0, 0,
   4955			CNTR_NORMAL,
   4956			access_tx_pkt_integrity_mem_cor_err_cnt),
   4957/* SendErrStatus */
   4958[C_SEND_CSR_WRITE_BAD_ADDR_ERR] = CNTR_ELEM("SendCsrWriteBadAddrErr", 0, 0,
   4959			CNTR_NORMAL,
   4960			access_send_csr_write_bad_addr_err_cnt),
   4961[C_SEND_CSR_READ_BAD_ADD_ERR] = CNTR_ELEM("SendCsrReadBadAddrErr", 0, 0,
   4962			CNTR_NORMAL,
   4963			access_send_csr_read_bad_addr_err_cnt),
   4964[C_SEND_CSR_PARITY_ERR] = CNTR_ELEM("SendCsrParityErr", 0, 0,
   4965			CNTR_NORMAL,
   4966			access_send_csr_parity_cnt),
   4967/* SendCtxtErrStatus */
   4968[C_PIO_WRITE_OUT_OF_BOUNDS_ERR] = CNTR_ELEM("PioWriteOutOfBoundsErr", 0, 0,
   4969			CNTR_NORMAL,
   4970			access_pio_write_out_of_bounds_err_cnt),
   4971[C_PIO_WRITE_OVERFLOW_ERR] = CNTR_ELEM("PioWriteOverflowErr", 0, 0,
   4972			CNTR_NORMAL,
   4973			access_pio_write_overflow_err_cnt),
   4974[C_PIO_WRITE_CROSSES_BOUNDARY_ERR] = CNTR_ELEM("PioWriteCrossesBoundaryErr",
   4975			0, 0, CNTR_NORMAL,
   4976			access_pio_write_crosses_boundary_err_cnt),
   4977[C_PIO_DISALLOWED_PACKET_ERR] = CNTR_ELEM("PioDisallowedPacketErr", 0, 0,
   4978			CNTR_NORMAL,
   4979			access_pio_disallowed_packet_err_cnt),
   4980[C_PIO_INCONSISTENT_SOP_ERR] = CNTR_ELEM("PioInconsistentSopErr", 0, 0,
   4981			CNTR_NORMAL,
   4982			access_pio_inconsistent_sop_err_cnt),
   4983/* SendDmaEngErrStatus */
   4984[C_SDMA_HEADER_REQUEST_FIFO_COR_ERR] = CNTR_ELEM("SDmaHeaderRequestFifoCorErr",
   4985			0, 0, CNTR_NORMAL,
   4986			access_sdma_header_request_fifo_cor_err_cnt),
   4987[C_SDMA_HEADER_STORAGE_COR_ERR] = CNTR_ELEM("SDmaHeaderStorageCorErr", 0, 0,
   4988			CNTR_NORMAL,
   4989			access_sdma_header_storage_cor_err_cnt),
   4990[C_SDMA_PACKET_TRACKING_COR_ERR] = CNTR_ELEM("SDmaPacketTrackingCorErr", 0, 0,
   4991			CNTR_NORMAL,
   4992			access_sdma_packet_tracking_cor_err_cnt),
   4993[C_SDMA_ASSEMBLY_COR_ERR] = CNTR_ELEM("SDmaAssemblyCorErr", 0, 0,
   4994			CNTR_NORMAL,
   4995			access_sdma_assembly_cor_err_cnt),
   4996[C_SDMA_DESC_TABLE_COR_ERR] = CNTR_ELEM("SDmaDescTableCorErr", 0, 0,
   4997			CNTR_NORMAL,
   4998			access_sdma_desc_table_cor_err_cnt),
   4999[C_SDMA_HEADER_REQUEST_FIFO_UNC_ERR] = CNTR_ELEM("SDmaHeaderRequestFifoUncErr",
   5000			0, 0, CNTR_NORMAL,
   5001			access_sdma_header_request_fifo_unc_err_cnt),
   5002[C_SDMA_HEADER_STORAGE_UNC_ERR] = CNTR_ELEM("SDmaHeaderStorageUncErr", 0, 0,
   5003			CNTR_NORMAL,
   5004			access_sdma_header_storage_unc_err_cnt),
   5005[C_SDMA_PACKET_TRACKING_UNC_ERR] = CNTR_ELEM("SDmaPacketTrackingUncErr", 0, 0,
   5006			CNTR_NORMAL,
   5007			access_sdma_packet_tracking_unc_err_cnt),
   5008[C_SDMA_ASSEMBLY_UNC_ERR] = CNTR_ELEM("SDmaAssemblyUncErr", 0, 0,
   5009			CNTR_NORMAL,
   5010			access_sdma_assembly_unc_err_cnt),
   5011[C_SDMA_DESC_TABLE_UNC_ERR] = CNTR_ELEM("SDmaDescTableUncErr", 0, 0,
   5012			CNTR_NORMAL,
   5013			access_sdma_desc_table_unc_err_cnt),
   5014[C_SDMA_TIMEOUT_ERR] = CNTR_ELEM("SDmaTimeoutErr", 0, 0,
   5015			CNTR_NORMAL,
   5016			access_sdma_timeout_err_cnt),
   5017[C_SDMA_HEADER_LENGTH_ERR] = CNTR_ELEM("SDmaHeaderLengthErr", 0, 0,
   5018			CNTR_NORMAL,
   5019			access_sdma_header_length_err_cnt),
   5020[C_SDMA_HEADER_ADDRESS_ERR] = CNTR_ELEM("SDmaHeaderAddressErr", 0, 0,
   5021			CNTR_NORMAL,
   5022			access_sdma_header_address_err_cnt),
   5023[C_SDMA_HEADER_SELECT_ERR] = CNTR_ELEM("SDmaHeaderSelectErr", 0, 0,
   5024			CNTR_NORMAL,
   5025			access_sdma_header_select_err_cnt),
   5026[C_SMDA_RESERVED_9] = CNTR_ELEM("SDma Reserved 9", 0, 0,
   5027			CNTR_NORMAL,
   5028			access_sdma_reserved_9_err_cnt),
   5029[C_SDMA_PACKET_DESC_OVERFLOW_ERR] = CNTR_ELEM("SDmaPacketDescOverflowErr", 0, 0,
   5030			CNTR_NORMAL,
   5031			access_sdma_packet_desc_overflow_err_cnt),
   5032[C_SDMA_LENGTH_MISMATCH_ERR] = CNTR_ELEM("SDmaLengthMismatchErr", 0, 0,
   5033			CNTR_NORMAL,
   5034			access_sdma_length_mismatch_err_cnt),
   5035[C_SDMA_HALT_ERR] = CNTR_ELEM("SDmaHaltErr", 0, 0,
   5036			CNTR_NORMAL,
   5037			access_sdma_halt_err_cnt),
   5038[C_SDMA_MEM_READ_ERR] = CNTR_ELEM("SDmaMemReadErr", 0, 0,
   5039			CNTR_NORMAL,
   5040			access_sdma_mem_read_err_cnt),
   5041[C_SDMA_FIRST_DESC_ERR] = CNTR_ELEM("SDmaFirstDescErr", 0, 0,
   5042			CNTR_NORMAL,
   5043			access_sdma_first_desc_err_cnt),
   5044[C_SDMA_TAIL_OUT_OF_BOUNDS_ERR] = CNTR_ELEM("SDmaTailOutOfBoundsErr", 0, 0,
   5045			CNTR_NORMAL,
   5046			access_sdma_tail_out_of_bounds_err_cnt),
   5047[C_SDMA_TOO_LONG_ERR] = CNTR_ELEM("SDmaTooLongErr", 0, 0,
   5048			CNTR_NORMAL,
   5049			access_sdma_too_long_err_cnt),
   5050[C_SDMA_GEN_MISMATCH_ERR] = CNTR_ELEM("SDmaGenMismatchErr", 0, 0,
   5051			CNTR_NORMAL,
   5052			access_sdma_gen_mismatch_err_cnt),
   5053[C_SDMA_WRONG_DW_ERR] = CNTR_ELEM("SDmaWrongDwErr", 0, 0,
   5054			CNTR_NORMAL,
   5055			access_sdma_wrong_dw_err_cnt),
   5056};
   5057
   5058static struct cntr_entry port_cntrs[PORT_CNTR_LAST] = {
   5059[C_TX_UNSUP_VL] = TXE32_PORT_CNTR_ELEM(TxUnVLErr, SEND_UNSUP_VL_ERR_CNT,
   5060			CNTR_NORMAL),
   5061[C_TX_INVAL_LEN] = TXE32_PORT_CNTR_ELEM(TxInvalLen, SEND_LEN_ERR_CNT,
   5062			CNTR_NORMAL),
   5063[C_TX_MM_LEN_ERR] = TXE32_PORT_CNTR_ELEM(TxMMLenErr, SEND_MAX_MIN_LEN_ERR_CNT,
   5064			CNTR_NORMAL),
   5065[C_TX_UNDERRUN] = TXE32_PORT_CNTR_ELEM(TxUnderrun, SEND_UNDERRUN_CNT,
   5066			CNTR_NORMAL),
   5067[C_TX_FLOW_STALL] = TXE32_PORT_CNTR_ELEM(TxFlowStall, SEND_FLOW_STALL_CNT,
   5068			CNTR_NORMAL),
   5069[C_TX_DROPPED] = TXE32_PORT_CNTR_ELEM(TxDropped, SEND_DROPPED_PKT_CNT,
   5070			CNTR_NORMAL),
   5071[C_TX_HDR_ERR] = TXE32_PORT_CNTR_ELEM(TxHdrErr, SEND_HEADERS_ERR_CNT,
   5072			CNTR_NORMAL),
   5073[C_TX_PKT] = TXE64_PORT_CNTR_ELEM(TxPkt, SEND_DATA_PKT_CNT, CNTR_NORMAL),
   5074[C_TX_WORDS] = TXE64_PORT_CNTR_ELEM(TxWords, SEND_DWORD_CNT, CNTR_NORMAL),
   5075[C_TX_WAIT] = TXE64_PORT_CNTR_ELEM(TxWait, SEND_WAIT_CNT, CNTR_SYNTH),
   5076[C_TX_FLIT_VL] = TXE64_PORT_CNTR_ELEM(TxFlitVL, SEND_DATA_VL0_CNT,
   5077				      CNTR_SYNTH | CNTR_VL),
   5078[C_TX_PKT_VL] = TXE64_PORT_CNTR_ELEM(TxPktVL, SEND_DATA_PKT_VL0_CNT,
   5079				     CNTR_SYNTH | CNTR_VL),
   5080[C_TX_WAIT_VL] = TXE64_PORT_CNTR_ELEM(TxWaitVL, SEND_WAIT_VL0_CNT,
   5081				      CNTR_SYNTH | CNTR_VL),
   5082[C_RX_PKT] = RXE64_PORT_CNTR_ELEM(RxPkt, RCV_DATA_PKT_CNT, CNTR_NORMAL),
   5083[C_RX_WORDS] = RXE64_PORT_CNTR_ELEM(RxWords, RCV_DWORD_CNT, CNTR_NORMAL),
   5084[C_SW_LINK_DOWN] = CNTR_ELEM("SwLinkDown", 0, 0, CNTR_SYNTH | CNTR_32BIT,
   5085			     access_sw_link_dn_cnt),
   5086[C_SW_LINK_UP] = CNTR_ELEM("SwLinkUp", 0, 0, CNTR_SYNTH | CNTR_32BIT,
   5087			   access_sw_link_up_cnt),
   5088[C_SW_UNKNOWN_FRAME] = CNTR_ELEM("UnknownFrame", 0, 0, CNTR_NORMAL,
   5089				 access_sw_unknown_frame_cnt),
   5090[C_SW_XMIT_DSCD] = CNTR_ELEM("XmitDscd", 0, 0, CNTR_SYNTH | CNTR_32BIT,
   5091			     access_sw_xmit_discards),
   5092[C_SW_XMIT_DSCD_VL] = CNTR_ELEM("XmitDscdVl", 0, 0,
   5093				CNTR_SYNTH | CNTR_32BIT | CNTR_VL,
   5094				access_sw_xmit_discards),
   5095[C_SW_XMIT_CSTR_ERR] = CNTR_ELEM("XmitCstrErr", 0, 0, CNTR_SYNTH,
   5096				 access_xmit_constraint_errs),
   5097[C_SW_RCV_CSTR_ERR] = CNTR_ELEM("RcvCstrErr", 0, 0, CNTR_SYNTH,
   5098				access_rcv_constraint_errs),
   5099[C_SW_IBP_LOOP_PKTS] = SW_IBP_CNTR(LoopPkts, loop_pkts),
   5100[C_SW_IBP_RC_RESENDS] = SW_IBP_CNTR(RcResend, rc_resends),
   5101[C_SW_IBP_RNR_NAKS] = SW_IBP_CNTR(RnrNak, rnr_naks),
   5102[C_SW_IBP_OTHER_NAKS] = SW_IBP_CNTR(OtherNak, other_naks),
   5103[C_SW_IBP_RC_TIMEOUTS] = SW_IBP_CNTR(RcTimeOut, rc_timeouts),
   5104[C_SW_IBP_PKT_DROPS] = SW_IBP_CNTR(PktDrop, pkt_drops),
   5105[C_SW_IBP_DMA_WAIT] = SW_IBP_CNTR(DmaWait, dmawait),
   5106[C_SW_IBP_RC_SEQNAK] = SW_IBP_CNTR(RcSeqNak, rc_seqnak),
   5107[C_SW_IBP_RC_DUPREQ] = SW_IBP_CNTR(RcDupRew, rc_dupreq),
   5108[C_SW_IBP_RDMA_SEQ] = SW_IBP_CNTR(RdmaSeq, rdma_seq),
   5109[C_SW_IBP_UNALIGNED] = SW_IBP_CNTR(Unaligned, unaligned),
   5110[C_SW_IBP_SEQ_NAK] = SW_IBP_CNTR(SeqNak, seq_naks),
   5111[C_SW_IBP_RC_CRWAITS] = SW_IBP_CNTR(RcCrWait, rc_crwaits),
   5112[C_SW_CPU_RC_ACKS] = CNTR_ELEM("RcAcks", 0, 0, CNTR_NORMAL,
   5113			       access_sw_cpu_rc_acks),
   5114[C_SW_CPU_RC_QACKS] = CNTR_ELEM("RcQacks", 0, 0, CNTR_NORMAL,
   5115				access_sw_cpu_rc_qacks),
   5116[C_SW_CPU_RC_DELAYED_COMP] = CNTR_ELEM("RcDelayComp", 0, 0, CNTR_NORMAL,
   5117				       access_sw_cpu_rc_delayed_comp),
   5118[OVR_LBL(0)] = OVR_ELM(0), [OVR_LBL(1)] = OVR_ELM(1),
   5119[OVR_LBL(2)] = OVR_ELM(2), [OVR_LBL(3)] = OVR_ELM(3),
   5120[OVR_LBL(4)] = OVR_ELM(4), [OVR_LBL(5)] = OVR_ELM(5),
   5121[OVR_LBL(6)] = OVR_ELM(6), [OVR_LBL(7)] = OVR_ELM(7),
   5122[OVR_LBL(8)] = OVR_ELM(8), [OVR_LBL(9)] = OVR_ELM(9),
   5123[OVR_LBL(10)] = OVR_ELM(10), [OVR_LBL(11)] = OVR_ELM(11),
   5124[OVR_LBL(12)] = OVR_ELM(12), [OVR_LBL(13)] = OVR_ELM(13),
   5125[OVR_LBL(14)] = OVR_ELM(14), [OVR_LBL(15)] = OVR_ELM(15),
   5126[OVR_LBL(16)] = OVR_ELM(16), [OVR_LBL(17)] = OVR_ELM(17),
   5127[OVR_LBL(18)] = OVR_ELM(18), [OVR_LBL(19)] = OVR_ELM(19),
   5128[OVR_LBL(20)] = OVR_ELM(20), [OVR_LBL(21)] = OVR_ELM(21),
   5129[OVR_LBL(22)] = OVR_ELM(22), [OVR_LBL(23)] = OVR_ELM(23),
   5130[OVR_LBL(24)] = OVR_ELM(24), [OVR_LBL(25)] = OVR_ELM(25),
   5131[OVR_LBL(26)] = OVR_ELM(26), [OVR_LBL(27)] = OVR_ELM(27),
   5132[OVR_LBL(28)] = OVR_ELM(28), [OVR_LBL(29)] = OVR_ELM(29),
   5133[OVR_LBL(30)] = OVR_ELM(30), [OVR_LBL(31)] = OVR_ELM(31),
   5134[OVR_LBL(32)] = OVR_ELM(32), [OVR_LBL(33)] = OVR_ELM(33),
   5135[OVR_LBL(34)] = OVR_ELM(34), [OVR_LBL(35)] = OVR_ELM(35),
   5136[OVR_LBL(36)] = OVR_ELM(36), [OVR_LBL(37)] = OVR_ELM(37),
   5137[OVR_LBL(38)] = OVR_ELM(38), [OVR_LBL(39)] = OVR_ELM(39),
   5138[OVR_LBL(40)] = OVR_ELM(40), [OVR_LBL(41)] = OVR_ELM(41),
   5139[OVR_LBL(42)] = OVR_ELM(42), [OVR_LBL(43)] = OVR_ELM(43),
   5140[OVR_LBL(44)] = OVR_ELM(44), [OVR_LBL(45)] = OVR_ELM(45),
   5141[OVR_LBL(46)] = OVR_ELM(46), [OVR_LBL(47)] = OVR_ELM(47),
   5142[OVR_LBL(48)] = OVR_ELM(48), [OVR_LBL(49)] = OVR_ELM(49),
   5143[OVR_LBL(50)] = OVR_ELM(50), [OVR_LBL(51)] = OVR_ELM(51),
   5144[OVR_LBL(52)] = OVR_ELM(52), [OVR_LBL(53)] = OVR_ELM(53),
   5145[OVR_LBL(54)] = OVR_ELM(54), [OVR_LBL(55)] = OVR_ELM(55),
   5146[OVR_LBL(56)] = OVR_ELM(56), [OVR_LBL(57)] = OVR_ELM(57),
   5147[OVR_LBL(58)] = OVR_ELM(58), [OVR_LBL(59)] = OVR_ELM(59),
   5148[OVR_LBL(60)] = OVR_ELM(60), [OVR_LBL(61)] = OVR_ELM(61),
   5149[OVR_LBL(62)] = OVR_ELM(62), [OVR_LBL(63)] = OVR_ELM(63),
   5150[OVR_LBL(64)] = OVR_ELM(64), [OVR_LBL(65)] = OVR_ELM(65),
   5151[OVR_LBL(66)] = OVR_ELM(66), [OVR_LBL(67)] = OVR_ELM(67),
   5152[OVR_LBL(68)] = OVR_ELM(68), [OVR_LBL(69)] = OVR_ELM(69),
   5153[OVR_LBL(70)] = OVR_ELM(70), [OVR_LBL(71)] = OVR_ELM(71),
   5154[OVR_LBL(72)] = OVR_ELM(72), [OVR_LBL(73)] = OVR_ELM(73),
   5155[OVR_LBL(74)] = OVR_ELM(74), [OVR_LBL(75)] = OVR_ELM(75),
   5156[OVR_LBL(76)] = OVR_ELM(76), [OVR_LBL(77)] = OVR_ELM(77),
   5157[OVR_LBL(78)] = OVR_ELM(78), [OVR_LBL(79)] = OVR_ELM(79),
   5158[OVR_LBL(80)] = OVR_ELM(80), [OVR_LBL(81)] = OVR_ELM(81),
   5159[OVR_LBL(82)] = OVR_ELM(82), [OVR_LBL(83)] = OVR_ELM(83),
   5160[OVR_LBL(84)] = OVR_ELM(84), [OVR_LBL(85)] = OVR_ELM(85),
   5161[OVR_LBL(86)] = OVR_ELM(86), [OVR_LBL(87)] = OVR_ELM(87),
   5162[OVR_LBL(88)] = OVR_ELM(88), [OVR_LBL(89)] = OVR_ELM(89),
   5163[OVR_LBL(90)] = OVR_ELM(90), [OVR_LBL(91)] = OVR_ELM(91),
   5164[OVR_LBL(92)] = OVR_ELM(92), [OVR_LBL(93)] = OVR_ELM(93),
   5165[OVR_LBL(94)] = OVR_ELM(94), [OVR_LBL(95)] = OVR_ELM(95),
   5166[OVR_LBL(96)] = OVR_ELM(96), [OVR_LBL(97)] = OVR_ELM(97),
   5167[OVR_LBL(98)] = OVR_ELM(98), [OVR_LBL(99)] = OVR_ELM(99),
   5168[OVR_LBL(100)] = OVR_ELM(100), [OVR_LBL(101)] = OVR_ELM(101),
   5169[OVR_LBL(102)] = OVR_ELM(102), [OVR_LBL(103)] = OVR_ELM(103),
   5170[OVR_LBL(104)] = OVR_ELM(104), [OVR_LBL(105)] = OVR_ELM(105),
   5171[OVR_LBL(106)] = OVR_ELM(106), [OVR_LBL(107)] = OVR_ELM(107),
   5172[OVR_LBL(108)] = OVR_ELM(108), [OVR_LBL(109)] = OVR_ELM(109),
   5173[OVR_LBL(110)] = OVR_ELM(110), [OVR_LBL(111)] = OVR_ELM(111),
   5174[OVR_LBL(112)] = OVR_ELM(112), [OVR_LBL(113)] = OVR_ELM(113),
   5175[OVR_LBL(114)] = OVR_ELM(114), [OVR_LBL(115)] = OVR_ELM(115),
   5176[OVR_LBL(116)] = OVR_ELM(116), [OVR_LBL(117)] = OVR_ELM(117),
   5177[OVR_LBL(118)] = OVR_ELM(118), [OVR_LBL(119)] = OVR_ELM(119),
   5178[OVR_LBL(120)] = OVR_ELM(120), [OVR_LBL(121)] = OVR_ELM(121),
   5179[OVR_LBL(122)] = OVR_ELM(122), [OVR_LBL(123)] = OVR_ELM(123),
   5180[OVR_LBL(124)] = OVR_ELM(124), [OVR_LBL(125)] = OVR_ELM(125),
   5181[OVR_LBL(126)] = OVR_ELM(126), [OVR_LBL(127)] = OVR_ELM(127),
   5182[OVR_LBL(128)] = OVR_ELM(128), [OVR_LBL(129)] = OVR_ELM(129),
   5183[OVR_LBL(130)] = OVR_ELM(130), [OVR_LBL(131)] = OVR_ELM(131),
   5184[OVR_LBL(132)] = OVR_ELM(132), [OVR_LBL(133)] = OVR_ELM(133),
   5185[OVR_LBL(134)] = OVR_ELM(134), [OVR_LBL(135)] = OVR_ELM(135),
   5186[OVR_LBL(136)] = OVR_ELM(136), [OVR_LBL(137)] = OVR_ELM(137),
   5187[OVR_LBL(138)] = OVR_ELM(138), [OVR_LBL(139)] = OVR_ELM(139),
   5188[OVR_LBL(140)] = OVR_ELM(140), [OVR_LBL(141)] = OVR_ELM(141),
   5189[OVR_LBL(142)] = OVR_ELM(142), [OVR_LBL(143)] = OVR_ELM(143),
   5190[OVR_LBL(144)] = OVR_ELM(144), [OVR_LBL(145)] = OVR_ELM(145),
   5191[OVR_LBL(146)] = OVR_ELM(146), [OVR_LBL(147)] = OVR_ELM(147),
   5192[OVR_LBL(148)] = OVR_ELM(148), [OVR_LBL(149)] = OVR_ELM(149),
   5193[OVR_LBL(150)] = OVR_ELM(150), [OVR_LBL(151)] = OVR_ELM(151),
   5194[OVR_LBL(152)] = OVR_ELM(152), [OVR_LBL(153)] = OVR_ELM(153),
   5195[OVR_LBL(154)] = OVR_ELM(154), [OVR_LBL(155)] = OVR_ELM(155),
   5196[OVR_LBL(156)] = OVR_ELM(156), [OVR_LBL(157)] = OVR_ELM(157),
   5197[OVR_LBL(158)] = OVR_ELM(158), [OVR_LBL(159)] = OVR_ELM(159),
   5198};
   5199
   5200/* ======================================================================== */
   5201
   5202/* return true if this is chip revision revision a */
   5203int is_ax(struct hfi1_devdata *dd)
   5204{
   5205	u8 chip_rev_minor =
   5206		dd->revision >> CCE_REVISION_CHIP_REV_MINOR_SHIFT
   5207			& CCE_REVISION_CHIP_REV_MINOR_MASK;
   5208	return (chip_rev_minor & 0xf0) == 0;
   5209}
   5210
   5211/* return true if this is chip revision revision b */
   5212int is_bx(struct hfi1_devdata *dd)
   5213{
   5214	u8 chip_rev_minor =
   5215		dd->revision >> CCE_REVISION_CHIP_REV_MINOR_SHIFT
   5216			& CCE_REVISION_CHIP_REV_MINOR_MASK;
   5217	return (chip_rev_minor & 0xF0) == 0x10;
   5218}
   5219
   5220/* return true is kernel urg disabled for rcd */
   5221bool is_urg_masked(struct hfi1_ctxtdata *rcd)
   5222{
   5223	u64 mask;
   5224	u32 is = IS_RCVURGENT_START + rcd->ctxt;
   5225	u8 bit = is % 64;
   5226
   5227	mask = read_csr(rcd->dd, CCE_INT_MASK + (8 * (is / 64)));
   5228	return !(mask & BIT_ULL(bit));
   5229}
   5230
   5231/*
   5232 * Append string s to buffer buf.  Arguments curp and len are the current
   5233 * position and remaining length, respectively.
   5234 *
   5235 * return 0 on success, 1 on out of room
   5236 */
   5237static int append_str(char *buf, char **curp, int *lenp, const char *s)
   5238{
   5239	char *p = *curp;
   5240	int len = *lenp;
   5241	int result = 0; /* success */
   5242	char c;
   5243
   5244	/* add a comma, if first in the buffer */
   5245	if (p != buf) {
   5246		if (len == 0) {
   5247			result = 1; /* out of room */
   5248			goto done;
   5249		}
   5250		*p++ = ',';
   5251		len--;
   5252	}
   5253
   5254	/* copy the string */
   5255	while ((c = *s++) != 0) {
   5256		if (len == 0) {
   5257			result = 1; /* out of room */
   5258			goto done;
   5259		}
   5260		*p++ = c;
   5261		len--;
   5262	}
   5263
   5264done:
   5265	/* write return values */
   5266	*curp = p;
   5267	*lenp = len;
   5268
   5269	return result;
   5270}
   5271
   5272/*
   5273 * Using the given flag table, print a comma separated string into
   5274 * the buffer.  End in '*' if the buffer is too short.
   5275 */
   5276static char *flag_string(char *buf, int buf_len, u64 flags,
   5277			 struct flag_table *table, int table_size)
   5278{
   5279	char extra[32];
   5280	char *p = buf;
   5281	int len = buf_len;
   5282	int no_room = 0;
   5283	int i;
   5284
   5285	/* make sure there is at least 2 so we can form "*" */
   5286	if (len < 2)
   5287		return "";
   5288
   5289	len--;	/* leave room for a nul */
   5290	for (i = 0; i < table_size; i++) {
   5291		if (flags & table[i].flag) {
   5292			no_room = append_str(buf, &p, &len, table[i].str);
   5293			if (no_room)
   5294				break;
   5295			flags &= ~table[i].flag;
   5296		}
   5297	}
   5298
   5299	/* any undocumented bits left? */
   5300	if (!no_room && flags) {
   5301		snprintf(extra, sizeof(extra), "bits 0x%llx", flags);
   5302		no_room = append_str(buf, &p, &len, extra);
   5303	}
   5304
   5305	/* add * if ran out of room */
   5306	if (no_room) {
   5307		/* may need to back up to add space for a '*' */
   5308		if (len == 0)
   5309			--p;
   5310		*p++ = '*';
   5311	}
   5312
   5313	/* add final nul - space already allocated above */
   5314	*p = 0;
   5315	return buf;
   5316}
   5317
   5318/* first 8 CCE error interrupt source names */
   5319static const char * const cce_misc_names[] = {
   5320	"CceErrInt",		/* 0 */
   5321	"RxeErrInt",		/* 1 */
   5322	"MiscErrInt",		/* 2 */
   5323	"Reserved3",		/* 3 */
   5324	"PioErrInt",		/* 4 */
   5325	"SDmaErrInt",		/* 5 */
   5326	"EgressErrInt",		/* 6 */
   5327	"TxeErrInt"		/* 7 */
   5328};
   5329
   5330/*
   5331 * Return the miscellaneous error interrupt name.
   5332 */
   5333static char *is_misc_err_name(char *buf, size_t bsize, unsigned int source)
   5334{
   5335	if (source < ARRAY_SIZE(cce_misc_names))
   5336		strncpy(buf, cce_misc_names[source], bsize);
   5337	else
   5338		snprintf(buf, bsize, "Reserved%u",
   5339			 source + IS_GENERAL_ERR_START);
   5340
   5341	return buf;
   5342}
   5343
   5344/*
   5345 * Return the SDMA engine error interrupt name.
   5346 */
   5347static char *is_sdma_eng_err_name(char *buf, size_t bsize, unsigned int source)
   5348{
   5349	snprintf(buf, bsize, "SDmaEngErrInt%u", source);
   5350	return buf;
   5351}
   5352
   5353/*
   5354 * Return the send context error interrupt name.
   5355 */
   5356static char *is_sendctxt_err_name(char *buf, size_t bsize, unsigned int source)
   5357{
   5358	snprintf(buf, bsize, "SendCtxtErrInt%u", source);
   5359	return buf;
   5360}
   5361
   5362static const char * const various_names[] = {
   5363	"PbcInt",
   5364	"GpioAssertInt",
   5365	"Qsfp1Int",
   5366	"Qsfp2Int",
   5367	"TCritInt"
   5368};
   5369
   5370/*
   5371 * Return the various interrupt name.
   5372 */
   5373static char *is_various_name(char *buf, size_t bsize, unsigned int source)
   5374{
   5375	if (source < ARRAY_SIZE(various_names))
   5376		strncpy(buf, various_names[source], bsize);
   5377	else
   5378		snprintf(buf, bsize, "Reserved%u", source + IS_VARIOUS_START);
   5379	return buf;
   5380}
   5381
   5382/*
   5383 * Return the DC interrupt name.
   5384 */
   5385static char *is_dc_name(char *buf, size_t bsize, unsigned int source)
   5386{
   5387	static const char * const dc_int_names[] = {
   5388		"common",
   5389		"lcb",
   5390		"8051",
   5391		"lbm"	/* local block merge */
   5392	};
   5393
   5394	if (source < ARRAY_SIZE(dc_int_names))
   5395		snprintf(buf, bsize, "dc_%s_int", dc_int_names[source]);
   5396	else
   5397		snprintf(buf, bsize, "DCInt%u", source);
   5398	return buf;
   5399}
   5400
   5401static const char * const sdma_int_names[] = {
   5402	"SDmaInt",
   5403	"SdmaIdleInt",
   5404	"SdmaProgressInt",
   5405};
   5406
   5407/*
   5408 * Return the SDMA engine interrupt name.
   5409 */
   5410static char *is_sdma_eng_name(char *buf, size_t bsize, unsigned int source)
   5411{
   5412	/* what interrupt */
   5413	unsigned int what  = source / TXE_NUM_SDMA_ENGINES;
   5414	/* which engine */
   5415	unsigned int which = source % TXE_NUM_SDMA_ENGINES;
   5416
   5417	if (likely(what < 3))
   5418		snprintf(buf, bsize, "%s%u", sdma_int_names[what], which);
   5419	else
   5420		snprintf(buf, bsize, "Invalid SDMA interrupt %u", source);
   5421	return buf;
   5422}
   5423
   5424/*
   5425 * Return the receive available interrupt name.
   5426 */
   5427static char *is_rcv_avail_name(char *buf, size_t bsize, unsigned int source)
   5428{
   5429	snprintf(buf, bsize, "RcvAvailInt%u", source);
   5430	return buf;
   5431}
   5432
   5433/*
   5434 * Return the receive urgent interrupt name.
   5435 */
   5436static char *is_rcv_urgent_name(char *buf, size_t bsize, unsigned int source)
   5437{
   5438	snprintf(buf, bsize, "RcvUrgentInt%u", source);
   5439	return buf;
   5440}
   5441
   5442/*
   5443 * Return the send credit interrupt name.
   5444 */
   5445static char *is_send_credit_name(char *buf, size_t bsize, unsigned int source)
   5446{
   5447	snprintf(buf, bsize, "SendCreditInt%u", source);
   5448	return buf;
   5449}
   5450
   5451/*
   5452 * Return the reserved interrupt name.
   5453 */
   5454static char *is_reserved_name(char *buf, size_t bsize, unsigned int source)
   5455{
   5456	snprintf(buf, bsize, "Reserved%u", source + IS_RESERVED_START);
   5457	return buf;
   5458}
   5459
   5460static char *cce_err_status_string(char *buf, int buf_len, u64 flags)
   5461{
   5462	return flag_string(buf, buf_len, flags,
   5463			   cce_err_status_flags,
   5464			   ARRAY_SIZE(cce_err_status_flags));
   5465}
   5466
   5467static char *rxe_err_status_string(char *buf, int buf_len, u64 flags)
   5468{
   5469	return flag_string(buf, buf_len, flags,
   5470			   rxe_err_status_flags,
   5471			   ARRAY_SIZE(rxe_err_status_flags));
   5472}
   5473
   5474static char *misc_err_status_string(char *buf, int buf_len, u64 flags)
   5475{
   5476	return flag_string(buf, buf_len, flags, misc_err_status_flags,
   5477			   ARRAY_SIZE(misc_err_status_flags));
   5478}
   5479
   5480static char *pio_err_status_string(char *buf, int buf_len, u64 flags)
   5481{
   5482	return flag_string(buf, buf_len, flags,
   5483			   pio_err_status_flags,
   5484			   ARRAY_SIZE(pio_err_status_flags));
   5485}
   5486
   5487static char *sdma_err_status_string(char *buf, int buf_len, u64 flags)
   5488{
   5489	return flag_string(buf, buf_len, flags,
   5490			   sdma_err_status_flags,
   5491			   ARRAY_SIZE(sdma_err_status_flags));
   5492}
   5493
   5494static char *egress_err_status_string(char *buf, int buf_len, u64 flags)
   5495{
   5496	return flag_string(buf, buf_len, flags,
   5497			   egress_err_status_flags,
   5498			   ARRAY_SIZE(egress_err_status_flags));
   5499}
   5500
   5501static char *egress_err_info_string(char *buf, int buf_len, u64 flags)
   5502{
   5503	return flag_string(buf, buf_len, flags,
   5504			   egress_err_info_flags,
   5505			   ARRAY_SIZE(egress_err_info_flags));
   5506}
   5507
   5508static char *send_err_status_string(char *buf, int buf_len, u64 flags)
   5509{
   5510	return flag_string(buf, buf_len, flags,
   5511			   send_err_status_flags,
   5512			   ARRAY_SIZE(send_err_status_flags));
   5513}
   5514
   5515static void handle_cce_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   5516{
   5517	char buf[96];
   5518	int i = 0;
   5519
   5520	/*
   5521	 * For most these errors, there is nothing that can be done except
   5522	 * report or record it.
   5523	 */
   5524	dd_dev_info(dd, "CCE Error: %s\n",
   5525		    cce_err_status_string(buf, sizeof(buf), reg));
   5526
   5527	if ((reg & CCE_ERR_STATUS_CCE_CLI2_ASYNC_FIFO_PARITY_ERR_SMASK) &&
   5528	    is_ax(dd) && (dd->icode != ICODE_FUNCTIONAL_SIMULATOR)) {
   5529		/* this error requires a manual drop into SPC freeze mode */
   5530		/* then a fix up */
   5531		start_freeze_handling(dd->pport, FREEZE_SELF);
   5532	}
   5533
   5534	for (i = 0; i < NUM_CCE_ERR_STATUS_COUNTERS; i++) {
   5535		if (reg & (1ull << i)) {
   5536			incr_cntr64(&dd->cce_err_status_cnt[i]);
   5537			/* maintain a counter over all cce_err_status errors */
   5538			incr_cntr64(&dd->sw_cce_err_status_aggregate);
   5539		}
   5540	}
   5541}
   5542
   5543/*
   5544 * Check counters for receive errors that do not have an interrupt
   5545 * associated with them.
   5546 */
   5547#define RCVERR_CHECK_TIME 10
   5548static void update_rcverr_timer(struct timer_list *t)
   5549{
   5550	struct hfi1_devdata *dd = from_timer(dd, t, rcverr_timer);
   5551	struct hfi1_pportdata *ppd = dd->pport;
   5552	u32 cur_ovfl_cnt = read_dev_cntr(dd, C_RCV_OVF, CNTR_INVALID_VL);
   5553
   5554	if (dd->rcv_ovfl_cnt < cur_ovfl_cnt &&
   5555	    ppd->port_error_action & OPA_PI_MASK_EX_BUFFER_OVERRUN) {
   5556		dd_dev_info(dd, "%s: PortErrorAction bounce\n", __func__);
   5557		set_link_down_reason(
   5558		ppd, OPA_LINKDOWN_REASON_EXCESSIVE_BUFFER_OVERRUN, 0,
   5559		OPA_LINKDOWN_REASON_EXCESSIVE_BUFFER_OVERRUN);
   5560		queue_work(ppd->link_wq, &ppd->link_bounce_work);
   5561	}
   5562	dd->rcv_ovfl_cnt = (u32)cur_ovfl_cnt;
   5563
   5564	mod_timer(&dd->rcverr_timer, jiffies + HZ * RCVERR_CHECK_TIME);
   5565}
   5566
   5567static int init_rcverr(struct hfi1_devdata *dd)
   5568{
   5569	timer_setup(&dd->rcverr_timer, update_rcverr_timer, 0);
   5570	/* Assume the hardware counter has been reset */
   5571	dd->rcv_ovfl_cnt = 0;
   5572	return mod_timer(&dd->rcverr_timer, jiffies + HZ * RCVERR_CHECK_TIME);
   5573}
   5574
   5575static void free_rcverr(struct hfi1_devdata *dd)
   5576{
   5577	if (dd->rcverr_timer.function)
   5578		del_timer_sync(&dd->rcverr_timer);
   5579}
   5580
   5581static void handle_rxe_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   5582{
   5583	char buf[96];
   5584	int i = 0;
   5585
   5586	dd_dev_info(dd, "Receive Error: %s\n",
   5587		    rxe_err_status_string(buf, sizeof(buf), reg));
   5588
   5589	if (reg & ALL_RXE_FREEZE_ERR) {
   5590		int flags = 0;
   5591
   5592		/*
   5593		 * Freeze mode recovery is disabled for the errors
   5594		 * in RXE_FREEZE_ABORT_MASK
   5595		 */
   5596		if (is_ax(dd) && (reg & RXE_FREEZE_ABORT_MASK))
   5597			flags = FREEZE_ABORT;
   5598
   5599		start_freeze_handling(dd->pport, flags);
   5600	}
   5601
   5602	for (i = 0; i < NUM_RCV_ERR_STATUS_COUNTERS; i++) {
   5603		if (reg & (1ull << i))
   5604			incr_cntr64(&dd->rcv_err_status_cnt[i]);
   5605	}
   5606}
   5607
   5608static void handle_misc_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   5609{
   5610	char buf[96];
   5611	int i = 0;
   5612
   5613	dd_dev_info(dd, "Misc Error: %s",
   5614		    misc_err_status_string(buf, sizeof(buf), reg));
   5615	for (i = 0; i < NUM_MISC_ERR_STATUS_COUNTERS; i++) {
   5616		if (reg & (1ull << i))
   5617			incr_cntr64(&dd->misc_err_status_cnt[i]);
   5618	}
   5619}
   5620
   5621static void handle_pio_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   5622{
   5623	char buf[96];
   5624	int i = 0;
   5625
   5626	dd_dev_info(dd, "PIO Error: %s\n",
   5627		    pio_err_status_string(buf, sizeof(buf), reg));
   5628
   5629	if (reg & ALL_PIO_FREEZE_ERR)
   5630		start_freeze_handling(dd->pport, 0);
   5631
   5632	for (i = 0; i < NUM_SEND_PIO_ERR_STATUS_COUNTERS; i++) {
   5633		if (reg & (1ull << i))
   5634			incr_cntr64(&dd->send_pio_err_status_cnt[i]);
   5635	}
   5636}
   5637
   5638static void handle_sdma_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   5639{
   5640	char buf[96];
   5641	int i = 0;
   5642
   5643	dd_dev_info(dd, "SDMA Error: %s\n",
   5644		    sdma_err_status_string(buf, sizeof(buf), reg));
   5645
   5646	if (reg & ALL_SDMA_FREEZE_ERR)
   5647		start_freeze_handling(dd->pport, 0);
   5648
   5649	for (i = 0; i < NUM_SEND_DMA_ERR_STATUS_COUNTERS; i++) {
   5650		if (reg & (1ull << i))
   5651			incr_cntr64(&dd->send_dma_err_status_cnt[i]);
   5652	}
   5653}
   5654
   5655static inline void __count_port_discards(struct hfi1_pportdata *ppd)
   5656{
   5657	incr_cntr64(&ppd->port_xmit_discards);
   5658}
   5659
   5660static void count_port_inactive(struct hfi1_devdata *dd)
   5661{
   5662	__count_port_discards(dd->pport);
   5663}
   5664
   5665/*
   5666 * We have had a "disallowed packet" error during egress. Determine the
   5667 * integrity check which failed, and update relevant error counter, etc.
   5668 *
   5669 * Note that the SEND_EGRESS_ERR_INFO register has only a single
   5670 * bit of state per integrity check, and so we can miss the reason for an
   5671 * egress error if more than one packet fails the same integrity check
   5672 * since we cleared the corresponding bit in SEND_EGRESS_ERR_INFO.
   5673 */
   5674static void handle_send_egress_err_info(struct hfi1_devdata *dd,
   5675					int vl)
   5676{
   5677	struct hfi1_pportdata *ppd = dd->pport;
   5678	u64 src = read_csr(dd, SEND_EGRESS_ERR_SOURCE); /* read first */
   5679	u64 info = read_csr(dd, SEND_EGRESS_ERR_INFO);
   5680	char buf[96];
   5681
   5682	/* clear down all observed info as quickly as possible after read */
   5683	write_csr(dd, SEND_EGRESS_ERR_INFO, info);
   5684
   5685	dd_dev_info(dd,
   5686		    "Egress Error Info: 0x%llx, %s Egress Error Src 0x%llx\n",
   5687		    info, egress_err_info_string(buf, sizeof(buf), info), src);
   5688
   5689	/* Eventually add other counters for each bit */
   5690	if (info & PORT_DISCARD_EGRESS_ERRS) {
   5691		int weight, i;
   5692
   5693		/*
   5694		 * Count all applicable bits as individual errors and
   5695		 * attribute them to the packet that triggered this handler.
   5696		 * This may not be completely accurate due to limitations
   5697		 * on the available hardware error information.  There is
   5698		 * a single information register and any number of error
   5699		 * packets may have occurred and contributed to it before
   5700		 * this routine is called.  This means that:
   5701		 * a) If multiple packets with the same error occur before
   5702		 *    this routine is called, earlier packets are missed.
   5703		 *    There is only a single bit for each error type.
   5704		 * b) Errors may not be attributed to the correct VL.
   5705		 *    The driver is attributing all bits in the info register
   5706		 *    to the packet that triggered this call, but bits
   5707		 *    could be an accumulation of different packets with
   5708		 *    different VLs.
   5709		 * c) A single error packet may have multiple counts attached
   5710		 *    to it.  There is no way for the driver to know if
   5711		 *    multiple bits set in the info register are due to a
   5712		 *    single packet or multiple packets.  The driver assumes
   5713		 *    multiple packets.
   5714		 */
   5715		weight = hweight64(info & PORT_DISCARD_EGRESS_ERRS);
   5716		for (i = 0; i < weight; i++) {
   5717			__count_port_discards(ppd);
   5718			if (vl >= 0 && vl < TXE_NUM_DATA_VL)
   5719				incr_cntr64(&ppd->port_xmit_discards_vl[vl]);
   5720			else if (vl == 15)
   5721				incr_cntr64(&ppd->port_xmit_discards_vl
   5722					    [C_VL_15]);
   5723		}
   5724	}
   5725}
   5726
   5727/*
   5728 * Input value is a bit position within the SEND_EGRESS_ERR_STATUS
   5729 * register. Does it represent a 'port inactive' error?
   5730 */
   5731static inline int port_inactive_err(u64 posn)
   5732{
   5733	return (posn >= SEES(TX_LINKDOWN) &&
   5734		posn <= SEES(TX_INCORRECT_LINK_STATE));
   5735}
   5736
   5737/*
   5738 * Input value is a bit position within the SEND_EGRESS_ERR_STATUS
   5739 * register. Does it represent a 'disallowed packet' error?
   5740 */
   5741static inline int disallowed_pkt_err(int posn)
   5742{
   5743	return (posn >= SEES(TX_SDMA0_DISALLOWED_PACKET) &&
   5744		posn <= SEES(TX_SDMA15_DISALLOWED_PACKET));
   5745}
   5746
   5747/*
   5748 * Input value is a bit position of one of the SDMA engine disallowed
   5749 * packet errors.  Return which engine.  Use of this must be guarded by
   5750 * disallowed_pkt_err().
   5751 */
   5752static inline int disallowed_pkt_engine(int posn)
   5753{
   5754	return posn - SEES(TX_SDMA0_DISALLOWED_PACKET);
   5755}
   5756
   5757/*
   5758 * Translate an SDMA engine to a VL.  Return -1 if the tranlation cannot
   5759 * be done.
   5760 */
   5761static int engine_to_vl(struct hfi1_devdata *dd, int engine)
   5762{
   5763	struct sdma_vl_map *m;
   5764	int vl;
   5765
   5766	/* range check */
   5767	if (engine < 0 || engine >= TXE_NUM_SDMA_ENGINES)
   5768		return -1;
   5769
   5770	rcu_read_lock();
   5771	m = rcu_dereference(dd->sdma_map);
   5772	vl = m->engine_to_vl[engine];
   5773	rcu_read_unlock();
   5774
   5775	return vl;
   5776}
   5777
   5778/*
   5779 * Translate the send context (sofware index) into a VL.  Return -1 if the
   5780 * translation cannot be done.
   5781 */
   5782static int sc_to_vl(struct hfi1_devdata *dd, int sw_index)
   5783{
   5784	struct send_context_info *sci;
   5785	struct send_context *sc;
   5786	int i;
   5787
   5788	sci = &dd->send_contexts[sw_index];
   5789
   5790	/* there is no information for user (PSM) and ack contexts */
   5791	if ((sci->type != SC_KERNEL) && (sci->type != SC_VL15))
   5792		return -1;
   5793
   5794	sc = sci->sc;
   5795	if (!sc)
   5796		return -1;
   5797	if (dd->vld[15].sc == sc)
   5798		return 15;
   5799	for (i = 0; i < num_vls; i++)
   5800		if (dd->vld[i].sc == sc)
   5801			return i;
   5802
   5803	return -1;
   5804}
   5805
   5806static void handle_egress_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   5807{
   5808	u64 reg_copy = reg, handled = 0;
   5809	char buf[96];
   5810	int i = 0;
   5811
   5812	if (reg & ALL_TXE_EGRESS_FREEZE_ERR)
   5813		start_freeze_handling(dd->pport, 0);
   5814	else if (is_ax(dd) &&
   5815		 (reg & SEND_EGRESS_ERR_STATUS_TX_CREDIT_RETURN_VL_ERR_SMASK) &&
   5816		 (dd->icode != ICODE_FUNCTIONAL_SIMULATOR))
   5817		start_freeze_handling(dd->pport, 0);
   5818
   5819	while (reg_copy) {
   5820		int posn = fls64(reg_copy);
   5821		/* fls64() returns a 1-based offset, we want it zero based */
   5822		int shift = posn - 1;
   5823		u64 mask = 1ULL << shift;
   5824
   5825		if (port_inactive_err(shift)) {
   5826			count_port_inactive(dd);
   5827			handled |= mask;
   5828		} else if (disallowed_pkt_err(shift)) {
   5829			int vl = engine_to_vl(dd, disallowed_pkt_engine(shift));
   5830
   5831			handle_send_egress_err_info(dd, vl);
   5832			handled |= mask;
   5833		}
   5834		reg_copy &= ~mask;
   5835	}
   5836
   5837	reg &= ~handled;
   5838
   5839	if (reg)
   5840		dd_dev_info(dd, "Egress Error: %s\n",
   5841			    egress_err_status_string(buf, sizeof(buf), reg));
   5842
   5843	for (i = 0; i < NUM_SEND_EGRESS_ERR_STATUS_COUNTERS; i++) {
   5844		if (reg & (1ull << i))
   5845			incr_cntr64(&dd->send_egress_err_status_cnt[i]);
   5846	}
   5847}
   5848
   5849static void handle_txe_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   5850{
   5851	char buf[96];
   5852	int i = 0;
   5853
   5854	dd_dev_info(dd, "Send Error: %s\n",
   5855		    send_err_status_string(buf, sizeof(buf), reg));
   5856
   5857	for (i = 0; i < NUM_SEND_ERR_STATUS_COUNTERS; i++) {
   5858		if (reg & (1ull << i))
   5859			incr_cntr64(&dd->send_err_status_cnt[i]);
   5860	}
   5861}
   5862
   5863/*
   5864 * The maximum number of times the error clear down will loop before
   5865 * blocking a repeating error.  This value is arbitrary.
   5866 */
   5867#define MAX_CLEAR_COUNT 20
   5868
   5869/*
   5870 * Clear and handle an error register.  All error interrupts are funneled
   5871 * through here to have a central location to correctly handle single-
   5872 * or multi-shot errors.
   5873 *
   5874 * For non per-context registers, call this routine with a context value
   5875 * of 0 so the per-context offset is zero.
   5876 *
   5877 * If the handler loops too many times, assume that something is wrong
   5878 * and can't be fixed, so mask the error bits.
   5879 */
   5880static void interrupt_clear_down(struct hfi1_devdata *dd,
   5881				 u32 context,
   5882				 const struct err_reg_info *eri)
   5883{
   5884	u64 reg;
   5885	u32 count;
   5886
   5887	/* read in a loop until no more errors are seen */
   5888	count = 0;
   5889	while (1) {
   5890		reg = read_kctxt_csr(dd, context, eri->status);
   5891		if (reg == 0)
   5892			break;
   5893		write_kctxt_csr(dd, context, eri->clear, reg);
   5894		if (likely(eri->handler))
   5895			eri->handler(dd, context, reg);
   5896		count++;
   5897		if (count > MAX_CLEAR_COUNT) {
   5898			u64 mask;
   5899
   5900			dd_dev_err(dd, "Repeating %s bits 0x%llx - masking\n",
   5901				   eri->desc, reg);
   5902			/*
   5903			 * Read-modify-write so any other masked bits
   5904			 * remain masked.
   5905			 */
   5906			mask = read_kctxt_csr(dd, context, eri->mask);
   5907			mask &= ~reg;
   5908			write_kctxt_csr(dd, context, eri->mask, mask);
   5909			break;
   5910		}
   5911	}
   5912}
   5913
   5914/*
   5915 * CCE block "misc" interrupt.  Source is < 16.
   5916 */
   5917static void is_misc_err_int(struct hfi1_devdata *dd, unsigned int source)
   5918{
   5919	const struct err_reg_info *eri = &misc_errs[source];
   5920
   5921	if (eri->handler) {
   5922		interrupt_clear_down(dd, 0, eri);
   5923	} else {
   5924		dd_dev_err(dd, "Unexpected misc interrupt (%u) - reserved\n",
   5925			   source);
   5926	}
   5927}
   5928
   5929static char *send_context_err_status_string(char *buf, int buf_len, u64 flags)
   5930{
   5931	return flag_string(buf, buf_len, flags,
   5932			   sc_err_status_flags,
   5933			   ARRAY_SIZE(sc_err_status_flags));
   5934}
   5935
   5936/*
   5937 * Send context error interrupt.  Source (hw_context) is < 160.
   5938 *
   5939 * All send context errors cause the send context to halt.  The normal
   5940 * clear-down mechanism cannot be used because we cannot clear the
   5941 * error bits until several other long-running items are done first.
   5942 * This is OK because with the context halted, nothing else is going
   5943 * to happen on it anyway.
   5944 */
   5945static void is_sendctxt_err_int(struct hfi1_devdata *dd,
   5946				unsigned int hw_context)
   5947{
   5948	struct send_context_info *sci;
   5949	struct send_context *sc;
   5950	char flags[96];
   5951	u64 status;
   5952	u32 sw_index;
   5953	int i = 0;
   5954	unsigned long irq_flags;
   5955
   5956	sw_index = dd->hw_to_sw[hw_context];
   5957	if (sw_index >= dd->num_send_contexts) {
   5958		dd_dev_err(dd,
   5959			   "out of range sw index %u for send context %u\n",
   5960			   sw_index, hw_context);
   5961		return;
   5962	}
   5963	sci = &dd->send_contexts[sw_index];
   5964	spin_lock_irqsave(&dd->sc_lock, irq_flags);
   5965	sc = sci->sc;
   5966	if (!sc) {
   5967		dd_dev_err(dd, "%s: context %u(%u): no sc?\n", __func__,
   5968			   sw_index, hw_context);
   5969		spin_unlock_irqrestore(&dd->sc_lock, irq_flags);
   5970		return;
   5971	}
   5972
   5973	/* tell the software that a halt has begun */
   5974	sc_stop(sc, SCF_HALTED);
   5975
   5976	status = read_kctxt_csr(dd, hw_context, SEND_CTXT_ERR_STATUS);
   5977
   5978	dd_dev_info(dd, "Send Context %u(%u) Error: %s\n", sw_index, hw_context,
   5979		    send_context_err_status_string(flags, sizeof(flags),
   5980						   status));
   5981
   5982	if (status & SEND_CTXT_ERR_STATUS_PIO_DISALLOWED_PACKET_ERR_SMASK)
   5983		handle_send_egress_err_info(dd, sc_to_vl(dd, sw_index));
   5984
   5985	/*
   5986	 * Automatically restart halted kernel contexts out of interrupt
   5987	 * context.  User contexts must ask the driver to restart the context.
   5988	 */
   5989	if (sc->type != SC_USER)
   5990		queue_work(dd->pport->hfi1_wq, &sc->halt_work);
   5991	spin_unlock_irqrestore(&dd->sc_lock, irq_flags);
   5992
   5993	/*
   5994	 * Update the counters for the corresponding status bits.
   5995	 * Note that these particular counters are aggregated over all
   5996	 * 160 contexts.
   5997	 */
   5998	for (i = 0; i < NUM_SEND_CTXT_ERR_STATUS_COUNTERS; i++) {
   5999		if (status & (1ull << i))
   6000			incr_cntr64(&dd->sw_ctxt_err_status_cnt[i]);
   6001	}
   6002}
   6003
   6004static void handle_sdma_eng_err(struct hfi1_devdata *dd,
   6005				unsigned int source, u64 status)
   6006{
   6007	struct sdma_engine *sde;
   6008	int i = 0;
   6009
   6010	sde = &dd->per_sdma[source];
   6011#ifdef CONFIG_SDMA_VERBOSITY
   6012	dd_dev_err(sde->dd, "CONFIG SDMA(%u) %s:%d %s()\n", sde->this_idx,
   6013		   slashstrip(__FILE__), __LINE__, __func__);
   6014	dd_dev_err(sde->dd, "CONFIG SDMA(%u) source: %u status 0x%llx\n",
   6015		   sde->this_idx, source, (unsigned long long)status);
   6016#endif
   6017	sde->err_cnt++;
   6018	sdma_engine_error(sde, status);
   6019
   6020	/*
   6021	* Update the counters for the corresponding status bits.
   6022	* Note that these particular counters are aggregated over
   6023	* all 16 DMA engines.
   6024	*/
   6025	for (i = 0; i < NUM_SEND_DMA_ENG_ERR_STATUS_COUNTERS; i++) {
   6026		if (status & (1ull << i))
   6027			incr_cntr64(&dd->sw_send_dma_eng_err_status_cnt[i]);
   6028	}
   6029}
   6030
   6031/*
   6032 * CCE block SDMA error interrupt.  Source is < 16.
   6033 */
   6034static void is_sdma_eng_err_int(struct hfi1_devdata *dd, unsigned int source)
   6035{
   6036#ifdef CONFIG_SDMA_VERBOSITY
   6037	struct sdma_engine *sde = &dd->per_sdma[source];
   6038
   6039	dd_dev_err(dd, "CONFIG SDMA(%u) %s:%d %s()\n", sde->this_idx,
   6040		   slashstrip(__FILE__), __LINE__, __func__);
   6041	dd_dev_err(dd, "CONFIG SDMA(%u) source: %u\n", sde->this_idx,
   6042		   source);
   6043	sdma_dumpstate(sde);
   6044#endif
   6045	interrupt_clear_down(dd, source, &sdma_eng_err);
   6046}
   6047
   6048/*
   6049 * CCE block "various" interrupt.  Source is < 8.
   6050 */
   6051static void is_various_int(struct hfi1_devdata *dd, unsigned int source)
   6052{
   6053	const struct err_reg_info *eri = &various_err[source];
   6054
   6055	/*
   6056	 * TCritInt cannot go through interrupt_clear_down()
   6057	 * because it is not a second tier interrupt. The handler
   6058	 * should be called directly.
   6059	 */
   6060	if (source == TCRIT_INT_SOURCE)
   6061		handle_temp_err(dd);
   6062	else if (eri->handler)
   6063		interrupt_clear_down(dd, 0, eri);
   6064	else
   6065		dd_dev_info(dd,
   6066			    "%s: Unimplemented/reserved interrupt %d\n",
   6067			    __func__, source);
   6068}
   6069
   6070static void handle_qsfp_int(struct hfi1_devdata *dd, u32 src_ctx, u64 reg)
   6071{
   6072	/* src_ctx is always zero */
   6073	struct hfi1_pportdata *ppd = dd->pport;
   6074	unsigned long flags;
   6075	u64 qsfp_int_mgmt = (u64)(QSFP_HFI0_INT_N | QSFP_HFI0_MODPRST_N);
   6076
   6077	if (reg & QSFP_HFI0_MODPRST_N) {
   6078		if (!qsfp_mod_present(ppd)) {
   6079			dd_dev_info(dd, "%s: QSFP module removed\n",
   6080				    __func__);
   6081
   6082			ppd->driver_link_ready = 0;
   6083			/*
   6084			 * Cable removed, reset all our information about the
   6085			 * cache and cable capabilities
   6086			 */
   6087
   6088			spin_lock_irqsave(&ppd->qsfp_info.qsfp_lock, flags);
   6089			/*
   6090			 * We don't set cache_refresh_required here as we expect
   6091			 * an interrupt when a cable is inserted
   6092			 */
   6093			ppd->qsfp_info.cache_valid = 0;
   6094			ppd->qsfp_info.reset_needed = 0;
   6095			ppd->qsfp_info.limiting_active = 0;
   6096			spin_unlock_irqrestore(&ppd->qsfp_info.qsfp_lock,
   6097					       flags);
   6098			/* Invert the ModPresent pin now to detect plug-in */
   6099			write_csr(dd, dd->hfi1_id ? ASIC_QSFP2_INVERT :
   6100				  ASIC_QSFP1_INVERT, qsfp_int_mgmt);
   6101
   6102			if ((ppd->offline_disabled_reason >
   6103			  HFI1_ODR_MASK(
   6104			  OPA_LINKDOWN_REASON_LOCAL_MEDIA_NOT_INSTALLED)) ||
   6105			  (ppd->offline_disabled_reason ==
   6106			  HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE)))
   6107				ppd->offline_disabled_reason =
   6108				HFI1_ODR_MASK(
   6109				OPA_LINKDOWN_REASON_LOCAL_MEDIA_NOT_INSTALLED);
   6110
   6111			if (ppd->host_link_state == HLS_DN_POLL) {
   6112				/*
   6113				 * The link is still in POLL. This means
   6114				 * that the normal link down processing
   6115				 * will not happen. We have to do it here
   6116				 * before turning the DC off.
   6117				 */
   6118				queue_work(ppd->link_wq, &ppd->link_down_work);
   6119			}
   6120		} else {
   6121			dd_dev_info(dd, "%s: QSFP module inserted\n",
   6122				    __func__);
   6123
   6124			spin_lock_irqsave(&ppd->qsfp_info.qsfp_lock, flags);
   6125			ppd->qsfp_info.cache_valid = 0;
   6126			ppd->qsfp_info.cache_refresh_required = 1;
   6127			spin_unlock_irqrestore(&ppd->qsfp_info.qsfp_lock,
   6128					       flags);
   6129
   6130			/*
   6131			 * Stop inversion of ModPresent pin to detect
   6132			 * removal of the cable
   6133			 */
   6134			qsfp_int_mgmt &= ~(u64)QSFP_HFI0_MODPRST_N;
   6135			write_csr(dd, dd->hfi1_id ? ASIC_QSFP2_INVERT :
   6136				  ASIC_QSFP1_INVERT, qsfp_int_mgmt);
   6137
   6138			ppd->offline_disabled_reason =
   6139				HFI1_ODR_MASK(OPA_LINKDOWN_REASON_TRANSIENT);
   6140		}
   6141	}
   6142
   6143	if (reg & QSFP_HFI0_INT_N) {
   6144		dd_dev_info(dd, "%s: Interrupt received from QSFP module\n",
   6145			    __func__);
   6146		spin_lock_irqsave(&ppd->qsfp_info.qsfp_lock, flags);
   6147		ppd->qsfp_info.check_interrupt_flags = 1;
   6148		spin_unlock_irqrestore(&ppd->qsfp_info.qsfp_lock, flags);
   6149	}
   6150
   6151	/* Schedule the QSFP work only if there is a cable attached. */
   6152	if (qsfp_mod_present(ppd))
   6153		queue_work(ppd->link_wq, &ppd->qsfp_info.qsfp_work);
   6154}
   6155
   6156static int request_host_lcb_access(struct hfi1_devdata *dd)
   6157{
   6158	int ret;
   6159
   6160	ret = do_8051_command(dd, HCMD_MISC,
   6161			      (u64)HCMD_MISC_REQUEST_LCB_ACCESS <<
   6162			      LOAD_DATA_FIELD_ID_SHIFT, NULL);
   6163	if (ret != HCMD_SUCCESS) {
   6164		dd_dev_err(dd, "%s: command failed with error %d\n",
   6165			   __func__, ret);
   6166	}
   6167	return ret == HCMD_SUCCESS ? 0 : -EBUSY;
   6168}
   6169
   6170static int request_8051_lcb_access(struct hfi1_devdata *dd)
   6171{
   6172	int ret;
   6173
   6174	ret = do_8051_command(dd, HCMD_MISC,
   6175			      (u64)HCMD_MISC_GRANT_LCB_ACCESS <<
   6176			      LOAD_DATA_FIELD_ID_SHIFT, NULL);
   6177	if (ret != HCMD_SUCCESS) {
   6178		dd_dev_err(dd, "%s: command failed with error %d\n",
   6179			   __func__, ret);
   6180	}
   6181	return ret == HCMD_SUCCESS ? 0 : -EBUSY;
   6182}
   6183
   6184/*
   6185 * Set the LCB selector - allow host access.  The DCC selector always
   6186 * points to the host.
   6187 */
   6188static inline void set_host_lcb_access(struct hfi1_devdata *dd)
   6189{
   6190	write_csr(dd, DC_DC8051_CFG_CSR_ACCESS_SEL,
   6191		  DC_DC8051_CFG_CSR_ACCESS_SEL_DCC_SMASK |
   6192		  DC_DC8051_CFG_CSR_ACCESS_SEL_LCB_SMASK);
   6193}
   6194
   6195/*
   6196 * Clear the LCB selector - allow 8051 access.  The DCC selector always
   6197 * points to the host.
   6198 */
   6199static inline void set_8051_lcb_access(struct hfi1_devdata *dd)
   6200{
   6201	write_csr(dd, DC_DC8051_CFG_CSR_ACCESS_SEL,
   6202		  DC_DC8051_CFG_CSR_ACCESS_SEL_DCC_SMASK);
   6203}
   6204
   6205/*
   6206 * Acquire LCB access from the 8051.  If the host already has access,
   6207 * just increment a counter.  Otherwise, inform the 8051 that the
   6208 * host is taking access.
   6209 *
   6210 * Returns:
   6211 *	0 on success
   6212 *	-EBUSY if the 8051 has control and cannot be disturbed
   6213 *	-errno if unable to acquire access from the 8051
   6214 */
   6215int acquire_lcb_access(struct hfi1_devdata *dd, int sleep_ok)
   6216{
   6217	struct hfi1_pportdata *ppd = dd->pport;
   6218	int ret = 0;
   6219
   6220	/*
   6221	 * Use the host link state lock so the operation of this routine
   6222	 * { link state check, selector change, count increment } can occur
   6223	 * as a unit against a link state change.  Otherwise there is a
   6224	 * race between the state change and the count increment.
   6225	 */
   6226	if (sleep_ok) {
   6227		mutex_lock(&ppd->hls_lock);
   6228	} else {
   6229		while (!mutex_trylock(&ppd->hls_lock))
   6230			udelay(1);
   6231	}
   6232
   6233	/* this access is valid only when the link is up */
   6234	if (ppd->host_link_state & HLS_DOWN) {
   6235		dd_dev_info(dd, "%s: link state %s not up\n",
   6236			    __func__, link_state_name(ppd->host_link_state));
   6237		ret = -EBUSY;
   6238		goto done;
   6239	}
   6240
   6241	if (dd->lcb_access_count == 0) {
   6242		ret = request_host_lcb_access(dd);
   6243		if (ret) {
   6244			dd_dev_err(dd,
   6245				   "%s: unable to acquire LCB access, err %d\n",
   6246				   __func__, ret);
   6247			goto done;
   6248		}
   6249		set_host_lcb_access(dd);
   6250	}
   6251	dd->lcb_access_count++;
   6252done:
   6253	mutex_unlock(&ppd->hls_lock);
   6254	return ret;
   6255}
   6256
   6257/*
   6258 * Release LCB access by decrementing the use count.  If the count is moving
   6259 * from 1 to 0, inform 8051 that it has control back.
   6260 *
   6261 * Returns:
   6262 *	0 on success
   6263 *	-errno if unable to release access to the 8051
   6264 */
   6265int release_lcb_access(struct hfi1_devdata *dd, int sleep_ok)
   6266{
   6267	int ret = 0;
   6268
   6269	/*
   6270	 * Use the host link state lock because the acquire needed it.
   6271	 * Here, we only need to keep { selector change, count decrement }
   6272	 * as a unit.
   6273	 */
   6274	if (sleep_ok) {
   6275		mutex_lock(&dd->pport->hls_lock);
   6276	} else {
   6277		while (!mutex_trylock(&dd->pport->hls_lock))
   6278			udelay(1);
   6279	}
   6280
   6281	if (dd->lcb_access_count == 0) {
   6282		dd_dev_err(dd, "%s: LCB access count is zero.  Skipping.\n",
   6283			   __func__);
   6284		goto done;
   6285	}
   6286
   6287	if (dd->lcb_access_count == 1) {
   6288		set_8051_lcb_access(dd);
   6289		ret = request_8051_lcb_access(dd);
   6290		if (ret) {
   6291			dd_dev_err(dd,
   6292				   "%s: unable to release LCB access, err %d\n",
   6293				   __func__, ret);
   6294			/* restore host access if the grant didn't work */
   6295			set_host_lcb_access(dd);
   6296			goto done;
   6297		}
   6298	}
   6299	dd->lcb_access_count--;
   6300done:
   6301	mutex_unlock(&dd->pport->hls_lock);
   6302	return ret;
   6303}
   6304
   6305/*
   6306 * Initialize LCB access variables and state.  Called during driver load,
   6307 * after most of the initialization is finished.
   6308 *
   6309 * The DC default is LCB access on for the host.  The driver defaults to
   6310 * leaving access to the 8051.  Assign access now - this constrains the call
   6311 * to this routine to be after all LCB set-up is done.  In particular, after
   6312 * hf1_init_dd() -> set_up_interrupts() -> clear_all_interrupts()
   6313 */
   6314static void init_lcb_access(struct hfi1_devdata *dd)
   6315{
   6316	dd->lcb_access_count = 0;
   6317}
   6318
   6319/*
   6320 * Write a response back to a 8051 request.
   6321 */
   6322static void hreq_response(struct hfi1_devdata *dd, u8 return_code, u16 rsp_data)
   6323{
   6324	write_csr(dd, DC_DC8051_CFG_EXT_DEV_0,
   6325		  DC_DC8051_CFG_EXT_DEV_0_COMPLETED_SMASK |
   6326		  (u64)return_code <<
   6327		  DC_DC8051_CFG_EXT_DEV_0_RETURN_CODE_SHIFT |
   6328		  (u64)rsp_data << DC_DC8051_CFG_EXT_DEV_0_RSP_DATA_SHIFT);
   6329}
   6330
   6331/*
   6332 * Handle host requests from the 8051.
   6333 */
   6334static void handle_8051_request(struct hfi1_pportdata *ppd)
   6335{
   6336	struct hfi1_devdata *dd = ppd->dd;
   6337	u64 reg;
   6338	u16 data = 0;
   6339	u8 type;
   6340
   6341	reg = read_csr(dd, DC_DC8051_CFG_EXT_DEV_1);
   6342	if ((reg & DC_DC8051_CFG_EXT_DEV_1_REQ_NEW_SMASK) == 0)
   6343		return;	/* no request */
   6344
   6345	/* zero out COMPLETED so the response is seen */
   6346	write_csr(dd, DC_DC8051_CFG_EXT_DEV_0, 0);
   6347
   6348	/* extract request details */
   6349	type = (reg >> DC_DC8051_CFG_EXT_DEV_1_REQ_TYPE_SHIFT)
   6350			& DC_DC8051_CFG_EXT_DEV_1_REQ_TYPE_MASK;
   6351	data = (reg >> DC_DC8051_CFG_EXT_DEV_1_REQ_DATA_SHIFT)
   6352			& DC_DC8051_CFG_EXT_DEV_1_REQ_DATA_MASK;
   6353
   6354	switch (type) {
   6355	case HREQ_LOAD_CONFIG:
   6356	case HREQ_SAVE_CONFIG:
   6357	case HREQ_READ_CONFIG:
   6358	case HREQ_SET_TX_EQ_ABS:
   6359	case HREQ_SET_TX_EQ_REL:
   6360	case HREQ_ENABLE:
   6361		dd_dev_info(dd, "8051 request: request 0x%x not supported\n",
   6362			    type);
   6363		hreq_response(dd, HREQ_NOT_SUPPORTED, 0);
   6364		break;
   6365	case HREQ_LCB_RESET:
   6366		/* Put the LCB, RX FPE and TX FPE into reset */
   6367		write_csr(dd, DCC_CFG_RESET, LCB_RX_FPE_TX_FPE_INTO_RESET);
   6368		/* Make sure the write completed */
   6369		(void)read_csr(dd, DCC_CFG_RESET);
   6370		/* Hold the reset long enough to take effect */
   6371		udelay(1);
   6372		/* Take the LCB, RX FPE and TX FPE out of reset */
   6373		write_csr(dd, DCC_CFG_RESET, LCB_RX_FPE_TX_FPE_OUT_OF_RESET);
   6374		hreq_response(dd, HREQ_SUCCESS, 0);
   6375
   6376		break;
   6377	case HREQ_CONFIG_DONE:
   6378		hreq_response(dd, HREQ_SUCCESS, 0);
   6379		break;
   6380
   6381	case HREQ_INTERFACE_TEST:
   6382		hreq_response(dd, HREQ_SUCCESS, data);
   6383		break;
   6384	default:
   6385		dd_dev_err(dd, "8051 request: unknown request 0x%x\n", type);
   6386		hreq_response(dd, HREQ_NOT_SUPPORTED, 0);
   6387		break;
   6388	}
   6389}
   6390
   6391/*
   6392 * Set up allocation unit vaulue.
   6393 */
   6394void set_up_vau(struct hfi1_devdata *dd, u8 vau)
   6395{
   6396	u64 reg = read_csr(dd, SEND_CM_GLOBAL_CREDIT);
   6397
   6398	/* do not modify other values in the register */
   6399	reg &= ~SEND_CM_GLOBAL_CREDIT_AU_SMASK;
   6400	reg |= (u64)vau << SEND_CM_GLOBAL_CREDIT_AU_SHIFT;
   6401	write_csr(dd, SEND_CM_GLOBAL_CREDIT, reg);
   6402}
   6403
   6404/*
   6405 * Set up initial VL15 credits of the remote.  Assumes the rest of
   6406 * the CM credit registers are zero from a previous global or credit reset.
   6407 * Shared limit for VL15 will always be 0.
   6408 */
   6409void set_up_vl15(struct hfi1_devdata *dd, u16 vl15buf)
   6410{
   6411	u64 reg = read_csr(dd, SEND_CM_GLOBAL_CREDIT);
   6412
   6413	/* set initial values for total and shared credit limit */
   6414	reg &= ~(SEND_CM_GLOBAL_CREDIT_TOTAL_CREDIT_LIMIT_SMASK |
   6415		 SEND_CM_GLOBAL_CREDIT_SHARED_LIMIT_SMASK);
   6416
   6417	/*
   6418	 * Set total limit to be equal to VL15 credits.
   6419	 * Leave shared limit at 0.
   6420	 */
   6421	reg |= (u64)vl15buf << SEND_CM_GLOBAL_CREDIT_TOTAL_CREDIT_LIMIT_SHIFT;
   6422	write_csr(dd, SEND_CM_GLOBAL_CREDIT, reg);
   6423
   6424	write_csr(dd, SEND_CM_CREDIT_VL15, (u64)vl15buf
   6425		  << SEND_CM_CREDIT_VL15_DEDICATED_LIMIT_VL_SHIFT);
   6426}
   6427
   6428/*
   6429 * Zero all credit details from the previous connection and
   6430 * reset the CM manager's internal counters.
   6431 */
   6432void reset_link_credits(struct hfi1_devdata *dd)
   6433{
   6434	int i;
   6435
   6436	/* remove all previous VL credit limits */
   6437	for (i = 0; i < TXE_NUM_DATA_VL; i++)
   6438		write_csr(dd, SEND_CM_CREDIT_VL + (8 * i), 0);
   6439	write_csr(dd, SEND_CM_CREDIT_VL15, 0);
   6440	write_csr(dd, SEND_CM_GLOBAL_CREDIT, 0);
   6441	/* reset the CM block */
   6442	pio_send_control(dd, PSC_CM_RESET);
   6443	/* reset cached value */
   6444	dd->vl15buf_cached = 0;
   6445}
   6446
   6447/* convert a vCU to a CU */
   6448static u32 vcu_to_cu(u8 vcu)
   6449{
   6450	return 1 << vcu;
   6451}
   6452
   6453/* convert a CU to a vCU */
   6454static u8 cu_to_vcu(u32 cu)
   6455{
   6456	return ilog2(cu);
   6457}
   6458
   6459/* convert a vAU to an AU */
   6460static u32 vau_to_au(u8 vau)
   6461{
   6462	return 8 * (1 << vau);
   6463}
   6464
   6465static void set_linkup_defaults(struct hfi1_pportdata *ppd)
   6466{
   6467	ppd->sm_trap_qp = 0x0;
   6468	ppd->sa_qp = 0x1;
   6469}
   6470
   6471/*
   6472 * Graceful LCB shutdown.  This leaves the LCB FIFOs in reset.
   6473 */
   6474static void lcb_shutdown(struct hfi1_devdata *dd, int abort)
   6475{
   6476	u64 reg;
   6477
   6478	/* clear lcb run: LCB_CFG_RUN.EN = 0 */
   6479	write_csr(dd, DC_LCB_CFG_RUN, 0);
   6480	/* set tx fifo reset: LCB_CFG_TX_FIFOS_RESET.VAL = 1 */
   6481	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET,
   6482		  1ull << DC_LCB_CFG_TX_FIFOS_RESET_VAL_SHIFT);
   6483	/* set dcc reset csr: DCC_CFG_RESET.{reset_lcb,reset_rx_fpe} = 1 */
   6484	dd->lcb_err_en = read_csr(dd, DC_LCB_ERR_EN);
   6485	reg = read_csr(dd, DCC_CFG_RESET);
   6486	write_csr(dd, DCC_CFG_RESET, reg |
   6487		  DCC_CFG_RESET_RESET_LCB | DCC_CFG_RESET_RESET_RX_FPE);
   6488	(void)read_csr(dd, DCC_CFG_RESET); /* make sure the write completed */
   6489	if (!abort) {
   6490		udelay(1);    /* must hold for the longer of 16cclks or 20ns */
   6491		write_csr(dd, DCC_CFG_RESET, reg);
   6492		write_csr(dd, DC_LCB_ERR_EN, dd->lcb_err_en);
   6493	}
   6494}
   6495
   6496/*
   6497 * This routine should be called after the link has been transitioned to
   6498 * OFFLINE (OFFLINE state has the side effect of putting the SerDes into
   6499 * reset).
   6500 *
   6501 * The expectation is that the caller of this routine would have taken
   6502 * care of properly transitioning the link into the correct state.
   6503 * NOTE: the caller needs to acquire the dd->dc8051_lock lock
   6504 *       before calling this function.
   6505 */
   6506static void _dc_shutdown(struct hfi1_devdata *dd)
   6507{
   6508	lockdep_assert_held(&dd->dc8051_lock);
   6509
   6510	if (dd->dc_shutdown)
   6511		return;
   6512
   6513	dd->dc_shutdown = 1;
   6514	/* Shutdown the LCB */
   6515	lcb_shutdown(dd, 1);
   6516	/*
   6517	 * Going to OFFLINE would have causes the 8051 to put the
   6518	 * SerDes into reset already. Just need to shut down the 8051,
   6519	 * itself.
   6520	 */
   6521	write_csr(dd, DC_DC8051_CFG_RST, 0x1);
   6522}
   6523
   6524static void dc_shutdown(struct hfi1_devdata *dd)
   6525{
   6526	mutex_lock(&dd->dc8051_lock);
   6527	_dc_shutdown(dd);
   6528	mutex_unlock(&dd->dc8051_lock);
   6529}
   6530
   6531/*
   6532 * Calling this after the DC has been brought out of reset should not
   6533 * do any damage.
   6534 * NOTE: the caller needs to acquire the dd->dc8051_lock lock
   6535 *       before calling this function.
   6536 */
   6537static void _dc_start(struct hfi1_devdata *dd)
   6538{
   6539	lockdep_assert_held(&dd->dc8051_lock);
   6540
   6541	if (!dd->dc_shutdown)
   6542		return;
   6543
   6544	/* Take the 8051 out of reset */
   6545	write_csr(dd, DC_DC8051_CFG_RST, 0ull);
   6546	/* Wait until 8051 is ready */
   6547	if (wait_fm_ready(dd, TIMEOUT_8051_START))
   6548		dd_dev_err(dd, "%s: timeout starting 8051 firmware\n",
   6549			   __func__);
   6550
   6551	/* Take away reset for LCB and RX FPE (set in lcb_shutdown). */
   6552	write_csr(dd, DCC_CFG_RESET, LCB_RX_FPE_TX_FPE_OUT_OF_RESET);
   6553	/* lcb_shutdown() with abort=1 does not restore these */
   6554	write_csr(dd, DC_LCB_ERR_EN, dd->lcb_err_en);
   6555	dd->dc_shutdown = 0;
   6556}
   6557
   6558static void dc_start(struct hfi1_devdata *dd)
   6559{
   6560	mutex_lock(&dd->dc8051_lock);
   6561	_dc_start(dd);
   6562	mutex_unlock(&dd->dc8051_lock);
   6563}
   6564
   6565/*
   6566 * These LCB adjustments are for the Aurora SerDes core in the FPGA.
   6567 */
   6568static void adjust_lcb_for_fpga_serdes(struct hfi1_devdata *dd)
   6569{
   6570	u64 rx_radr, tx_radr;
   6571	u32 version;
   6572
   6573	if (dd->icode != ICODE_FPGA_EMULATION)
   6574		return;
   6575
   6576	/*
   6577	 * These LCB defaults on emulator _s are good, nothing to do here:
   6578	 *	LCB_CFG_TX_FIFOS_RADR
   6579	 *	LCB_CFG_RX_FIFOS_RADR
   6580	 *	LCB_CFG_LN_DCLK
   6581	 *	LCB_CFG_IGNORE_LOST_RCLK
   6582	 */
   6583	if (is_emulator_s(dd))
   6584		return;
   6585	/* else this is _p */
   6586
   6587	version = emulator_rev(dd);
   6588	if (!is_ax(dd))
   6589		version = 0x2d;	/* all B0 use 0x2d or higher settings */
   6590
   6591	if (version <= 0x12) {
   6592		/* release 0x12 and below */
   6593
   6594		/*
   6595		 * LCB_CFG_RX_FIFOS_RADR.RST_VAL = 0x9
   6596		 * LCB_CFG_RX_FIFOS_RADR.OK_TO_JUMP_VAL = 0x9
   6597		 * LCB_CFG_RX_FIFOS_RADR.DO_NOT_JUMP_VAL = 0xa
   6598		 */
   6599		rx_radr =
   6600		      0xaull << DC_LCB_CFG_RX_FIFOS_RADR_DO_NOT_JUMP_VAL_SHIFT
   6601		    | 0x9ull << DC_LCB_CFG_RX_FIFOS_RADR_OK_TO_JUMP_VAL_SHIFT
   6602		    | 0x9ull << DC_LCB_CFG_RX_FIFOS_RADR_RST_VAL_SHIFT;
   6603		/*
   6604		 * LCB_CFG_TX_FIFOS_RADR.ON_REINIT = 0 (default)
   6605		 * LCB_CFG_TX_FIFOS_RADR.RST_VAL = 6
   6606		 */
   6607		tx_radr = 6ull << DC_LCB_CFG_TX_FIFOS_RADR_RST_VAL_SHIFT;
   6608	} else if (version <= 0x18) {
   6609		/* release 0x13 up to 0x18 */
   6610		/* LCB_CFG_RX_FIFOS_RADR = 0x988 */
   6611		rx_radr =
   6612		      0x9ull << DC_LCB_CFG_RX_FIFOS_RADR_DO_NOT_JUMP_VAL_SHIFT
   6613		    | 0x8ull << DC_LCB_CFG_RX_FIFOS_RADR_OK_TO_JUMP_VAL_SHIFT
   6614		    | 0x8ull << DC_LCB_CFG_RX_FIFOS_RADR_RST_VAL_SHIFT;
   6615		tx_radr = 7ull << DC_LCB_CFG_TX_FIFOS_RADR_RST_VAL_SHIFT;
   6616	} else if (version == 0x19) {
   6617		/* release 0x19 */
   6618		/* LCB_CFG_RX_FIFOS_RADR = 0xa99 */
   6619		rx_radr =
   6620		      0xAull << DC_LCB_CFG_RX_FIFOS_RADR_DO_NOT_JUMP_VAL_SHIFT
   6621		    | 0x9ull << DC_LCB_CFG_RX_FIFOS_RADR_OK_TO_JUMP_VAL_SHIFT
   6622		    | 0x9ull << DC_LCB_CFG_RX_FIFOS_RADR_RST_VAL_SHIFT;
   6623		tx_radr = 3ull << DC_LCB_CFG_TX_FIFOS_RADR_RST_VAL_SHIFT;
   6624	} else if (version == 0x1a) {
   6625		/* release 0x1a */
   6626		/* LCB_CFG_RX_FIFOS_RADR = 0x988 */
   6627		rx_radr =
   6628		      0x9ull << DC_LCB_CFG_RX_FIFOS_RADR_DO_NOT_JUMP_VAL_SHIFT
   6629		    | 0x8ull << DC_LCB_CFG_RX_FIFOS_RADR_OK_TO_JUMP_VAL_SHIFT
   6630		    | 0x8ull << DC_LCB_CFG_RX_FIFOS_RADR_RST_VAL_SHIFT;
   6631		tx_radr = 7ull << DC_LCB_CFG_TX_FIFOS_RADR_RST_VAL_SHIFT;
   6632		write_csr(dd, DC_LCB_CFG_LN_DCLK, 1ull);
   6633	} else {
   6634		/* release 0x1b and higher */
   6635		/* LCB_CFG_RX_FIFOS_RADR = 0x877 */
   6636		rx_radr =
   6637		      0x8ull << DC_LCB_CFG_RX_FIFOS_RADR_DO_NOT_JUMP_VAL_SHIFT
   6638		    | 0x7ull << DC_LCB_CFG_RX_FIFOS_RADR_OK_TO_JUMP_VAL_SHIFT
   6639		    | 0x7ull << DC_LCB_CFG_RX_FIFOS_RADR_RST_VAL_SHIFT;
   6640		tx_radr = 3ull << DC_LCB_CFG_TX_FIFOS_RADR_RST_VAL_SHIFT;
   6641	}
   6642
   6643	write_csr(dd, DC_LCB_CFG_RX_FIFOS_RADR, rx_radr);
   6644	/* LCB_CFG_IGNORE_LOST_RCLK.EN = 1 */
   6645	write_csr(dd, DC_LCB_CFG_IGNORE_LOST_RCLK,
   6646		  DC_LCB_CFG_IGNORE_LOST_RCLK_EN_SMASK);
   6647	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RADR, tx_radr);
   6648}
   6649
   6650/*
   6651 * Handle a SMA idle message
   6652 *
   6653 * This is a work-queue function outside of the interrupt.
   6654 */
   6655void handle_sma_message(struct work_struct *work)
   6656{
   6657	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   6658							sma_message_work);
   6659	struct hfi1_devdata *dd = ppd->dd;
   6660	u64 msg;
   6661	int ret;
   6662
   6663	/*
   6664	 * msg is bytes 1-4 of the 40-bit idle message - the command code
   6665	 * is stripped off
   6666	 */
   6667	ret = read_idle_sma(dd, &msg);
   6668	if (ret)
   6669		return;
   6670	dd_dev_info(dd, "%s: SMA message 0x%llx\n", __func__, msg);
   6671	/*
   6672	 * React to the SMA message.  Byte[1] (0 for us) is the command.
   6673	 */
   6674	switch (msg & 0xff) {
   6675	case SMA_IDLE_ARM:
   6676		/*
   6677		 * See OPAv1 table 9-14 - HFI and External Switch Ports Key
   6678		 * State Transitions
   6679		 *
   6680		 * Only expected in INIT or ARMED, discard otherwise.
   6681		 */
   6682		if (ppd->host_link_state & (HLS_UP_INIT | HLS_UP_ARMED))
   6683			ppd->neighbor_normal = 1;
   6684		break;
   6685	case SMA_IDLE_ACTIVE:
   6686		/*
   6687		 * See OPAv1 table 9-14 - HFI and External Switch Ports Key
   6688		 * State Transitions
   6689		 *
   6690		 * Can activate the node.  Discard otherwise.
   6691		 */
   6692		if (ppd->host_link_state == HLS_UP_ARMED &&
   6693		    ppd->is_active_optimize_enabled) {
   6694			ppd->neighbor_normal = 1;
   6695			ret = set_link_state(ppd, HLS_UP_ACTIVE);
   6696			if (ret)
   6697				dd_dev_err(
   6698					dd,
   6699					"%s: received Active SMA idle message, couldn't set link to Active\n",
   6700					__func__);
   6701		}
   6702		break;
   6703	default:
   6704		dd_dev_err(dd,
   6705			   "%s: received unexpected SMA idle message 0x%llx\n",
   6706			   __func__, msg);
   6707		break;
   6708	}
   6709}
   6710
   6711static void adjust_rcvctrl(struct hfi1_devdata *dd, u64 add, u64 clear)
   6712{
   6713	u64 rcvctrl;
   6714	unsigned long flags;
   6715
   6716	spin_lock_irqsave(&dd->rcvctrl_lock, flags);
   6717	rcvctrl = read_csr(dd, RCV_CTRL);
   6718	rcvctrl |= add;
   6719	rcvctrl &= ~clear;
   6720	write_csr(dd, RCV_CTRL, rcvctrl);
   6721	spin_unlock_irqrestore(&dd->rcvctrl_lock, flags);
   6722}
   6723
   6724static inline void add_rcvctrl(struct hfi1_devdata *dd, u64 add)
   6725{
   6726	adjust_rcvctrl(dd, add, 0);
   6727}
   6728
   6729static inline void clear_rcvctrl(struct hfi1_devdata *dd, u64 clear)
   6730{
   6731	adjust_rcvctrl(dd, 0, clear);
   6732}
   6733
   6734/*
   6735 * Called from all interrupt handlers to start handling an SPC freeze.
   6736 */
   6737void start_freeze_handling(struct hfi1_pportdata *ppd, int flags)
   6738{
   6739	struct hfi1_devdata *dd = ppd->dd;
   6740	struct send_context *sc;
   6741	int i;
   6742	int sc_flags;
   6743
   6744	if (flags & FREEZE_SELF)
   6745		write_csr(dd, CCE_CTRL, CCE_CTRL_SPC_FREEZE_SMASK);
   6746
   6747	/* enter frozen mode */
   6748	dd->flags |= HFI1_FROZEN;
   6749
   6750	/* notify all SDMA engines that they are going into a freeze */
   6751	sdma_freeze_notify(dd, !!(flags & FREEZE_LINK_DOWN));
   6752
   6753	sc_flags = SCF_FROZEN | SCF_HALTED | (flags & FREEZE_LINK_DOWN ?
   6754					      SCF_LINK_DOWN : 0);
   6755	/* do halt pre-handling on all enabled send contexts */
   6756	for (i = 0; i < dd->num_send_contexts; i++) {
   6757		sc = dd->send_contexts[i].sc;
   6758		if (sc && (sc->flags & SCF_ENABLED))
   6759			sc_stop(sc, sc_flags);
   6760	}
   6761
   6762	/* Send context are frozen. Notify user space */
   6763	hfi1_set_uevent_bits(ppd, _HFI1_EVENT_FROZEN_BIT);
   6764
   6765	if (flags & FREEZE_ABORT) {
   6766		dd_dev_err(dd,
   6767			   "Aborted freeze recovery. Please REBOOT system\n");
   6768		return;
   6769	}
   6770	/* queue non-interrupt handler */
   6771	queue_work(ppd->hfi1_wq, &ppd->freeze_work);
   6772}
   6773
   6774/*
   6775 * Wait until all 4 sub-blocks indicate that they have frozen or unfrozen,
   6776 * depending on the "freeze" parameter.
   6777 *
   6778 * No need to return an error if it times out, our only option
   6779 * is to proceed anyway.
   6780 */
   6781static void wait_for_freeze_status(struct hfi1_devdata *dd, int freeze)
   6782{
   6783	unsigned long timeout;
   6784	u64 reg;
   6785
   6786	timeout = jiffies + msecs_to_jiffies(FREEZE_STATUS_TIMEOUT);
   6787	while (1) {
   6788		reg = read_csr(dd, CCE_STATUS);
   6789		if (freeze) {
   6790			/* waiting until all indicators are set */
   6791			if ((reg & ALL_FROZE) == ALL_FROZE)
   6792				return;	/* all done */
   6793		} else {
   6794			/* waiting until all indicators are clear */
   6795			if ((reg & ALL_FROZE) == 0)
   6796				return; /* all done */
   6797		}
   6798
   6799		if (time_after(jiffies, timeout)) {
   6800			dd_dev_err(dd,
   6801				   "Time out waiting for SPC %sfreeze, bits 0x%llx, expecting 0x%llx, continuing",
   6802				   freeze ? "" : "un", reg & ALL_FROZE,
   6803				   freeze ? ALL_FROZE : 0ull);
   6804			return;
   6805		}
   6806		usleep_range(80, 120);
   6807	}
   6808}
   6809
   6810/*
   6811 * Do all freeze handling for the RXE block.
   6812 */
   6813static void rxe_freeze(struct hfi1_devdata *dd)
   6814{
   6815	int i;
   6816	struct hfi1_ctxtdata *rcd;
   6817
   6818	/* disable port */
   6819	clear_rcvctrl(dd, RCV_CTRL_RCV_PORT_ENABLE_SMASK);
   6820
   6821	/* disable all receive contexts */
   6822	for (i = 0; i < dd->num_rcv_contexts; i++) {
   6823		rcd = hfi1_rcd_get_by_index(dd, i);
   6824		hfi1_rcvctrl(dd, HFI1_RCVCTRL_CTXT_DIS, rcd);
   6825		hfi1_rcd_put(rcd);
   6826	}
   6827}
   6828
   6829/*
   6830 * Unfreeze handling for the RXE block - kernel contexts only.
   6831 * This will also enable the port.  User contexts will do unfreeze
   6832 * handling on a per-context basis as they call into the driver.
   6833 *
   6834 */
   6835static void rxe_kernel_unfreeze(struct hfi1_devdata *dd)
   6836{
   6837	u32 rcvmask;
   6838	u16 i;
   6839	struct hfi1_ctxtdata *rcd;
   6840
   6841	/* enable all kernel contexts */
   6842	for (i = 0; i < dd->num_rcv_contexts; i++) {
   6843		rcd = hfi1_rcd_get_by_index(dd, i);
   6844
   6845		/* Ensure all non-user contexts(including vnic) are enabled */
   6846		if (!rcd ||
   6847		    (i >= dd->first_dyn_alloc_ctxt && !rcd->is_vnic)) {
   6848			hfi1_rcd_put(rcd);
   6849			continue;
   6850		}
   6851		rcvmask = HFI1_RCVCTRL_CTXT_ENB;
   6852		/* HFI1_RCVCTRL_TAILUPD_[ENB|DIS] needs to be set explicitly */
   6853		rcvmask |= hfi1_rcvhdrtail_kvaddr(rcd) ?
   6854			HFI1_RCVCTRL_TAILUPD_ENB : HFI1_RCVCTRL_TAILUPD_DIS;
   6855		hfi1_rcvctrl(dd, rcvmask, rcd);
   6856		hfi1_rcd_put(rcd);
   6857	}
   6858
   6859	/* enable port */
   6860	add_rcvctrl(dd, RCV_CTRL_RCV_PORT_ENABLE_SMASK);
   6861}
   6862
   6863/*
   6864 * Non-interrupt SPC freeze handling.
   6865 *
   6866 * This is a work-queue function outside of the triggering interrupt.
   6867 */
   6868void handle_freeze(struct work_struct *work)
   6869{
   6870	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   6871								freeze_work);
   6872	struct hfi1_devdata *dd = ppd->dd;
   6873
   6874	/* wait for freeze indicators on all affected blocks */
   6875	wait_for_freeze_status(dd, 1);
   6876
   6877	/* SPC is now frozen */
   6878
   6879	/* do send PIO freeze steps */
   6880	pio_freeze(dd);
   6881
   6882	/* do send DMA freeze steps */
   6883	sdma_freeze(dd);
   6884
   6885	/* do send egress freeze steps - nothing to do */
   6886
   6887	/* do receive freeze steps */
   6888	rxe_freeze(dd);
   6889
   6890	/*
   6891	 * Unfreeze the hardware - clear the freeze, wait for each
   6892	 * block's frozen bit to clear, then clear the frozen flag.
   6893	 */
   6894	write_csr(dd, CCE_CTRL, CCE_CTRL_SPC_UNFREEZE_SMASK);
   6895	wait_for_freeze_status(dd, 0);
   6896
   6897	if (is_ax(dd)) {
   6898		write_csr(dd, CCE_CTRL, CCE_CTRL_SPC_FREEZE_SMASK);
   6899		wait_for_freeze_status(dd, 1);
   6900		write_csr(dd, CCE_CTRL, CCE_CTRL_SPC_UNFREEZE_SMASK);
   6901		wait_for_freeze_status(dd, 0);
   6902	}
   6903
   6904	/* do send PIO unfreeze steps for kernel contexts */
   6905	pio_kernel_unfreeze(dd);
   6906
   6907	/* do send DMA unfreeze steps */
   6908	sdma_unfreeze(dd);
   6909
   6910	/* do send egress unfreeze steps - nothing to do */
   6911
   6912	/* do receive unfreeze steps for kernel contexts */
   6913	rxe_kernel_unfreeze(dd);
   6914
   6915	/*
   6916	 * The unfreeze procedure touches global device registers when
   6917	 * it disables and re-enables RXE. Mark the device unfrozen
   6918	 * after all that is done so other parts of the driver waiting
   6919	 * for the device to unfreeze don't do things out of order.
   6920	 *
   6921	 * The above implies that the meaning of HFI1_FROZEN flag is
   6922	 * "Device has gone into freeze mode and freeze mode handling
   6923	 * is still in progress."
   6924	 *
   6925	 * The flag will be removed when freeze mode processing has
   6926	 * completed.
   6927	 */
   6928	dd->flags &= ~HFI1_FROZEN;
   6929	wake_up(&dd->event_queue);
   6930
   6931	/* no longer frozen */
   6932}
   6933
   6934/**
   6935 * update_xmit_counters - update PortXmitWait/PortVlXmitWait
   6936 * counters.
   6937 * @ppd: info of physical Hfi port
   6938 * @link_width: new link width after link up or downgrade
   6939 *
   6940 * Update the PortXmitWait and PortVlXmitWait counters after
   6941 * a link up or downgrade event to reflect a link width change.
   6942 */
   6943static void update_xmit_counters(struct hfi1_pportdata *ppd, u16 link_width)
   6944{
   6945	int i;
   6946	u16 tx_width;
   6947	u16 link_speed;
   6948
   6949	tx_width = tx_link_width(link_width);
   6950	link_speed = get_link_speed(ppd->link_speed_active);
   6951
   6952	/*
   6953	 * There are C_VL_COUNT number of PortVLXmitWait counters.
   6954	 * Adding 1 to C_VL_COUNT to include the PortXmitWait counter.
   6955	 */
   6956	for (i = 0; i < C_VL_COUNT + 1; i++)
   6957		get_xmit_wait_counters(ppd, tx_width, link_speed, i);
   6958}
   6959
   6960/*
   6961 * Handle a link up interrupt from the 8051.
   6962 *
   6963 * This is a work-queue function outside of the interrupt.
   6964 */
   6965void handle_link_up(struct work_struct *work)
   6966{
   6967	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   6968						  link_up_work);
   6969	struct hfi1_devdata *dd = ppd->dd;
   6970
   6971	set_link_state(ppd, HLS_UP_INIT);
   6972
   6973	/* cache the read of DC_LCB_STS_ROUND_TRIP_LTP_CNT */
   6974	read_ltp_rtt(dd);
   6975	/*
   6976	 * OPA specifies that certain counters are cleared on a transition
   6977	 * to link up, so do that.
   6978	 */
   6979	clear_linkup_counters(dd);
   6980	/*
   6981	 * And (re)set link up default values.
   6982	 */
   6983	set_linkup_defaults(ppd);
   6984
   6985	/*
   6986	 * Set VL15 credits. Use cached value from verify cap interrupt.
   6987	 * In case of quick linkup or simulator, vl15 value will be set by
   6988	 * handle_linkup_change. VerifyCap interrupt handler will not be
   6989	 * called in those scenarios.
   6990	 */
   6991	if (!(quick_linkup || dd->icode == ICODE_FUNCTIONAL_SIMULATOR))
   6992		set_up_vl15(dd, dd->vl15buf_cached);
   6993
   6994	/* enforce link speed enabled */
   6995	if ((ppd->link_speed_active & ppd->link_speed_enabled) == 0) {
   6996		/* oops - current speed is not enabled, bounce */
   6997		dd_dev_err(dd,
   6998			   "Link speed active 0x%x is outside enabled 0x%x, downing link\n",
   6999			   ppd->link_speed_active, ppd->link_speed_enabled);
   7000		set_link_down_reason(ppd, OPA_LINKDOWN_REASON_SPEED_POLICY, 0,
   7001				     OPA_LINKDOWN_REASON_SPEED_POLICY);
   7002		set_link_state(ppd, HLS_DN_OFFLINE);
   7003		start_link(ppd);
   7004	}
   7005}
   7006
   7007/*
   7008 * Several pieces of LNI information were cached for SMA in ppd.
   7009 * Reset these on link down
   7010 */
   7011static void reset_neighbor_info(struct hfi1_pportdata *ppd)
   7012{
   7013	ppd->neighbor_guid = 0;
   7014	ppd->neighbor_port_number = 0;
   7015	ppd->neighbor_type = 0;
   7016	ppd->neighbor_fm_security = 0;
   7017}
   7018
   7019static const char * const link_down_reason_strs[] = {
   7020	[OPA_LINKDOWN_REASON_NONE] = "None",
   7021	[OPA_LINKDOWN_REASON_RCV_ERROR_0] = "Receive error 0",
   7022	[OPA_LINKDOWN_REASON_BAD_PKT_LEN] = "Bad packet length",
   7023	[OPA_LINKDOWN_REASON_PKT_TOO_LONG] = "Packet too long",
   7024	[OPA_LINKDOWN_REASON_PKT_TOO_SHORT] = "Packet too short",
   7025	[OPA_LINKDOWN_REASON_BAD_SLID] = "Bad SLID",
   7026	[OPA_LINKDOWN_REASON_BAD_DLID] = "Bad DLID",
   7027	[OPA_LINKDOWN_REASON_BAD_L2] = "Bad L2",
   7028	[OPA_LINKDOWN_REASON_BAD_SC] = "Bad SC",
   7029	[OPA_LINKDOWN_REASON_RCV_ERROR_8] = "Receive error 8",
   7030	[OPA_LINKDOWN_REASON_BAD_MID_TAIL] = "Bad mid tail",
   7031	[OPA_LINKDOWN_REASON_RCV_ERROR_10] = "Receive error 10",
   7032	[OPA_LINKDOWN_REASON_PREEMPT_ERROR] = "Preempt error",
   7033	[OPA_LINKDOWN_REASON_PREEMPT_VL15] = "Preempt vl15",
   7034	[OPA_LINKDOWN_REASON_BAD_VL_MARKER] = "Bad VL marker",
   7035	[OPA_LINKDOWN_REASON_RCV_ERROR_14] = "Receive error 14",
   7036	[OPA_LINKDOWN_REASON_RCV_ERROR_15] = "Receive error 15",
   7037	[OPA_LINKDOWN_REASON_BAD_HEAD_DIST] = "Bad head distance",
   7038	[OPA_LINKDOWN_REASON_BAD_TAIL_DIST] = "Bad tail distance",
   7039	[OPA_LINKDOWN_REASON_BAD_CTRL_DIST] = "Bad control distance",
   7040	[OPA_LINKDOWN_REASON_BAD_CREDIT_ACK] = "Bad credit ack",
   7041	[OPA_LINKDOWN_REASON_UNSUPPORTED_VL_MARKER] = "Unsupported VL marker",
   7042	[OPA_LINKDOWN_REASON_BAD_PREEMPT] = "Bad preempt",
   7043	[OPA_LINKDOWN_REASON_BAD_CONTROL_FLIT] = "Bad control flit",
   7044	[OPA_LINKDOWN_REASON_EXCEED_MULTICAST_LIMIT] = "Exceed multicast limit",
   7045	[OPA_LINKDOWN_REASON_RCV_ERROR_24] = "Receive error 24",
   7046	[OPA_LINKDOWN_REASON_RCV_ERROR_25] = "Receive error 25",
   7047	[OPA_LINKDOWN_REASON_RCV_ERROR_26] = "Receive error 26",
   7048	[OPA_LINKDOWN_REASON_RCV_ERROR_27] = "Receive error 27",
   7049	[OPA_LINKDOWN_REASON_RCV_ERROR_28] = "Receive error 28",
   7050	[OPA_LINKDOWN_REASON_RCV_ERROR_29] = "Receive error 29",
   7051	[OPA_LINKDOWN_REASON_RCV_ERROR_30] = "Receive error 30",
   7052	[OPA_LINKDOWN_REASON_EXCESSIVE_BUFFER_OVERRUN] =
   7053					"Excessive buffer overrun",
   7054	[OPA_LINKDOWN_REASON_UNKNOWN] = "Unknown",
   7055	[OPA_LINKDOWN_REASON_REBOOT] = "Reboot",
   7056	[OPA_LINKDOWN_REASON_NEIGHBOR_UNKNOWN] = "Neighbor unknown",
   7057	[OPA_LINKDOWN_REASON_FM_BOUNCE] = "FM bounce",
   7058	[OPA_LINKDOWN_REASON_SPEED_POLICY] = "Speed policy",
   7059	[OPA_LINKDOWN_REASON_WIDTH_POLICY] = "Width policy",
   7060	[OPA_LINKDOWN_REASON_DISCONNECTED] = "Disconnected",
   7061	[OPA_LINKDOWN_REASON_LOCAL_MEDIA_NOT_INSTALLED] =
   7062					"Local media not installed",
   7063	[OPA_LINKDOWN_REASON_NOT_INSTALLED] = "Not installed",
   7064	[OPA_LINKDOWN_REASON_CHASSIS_CONFIG] = "Chassis config",
   7065	[OPA_LINKDOWN_REASON_END_TO_END_NOT_INSTALLED] =
   7066					"End to end not installed",
   7067	[OPA_LINKDOWN_REASON_POWER_POLICY] = "Power policy",
   7068	[OPA_LINKDOWN_REASON_LINKSPEED_POLICY] = "Link speed policy",
   7069	[OPA_LINKDOWN_REASON_LINKWIDTH_POLICY] = "Link width policy",
   7070	[OPA_LINKDOWN_REASON_SWITCH_MGMT] = "Switch management",
   7071	[OPA_LINKDOWN_REASON_SMA_DISABLED] = "SMA disabled",
   7072	[OPA_LINKDOWN_REASON_TRANSIENT] = "Transient"
   7073};
   7074
   7075/* return the neighbor link down reason string */
   7076static const char *link_down_reason_str(u8 reason)
   7077{
   7078	const char *str = NULL;
   7079
   7080	if (reason < ARRAY_SIZE(link_down_reason_strs))
   7081		str = link_down_reason_strs[reason];
   7082	if (!str)
   7083		str = "(invalid)";
   7084
   7085	return str;
   7086}
   7087
   7088/*
   7089 * Handle a link down interrupt from the 8051.
   7090 *
   7091 * This is a work-queue function outside of the interrupt.
   7092 */
   7093void handle_link_down(struct work_struct *work)
   7094{
   7095	u8 lcl_reason, neigh_reason = 0;
   7096	u8 link_down_reason;
   7097	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   7098						  link_down_work);
   7099	int was_up;
   7100	static const char ldr_str[] = "Link down reason: ";
   7101
   7102	if ((ppd->host_link_state &
   7103	     (HLS_DN_POLL | HLS_VERIFY_CAP | HLS_GOING_UP)) &&
   7104	     ppd->port_type == PORT_TYPE_FIXED)
   7105		ppd->offline_disabled_reason =
   7106			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NOT_INSTALLED);
   7107
   7108	/* Go offline first, then deal with reading/writing through 8051 */
   7109	was_up = !!(ppd->host_link_state & HLS_UP);
   7110	set_link_state(ppd, HLS_DN_OFFLINE);
   7111	xchg(&ppd->is_link_down_queued, 0);
   7112
   7113	if (was_up) {
   7114		lcl_reason = 0;
   7115		/* link down reason is only valid if the link was up */
   7116		read_link_down_reason(ppd->dd, &link_down_reason);
   7117		switch (link_down_reason) {
   7118		case LDR_LINK_TRANSFER_ACTIVE_LOW:
   7119			/* the link went down, no idle message reason */
   7120			dd_dev_info(ppd->dd, "%sUnexpected link down\n",
   7121				    ldr_str);
   7122			break;
   7123		case LDR_RECEIVED_LINKDOWN_IDLE_MSG:
   7124			/*
   7125			 * The neighbor reason is only valid if an idle message
   7126			 * was received for it.
   7127			 */
   7128			read_planned_down_reason_code(ppd->dd, &neigh_reason);
   7129			dd_dev_info(ppd->dd,
   7130				    "%sNeighbor link down message %d, %s\n",
   7131				    ldr_str, neigh_reason,
   7132				    link_down_reason_str(neigh_reason));
   7133			break;
   7134		case LDR_RECEIVED_HOST_OFFLINE_REQ:
   7135			dd_dev_info(ppd->dd,
   7136				    "%sHost requested link to go offline\n",
   7137				    ldr_str);
   7138			break;
   7139		default:
   7140			dd_dev_info(ppd->dd, "%sUnknown reason 0x%x\n",
   7141				    ldr_str, link_down_reason);
   7142			break;
   7143		}
   7144
   7145		/*
   7146		 * If no reason, assume peer-initiated but missed
   7147		 * LinkGoingDown idle flits.
   7148		 */
   7149		if (neigh_reason == 0)
   7150			lcl_reason = OPA_LINKDOWN_REASON_NEIGHBOR_UNKNOWN;
   7151	} else {
   7152		/* went down while polling or going up */
   7153		lcl_reason = OPA_LINKDOWN_REASON_TRANSIENT;
   7154	}
   7155
   7156	set_link_down_reason(ppd, lcl_reason, neigh_reason, 0);
   7157
   7158	/* inform the SMA when the link transitions from up to down */
   7159	if (was_up && ppd->local_link_down_reason.sma == 0 &&
   7160	    ppd->neigh_link_down_reason.sma == 0) {
   7161		ppd->local_link_down_reason.sma =
   7162					ppd->local_link_down_reason.latest;
   7163		ppd->neigh_link_down_reason.sma =
   7164					ppd->neigh_link_down_reason.latest;
   7165	}
   7166
   7167	reset_neighbor_info(ppd);
   7168
   7169	/* disable the port */
   7170	clear_rcvctrl(ppd->dd, RCV_CTRL_RCV_PORT_ENABLE_SMASK);
   7171
   7172	/*
   7173	 * If there is no cable attached, turn the DC off. Otherwise,
   7174	 * start the link bring up.
   7175	 */
   7176	if (ppd->port_type == PORT_TYPE_QSFP && !qsfp_mod_present(ppd))
   7177		dc_shutdown(ppd->dd);
   7178	else
   7179		start_link(ppd);
   7180}
   7181
   7182void handle_link_bounce(struct work_struct *work)
   7183{
   7184	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   7185							link_bounce_work);
   7186
   7187	/*
   7188	 * Only do something if the link is currently up.
   7189	 */
   7190	if (ppd->host_link_state & HLS_UP) {
   7191		set_link_state(ppd, HLS_DN_OFFLINE);
   7192		start_link(ppd);
   7193	} else {
   7194		dd_dev_info(ppd->dd, "%s: link not up (%s), nothing to do\n",
   7195			    __func__, link_state_name(ppd->host_link_state));
   7196	}
   7197}
   7198
   7199/*
   7200 * Mask conversion: Capability exchange to Port LTP.  The capability
   7201 * exchange has an implicit 16b CRC that is mandatory.
   7202 */
   7203static int cap_to_port_ltp(int cap)
   7204{
   7205	int port_ltp = PORT_LTP_CRC_MODE_16; /* this mode is mandatory */
   7206
   7207	if (cap & CAP_CRC_14B)
   7208		port_ltp |= PORT_LTP_CRC_MODE_14;
   7209	if (cap & CAP_CRC_48B)
   7210		port_ltp |= PORT_LTP_CRC_MODE_48;
   7211	if (cap & CAP_CRC_12B_16B_PER_LANE)
   7212		port_ltp |= PORT_LTP_CRC_MODE_PER_LANE;
   7213
   7214	return port_ltp;
   7215}
   7216
   7217/*
   7218 * Convert an OPA Port LTP mask to capability mask
   7219 */
   7220int port_ltp_to_cap(int port_ltp)
   7221{
   7222	int cap_mask = 0;
   7223
   7224	if (port_ltp & PORT_LTP_CRC_MODE_14)
   7225		cap_mask |= CAP_CRC_14B;
   7226	if (port_ltp & PORT_LTP_CRC_MODE_48)
   7227		cap_mask |= CAP_CRC_48B;
   7228	if (port_ltp & PORT_LTP_CRC_MODE_PER_LANE)
   7229		cap_mask |= CAP_CRC_12B_16B_PER_LANE;
   7230
   7231	return cap_mask;
   7232}
   7233
   7234/*
   7235 * Convert a single DC LCB CRC mode to an OPA Port LTP mask.
   7236 */
   7237static int lcb_to_port_ltp(int lcb_crc)
   7238{
   7239	int port_ltp = 0;
   7240
   7241	if (lcb_crc == LCB_CRC_12B_16B_PER_LANE)
   7242		port_ltp = PORT_LTP_CRC_MODE_PER_LANE;
   7243	else if (lcb_crc == LCB_CRC_48B)
   7244		port_ltp = PORT_LTP_CRC_MODE_48;
   7245	else if (lcb_crc == LCB_CRC_14B)
   7246		port_ltp = PORT_LTP_CRC_MODE_14;
   7247	else
   7248		port_ltp = PORT_LTP_CRC_MODE_16;
   7249
   7250	return port_ltp;
   7251}
   7252
   7253static void clear_full_mgmt_pkey(struct hfi1_pportdata *ppd)
   7254{
   7255	if (ppd->pkeys[2] != 0) {
   7256		ppd->pkeys[2] = 0;
   7257		(void)hfi1_set_ib_cfg(ppd, HFI1_IB_CFG_PKEYS, 0);
   7258		hfi1_event_pkey_change(ppd->dd, ppd->port);
   7259	}
   7260}
   7261
   7262/*
   7263 * Convert the given link width to the OPA link width bitmask.
   7264 */
   7265static u16 link_width_to_bits(struct hfi1_devdata *dd, u16 width)
   7266{
   7267	switch (width) {
   7268	case 0:
   7269		/*
   7270		 * Simulator and quick linkup do not set the width.
   7271		 * Just set it to 4x without complaint.
   7272		 */
   7273		if (dd->icode == ICODE_FUNCTIONAL_SIMULATOR || quick_linkup)
   7274			return OPA_LINK_WIDTH_4X;
   7275		return 0; /* no lanes up */
   7276	case 1: return OPA_LINK_WIDTH_1X;
   7277	case 2: return OPA_LINK_WIDTH_2X;
   7278	case 3: return OPA_LINK_WIDTH_3X;
   7279	case 4: return OPA_LINK_WIDTH_4X;
   7280	default:
   7281		dd_dev_info(dd, "%s: invalid width %d, using 4\n",
   7282			    __func__, width);
   7283		return OPA_LINK_WIDTH_4X;
   7284	}
   7285}
   7286
   7287/*
   7288 * Do a population count on the bottom nibble.
   7289 */
   7290static const u8 bit_counts[16] = {
   7291	0, 1, 1, 2, 1, 2, 2, 3, 1, 2, 2, 3, 2, 3, 3, 4
   7292};
   7293
   7294static inline u8 nibble_to_count(u8 nibble)
   7295{
   7296	return bit_counts[nibble & 0xf];
   7297}
   7298
   7299/*
   7300 * Read the active lane information from the 8051 registers and return
   7301 * their widths.
   7302 *
   7303 * Active lane information is found in these 8051 registers:
   7304 *	enable_lane_tx
   7305 *	enable_lane_rx
   7306 */
   7307static void get_link_widths(struct hfi1_devdata *dd, u16 *tx_width,
   7308			    u16 *rx_width)
   7309{
   7310	u16 tx, rx;
   7311	u8 enable_lane_rx;
   7312	u8 enable_lane_tx;
   7313	u8 tx_polarity_inversion;
   7314	u8 rx_polarity_inversion;
   7315	u8 max_rate;
   7316
   7317	/* read the active lanes */
   7318	read_tx_settings(dd, &enable_lane_tx, &tx_polarity_inversion,
   7319			 &rx_polarity_inversion, &max_rate);
   7320	read_local_lni(dd, &enable_lane_rx);
   7321
   7322	/* convert to counts */
   7323	tx = nibble_to_count(enable_lane_tx);
   7324	rx = nibble_to_count(enable_lane_rx);
   7325
   7326	/*
   7327	 * Set link_speed_active here, overriding what was set in
   7328	 * handle_verify_cap().  The ASIC 8051 firmware does not correctly
   7329	 * set the max_rate field in handle_verify_cap until v0.19.
   7330	 */
   7331	if ((dd->icode == ICODE_RTL_SILICON) &&
   7332	    (dd->dc8051_ver < dc8051_ver(0, 19, 0))) {
   7333		/* max_rate: 0 = 12.5G, 1 = 25G */
   7334		switch (max_rate) {
   7335		case 0:
   7336			dd->pport[0].link_speed_active = OPA_LINK_SPEED_12_5G;
   7337			break;
   7338		case 1:
   7339			dd->pport[0].link_speed_active = OPA_LINK_SPEED_25G;
   7340			break;
   7341		default:
   7342			dd_dev_err(dd,
   7343				   "%s: unexpected max rate %d, using 25Gb\n",
   7344				   __func__, (int)max_rate);
   7345			dd->pport[0].link_speed_active = OPA_LINK_SPEED_25G;
   7346			break;
   7347		}
   7348	}
   7349
   7350	dd_dev_info(dd,
   7351		    "Fabric active lanes (width): tx 0x%x (%d), rx 0x%x (%d)\n",
   7352		    enable_lane_tx, tx, enable_lane_rx, rx);
   7353	*tx_width = link_width_to_bits(dd, tx);
   7354	*rx_width = link_width_to_bits(dd, rx);
   7355}
   7356
   7357/*
   7358 * Read verify_cap_local_fm_link_width[1] to obtain the link widths.
   7359 * Valid after the end of VerifyCap and during LinkUp.  Does not change
   7360 * after link up.  I.e. look elsewhere for downgrade information.
   7361 *
   7362 * Bits are:
   7363 *	+ bits [7:4] contain the number of active transmitters
   7364 *	+ bits [3:0] contain the number of active receivers
   7365 * These are numbers 1 through 4 and can be different values if the
   7366 * link is asymmetric.
   7367 *
   7368 * verify_cap_local_fm_link_width[0] retains its original value.
   7369 */
   7370static void get_linkup_widths(struct hfi1_devdata *dd, u16 *tx_width,
   7371			      u16 *rx_width)
   7372{
   7373	u16 widths, tx, rx;
   7374	u8 misc_bits, local_flags;
   7375	u16 active_tx, active_rx;
   7376
   7377	read_vc_local_link_mode(dd, &misc_bits, &local_flags, &widths);
   7378	tx = widths >> 12;
   7379	rx = (widths >> 8) & 0xf;
   7380
   7381	*tx_width = link_width_to_bits(dd, tx);
   7382	*rx_width = link_width_to_bits(dd, rx);
   7383
   7384	/* print the active widths */
   7385	get_link_widths(dd, &active_tx, &active_rx);
   7386}
   7387
   7388/*
   7389 * Set ppd->link_width_active and ppd->link_width_downgrade_active using
   7390 * hardware information when the link first comes up.
   7391 *
   7392 * The link width is not available until after VerifyCap.AllFramesReceived
   7393 * (the trigger for handle_verify_cap), so this is outside that routine
   7394 * and should be called when the 8051 signals linkup.
   7395 */
   7396void get_linkup_link_widths(struct hfi1_pportdata *ppd)
   7397{
   7398	u16 tx_width, rx_width;
   7399
   7400	/* get end-of-LNI link widths */
   7401	get_linkup_widths(ppd->dd, &tx_width, &rx_width);
   7402
   7403	/* use tx_width as the link is supposed to be symmetric on link up */
   7404	ppd->link_width_active = tx_width;
   7405	/* link width downgrade active (LWD.A) starts out matching LW.A */
   7406	ppd->link_width_downgrade_tx_active = ppd->link_width_active;
   7407	ppd->link_width_downgrade_rx_active = ppd->link_width_active;
   7408	/* per OPA spec, on link up LWD.E resets to LWD.S */
   7409	ppd->link_width_downgrade_enabled = ppd->link_width_downgrade_supported;
   7410	/* cache the active egress rate (units {10^6 bits/sec]) */
   7411	ppd->current_egress_rate = active_egress_rate(ppd);
   7412}
   7413
   7414/*
   7415 * Handle a verify capabilities interrupt from the 8051.
   7416 *
   7417 * This is a work-queue function outside of the interrupt.
   7418 */
   7419void handle_verify_cap(struct work_struct *work)
   7420{
   7421	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   7422								link_vc_work);
   7423	struct hfi1_devdata *dd = ppd->dd;
   7424	u64 reg;
   7425	u8 power_management;
   7426	u8 continuous;
   7427	u8 vcu;
   7428	u8 vau;
   7429	u8 z;
   7430	u16 vl15buf;
   7431	u16 link_widths;
   7432	u16 crc_mask;
   7433	u16 crc_val;
   7434	u16 device_id;
   7435	u16 active_tx, active_rx;
   7436	u8 partner_supported_crc;
   7437	u8 remote_tx_rate;
   7438	u8 device_rev;
   7439
   7440	set_link_state(ppd, HLS_VERIFY_CAP);
   7441
   7442	lcb_shutdown(dd, 0);
   7443	adjust_lcb_for_fpga_serdes(dd);
   7444
   7445	read_vc_remote_phy(dd, &power_management, &continuous);
   7446	read_vc_remote_fabric(dd, &vau, &z, &vcu, &vl15buf,
   7447			      &partner_supported_crc);
   7448	read_vc_remote_link_width(dd, &remote_tx_rate, &link_widths);
   7449	read_remote_device_id(dd, &device_id, &device_rev);
   7450
   7451	/* print the active widths */
   7452	get_link_widths(dd, &active_tx, &active_rx);
   7453	dd_dev_info(dd,
   7454		    "Peer PHY: power management 0x%x, continuous updates 0x%x\n",
   7455		    (int)power_management, (int)continuous);
   7456	dd_dev_info(dd,
   7457		    "Peer Fabric: vAU %d, Z %d, vCU %d, vl15 credits 0x%x, CRC sizes 0x%x\n",
   7458		    (int)vau, (int)z, (int)vcu, (int)vl15buf,
   7459		    (int)partner_supported_crc);
   7460	dd_dev_info(dd, "Peer Link Width: tx rate 0x%x, widths 0x%x\n",
   7461		    (u32)remote_tx_rate, (u32)link_widths);
   7462	dd_dev_info(dd, "Peer Device ID: 0x%04x, Revision 0x%02x\n",
   7463		    (u32)device_id, (u32)device_rev);
   7464	/*
   7465	 * The peer vAU value just read is the peer receiver value.  HFI does
   7466	 * not support a transmit vAU of 0 (AU == 8).  We advertised that
   7467	 * with Z=1 in the fabric capabilities sent to the peer.  The peer
   7468	 * will see our Z=1, and, if it advertised a vAU of 0, will move its
   7469	 * receive to vAU of 1 (AU == 16).  Do the same here.  We do not care
   7470	 * about the peer Z value - our sent vAU is 3 (hardwired) and is not
   7471	 * subject to the Z value exception.
   7472	 */
   7473	if (vau == 0)
   7474		vau = 1;
   7475	set_up_vau(dd, vau);
   7476
   7477	/*
   7478	 * Set VL15 credits to 0 in global credit register. Cache remote VL15
   7479	 * credits value and wait for link-up interrupt ot set it.
   7480	 */
   7481	set_up_vl15(dd, 0);
   7482	dd->vl15buf_cached = vl15buf;
   7483
   7484	/* set up the LCB CRC mode */
   7485	crc_mask = ppd->port_crc_mode_enabled & partner_supported_crc;
   7486
   7487	/* order is important: use the lowest bit in common */
   7488	if (crc_mask & CAP_CRC_14B)
   7489		crc_val = LCB_CRC_14B;
   7490	else if (crc_mask & CAP_CRC_48B)
   7491		crc_val = LCB_CRC_48B;
   7492	else if (crc_mask & CAP_CRC_12B_16B_PER_LANE)
   7493		crc_val = LCB_CRC_12B_16B_PER_LANE;
   7494	else
   7495		crc_val = LCB_CRC_16B;
   7496
   7497	dd_dev_info(dd, "Final LCB CRC mode: %d\n", (int)crc_val);
   7498	write_csr(dd, DC_LCB_CFG_CRC_MODE,
   7499		  (u64)crc_val << DC_LCB_CFG_CRC_MODE_TX_VAL_SHIFT);
   7500
   7501	/* set (14b only) or clear sideband credit */
   7502	reg = read_csr(dd, SEND_CM_CTRL);
   7503	if (crc_val == LCB_CRC_14B && crc_14b_sideband) {
   7504		write_csr(dd, SEND_CM_CTRL,
   7505			  reg | SEND_CM_CTRL_FORCE_CREDIT_MODE_SMASK);
   7506	} else {
   7507		write_csr(dd, SEND_CM_CTRL,
   7508			  reg & ~SEND_CM_CTRL_FORCE_CREDIT_MODE_SMASK);
   7509	}
   7510
   7511	ppd->link_speed_active = 0;	/* invalid value */
   7512	if (dd->dc8051_ver < dc8051_ver(0, 20, 0)) {
   7513		/* remote_tx_rate: 0 = 12.5G, 1 = 25G */
   7514		switch (remote_tx_rate) {
   7515		case 0:
   7516			ppd->link_speed_active = OPA_LINK_SPEED_12_5G;
   7517			break;
   7518		case 1:
   7519			ppd->link_speed_active = OPA_LINK_SPEED_25G;
   7520			break;
   7521		}
   7522	} else {
   7523		/* actual rate is highest bit of the ANDed rates */
   7524		u8 rate = remote_tx_rate & ppd->local_tx_rate;
   7525
   7526		if (rate & 2)
   7527			ppd->link_speed_active = OPA_LINK_SPEED_25G;
   7528		else if (rate & 1)
   7529			ppd->link_speed_active = OPA_LINK_SPEED_12_5G;
   7530	}
   7531	if (ppd->link_speed_active == 0) {
   7532		dd_dev_err(dd, "%s: unexpected remote tx rate %d, using 25Gb\n",
   7533			   __func__, (int)remote_tx_rate);
   7534		ppd->link_speed_active = OPA_LINK_SPEED_25G;
   7535	}
   7536
   7537	/*
   7538	 * Cache the values of the supported, enabled, and active
   7539	 * LTP CRC modes to return in 'portinfo' queries. But the bit
   7540	 * flags that are returned in the portinfo query differ from
   7541	 * what's in the link_crc_mask, crc_sizes, and crc_val
   7542	 * variables. Convert these here.
   7543	 */
   7544	ppd->port_ltp_crc_mode = cap_to_port_ltp(link_crc_mask) << 8;
   7545		/* supported crc modes */
   7546	ppd->port_ltp_crc_mode |=
   7547		cap_to_port_ltp(ppd->port_crc_mode_enabled) << 4;
   7548		/* enabled crc modes */
   7549	ppd->port_ltp_crc_mode |= lcb_to_port_ltp(crc_val);
   7550		/* active crc mode */
   7551
   7552	/* set up the remote credit return table */
   7553	assign_remote_cm_au_table(dd, vcu);
   7554
   7555	/*
   7556	 * The LCB is reset on entry to handle_verify_cap(), so this must
   7557	 * be applied on every link up.
   7558	 *
   7559	 * Adjust LCB error kill enable to kill the link if
   7560	 * these RBUF errors are seen:
   7561	 *	REPLAY_BUF_MBE_SMASK
   7562	 *	FLIT_INPUT_BUF_MBE_SMASK
   7563	 */
   7564	if (is_ax(dd)) {			/* fixed in B0 */
   7565		reg = read_csr(dd, DC_LCB_CFG_LINK_KILL_EN);
   7566		reg |= DC_LCB_CFG_LINK_KILL_EN_REPLAY_BUF_MBE_SMASK
   7567			| DC_LCB_CFG_LINK_KILL_EN_FLIT_INPUT_BUF_MBE_SMASK;
   7568		write_csr(dd, DC_LCB_CFG_LINK_KILL_EN, reg);
   7569	}
   7570
   7571	/* pull LCB fifos out of reset - all fifo clocks must be stable */
   7572	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET, 0);
   7573
   7574	/* give 8051 access to the LCB CSRs */
   7575	write_csr(dd, DC_LCB_ERR_EN, 0); /* mask LCB errors */
   7576	set_8051_lcb_access(dd);
   7577
   7578	/* tell the 8051 to go to LinkUp */
   7579	set_link_state(ppd, HLS_GOING_UP);
   7580}
   7581
   7582/**
   7583 * apply_link_downgrade_policy - Apply the link width downgrade enabled
   7584 * policy against the current active link widths.
   7585 * @ppd: info of physical Hfi port
   7586 * @refresh_widths: True indicates link downgrade event
   7587 * @return: True indicates a successful link downgrade. False indicates
   7588 *	    link downgrade event failed and the link will bounce back to
   7589 *	    default link width.
   7590 *
   7591 * Called when the enabled policy changes or the active link widths
   7592 * change.
   7593 * Refresh_widths indicates that a link downgrade occurred. The
   7594 * link_downgraded variable is set by refresh_widths and
   7595 * determines the success/failure of the policy application.
   7596 */
   7597bool apply_link_downgrade_policy(struct hfi1_pportdata *ppd,
   7598				 bool refresh_widths)
   7599{
   7600	int do_bounce = 0;
   7601	int tries;
   7602	u16 lwde;
   7603	u16 tx, rx;
   7604	bool link_downgraded = refresh_widths;
   7605
   7606	/* use the hls lock to avoid a race with actual link up */
   7607	tries = 0;
   7608retry:
   7609	mutex_lock(&ppd->hls_lock);
   7610	/* only apply if the link is up */
   7611	if (ppd->host_link_state & HLS_DOWN) {
   7612		/* still going up..wait and retry */
   7613		if (ppd->host_link_state & HLS_GOING_UP) {
   7614			if (++tries < 1000) {
   7615				mutex_unlock(&ppd->hls_lock);
   7616				usleep_range(100, 120); /* arbitrary */
   7617				goto retry;
   7618			}
   7619			dd_dev_err(ppd->dd,
   7620				   "%s: giving up waiting for link state change\n",
   7621				   __func__);
   7622		}
   7623		goto done;
   7624	}
   7625
   7626	lwde = ppd->link_width_downgrade_enabled;
   7627
   7628	if (refresh_widths) {
   7629		get_link_widths(ppd->dd, &tx, &rx);
   7630		ppd->link_width_downgrade_tx_active = tx;
   7631		ppd->link_width_downgrade_rx_active = rx;
   7632	}
   7633
   7634	if (ppd->link_width_downgrade_tx_active == 0 ||
   7635	    ppd->link_width_downgrade_rx_active == 0) {
   7636		/* the 8051 reported a dead link as a downgrade */
   7637		dd_dev_err(ppd->dd, "Link downgrade is really a link down, ignoring\n");
   7638		link_downgraded = false;
   7639	} else if (lwde == 0) {
   7640		/* downgrade is disabled */
   7641
   7642		/* bounce if not at starting active width */
   7643		if ((ppd->link_width_active !=
   7644		     ppd->link_width_downgrade_tx_active) ||
   7645		    (ppd->link_width_active !=
   7646		     ppd->link_width_downgrade_rx_active)) {
   7647			dd_dev_err(ppd->dd,
   7648				   "Link downgrade is disabled and link has downgraded, downing link\n");
   7649			dd_dev_err(ppd->dd,
   7650				   "  original 0x%x, tx active 0x%x, rx active 0x%x\n",
   7651				   ppd->link_width_active,
   7652				   ppd->link_width_downgrade_tx_active,
   7653				   ppd->link_width_downgrade_rx_active);
   7654			do_bounce = 1;
   7655			link_downgraded = false;
   7656		}
   7657	} else if ((lwde & ppd->link_width_downgrade_tx_active) == 0 ||
   7658		   (lwde & ppd->link_width_downgrade_rx_active) == 0) {
   7659		/* Tx or Rx is outside the enabled policy */
   7660		dd_dev_err(ppd->dd,
   7661			   "Link is outside of downgrade allowed, downing link\n");
   7662		dd_dev_err(ppd->dd,
   7663			   "  enabled 0x%x, tx active 0x%x, rx active 0x%x\n",
   7664			   lwde, ppd->link_width_downgrade_tx_active,
   7665			   ppd->link_width_downgrade_rx_active);
   7666		do_bounce = 1;
   7667		link_downgraded = false;
   7668	}
   7669
   7670done:
   7671	mutex_unlock(&ppd->hls_lock);
   7672
   7673	if (do_bounce) {
   7674		set_link_down_reason(ppd, OPA_LINKDOWN_REASON_WIDTH_POLICY, 0,
   7675				     OPA_LINKDOWN_REASON_WIDTH_POLICY);
   7676		set_link_state(ppd, HLS_DN_OFFLINE);
   7677		start_link(ppd);
   7678	}
   7679
   7680	return link_downgraded;
   7681}
   7682
   7683/*
   7684 * Handle a link downgrade interrupt from the 8051.
   7685 *
   7686 * This is a work-queue function outside of the interrupt.
   7687 */
   7688void handle_link_downgrade(struct work_struct *work)
   7689{
   7690	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   7691							link_downgrade_work);
   7692
   7693	dd_dev_info(ppd->dd, "8051: Link width downgrade\n");
   7694	if (apply_link_downgrade_policy(ppd, true))
   7695		update_xmit_counters(ppd, ppd->link_width_downgrade_tx_active);
   7696}
   7697
   7698static char *dcc_err_string(char *buf, int buf_len, u64 flags)
   7699{
   7700	return flag_string(buf, buf_len, flags, dcc_err_flags,
   7701		ARRAY_SIZE(dcc_err_flags));
   7702}
   7703
   7704static char *lcb_err_string(char *buf, int buf_len, u64 flags)
   7705{
   7706	return flag_string(buf, buf_len, flags, lcb_err_flags,
   7707		ARRAY_SIZE(lcb_err_flags));
   7708}
   7709
   7710static char *dc8051_err_string(char *buf, int buf_len, u64 flags)
   7711{
   7712	return flag_string(buf, buf_len, flags, dc8051_err_flags,
   7713		ARRAY_SIZE(dc8051_err_flags));
   7714}
   7715
   7716static char *dc8051_info_err_string(char *buf, int buf_len, u64 flags)
   7717{
   7718	return flag_string(buf, buf_len, flags, dc8051_info_err_flags,
   7719		ARRAY_SIZE(dc8051_info_err_flags));
   7720}
   7721
   7722static char *dc8051_info_host_msg_string(char *buf, int buf_len, u64 flags)
   7723{
   7724	return flag_string(buf, buf_len, flags, dc8051_info_host_msg_flags,
   7725		ARRAY_SIZE(dc8051_info_host_msg_flags));
   7726}
   7727
   7728static void handle_8051_interrupt(struct hfi1_devdata *dd, u32 unused, u64 reg)
   7729{
   7730	struct hfi1_pportdata *ppd = dd->pport;
   7731	u64 info, err, host_msg;
   7732	int queue_link_down = 0;
   7733	char buf[96];
   7734
   7735	/* look at the flags */
   7736	if (reg & DC_DC8051_ERR_FLG_SET_BY_8051_SMASK) {
   7737		/* 8051 information set by firmware */
   7738		/* read DC8051_DBG_ERR_INFO_SET_BY_8051 for details */
   7739		info = read_csr(dd, DC_DC8051_DBG_ERR_INFO_SET_BY_8051);
   7740		err = (info >> DC_DC8051_DBG_ERR_INFO_SET_BY_8051_ERROR_SHIFT)
   7741			& DC_DC8051_DBG_ERR_INFO_SET_BY_8051_ERROR_MASK;
   7742		host_msg = (info >>
   7743			DC_DC8051_DBG_ERR_INFO_SET_BY_8051_HOST_MSG_SHIFT)
   7744			& DC_DC8051_DBG_ERR_INFO_SET_BY_8051_HOST_MSG_MASK;
   7745
   7746		/*
   7747		 * Handle error flags.
   7748		 */
   7749		if (err & FAILED_LNI) {
   7750			/*
   7751			 * LNI error indications are cleared by the 8051
   7752			 * only when starting polling.  Only pay attention
   7753			 * to them when in the states that occur during
   7754			 * LNI.
   7755			 */
   7756			if (ppd->host_link_state
   7757			    & (HLS_DN_POLL | HLS_VERIFY_CAP | HLS_GOING_UP)) {
   7758				queue_link_down = 1;
   7759				dd_dev_info(dd, "Link error: %s\n",
   7760					    dc8051_info_err_string(buf,
   7761								   sizeof(buf),
   7762								   err &
   7763								   FAILED_LNI));
   7764			}
   7765			err &= ~(u64)FAILED_LNI;
   7766		}
   7767		/* unknown frames can happen durning LNI, just count */
   7768		if (err & UNKNOWN_FRAME) {
   7769			ppd->unknown_frame_count++;
   7770			err &= ~(u64)UNKNOWN_FRAME;
   7771		}
   7772		if (err) {
   7773			/* report remaining errors, but do not do anything */
   7774			dd_dev_err(dd, "8051 info error: %s\n",
   7775				   dc8051_info_err_string(buf, sizeof(buf),
   7776							  err));
   7777		}
   7778
   7779		/*
   7780		 * Handle host message flags.
   7781		 */
   7782		if (host_msg & HOST_REQ_DONE) {
   7783			/*
   7784			 * Presently, the driver does a busy wait for
   7785			 * host requests to complete.  This is only an
   7786			 * informational message.
   7787			 * NOTE: The 8051 clears the host message
   7788			 * information *on the next 8051 command*.
   7789			 * Therefore, when linkup is achieved,
   7790			 * this flag will still be set.
   7791			 */
   7792			host_msg &= ~(u64)HOST_REQ_DONE;
   7793		}
   7794		if (host_msg & BC_SMA_MSG) {
   7795			queue_work(ppd->link_wq, &ppd->sma_message_work);
   7796			host_msg &= ~(u64)BC_SMA_MSG;
   7797		}
   7798		if (host_msg & LINKUP_ACHIEVED) {
   7799			dd_dev_info(dd, "8051: Link up\n");
   7800			queue_work(ppd->link_wq, &ppd->link_up_work);
   7801			host_msg &= ~(u64)LINKUP_ACHIEVED;
   7802		}
   7803		if (host_msg & EXT_DEVICE_CFG_REQ) {
   7804			handle_8051_request(ppd);
   7805			host_msg &= ~(u64)EXT_DEVICE_CFG_REQ;
   7806		}
   7807		if (host_msg & VERIFY_CAP_FRAME) {
   7808			queue_work(ppd->link_wq, &ppd->link_vc_work);
   7809			host_msg &= ~(u64)VERIFY_CAP_FRAME;
   7810		}
   7811		if (host_msg & LINK_GOING_DOWN) {
   7812			const char *extra = "";
   7813			/* no downgrade action needed if going down */
   7814			if (host_msg & LINK_WIDTH_DOWNGRADED) {
   7815				host_msg &= ~(u64)LINK_WIDTH_DOWNGRADED;
   7816				extra = " (ignoring downgrade)";
   7817			}
   7818			dd_dev_info(dd, "8051: Link down%s\n", extra);
   7819			queue_link_down = 1;
   7820			host_msg &= ~(u64)LINK_GOING_DOWN;
   7821		}
   7822		if (host_msg & LINK_WIDTH_DOWNGRADED) {
   7823			queue_work(ppd->link_wq, &ppd->link_downgrade_work);
   7824			host_msg &= ~(u64)LINK_WIDTH_DOWNGRADED;
   7825		}
   7826		if (host_msg) {
   7827			/* report remaining messages, but do not do anything */
   7828			dd_dev_info(dd, "8051 info host message: %s\n",
   7829				    dc8051_info_host_msg_string(buf,
   7830								sizeof(buf),
   7831								host_msg));
   7832		}
   7833
   7834		reg &= ~DC_DC8051_ERR_FLG_SET_BY_8051_SMASK;
   7835	}
   7836	if (reg & DC_DC8051_ERR_FLG_LOST_8051_HEART_BEAT_SMASK) {
   7837		/*
   7838		 * Lost the 8051 heartbeat.  If this happens, we
   7839		 * receive constant interrupts about it.  Disable
   7840		 * the interrupt after the first.
   7841		 */
   7842		dd_dev_err(dd, "Lost 8051 heartbeat\n");
   7843		write_csr(dd, DC_DC8051_ERR_EN,
   7844			  read_csr(dd, DC_DC8051_ERR_EN) &
   7845			  ~DC_DC8051_ERR_EN_LOST_8051_HEART_BEAT_SMASK);
   7846
   7847		reg &= ~DC_DC8051_ERR_FLG_LOST_8051_HEART_BEAT_SMASK;
   7848	}
   7849	if (reg) {
   7850		/* report the error, but do not do anything */
   7851		dd_dev_err(dd, "8051 error: %s\n",
   7852			   dc8051_err_string(buf, sizeof(buf), reg));
   7853	}
   7854
   7855	if (queue_link_down) {
   7856		/*
   7857		 * if the link is already going down or disabled, do not
   7858		 * queue another. If there's a link down entry already
   7859		 * queued, don't queue another one.
   7860		 */
   7861		if ((ppd->host_link_state &
   7862		    (HLS_GOING_OFFLINE | HLS_LINK_COOLDOWN)) ||
   7863		    ppd->link_enabled == 0) {
   7864			dd_dev_info(dd, "%s: not queuing link down. host_link_state %x, link_enabled %x\n",
   7865				    __func__, ppd->host_link_state,
   7866				    ppd->link_enabled);
   7867		} else {
   7868			if (xchg(&ppd->is_link_down_queued, 1) == 1)
   7869				dd_dev_info(dd,
   7870					    "%s: link down request already queued\n",
   7871					    __func__);
   7872			else
   7873				queue_work(ppd->link_wq, &ppd->link_down_work);
   7874		}
   7875	}
   7876}
   7877
   7878static const char * const fm_config_txt[] = {
   7879[0] =
   7880	"BadHeadDist: Distance violation between two head flits",
   7881[1] =
   7882	"BadTailDist: Distance violation between two tail flits",
   7883[2] =
   7884	"BadCtrlDist: Distance violation between two credit control flits",
   7885[3] =
   7886	"BadCrdAck: Credits return for unsupported VL",
   7887[4] =
   7888	"UnsupportedVLMarker: Received VL Marker",
   7889[5] =
   7890	"BadPreempt: Exceeded the preemption nesting level",
   7891[6] =
   7892	"BadControlFlit: Received unsupported control flit",
   7893/* no 7 */
   7894[8] =
   7895	"UnsupportedVLMarker: Received VL Marker for unconfigured or disabled VL",
   7896};
   7897
   7898static const char * const port_rcv_txt[] = {
   7899[1] =
   7900	"BadPktLen: Illegal PktLen",
   7901[2] =
   7902	"PktLenTooLong: Packet longer than PktLen",
   7903[3] =
   7904	"PktLenTooShort: Packet shorter than PktLen",
   7905[4] =
   7906	"BadSLID: Illegal SLID (0, using multicast as SLID, does not include security validation of SLID)",
   7907[5] =
   7908	"BadDLID: Illegal DLID (0, doesn't match HFI)",
   7909[6] =
   7910	"BadL2: Illegal L2 opcode",
   7911[7] =
   7912	"BadSC: Unsupported SC",
   7913[9] =
   7914	"BadRC: Illegal RC",
   7915[11] =
   7916	"PreemptError: Preempting with same VL",
   7917[12] =
   7918	"PreemptVL15: Preempting a VL15 packet",
   7919};
   7920
   7921#define OPA_LDR_FMCONFIG_OFFSET 16
   7922#define OPA_LDR_PORTRCV_OFFSET 0
   7923static void handle_dcc_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   7924{
   7925	u64 info, hdr0, hdr1;
   7926	const char *extra;
   7927	char buf[96];
   7928	struct hfi1_pportdata *ppd = dd->pport;
   7929	u8 lcl_reason = 0;
   7930	int do_bounce = 0;
   7931
   7932	if (reg & DCC_ERR_FLG_UNCORRECTABLE_ERR_SMASK) {
   7933		if (!(dd->err_info_uncorrectable & OPA_EI_STATUS_SMASK)) {
   7934			info = read_csr(dd, DCC_ERR_INFO_UNCORRECTABLE);
   7935			dd->err_info_uncorrectable = info & OPA_EI_CODE_SMASK;
   7936			/* set status bit */
   7937			dd->err_info_uncorrectable |= OPA_EI_STATUS_SMASK;
   7938		}
   7939		reg &= ~DCC_ERR_FLG_UNCORRECTABLE_ERR_SMASK;
   7940	}
   7941
   7942	if (reg & DCC_ERR_FLG_LINK_ERR_SMASK) {
   7943		struct hfi1_pportdata *ppd = dd->pport;
   7944		/* this counter saturates at (2^32) - 1 */
   7945		if (ppd->link_downed < (u32)UINT_MAX)
   7946			ppd->link_downed++;
   7947		reg &= ~DCC_ERR_FLG_LINK_ERR_SMASK;
   7948	}
   7949
   7950	if (reg & DCC_ERR_FLG_FMCONFIG_ERR_SMASK) {
   7951		u8 reason_valid = 1;
   7952
   7953		info = read_csr(dd, DCC_ERR_INFO_FMCONFIG);
   7954		if (!(dd->err_info_fmconfig & OPA_EI_STATUS_SMASK)) {
   7955			dd->err_info_fmconfig = info & OPA_EI_CODE_SMASK;
   7956			/* set status bit */
   7957			dd->err_info_fmconfig |= OPA_EI_STATUS_SMASK;
   7958		}
   7959		switch (info) {
   7960		case 0:
   7961		case 1:
   7962		case 2:
   7963		case 3:
   7964		case 4:
   7965		case 5:
   7966		case 6:
   7967			extra = fm_config_txt[info];
   7968			break;
   7969		case 8:
   7970			extra = fm_config_txt[info];
   7971			if (ppd->port_error_action &
   7972			    OPA_PI_MASK_FM_CFG_UNSUPPORTED_VL_MARKER) {
   7973				do_bounce = 1;
   7974				/*
   7975				 * lcl_reason cannot be derived from info
   7976				 * for this error
   7977				 */
   7978				lcl_reason =
   7979				  OPA_LINKDOWN_REASON_UNSUPPORTED_VL_MARKER;
   7980			}
   7981			break;
   7982		default:
   7983			reason_valid = 0;
   7984			snprintf(buf, sizeof(buf), "reserved%lld", info);
   7985			extra = buf;
   7986			break;
   7987		}
   7988
   7989		if (reason_valid && !do_bounce) {
   7990			do_bounce = ppd->port_error_action &
   7991					(1 << (OPA_LDR_FMCONFIG_OFFSET + info));
   7992			lcl_reason = info + OPA_LINKDOWN_REASON_BAD_HEAD_DIST;
   7993		}
   7994
   7995		/* just report this */
   7996		dd_dev_info_ratelimited(dd, "DCC Error: fmconfig error: %s\n",
   7997					extra);
   7998		reg &= ~DCC_ERR_FLG_FMCONFIG_ERR_SMASK;
   7999	}
   8000
   8001	if (reg & DCC_ERR_FLG_RCVPORT_ERR_SMASK) {
   8002		u8 reason_valid = 1;
   8003
   8004		info = read_csr(dd, DCC_ERR_INFO_PORTRCV);
   8005		hdr0 = read_csr(dd, DCC_ERR_INFO_PORTRCV_HDR0);
   8006		hdr1 = read_csr(dd, DCC_ERR_INFO_PORTRCV_HDR1);
   8007		if (!(dd->err_info_rcvport.status_and_code &
   8008		      OPA_EI_STATUS_SMASK)) {
   8009			dd->err_info_rcvport.status_and_code =
   8010				info & OPA_EI_CODE_SMASK;
   8011			/* set status bit */
   8012			dd->err_info_rcvport.status_and_code |=
   8013				OPA_EI_STATUS_SMASK;
   8014			/*
   8015			 * save first 2 flits in the packet that caused
   8016			 * the error
   8017			 */
   8018			dd->err_info_rcvport.packet_flit1 = hdr0;
   8019			dd->err_info_rcvport.packet_flit2 = hdr1;
   8020		}
   8021		switch (info) {
   8022		case 1:
   8023		case 2:
   8024		case 3:
   8025		case 4:
   8026		case 5:
   8027		case 6:
   8028		case 7:
   8029		case 9:
   8030		case 11:
   8031		case 12:
   8032			extra = port_rcv_txt[info];
   8033			break;
   8034		default:
   8035			reason_valid = 0;
   8036			snprintf(buf, sizeof(buf), "reserved%lld", info);
   8037			extra = buf;
   8038			break;
   8039		}
   8040
   8041		if (reason_valid && !do_bounce) {
   8042			do_bounce = ppd->port_error_action &
   8043					(1 << (OPA_LDR_PORTRCV_OFFSET + info));
   8044			lcl_reason = info + OPA_LINKDOWN_REASON_RCV_ERROR_0;
   8045		}
   8046
   8047		/* just report this */
   8048		dd_dev_info_ratelimited(dd, "DCC Error: PortRcv error: %s\n"
   8049					"               hdr0 0x%llx, hdr1 0x%llx\n",
   8050					extra, hdr0, hdr1);
   8051
   8052		reg &= ~DCC_ERR_FLG_RCVPORT_ERR_SMASK;
   8053	}
   8054
   8055	if (reg & DCC_ERR_FLG_EN_CSR_ACCESS_BLOCKED_UC_SMASK) {
   8056		/* informative only */
   8057		dd_dev_info_ratelimited(dd, "8051 access to LCB blocked\n");
   8058		reg &= ~DCC_ERR_FLG_EN_CSR_ACCESS_BLOCKED_UC_SMASK;
   8059	}
   8060	if (reg & DCC_ERR_FLG_EN_CSR_ACCESS_BLOCKED_HOST_SMASK) {
   8061		/* informative only */
   8062		dd_dev_info_ratelimited(dd, "host access to LCB blocked\n");
   8063		reg &= ~DCC_ERR_FLG_EN_CSR_ACCESS_BLOCKED_HOST_SMASK;
   8064	}
   8065
   8066	if (unlikely(hfi1_dbg_fault_suppress_err(&dd->verbs_dev)))
   8067		reg &= ~DCC_ERR_FLG_LATE_EBP_ERR_SMASK;
   8068
   8069	/* report any remaining errors */
   8070	if (reg)
   8071		dd_dev_info_ratelimited(dd, "DCC Error: %s\n",
   8072					dcc_err_string(buf, sizeof(buf), reg));
   8073
   8074	if (lcl_reason == 0)
   8075		lcl_reason = OPA_LINKDOWN_REASON_UNKNOWN;
   8076
   8077	if (do_bounce) {
   8078		dd_dev_info_ratelimited(dd, "%s: PortErrorAction bounce\n",
   8079					__func__);
   8080		set_link_down_reason(ppd, lcl_reason, 0, lcl_reason);
   8081		queue_work(ppd->link_wq, &ppd->link_bounce_work);
   8082	}
   8083}
   8084
   8085static void handle_lcb_err(struct hfi1_devdata *dd, u32 unused, u64 reg)
   8086{
   8087	char buf[96];
   8088
   8089	dd_dev_info(dd, "LCB Error: %s\n",
   8090		    lcb_err_string(buf, sizeof(buf), reg));
   8091}
   8092
   8093/*
   8094 * CCE block DC interrupt.  Source is < 8.
   8095 */
   8096static void is_dc_int(struct hfi1_devdata *dd, unsigned int source)
   8097{
   8098	const struct err_reg_info *eri = &dc_errs[source];
   8099
   8100	if (eri->handler) {
   8101		interrupt_clear_down(dd, 0, eri);
   8102	} else if (source == 3 /* dc_lbm_int */) {
   8103		/*
   8104		 * This indicates that a parity error has occurred on the
   8105		 * address/control lines presented to the LBM.  The error
   8106		 * is a single pulse, there is no associated error flag,
   8107		 * and it is non-maskable.  This is because if a parity
   8108		 * error occurs on the request the request is dropped.
   8109		 * This should never occur, but it is nice to know if it
   8110		 * ever does.
   8111		 */
   8112		dd_dev_err(dd, "Parity error in DC LBM block\n");
   8113	} else {
   8114		dd_dev_err(dd, "Invalid DC interrupt %u\n", source);
   8115	}
   8116}
   8117
   8118/*
   8119 * TX block send credit interrupt.  Source is < 160.
   8120 */
   8121static void is_send_credit_int(struct hfi1_devdata *dd, unsigned int source)
   8122{
   8123	sc_group_release_update(dd, source);
   8124}
   8125
   8126/*
   8127 * TX block SDMA interrupt.  Source is < 48.
   8128 *
   8129 * SDMA interrupts are grouped by type:
   8130 *
   8131 *	 0 -  N-1 = SDma
   8132 *	 N - 2N-1 = SDmaProgress
   8133 *	2N - 3N-1 = SDmaIdle
   8134 */
   8135static void is_sdma_eng_int(struct hfi1_devdata *dd, unsigned int source)
   8136{
   8137	/* what interrupt */
   8138	unsigned int what  = source / TXE_NUM_SDMA_ENGINES;
   8139	/* which engine */
   8140	unsigned int which = source % TXE_NUM_SDMA_ENGINES;
   8141
   8142#ifdef CONFIG_SDMA_VERBOSITY
   8143	dd_dev_err(dd, "CONFIG SDMA(%u) %s:%d %s()\n", which,
   8144		   slashstrip(__FILE__), __LINE__, __func__);
   8145	sdma_dumpstate(&dd->per_sdma[which]);
   8146#endif
   8147
   8148	if (likely(what < 3 && which < dd->num_sdma)) {
   8149		sdma_engine_interrupt(&dd->per_sdma[which], 1ull << source);
   8150	} else {
   8151		/* should not happen */
   8152		dd_dev_err(dd, "Invalid SDMA interrupt 0x%x\n", source);
   8153	}
   8154}
   8155
   8156/**
   8157 * is_rcv_avail_int() - User receive context available IRQ handler
   8158 * @dd: valid dd
   8159 * @source: logical IRQ source (offset from IS_RCVAVAIL_START)
   8160 *
   8161 * RX block receive available interrupt.  Source is < 160.
   8162 *
   8163 * This is the general interrupt handler for user (PSM) receive contexts,
   8164 * and can only be used for non-threaded IRQs.
   8165 */
   8166static void is_rcv_avail_int(struct hfi1_devdata *dd, unsigned int source)
   8167{
   8168	struct hfi1_ctxtdata *rcd;
   8169	char *err_detail;
   8170
   8171	if (likely(source < dd->num_rcv_contexts)) {
   8172		rcd = hfi1_rcd_get_by_index(dd, source);
   8173		if (rcd) {
   8174			handle_user_interrupt(rcd);
   8175			hfi1_rcd_put(rcd);
   8176			return;	/* OK */
   8177		}
   8178		/* received an interrupt, but no rcd */
   8179		err_detail = "dataless";
   8180	} else {
   8181		/* received an interrupt, but are not using that context */
   8182		err_detail = "out of range";
   8183	}
   8184	dd_dev_err(dd, "unexpected %s receive available context interrupt %u\n",
   8185		   err_detail, source);
   8186}
   8187
   8188/**
   8189 * is_rcv_urgent_int() - User receive context urgent IRQ handler
   8190 * @dd: valid dd
   8191 * @source: logical IRQ source (offset from IS_RCVURGENT_START)
   8192 *
   8193 * RX block receive urgent interrupt.  Source is < 160.
   8194 *
   8195 * NOTE: kernel receive contexts specifically do NOT enable this IRQ.
   8196 */
   8197static void is_rcv_urgent_int(struct hfi1_devdata *dd, unsigned int source)
   8198{
   8199	struct hfi1_ctxtdata *rcd;
   8200	char *err_detail;
   8201
   8202	if (likely(source < dd->num_rcv_contexts)) {
   8203		rcd = hfi1_rcd_get_by_index(dd, source);
   8204		if (rcd) {
   8205			handle_user_interrupt(rcd);
   8206			hfi1_rcd_put(rcd);
   8207			return;	/* OK */
   8208		}
   8209		/* received an interrupt, but no rcd */
   8210		err_detail = "dataless";
   8211	} else {
   8212		/* received an interrupt, but are not using that context */
   8213		err_detail = "out of range";
   8214	}
   8215	dd_dev_err(dd, "unexpected %s receive urgent context interrupt %u\n",
   8216		   err_detail, source);
   8217}
   8218
   8219/*
   8220 * Reserved range interrupt.  Should not be called in normal operation.
   8221 */
   8222static void is_reserved_int(struct hfi1_devdata *dd, unsigned int source)
   8223{
   8224	char name[64];
   8225
   8226	dd_dev_err(dd, "unexpected %s interrupt\n",
   8227		   is_reserved_name(name, sizeof(name), source));
   8228}
   8229
   8230static const struct is_table is_table[] = {
   8231/*
   8232 * start		 end
   8233 *				name func		interrupt func
   8234 */
   8235{ IS_GENERAL_ERR_START,  IS_GENERAL_ERR_END,
   8236				is_misc_err_name,	is_misc_err_int },
   8237{ IS_SDMAENG_ERR_START,  IS_SDMAENG_ERR_END,
   8238				is_sdma_eng_err_name,	is_sdma_eng_err_int },
   8239{ IS_SENDCTXT_ERR_START, IS_SENDCTXT_ERR_END,
   8240				is_sendctxt_err_name,	is_sendctxt_err_int },
   8241{ IS_SDMA_START,	     IS_SDMA_IDLE_END,
   8242				is_sdma_eng_name,	is_sdma_eng_int },
   8243{ IS_VARIOUS_START,	     IS_VARIOUS_END,
   8244				is_various_name,	is_various_int },
   8245{ IS_DC_START,	     IS_DC_END,
   8246				is_dc_name,		is_dc_int },
   8247{ IS_RCVAVAIL_START,     IS_RCVAVAIL_END,
   8248				is_rcv_avail_name,	is_rcv_avail_int },
   8249{ IS_RCVURGENT_START,    IS_RCVURGENT_END,
   8250				is_rcv_urgent_name,	is_rcv_urgent_int },
   8251{ IS_SENDCREDIT_START,   IS_SENDCREDIT_END,
   8252				is_send_credit_name,	is_send_credit_int},
   8253{ IS_RESERVED_START,     IS_RESERVED_END,
   8254				is_reserved_name,	is_reserved_int},
   8255};
   8256
   8257/*
   8258 * Interrupt source interrupt - called when the given source has an interrupt.
   8259 * Source is a bit index into an array of 64-bit integers.
   8260 */
   8261static void is_interrupt(struct hfi1_devdata *dd, unsigned int source)
   8262{
   8263	const struct is_table *entry;
   8264
   8265	/* avoids a double compare by walking the table in-order */
   8266	for (entry = &is_table[0]; entry->is_name; entry++) {
   8267		if (source <= entry->end) {
   8268			trace_hfi1_interrupt(dd, entry, source);
   8269			entry->is_int(dd, source - entry->start);
   8270			return;
   8271		}
   8272	}
   8273	/* fell off the end */
   8274	dd_dev_err(dd, "invalid interrupt source %u\n", source);
   8275}
   8276
   8277/**
   8278 * general_interrupt -  General interrupt handler
   8279 * @irq: MSIx IRQ vector
   8280 * @data: hfi1 devdata
   8281 *
   8282 * This is able to correctly handle all non-threaded interrupts.  Receive
   8283 * context DATA IRQs are threaded and are not supported by this handler.
   8284 *
   8285 */
   8286irqreturn_t general_interrupt(int irq, void *data)
   8287{
   8288	struct hfi1_devdata *dd = data;
   8289	u64 regs[CCE_NUM_INT_CSRS];
   8290	u32 bit;
   8291	int i;
   8292	irqreturn_t handled = IRQ_NONE;
   8293
   8294	this_cpu_inc(*dd->int_counter);
   8295
   8296	/* phase 1: scan and clear all handled interrupts */
   8297	for (i = 0; i < CCE_NUM_INT_CSRS; i++) {
   8298		if (dd->gi_mask[i] == 0) {
   8299			regs[i] = 0;	/* used later */
   8300			continue;
   8301		}
   8302		regs[i] = read_csr(dd, CCE_INT_STATUS + (8 * i)) &
   8303				dd->gi_mask[i];
   8304		/* only clear if anything is set */
   8305		if (regs[i])
   8306			write_csr(dd, CCE_INT_CLEAR + (8 * i), regs[i]);
   8307	}
   8308
   8309	/* phase 2: call the appropriate handler */
   8310	for_each_set_bit(bit, (unsigned long *)&regs[0],
   8311			 CCE_NUM_INT_CSRS * 64) {
   8312		is_interrupt(dd, bit);
   8313		handled = IRQ_HANDLED;
   8314	}
   8315
   8316	return handled;
   8317}
   8318
   8319irqreturn_t sdma_interrupt(int irq, void *data)
   8320{
   8321	struct sdma_engine *sde = data;
   8322	struct hfi1_devdata *dd = sde->dd;
   8323	u64 status;
   8324
   8325#ifdef CONFIG_SDMA_VERBOSITY
   8326	dd_dev_err(dd, "CONFIG SDMA(%u) %s:%d %s()\n", sde->this_idx,
   8327		   slashstrip(__FILE__), __LINE__, __func__);
   8328	sdma_dumpstate(sde);
   8329#endif
   8330
   8331	this_cpu_inc(*dd->int_counter);
   8332
   8333	/* This read_csr is really bad in the hot path */
   8334	status = read_csr(dd,
   8335			  CCE_INT_STATUS + (8 * (IS_SDMA_START / 64)))
   8336			  & sde->imask;
   8337	if (likely(status)) {
   8338		/* clear the interrupt(s) */
   8339		write_csr(dd,
   8340			  CCE_INT_CLEAR + (8 * (IS_SDMA_START / 64)),
   8341			  status);
   8342
   8343		/* handle the interrupt(s) */
   8344		sdma_engine_interrupt(sde, status);
   8345	} else {
   8346		dd_dev_info_ratelimited(dd, "SDMA engine %u interrupt, but no status bits set\n",
   8347					sde->this_idx);
   8348	}
   8349	return IRQ_HANDLED;
   8350}
   8351
   8352/*
   8353 * Clear the receive interrupt.  Use a read of the interrupt clear CSR
   8354 * to insure that the write completed.  This does NOT guarantee that
   8355 * queued DMA writes to memory from the chip are pushed.
   8356 */
   8357static inline void clear_recv_intr(struct hfi1_ctxtdata *rcd)
   8358{
   8359	struct hfi1_devdata *dd = rcd->dd;
   8360	u32 addr = CCE_INT_CLEAR + (8 * rcd->ireg);
   8361
   8362	write_csr(dd, addr, rcd->imask);
   8363	/* force the above write on the chip and get a value back */
   8364	(void)read_csr(dd, addr);
   8365}
   8366
   8367/* force the receive interrupt */
   8368void force_recv_intr(struct hfi1_ctxtdata *rcd)
   8369{
   8370	write_csr(rcd->dd, CCE_INT_FORCE + (8 * rcd->ireg), rcd->imask);
   8371}
   8372
   8373/*
   8374 * Return non-zero if a packet is present.
   8375 *
   8376 * This routine is called when rechecking for packets after the RcvAvail
   8377 * interrupt has been cleared down.  First, do a quick check of memory for
   8378 * a packet present.  If not found, use an expensive CSR read of the context
   8379 * tail to determine the actual tail.  The CSR read is necessary because there
   8380 * is no method to push pending DMAs to memory other than an interrupt and we
   8381 * are trying to determine if we need to force an interrupt.
   8382 */
   8383static inline int check_packet_present(struct hfi1_ctxtdata *rcd)
   8384{
   8385	u32 tail;
   8386
   8387	if (hfi1_packet_present(rcd))
   8388		return 1;
   8389
   8390	/* fall back to a CSR read, correct indpendent of DMA_RTAIL */
   8391	tail = (u32)read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_TAIL);
   8392	return hfi1_rcd_head(rcd) != tail;
   8393}
   8394
   8395/*
   8396 * Common code for receive contexts interrupt handlers.
   8397 * Update traces, increment kernel IRQ counter and
   8398 * setup ASPM when needed.
   8399 */
   8400static void receive_interrupt_common(struct hfi1_ctxtdata *rcd)
   8401{
   8402	struct hfi1_devdata *dd = rcd->dd;
   8403
   8404	trace_hfi1_receive_interrupt(dd, rcd);
   8405	this_cpu_inc(*dd->int_counter);
   8406	aspm_ctx_disable(rcd);
   8407}
   8408
   8409/*
   8410 * __hfi1_rcd_eoi_intr() - Make HW issue receive interrupt
   8411 * when there are packets present in the queue. When calling
   8412 * with interrupts enabled please use hfi1_rcd_eoi_intr.
   8413 *
   8414 * @rcd: valid receive context
   8415 */
   8416static void __hfi1_rcd_eoi_intr(struct hfi1_ctxtdata *rcd)
   8417{
   8418	if (!rcd->rcvhdrq)
   8419		return;
   8420	clear_recv_intr(rcd);
   8421	if (check_packet_present(rcd))
   8422		force_recv_intr(rcd);
   8423}
   8424
   8425/**
   8426 * hfi1_rcd_eoi_intr() - End of Interrupt processing action
   8427 *
   8428 * @rcd: Ptr to hfi1_ctxtdata of receive context
   8429 *
   8430 *  Hold IRQs so we can safely clear the interrupt and
   8431 *  recheck for a packet that may have arrived after the previous
   8432 *  check and the interrupt clear.  If a packet arrived, force another
   8433 *  interrupt. This routine can be called at the end of receive packet
   8434 *  processing in interrupt service routines, interrupt service thread
   8435 *  and softirqs
   8436 */
   8437static void hfi1_rcd_eoi_intr(struct hfi1_ctxtdata *rcd)
   8438{
   8439	unsigned long flags;
   8440
   8441	local_irq_save(flags);
   8442	__hfi1_rcd_eoi_intr(rcd);
   8443	local_irq_restore(flags);
   8444}
   8445
   8446/**
   8447 * hfi1_netdev_rx_napi - napi poll function to move eoi inline
   8448 * @napi: pointer to napi object
   8449 * @budget: netdev budget
   8450 */
   8451int hfi1_netdev_rx_napi(struct napi_struct *napi, int budget)
   8452{
   8453	struct hfi1_netdev_rxq *rxq = container_of(napi,
   8454			struct hfi1_netdev_rxq, napi);
   8455	struct hfi1_ctxtdata *rcd = rxq->rcd;
   8456	int work_done = 0;
   8457
   8458	work_done = rcd->do_interrupt(rcd, budget);
   8459
   8460	if (work_done < budget) {
   8461		napi_complete_done(napi, work_done);
   8462		hfi1_rcd_eoi_intr(rcd);
   8463	}
   8464
   8465	return work_done;
   8466}
   8467
   8468/* Receive packet napi handler for netdevs VNIC and AIP  */
   8469irqreturn_t receive_context_interrupt_napi(int irq, void *data)
   8470{
   8471	struct hfi1_ctxtdata *rcd = data;
   8472
   8473	receive_interrupt_common(rcd);
   8474
   8475	if (likely(rcd->napi)) {
   8476		if (likely(napi_schedule_prep(rcd->napi)))
   8477			__napi_schedule_irqoff(rcd->napi);
   8478		else
   8479			__hfi1_rcd_eoi_intr(rcd);
   8480	} else {
   8481		WARN_ONCE(1, "Napi IRQ handler without napi set up ctxt=%d\n",
   8482			  rcd->ctxt);
   8483		__hfi1_rcd_eoi_intr(rcd);
   8484	}
   8485
   8486	return IRQ_HANDLED;
   8487}
   8488
   8489/*
   8490 * Receive packet IRQ handler.  This routine expects to be on its own IRQ.
   8491 * This routine will try to handle packets immediately (latency), but if
   8492 * it finds too many, it will invoke the thread handler (bandwitdh).  The
   8493 * chip receive interrupt is *not* cleared down until this or the thread (if
   8494 * invoked) is finished.  The intent is to avoid extra interrupts while we
   8495 * are processing packets anyway.
   8496 */
   8497irqreturn_t receive_context_interrupt(int irq, void *data)
   8498{
   8499	struct hfi1_ctxtdata *rcd = data;
   8500	int disposition;
   8501
   8502	receive_interrupt_common(rcd);
   8503
   8504	/* receive interrupt remains blocked while processing packets */
   8505	disposition = rcd->do_interrupt(rcd, 0);
   8506
   8507	/*
   8508	 * Too many packets were seen while processing packets in this
   8509	 * IRQ handler.  Invoke the handler thread.  The receive interrupt
   8510	 * remains blocked.
   8511	 */
   8512	if (disposition == RCV_PKT_LIMIT)
   8513		return IRQ_WAKE_THREAD;
   8514
   8515	__hfi1_rcd_eoi_intr(rcd);
   8516	return IRQ_HANDLED;
   8517}
   8518
   8519/*
   8520 * Receive packet thread handler.  This expects to be invoked with the
   8521 * receive interrupt still blocked.
   8522 */
   8523irqreturn_t receive_context_thread(int irq, void *data)
   8524{
   8525	struct hfi1_ctxtdata *rcd = data;
   8526
   8527	/* receive interrupt is still blocked from the IRQ handler */
   8528	(void)rcd->do_interrupt(rcd, 1);
   8529
   8530	hfi1_rcd_eoi_intr(rcd);
   8531
   8532	return IRQ_HANDLED;
   8533}
   8534
   8535/* ========================================================================= */
   8536
   8537u32 read_physical_state(struct hfi1_devdata *dd)
   8538{
   8539	u64 reg;
   8540
   8541	reg = read_csr(dd, DC_DC8051_STS_CUR_STATE);
   8542	return (reg >> DC_DC8051_STS_CUR_STATE_PORT_SHIFT)
   8543				& DC_DC8051_STS_CUR_STATE_PORT_MASK;
   8544}
   8545
   8546u32 read_logical_state(struct hfi1_devdata *dd)
   8547{
   8548	u64 reg;
   8549
   8550	reg = read_csr(dd, DCC_CFG_PORT_CONFIG);
   8551	return (reg >> DCC_CFG_PORT_CONFIG_LINK_STATE_SHIFT)
   8552				& DCC_CFG_PORT_CONFIG_LINK_STATE_MASK;
   8553}
   8554
   8555static void set_logical_state(struct hfi1_devdata *dd, u32 chip_lstate)
   8556{
   8557	u64 reg;
   8558
   8559	reg = read_csr(dd, DCC_CFG_PORT_CONFIG);
   8560	/* clear current state, set new state */
   8561	reg &= ~DCC_CFG_PORT_CONFIG_LINK_STATE_SMASK;
   8562	reg |= (u64)chip_lstate << DCC_CFG_PORT_CONFIG_LINK_STATE_SHIFT;
   8563	write_csr(dd, DCC_CFG_PORT_CONFIG, reg);
   8564}
   8565
   8566/*
   8567 * Use the 8051 to read a LCB CSR.
   8568 */
   8569static int read_lcb_via_8051(struct hfi1_devdata *dd, u32 addr, u64 *data)
   8570{
   8571	u32 regno;
   8572	int ret;
   8573
   8574	if (dd->icode == ICODE_FUNCTIONAL_SIMULATOR) {
   8575		if (acquire_lcb_access(dd, 0) == 0) {
   8576			*data = read_csr(dd, addr);
   8577			release_lcb_access(dd, 0);
   8578			return 0;
   8579		}
   8580		return -EBUSY;
   8581	}
   8582
   8583	/* register is an index of LCB registers: (offset - base) / 8 */
   8584	regno = (addr - DC_LCB_CFG_RUN) >> 3;
   8585	ret = do_8051_command(dd, HCMD_READ_LCB_CSR, regno, data);
   8586	if (ret != HCMD_SUCCESS)
   8587		return -EBUSY;
   8588	return 0;
   8589}
   8590
   8591/*
   8592 * Provide a cache for some of the LCB registers in case the LCB is
   8593 * unavailable.
   8594 * (The LCB is unavailable in certain link states, for example.)
   8595 */
   8596struct lcb_datum {
   8597	u32 off;
   8598	u64 val;
   8599};
   8600
   8601static struct lcb_datum lcb_cache[] = {
   8602	{ DC_LCB_ERR_INFO_RX_REPLAY_CNT, 0},
   8603	{ DC_LCB_ERR_INFO_SEQ_CRC_CNT, 0 },
   8604	{ DC_LCB_ERR_INFO_REINIT_FROM_PEER_CNT, 0 },
   8605};
   8606
   8607static void update_lcb_cache(struct hfi1_devdata *dd)
   8608{
   8609	int i;
   8610	int ret;
   8611	u64 val;
   8612
   8613	for (i = 0; i < ARRAY_SIZE(lcb_cache); i++) {
   8614		ret = read_lcb_csr(dd, lcb_cache[i].off, &val);
   8615
   8616		/* Update if we get good data */
   8617		if (likely(ret != -EBUSY))
   8618			lcb_cache[i].val = val;
   8619	}
   8620}
   8621
   8622static int read_lcb_cache(u32 off, u64 *val)
   8623{
   8624	int i;
   8625
   8626	for (i = 0; i < ARRAY_SIZE(lcb_cache); i++) {
   8627		if (lcb_cache[i].off == off) {
   8628			*val = lcb_cache[i].val;
   8629			return 0;
   8630		}
   8631	}
   8632
   8633	pr_warn("%s bad offset 0x%x\n", __func__, off);
   8634	return -1;
   8635}
   8636
   8637/*
   8638 * Read an LCB CSR.  Access may not be in host control, so check.
   8639 * Return 0 on success, -EBUSY on failure.
   8640 */
   8641int read_lcb_csr(struct hfi1_devdata *dd, u32 addr, u64 *data)
   8642{
   8643	struct hfi1_pportdata *ppd = dd->pport;
   8644
   8645	/* if up, go through the 8051 for the value */
   8646	if (ppd->host_link_state & HLS_UP)
   8647		return read_lcb_via_8051(dd, addr, data);
   8648	/* if going up or down, check the cache, otherwise, no access */
   8649	if (ppd->host_link_state & (HLS_GOING_UP | HLS_GOING_OFFLINE)) {
   8650		if (read_lcb_cache(addr, data))
   8651			return -EBUSY;
   8652		return 0;
   8653	}
   8654
   8655	/* otherwise, host has access */
   8656	*data = read_csr(dd, addr);
   8657	return 0;
   8658}
   8659
   8660/*
   8661 * Use the 8051 to write a LCB CSR.
   8662 */
   8663static int write_lcb_via_8051(struct hfi1_devdata *dd, u32 addr, u64 data)
   8664{
   8665	u32 regno;
   8666	int ret;
   8667
   8668	if (dd->icode == ICODE_FUNCTIONAL_SIMULATOR ||
   8669	    (dd->dc8051_ver < dc8051_ver(0, 20, 0))) {
   8670		if (acquire_lcb_access(dd, 0) == 0) {
   8671			write_csr(dd, addr, data);
   8672			release_lcb_access(dd, 0);
   8673			return 0;
   8674		}
   8675		return -EBUSY;
   8676	}
   8677
   8678	/* register is an index of LCB registers: (offset - base) / 8 */
   8679	regno = (addr - DC_LCB_CFG_RUN) >> 3;
   8680	ret = do_8051_command(dd, HCMD_WRITE_LCB_CSR, regno, &data);
   8681	if (ret != HCMD_SUCCESS)
   8682		return -EBUSY;
   8683	return 0;
   8684}
   8685
   8686/*
   8687 * Write an LCB CSR.  Access may not be in host control, so check.
   8688 * Return 0 on success, -EBUSY on failure.
   8689 */
   8690int write_lcb_csr(struct hfi1_devdata *dd, u32 addr, u64 data)
   8691{
   8692	struct hfi1_pportdata *ppd = dd->pport;
   8693
   8694	/* if up, go through the 8051 for the value */
   8695	if (ppd->host_link_state & HLS_UP)
   8696		return write_lcb_via_8051(dd, addr, data);
   8697	/* if going up or down, no access */
   8698	if (ppd->host_link_state & (HLS_GOING_UP | HLS_GOING_OFFLINE))
   8699		return -EBUSY;
   8700	/* otherwise, host has access */
   8701	write_csr(dd, addr, data);
   8702	return 0;
   8703}
   8704
   8705/*
   8706 * Returns:
   8707 *	< 0 = Linux error, not able to get access
   8708 *	> 0 = 8051 command RETURN_CODE
   8709 */
   8710static int do_8051_command(struct hfi1_devdata *dd, u32 type, u64 in_data,
   8711			   u64 *out_data)
   8712{
   8713	u64 reg, completed;
   8714	int return_code;
   8715	unsigned long timeout;
   8716
   8717	hfi1_cdbg(DC8051, "type %d, data 0x%012llx", type, in_data);
   8718
   8719	mutex_lock(&dd->dc8051_lock);
   8720
   8721	/* We can't send any commands to the 8051 if it's in reset */
   8722	if (dd->dc_shutdown) {
   8723		return_code = -ENODEV;
   8724		goto fail;
   8725	}
   8726
   8727	/*
   8728	 * If an 8051 host command timed out previously, then the 8051 is
   8729	 * stuck.
   8730	 *
   8731	 * On first timeout, attempt to reset and restart the entire DC
   8732	 * block (including 8051). (Is this too big of a hammer?)
   8733	 *
   8734	 * If the 8051 times out a second time, the reset did not bring it
   8735	 * back to healthy life. In that case, fail any subsequent commands.
   8736	 */
   8737	if (dd->dc8051_timed_out) {
   8738		if (dd->dc8051_timed_out > 1) {
   8739			dd_dev_err(dd,
   8740				   "Previous 8051 host command timed out, skipping command %u\n",
   8741				   type);
   8742			return_code = -ENXIO;
   8743			goto fail;
   8744		}
   8745		_dc_shutdown(dd);
   8746		_dc_start(dd);
   8747	}
   8748
   8749	/*
   8750	 * If there is no timeout, then the 8051 command interface is
   8751	 * waiting for a command.
   8752	 */
   8753
   8754	/*
   8755	 * When writing a LCB CSR, out_data contains the full value to
   8756	 * to be written, while in_data contains the relative LCB
   8757	 * address in 7:0.  Do the work here, rather than the caller,
   8758	 * of distrubting the write data to where it needs to go:
   8759	 *
   8760	 * Write data
   8761	 *   39:00 -> in_data[47:8]
   8762	 *   47:40 -> DC8051_CFG_EXT_DEV_0.RETURN_CODE
   8763	 *   63:48 -> DC8051_CFG_EXT_DEV_0.RSP_DATA
   8764	 */
   8765	if (type == HCMD_WRITE_LCB_CSR) {
   8766		in_data |= ((*out_data) & 0xffffffffffull) << 8;
   8767		/* must preserve COMPLETED - it is tied to hardware */
   8768		reg = read_csr(dd, DC_DC8051_CFG_EXT_DEV_0);
   8769		reg &= DC_DC8051_CFG_EXT_DEV_0_COMPLETED_SMASK;
   8770		reg |= ((((*out_data) >> 40) & 0xff) <<
   8771				DC_DC8051_CFG_EXT_DEV_0_RETURN_CODE_SHIFT)
   8772		      | ((((*out_data) >> 48) & 0xffff) <<
   8773				DC_DC8051_CFG_EXT_DEV_0_RSP_DATA_SHIFT);
   8774		write_csr(dd, DC_DC8051_CFG_EXT_DEV_0, reg);
   8775	}
   8776
   8777	/*
   8778	 * Do two writes: the first to stabilize the type and req_data, the
   8779	 * second to activate.
   8780	 */
   8781	reg = ((u64)type & DC_DC8051_CFG_HOST_CMD_0_REQ_TYPE_MASK)
   8782			<< DC_DC8051_CFG_HOST_CMD_0_REQ_TYPE_SHIFT
   8783		| (in_data & DC_DC8051_CFG_HOST_CMD_0_REQ_DATA_MASK)
   8784			<< DC_DC8051_CFG_HOST_CMD_0_REQ_DATA_SHIFT;
   8785	write_csr(dd, DC_DC8051_CFG_HOST_CMD_0, reg);
   8786	reg |= DC_DC8051_CFG_HOST_CMD_0_REQ_NEW_SMASK;
   8787	write_csr(dd, DC_DC8051_CFG_HOST_CMD_0, reg);
   8788
   8789	/* wait for completion, alternate: interrupt */
   8790	timeout = jiffies + msecs_to_jiffies(DC8051_COMMAND_TIMEOUT);
   8791	while (1) {
   8792		reg = read_csr(dd, DC_DC8051_CFG_HOST_CMD_1);
   8793		completed = reg & DC_DC8051_CFG_HOST_CMD_1_COMPLETED_SMASK;
   8794		if (completed)
   8795			break;
   8796		if (time_after(jiffies, timeout)) {
   8797			dd->dc8051_timed_out++;
   8798			dd_dev_err(dd, "8051 host command %u timeout\n", type);
   8799			if (out_data)
   8800				*out_data = 0;
   8801			return_code = -ETIMEDOUT;
   8802			goto fail;
   8803		}
   8804		udelay(2);
   8805	}
   8806
   8807	if (out_data) {
   8808		*out_data = (reg >> DC_DC8051_CFG_HOST_CMD_1_RSP_DATA_SHIFT)
   8809				& DC_DC8051_CFG_HOST_CMD_1_RSP_DATA_MASK;
   8810		if (type == HCMD_READ_LCB_CSR) {
   8811			/* top 16 bits are in a different register */
   8812			*out_data |= (read_csr(dd, DC_DC8051_CFG_EXT_DEV_1)
   8813				& DC_DC8051_CFG_EXT_DEV_1_REQ_DATA_SMASK)
   8814				<< (48
   8815				    - DC_DC8051_CFG_EXT_DEV_1_REQ_DATA_SHIFT);
   8816		}
   8817	}
   8818	return_code = (reg >> DC_DC8051_CFG_HOST_CMD_1_RETURN_CODE_SHIFT)
   8819				& DC_DC8051_CFG_HOST_CMD_1_RETURN_CODE_MASK;
   8820	dd->dc8051_timed_out = 0;
   8821	/*
   8822	 * Clear command for next user.
   8823	 */
   8824	write_csr(dd, DC_DC8051_CFG_HOST_CMD_0, 0);
   8825
   8826fail:
   8827	mutex_unlock(&dd->dc8051_lock);
   8828	return return_code;
   8829}
   8830
   8831static int set_physical_link_state(struct hfi1_devdata *dd, u64 state)
   8832{
   8833	return do_8051_command(dd, HCMD_CHANGE_PHY_STATE, state, NULL);
   8834}
   8835
   8836int load_8051_config(struct hfi1_devdata *dd, u8 field_id,
   8837		     u8 lane_id, u32 config_data)
   8838{
   8839	u64 data;
   8840	int ret;
   8841
   8842	data = (u64)field_id << LOAD_DATA_FIELD_ID_SHIFT
   8843		| (u64)lane_id << LOAD_DATA_LANE_ID_SHIFT
   8844		| (u64)config_data << LOAD_DATA_DATA_SHIFT;
   8845	ret = do_8051_command(dd, HCMD_LOAD_CONFIG_DATA, data, NULL);
   8846	if (ret != HCMD_SUCCESS) {
   8847		dd_dev_err(dd,
   8848			   "load 8051 config: field id %d, lane %d, err %d\n",
   8849			   (int)field_id, (int)lane_id, ret);
   8850	}
   8851	return ret;
   8852}
   8853
   8854/*
   8855 * Read the 8051 firmware "registers".  Use the RAM directly.  Always
   8856 * set the result, even on error.
   8857 * Return 0 on success, -errno on failure
   8858 */
   8859int read_8051_config(struct hfi1_devdata *dd, u8 field_id, u8 lane_id,
   8860		     u32 *result)
   8861{
   8862	u64 big_data;
   8863	u32 addr;
   8864	int ret;
   8865
   8866	/* address start depends on the lane_id */
   8867	if (lane_id < 4)
   8868		addr = (4 * NUM_GENERAL_FIELDS)
   8869			+ (lane_id * 4 * NUM_LANE_FIELDS);
   8870	else
   8871		addr = 0;
   8872	addr += field_id * 4;
   8873
   8874	/* read is in 8-byte chunks, hardware will truncate the address down */
   8875	ret = read_8051_data(dd, addr, 8, &big_data);
   8876
   8877	if (ret == 0) {
   8878		/* extract the 4 bytes we want */
   8879		if (addr & 0x4)
   8880			*result = (u32)(big_data >> 32);
   8881		else
   8882			*result = (u32)big_data;
   8883	} else {
   8884		*result = 0;
   8885		dd_dev_err(dd, "%s: direct read failed, lane %d, field %d!\n",
   8886			   __func__, lane_id, field_id);
   8887	}
   8888
   8889	return ret;
   8890}
   8891
   8892static int write_vc_local_phy(struct hfi1_devdata *dd, u8 power_management,
   8893			      u8 continuous)
   8894{
   8895	u32 frame;
   8896
   8897	frame = continuous << CONTINIOUS_REMOTE_UPDATE_SUPPORT_SHIFT
   8898		| power_management << POWER_MANAGEMENT_SHIFT;
   8899	return load_8051_config(dd, VERIFY_CAP_LOCAL_PHY,
   8900				GENERAL_CONFIG, frame);
   8901}
   8902
   8903static int write_vc_local_fabric(struct hfi1_devdata *dd, u8 vau, u8 z, u8 vcu,
   8904				 u16 vl15buf, u8 crc_sizes)
   8905{
   8906	u32 frame;
   8907
   8908	frame = (u32)vau << VAU_SHIFT
   8909		| (u32)z << Z_SHIFT
   8910		| (u32)vcu << VCU_SHIFT
   8911		| (u32)vl15buf << VL15BUF_SHIFT
   8912		| (u32)crc_sizes << CRC_SIZES_SHIFT;
   8913	return load_8051_config(dd, VERIFY_CAP_LOCAL_FABRIC,
   8914				GENERAL_CONFIG, frame);
   8915}
   8916
   8917static void read_vc_local_link_mode(struct hfi1_devdata *dd, u8 *misc_bits,
   8918				    u8 *flag_bits, u16 *link_widths)
   8919{
   8920	u32 frame;
   8921
   8922	read_8051_config(dd, VERIFY_CAP_LOCAL_LINK_MODE, GENERAL_CONFIG,
   8923			 &frame);
   8924	*misc_bits = (frame >> MISC_CONFIG_BITS_SHIFT) & MISC_CONFIG_BITS_MASK;
   8925	*flag_bits = (frame >> LOCAL_FLAG_BITS_SHIFT) & LOCAL_FLAG_BITS_MASK;
   8926	*link_widths = (frame >> LINK_WIDTH_SHIFT) & LINK_WIDTH_MASK;
   8927}
   8928
   8929static int write_vc_local_link_mode(struct hfi1_devdata *dd,
   8930				    u8 misc_bits,
   8931				    u8 flag_bits,
   8932				    u16 link_widths)
   8933{
   8934	u32 frame;
   8935
   8936	frame = (u32)misc_bits << MISC_CONFIG_BITS_SHIFT
   8937		| (u32)flag_bits << LOCAL_FLAG_BITS_SHIFT
   8938		| (u32)link_widths << LINK_WIDTH_SHIFT;
   8939	return load_8051_config(dd, VERIFY_CAP_LOCAL_LINK_MODE, GENERAL_CONFIG,
   8940		     frame);
   8941}
   8942
   8943static int write_local_device_id(struct hfi1_devdata *dd, u16 device_id,
   8944				 u8 device_rev)
   8945{
   8946	u32 frame;
   8947
   8948	frame = ((u32)device_id << LOCAL_DEVICE_ID_SHIFT)
   8949		| ((u32)device_rev << LOCAL_DEVICE_REV_SHIFT);
   8950	return load_8051_config(dd, LOCAL_DEVICE_ID, GENERAL_CONFIG, frame);
   8951}
   8952
   8953static void read_remote_device_id(struct hfi1_devdata *dd, u16 *device_id,
   8954				  u8 *device_rev)
   8955{
   8956	u32 frame;
   8957
   8958	read_8051_config(dd, REMOTE_DEVICE_ID, GENERAL_CONFIG, &frame);
   8959	*device_id = (frame >> REMOTE_DEVICE_ID_SHIFT) & REMOTE_DEVICE_ID_MASK;
   8960	*device_rev = (frame >> REMOTE_DEVICE_REV_SHIFT)
   8961			& REMOTE_DEVICE_REV_MASK;
   8962}
   8963
   8964int write_host_interface_version(struct hfi1_devdata *dd, u8 version)
   8965{
   8966	u32 frame;
   8967	u32 mask;
   8968
   8969	mask = (HOST_INTERFACE_VERSION_MASK << HOST_INTERFACE_VERSION_SHIFT);
   8970	read_8051_config(dd, RESERVED_REGISTERS, GENERAL_CONFIG, &frame);
   8971	/* Clear, then set field */
   8972	frame &= ~mask;
   8973	frame |= ((u32)version << HOST_INTERFACE_VERSION_SHIFT);
   8974	return load_8051_config(dd, RESERVED_REGISTERS, GENERAL_CONFIG,
   8975				frame);
   8976}
   8977
   8978void read_misc_status(struct hfi1_devdata *dd, u8 *ver_major, u8 *ver_minor,
   8979		      u8 *ver_patch)
   8980{
   8981	u32 frame;
   8982
   8983	read_8051_config(dd, MISC_STATUS, GENERAL_CONFIG, &frame);
   8984	*ver_major = (frame >> STS_FM_VERSION_MAJOR_SHIFT) &
   8985		STS_FM_VERSION_MAJOR_MASK;
   8986	*ver_minor = (frame >> STS_FM_VERSION_MINOR_SHIFT) &
   8987		STS_FM_VERSION_MINOR_MASK;
   8988
   8989	read_8051_config(dd, VERSION_PATCH, GENERAL_CONFIG, &frame);
   8990	*ver_patch = (frame >> STS_FM_VERSION_PATCH_SHIFT) &
   8991		STS_FM_VERSION_PATCH_MASK;
   8992}
   8993
   8994static void read_vc_remote_phy(struct hfi1_devdata *dd, u8 *power_management,
   8995			       u8 *continuous)
   8996{
   8997	u32 frame;
   8998
   8999	read_8051_config(dd, VERIFY_CAP_REMOTE_PHY, GENERAL_CONFIG, &frame);
   9000	*power_management = (frame >> POWER_MANAGEMENT_SHIFT)
   9001					& POWER_MANAGEMENT_MASK;
   9002	*continuous = (frame >> CONTINIOUS_REMOTE_UPDATE_SUPPORT_SHIFT)
   9003					& CONTINIOUS_REMOTE_UPDATE_SUPPORT_MASK;
   9004}
   9005
   9006static void read_vc_remote_fabric(struct hfi1_devdata *dd, u8 *vau, u8 *z,
   9007				  u8 *vcu, u16 *vl15buf, u8 *crc_sizes)
   9008{
   9009	u32 frame;
   9010
   9011	read_8051_config(dd, VERIFY_CAP_REMOTE_FABRIC, GENERAL_CONFIG, &frame);
   9012	*vau = (frame >> VAU_SHIFT) & VAU_MASK;
   9013	*z = (frame >> Z_SHIFT) & Z_MASK;
   9014	*vcu = (frame >> VCU_SHIFT) & VCU_MASK;
   9015	*vl15buf = (frame >> VL15BUF_SHIFT) & VL15BUF_MASK;
   9016	*crc_sizes = (frame >> CRC_SIZES_SHIFT) & CRC_SIZES_MASK;
   9017}
   9018
   9019static void read_vc_remote_link_width(struct hfi1_devdata *dd,
   9020				      u8 *remote_tx_rate,
   9021				      u16 *link_widths)
   9022{
   9023	u32 frame;
   9024
   9025	read_8051_config(dd, VERIFY_CAP_REMOTE_LINK_WIDTH, GENERAL_CONFIG,
   9026			 &frame);
   9027	*remote_tx_rate = (frame >> REMOTE_TX_RATE_SHIFT)
   9028				& REMOTE_TX_RATE_MASK;
   9029	*link_widths = (frame >> LINK_WIDTH_SHIFT) & LINK_WIDTH_MASK;
   9030}
   9031
   9032static void read_local_lni(struct hfi1_devdata *dd, u8 *enable_lane_rx)
   9033{
   9034	u32 frame;
   9035
   9036	read_8051_config(dd, LOCAL_LNI_INFO, GENERAL_CONFIG, &frame);
   9037	*enable_lane_rx = (frame >> ENABLE_LANE_RX_SHIFT) & ENABLE_LANE_RX_MASK;
   9038}
   9039
   9040static void read_last_local_state(struct hfi1_devdata *dd, u32 *lls)
   9041{
   9042	read_8051_config(dd, LAST_LOCAL_STATE_COMPLETE, GENERAL_CONFIG, lls);
   9043}
   9044
   9045static void read_last_remote_state(struct hfi1_devdata *dd, u32 *lrs)
   9046{
   9047	read_8051_config(dd, LAST_REMOTE_STATE_COMPLETE, GENERAL_CONFIG, lrs);
   9048}
   9049
   9050void hfi1_read_link_quality(struct hfi1_devdata *dd, u8 *link_quality)
   9051{
   9052	u32 frame;
   9053	int ret;
   9054
   9055	*link_quality = 0;
   9056	if (dd->pport->host_link_state & HLS_UP) {
   9057		ret = read_8051_config(dd, LINK_QUALITY_INFO, GENERAL_CONFIG,
   9058				       &frame);
   9059		if (ret == 0)
   9060			*link_quality = (frame >> LINK_QUALITY_SHIFT)
   9061						& LINK_QUALITY_MASK;
   9062	}
   9063}
   9064
   9065static void read_planned_down_reason_code(struct hfi1_devdata *dd, u8 *pdrrc)
   9066{
   9067	u32 frame;
   9068
   9069	read_8051_config(dd, LINK_QUALITY_INFO, GENERAL_CONFIG, &frame);
   9070	*pdrrc = (frame >> DOWN_REMOTE_REASON_SHIFT) & DOWN_REMOTE_REASON_MASK;
   9071}
   9072
   9073static void read_link_down_reason(struct hfi1_devdata *dd, u8 *ldr)
   9074{
   9075	u32 frame;
   9076
   9077	read_8051_config(dd, LINK_DOWN_REASON, GENERAL_CONFIG, &frame);
   9078	*ldr = (frame & 0xff);
   9079}
   9080
   9081static int read_tx_settings(struct hfi1_devdata *dd,
   9082			    u8 *enable_lane_tx,
   9083			    u8 *tx_polarity_inversion,
   9084			    u8 *rx_polarity_inversion,
   9085			    u8 *max_rate)
   9086{
   9087	u32 frame;
   9088	int ret;
   9089
   9090	ret = read_8051_config(dd, TX_SETTINGS, GENERAL_CONFIG, &frame);
   9091	*enable_lane_tx = (frame >> ENABLE_LANE_TX_SHIFT)
   9092				& ENABLE_LANE_TX_MASK;
   9093	*tx_polarity_inversion = (frame >> TX_POLARITY_INVERSION_SHIFT)
   9094				& TX_POLARITY_INVERSION_MASK;
   9095	*rx_polarity_inversion = (frame >> RX_POLARITY_INVERSION_SHIFT)
   9096				& RX_POLARITY_INVERSION_MASK;
   9097	*max_rate = (frame >> MAX_RATE_SHIFT) & MAX_RATE_MASK;
   9098	return ret;
   9099}
   9100
   9101static int write_tx_settings(struct hfi1_devdata *dd,
   9102			     u8 enable_lane_tx,
   9103			     u8 tx_polarity_inversion,
   9104			     u8 rx_polarity_inversion,
   9105			     u8 max_rate)
   9106{
   9107	u32 frame;
   9108
   9109	/* no need to mask, all variable sizes match field widths */
   9110	frame = enable_lane_tx << ENABLE_LANE_TX_SHIFT
   9111		| tx_polarity_inversion << TX_POLARITY_INVERSION_SHIFT
   9112		| rx_polarity_inversion << RX_POLARITY_INVERSION_SHIFT
   9113		| max_rate << MAX_RATE_SHIFT;
   9114	return load_8051_config(dd, TX_SETTINGS, GENERAL_CONFIG, frame);
   9115}
   9116
   9117/*
   9118 * Read an idle LCB message.
   9119 *
   9120 * Returns 0 on success, -EINVAL on error
   9121 */
   9122static int read_idle_message(struct hfi1_devdata *dd, u64 type, u64 *data_out)
   9123{
   9124	int ret;
   9125
   9126	ret = do_8051_command(dd, HCMD_READ_LCB_IDLE_MSG, type, data_out);
   9127	if (ret != HCMD_SUCCESS) {
   9128		dd_dev_err(dd, "read idle message: type %d, err %d\n",
   9129			   (u32)type, ret);
   9130		return -EINVAL;
   9131	}
   9132	dd_dev_info(dd, "%s: read idle message 0x%llx\n", __func__, *data_out);
   9133	/* return only the payload as we already know the type */
   9134	*data_out >>= IDLE_PAYLOAD_SHIFT;
   9135	return 0;
   9136}
   9137
   9138/*
   9139 * Read an idle SMA message.  To be done in response to a notification from
   9140 * the 8051.
   9141 *
   9142 * Returns 0 on success, -EINVAL on error
   9143 */
   9144static int read_idle_sma(struct hfi1_devdata *dd, u64 *data)
   9145{
   9146	return read_idle_message(dd, (u64)IDLE_SMA << IDLE_MSG_TYPE_SHIFT,
   9147				 data);
   9148}
   9149
   9150/*
   9151 * Send an idle LCB message.
   9152 *
   9153 * Returns 0 on success, -EINVAL on error
   9154 */
   9155static int send_idle_message(struct hfi1_devdata *dd, u64 data)
   9156{
   9157	int ret;
   9158
   9159	dd_dev_info(dd, "%s: sending idle message 0x%llx\n", __func__, data);
   9160	ret = do_8051_command(dd, HCMD_SEND_LCB_IDLE_MSG, data, NULL);
   9161	if (ret != HCMD_SUCCESS) {
   9162		dd_dev_err(dd, "send idle message: data 0x%llx, err %d\n",
   9163			   data, ret);
   9164		return -EINVAL;
   9165	}
   9166	return 0;
   9167}
   9168
   9169/*
   9170 * Send an idle SMA message.
   9171 *
   9172 * Returns 0 on success, -EINVAL on error
   9173 */
   9174int send_idle_sma(struct hfi1_devdata *dd, u64 message)
   9175{
   9176	u64 data;
   9177
   9178	data = ((message & IDLE_PAYLOAD_MASK) << IDLE_PAYLOAD_SHIFT) |
   9179		((u64)IDLE_SMA << IDLE_MSG_TYPE_SHIFT);
   9180	return send_idle_message(dd, data);
   9181}
   9182
   9183/*
   9184 * Initialize the LCB then do a quick link up.  This may or may not be
   9185 * in loopback.
   9186 *
   9187 * return 0 on success, -errno on error
   9188 */
   9189static int do_quick_linkup(struct hfi1_devdata *dd)
   9190{
   9191	int ret;
   9192
   9193	lcb_shutdown(dd, 0);
   9194
   9195	if (loopback) {
   9196		/* LCB_CFG_LOOPBACK.VAL = 2 */
   9197		/* LCB_CFG_LANE_WIDTH.VAL = 0 */
   9198		write_csr(dd, DC_LCB_CFG_LOOPBACK,
   9199			  IB_PACKET_TYPE << DC_LCB_CFG_LOOPBACK_VAL_SHIFT);
   9200		write_csr(dd, DC_LCB_CFG_LANE_WIDTH, 0);
   9201	}
   9202
   9203	/* start the LCBs */
   9204	/* LCB_CFG_TX_FIFOS_RESET.VAL = 0 */
   9205	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET, 0);
   9206
   9207	/* simulator only loopback steps */
   9208	if (loopback && dd->icode == ICODE_FUNCTIONAL_SIMULATOR) {
   9209		/* LCB_CFG_RUN.EN = 1 */
   9210		write_csr(dd, DC_LCB_CFG_RUN,
   9211			  1ull << DC_LCB_CFG_RUN_EN_SHIFT);
   9212
   9213		ret = wait_link_transfer_active(dd, 10);
   9214		if (ret)
   9215			return ret;
   9216
   9217		write_csr(dd, DC_LCB_CFG_ALLOW_LINK_UP,
   9218			  1ull << DC_LCB_CFG_ALLOW_LINK_UP_VAL_SHIFT);
   9219	}
   9220
   9221	if (!loopback) {
   9222		/*
   9223		 * When doing quick linkup and not in loopback, both
   9224		 * sides must be done with LCB set-up before either
   9225		 * starts the quick linkup.  Put a delay here so that
   9226		 * both sides can be started and have a chance to be
   9227		 * done with LCB set up before resuming.
   9228		 */
   9229		dd_dev_err(dd,
   9230			   "Pausing for peer to be finished with LCB set up\n");
   9231		msleep(5000);
   9232		dd_dev_err(dd, "Continuing with quick linkup\n");
   9233	}
   9234
   9235	write_csr(dd, DC_LCB_ERR_EN, 0); /* mask LCB errors */
   9236	set_8051_lcb_access(dd);
   9237
   9238	/*
   9239	 * State "quick" LinkUp request sets the physical link state to
   9240	 * LinkUp without a verify capability sequence.
   9241	 * This state is in simulator v37 and later.
   9242	 */
   9243	ret = set_physical_link_state(dd, PLS_QUICK_LINKUP);
   9244	if (ret != HCMD_SUCCESS) {
   9245		dd_dev_err(dd,
   9246			   "%s: set physical link state to quick LinkUp failed with return %d\n",
   9247			   __func__, ret);
   9248
   9249		set_host_lcb_access(dd);
   9250		write_csr(dd, DC_LCB_ERR_EN, ~0ull); /* watch LCB errors */
   9251
   9252		if (ret >= 0)
   9253			ret = -EINVAL;
   9254		return ret;
   9255	}
   9256
   9257	return 0; /* success */
   9258}
   9259
   9260/*
   9261 * Do all special steps to set up loopback.
   9262 */
   9263static int init_loopback(struct hfi1_devdata *dd)
   9264{
   9265	dd_dev_info(dd, "Entering loopback mode\n");
   9266
   9267	/* all loopbacks should disable self GUID check */
   9268	write_csr(dd, DC_DC8051_CFG_MODE,
   9269		  (read_csr(dd, DC_DC8051_CFG_MODE) | DISABLE_SELF_GUID_CHECK));
   9270
   9271	/*
   9272	 * The simulator has only one loopback option - LCB.  Switch
   9273	 * to that option, which includes quick link up.
   9274	 *
   9275	 * Accept all valid loopback values.
   9276	 */
   9277	if ((dd->icode == ICODE_FUNCTIONAL_SIMULATOR) &&
   9278	    (loopback == LOOPBACK_SERDES || loopback == LOOPBACK_LCB ||
   9279	     loopback == LOOPBACK_CABLE)) {
   9280		loopback = LOOPBACK_LCB;
   9281		quick_linkup = 1;
   9282		return 0;
   9283	}
   9284
   9285	/*
   9286	 * SerDes loopback init sequence is handled in set_local_link_attributes
   9287	 */
   9288	if (loopback == LOOPBACK_SERDES)
   9289		return 0;
   9290
   9291	/* LCB loopback - handled at poll time */
   9292	if (loopback == LOOPBACK_LCB) {
   9293		quick_linkup = 1; /* LCB is always quick linkup */
   9294
   9295		/* not supported in emulation due to emulation RTL changes */
   9296		if (dd->icode == ICODE_FPGA_EMULATION) {
   9297			dd_dev_err(dd,
   9298				   "LCB loopback not supported in emulation\n");
   9299			return -EINVAL;
   9300		}
   9301		return 0;
   9302	}
   9303
   9304	/* external cable loopback requires no extra steps */
   9305	if (loopback == LOOPBACK_CABLE)
   9306		return 0;
   9307
   9308	dd_dev_err(dd, "Invalid loopback mode %d\n", loopback);
   9309	return -EINVAL;
   9310}
   9311
   9312/*
   9313 * Translate from the OPA_LINK_WIDTH handed to us by the FM to bits
   9314 * used in the Verify Capability link width attribute.
   9315 */
   9316static u16 opa_to_vc_link_widths(u16 opa_widths)
   9317{
   9318	int i;
   9319	u16 result = 0;
   9320
   9321	static const struct link_bits {
   9322		u16 from;
   9323		u16 to;
   9324	} opa_link_xlate[] = {
   9325		{ OPA_LINK_WIDTH_1X, 1 << (1 - 1)  },
   9326		{ OPA_LINK_WIDTH_2X, 1 << (2 - 1)  },
   9327		{ OPA_LINK_WIDTH_3X, 1 << (3 - 1)  },
   9328		{ OPA_LINK_WIDTH_4X, 1 << (4 - 1)  },
   9329	};
   9330
   9331	for (i = 0; i < ARRAY_SIZE(opa_link_xlate); i++) {
   9332		if (opa_widths & opa_link_xlate[i].from)
   9333			result |= opa_link_xlate[i].to;
   9334	}
   9335	return result;
   9336}
   9337
   9338/*
   9339 * Set link attributes before moving to polling.
   9340 */
   9341static int set_local_link_attributes(struct hfi1_pportdata *ppd)
   9342{
   9343	struct hfi1_devdata *dd = ppd->dd;
   9344	u8 enable_lane_tx;
   9345	u8 tx_polarity_inversion;
   9346	u8 rx_polarity_inversion;
   9347	int ret;
   9348	u32 misc_bits = 0;
   9349	/* reset our fabric serdes to clear any lingering problems */
   9350	fabric_serdes_reset(dd);
   9351
   9352	/* set the local tx rate - need to read-modify-write */
   9353	ret = read_tx_settings(dd, &enable_lane_tx, &tx_polarity_inversion,
   9354			       &rx_polarity_inversion, &ppd->local_tx_rate);
   9355	if (ret)
   9356		goto set_local_link_attributes_fail;
   9357
   9358	if (dd->dc8051_ver < dc8051_ver(0, 20, 0)) {
   9359		/* set the tx rate to the fastest enabled */
   9360		if (ppd->link_speed_enabled & OPA_LINK_SPEED_25G)
   9361			ppd->local_tx_rate = 1;
   9362		else
   9363			ppd->local_tx_rate = 0;
   9364	} else {
   9365		/* set the tx rate to all enabled */
   9366		ppd->local_tx_rate = 0;
   9367		if (ppd->link_speed_enabled & OPA_LINK_SPEED_25G)
   9368			ppd->local_tx_rate |= 2;
   9369		if (ppd->link_speed_enabled & OPA_LINK_SPEED_12_5G)
   9370			ppd->local_tx_rate |= 1;
   9371	}
   9372
   9373	enable_lane_tx = 0xF; /* enable all four lanes */
   9374	ret = write_tx_settings(dd, enable_lane_tx, tx_polarity_inversion,
   9375				rx_polarity_inversion, ppd->local_tx_rate);
   9376	if (ret != HCMD_SUCCESS)
   9377		goto set_local_link_attributes_fail;
   9378
   9379	ret = write_host_interface_version(dd, HOST_INTERFACE_VERSION);
   9380	if (ret != HCMD_SUCCESS) {
   9381		dd_dev_err(dd,
   9382			   "Failed to set host interface version, return 0x%x\n",
   9383			   ret);
   9384		goto set_local_link_attributes_fail;
   9385	}
   9386
   9387	/*
   9388	 * DC supports continuous updates.
   9389	 */
   9390	ret = write_vc_local_phy(dd,
   9391				 0 /* no power management */,
   9392				 1 /* continuous updates */);
   9393	if (ret != HCMD_SUCCESS)
   9394		goto set_local_link_attributes_fail;
   9395
   9396	/* z=1 in the next call: AU of 0 is not supported by the hardware */
   9397	ret = write_vc_local_fabric(dd, dd->vau, 1, dd->vcu, dd->vl15_init,
   9398				    ppd->port_crc_mode_enabled);
   9399	if (ret != HCMD_SUCCESS)
   9400		goto set_local_link_attributes_fail;
   9401
   9402	/*
   9403	 * SerDes loopback init sequence requires
   9404	 * setting bit 0 of MISC_CONFIG_BITS
   9405	 */
   9406	if (loopback == LOOPBACK_SERDES)
   9407		misc_bits |= 1 << LOOPBACK_SERDES_CONFIG_BIT_MASK_SHIFT;
   9408
   9409	/*
   9410	 * An external device configuration request is used to reset the LCB
   9411	 * to retry to obtain operational lanes when the first attempt is
   9412	 * unsuccesful.
   9413	 */
   9414	if (dd->dc8051_ver >= dc8051_ver(1, 25, 0))
   9415		misc_bits |= 1 << EXT_CFG_LCB_RESET_SUPPORTED_SHIFT;
   9416
   9417	ret = write_vc_local_link_mode(dd, misc_bits, 0,
   9418				       opa_to_vc_link_widths(
   9419						ppd->link_width_enabled));
   9420	if (ret != HCMD_SUCCESS)
   9421		goto set_local_link_attributes_fail;
   9422
   9423	/* let peer know who we are */
   9424	ret = write_local_device_id(dd, dd->pcidev->device, dd->minrev);
   9425	if (ret == HCMD_SUCCESS)
   9426		return 0;
   9427
   9428set_local_link_attributes_fail:
   9429	dd_dev_err(dd,
   9430		   "Failed to set local link attributes, return 0x%x\n",
   9431		   ret);
   9432	return ret;
   9433}
   9434
   9435/*
   9436 * Call this to start the link.
   9437 * Do not do anything if the link is disabled.
   9438 * Returns 0 if link is disabled, moved to polling, or the driver is not ready.
   9439 */
   9440int start_link(struct hfi1_pportdata *ppd)
   9441{
   9442	/*
   9443	 * Tune the SerDes to a ballpark setting for optimal signal and bit
   9444	 * error rate.  Needs to be done before starting the link.
   9445	 */
   9446	tune_serdes(ppd);
   9447
   9448	if (!ppd->driver_link_ready) {
   9449		dd_dev_info(ppd->dd,
   9450			    "%s: stopping link start because driver is not ready\n",
   9451			    __func__);
   9452		return 0;
   9453	}
   9454
   9455	/*
   9456	 * FULL_MGMT_P_KEY is cleared from the pkey table, so that the
   9457	 * pkey table can be configured properly if the HFI unit is connected
   9458	 * to switch port with MgmtAllowed=NO
   9459	 */
   9460	clear_full_mgmt_pkey(ppd);
   9461
   9462	return set_link_state(ppd, HLS_DN_POLL);
   9463}
   9464
   9465static void wait_for_qsfp_init(struct hfi1_pportdata *ppd)
   9466{
   9467	struct hfi1_devdata *dd = ppd->dd;
   9468	u64 mask;
   9469	unsigned long timeout;
   9470
   9471	/*
   9472	 * Some QSFP cables have a quirk that asserts the IntN line as a side
   9473	 * effect of power up on plug-in. We ignore this false positive
   9474	 * interrupt until the module has finished powering up by waiting for
   9475	 * a minimum timeout of the module inrush initialization time of
   9476	 * 500 ms (SFF 8679 Table 5-6) to ensure the voltage rails in the
   9477	 * module have stabilized.
   9478	 */
   9479	msleep(500);
   9480
   9481	/*
   9482	 * Check for QSFP interrupt for t_init (SFF 8679 Table 8-1)
   9483	 */
   9484	timeout = jiffies + msecs_to_jiffies(2000);
   9485	while (1) {
   9486		mask = read_csr(dd, dd->hfi1_id ?
   9487				ASIC_QSFP2_IN : ASIC_QSFP1_IN);
   9488		if (!(mask & QSFP_HFI0_INT_N))
   9489			break;
   9490		if (time_after(jiffies, timeout)) {
   9491			dd_dev_info(dd, "%s: No IntN detected, reset complete\n",
   9492				    __func__);
   9493			break;
   9494		}
   9495		udelay(2);
   9496	}
   9497}
   9498
   9499static void set_qsfp_int_n(struct hfi1_pportdata *ppd, u8 enable)
   9500{
   9501	struct hfi1_devdata *dd = ppd->dd;
   9502	u64 mask;
   9503
   9504	mask = read_csr(dd, dd->hfi1_id ? ASIC_QSFP2_MASK : ASIC_QSFP1_MASK);
   9505	if (enable) {
   9506		/*
   9507		 * Clear the status register to avoid an immediate interrupt
   9508		 * when we re-enable the IntN pin
   9509		 */
   9510		write_csr(dd, dd->hfi1_id ? ASIC_QSFP2_CLEAR : ASIC_QSFP1_CLEAR,
   9511			  QSFP_HFI0_INT_N);
   9512		mask |= (u64)QSFP_HFI0_INT_N;
   9513	} else {
   9514		mask &= ~(u64)QSFP_HFI0_INT_N;
   9515	}
   9516	write_csr(dd, dd->hfi1_id ? ASIC_QSFP2_MASK : ASIC_QSFP1_MASK, mask);
   9517}
   9518
   9519int reset_qsfp(struct hfi1_pportdata *ppd)
   9520{
   9521	struct hfi1_devdata *dd = ppd->dd;
   9522	u64 mask, qsfp_mask;
   9523
   9524	/* Disable INT_N from triggering QSFP interrupts */
   9525	set_qsfp_int_n(ppd, 0);
   9526
   9527	/* Reset the QSFP */
   9528	mask = (u64)QSFP_HFI0_RESET_N;
   9529
   9530	qsfp_mask = read_csr(dd,
   9531			     dd->hfi1_id ? ASIC_QSFP2_OUT : ASIC_QSFP1_OUT);
   9532	qsfp_mask &= ~mask;
   9533	write_csr(dd,
   9534		  dd->hfi1_id ? ASIC_QSFP2_OUT : ASIC_QSFP1_OUT, qsfp_mask);
   9535
   9536	udelay(10);
   9537
   9538	qsfp_mask |= mask;
   9539	write_csr(dd,
   9540		  dd->hfi1_id ? ASIC_QSFP2_OUT : ASIC_QSFP1_OUT, qsfp_mask);
   9541
   9542	wait_for_qsfp_init(ppd);
   9543
   9544	/*
   9545	 * Allow INT_N to trigger the QSFP interrupt to watch
   9546	 * for alarms and warnings
   9547	 */
   9548	set_qsfp_int_n(ppd, 1);
   9549
   9550	/*
   9551	 * After the reset, AOC transmitters are enabled by default. They need
   9552	 * to be turned off to complete the QSFP setup before they can be
   9553	 * enabled again.
   9554	 */
   9555	return set_qsfp_tx(ppd, 0);
   9556}
   9557
   9558static int handle_qsfp_error_conditions(struct hfi1_pportdata *ppd,
   9559					u8 *qsfp_interrupt_status)
   9560{
   9561	struct hfi1_devdata *dd = ppd->dd;
   9562
   9563	if ((qsfp_interrupt_status[0] & QSFP_HIGH_TEMP_ALARM) ||
   9564	    (qsfp_interrupt_status[0] & QSFP_HIGH_TEMP_WARNING))
   9565		dd_dev_err(dd, "%s: QSFP cable temperature too high\n",
   9566			   __func__);
   9567
   9568	if ((qsfp_interrupt_status[0] & QSFP_LOW_TEMP_ALARM) ||
   9569	    (qsfp_interrupt_status[0] & QSFP_LOW_TEMP_WARNING))
   9570		dd_dev_err(dd, "%s: QSFP cable temperature too low\n",
   9571			   __func__);
   9572
   9573	/*
   9574	 * The remaining alarms/warnings don't matter if the link is down.
   9575	 */
   9576	if (ppd->host_link_state & HLS_DOWN)
   9577		return 0;
   9578
   9579	if ((qsfp_interrupt_status[1] & QSFP_HIGH_VCC_ALARM) ||
   9580	    (qsfp_interrupt_status[1] & QSFP_HIGH_VCC_WARNING))
   9581		dd_dev_err(dd, "%s: QSFP supply voltage too high\n",
   9582			   __func__);
   9583
   9584	if ((qsfp_interrupt_status[1] & QSFP_LOW_VCC_ALARM) ||
   9585	    (qsfp_interrupt_status[1] & QSFP_LOW_VCC_WARNING))
   9586		dd_dev_err(dd, "%s: QSFP supply voltage too low\n",
   9587			   __func__);
   9588
   9589	/* Byte 2 is vendor specific */
   9590
   9591	if ((qsfp_interrupt_status[3] & QSFP_HIGH_POWER_ALARM) ||
   9592	    (qsfp_interrupt_status[3] & QSFP_HIGH_POWER_WARNING))
   9593		dd_dev_err(dd, "%s: Cable RX channel 1/2 power too high\n",
   9594			   __func__);
   9595
   9596	if ((qsfp_interrupt_status[3] & QSFP_LOW_POWER_ALARM) ||
   9597	    (qsfp_interrupt_status[3] & QSFP_LOW_POWER_WARNING))
   9598		dd_dev_err(dd, "%s: Cable RX channel 1/2 power too low\n",
   9599			   __func__);
   9600
   9601	if ((qsfp_interrupt_status[4] & QSFP_HIGH_POWER_ALARM) ||
   9602	    (qsfp_interrupt_status[4] & QSFP_HIGH_POWER_WARNING))
   9603		dd_dev_err(dd, "%s: Cable RX channel 3/4 power too high\n",
   9604			   __func__);
   9605
   9606	if ((qsfp_interrupt_status[4] & QSFP_LOW_POWER_ALARM) ||
   9607	    (qsfp_interrupt_status[4] & QSFP_LOW_POWER_WARNING))
   9608		dd_dev_err(dd, "%s: Cable RX channel 3/4 power too low\n",
   9609			   __func__);
   9610
   9611	if ((qsfp_interrupt_status[5] & QSFP_HIGH_BIAS_ALARM) ||
   9612	    (qsfp_interrupt_status[5] & QSFP_HIGH_BIAS_WARNING))
   9613		dd_dev_err(dd, "%s: Cable TX channel 1/2 bias too high\n",
   9614			   __func__);
   9615
   9616	if ((qsfp_interrupt_status[5] & QSFP_LOW_BIAS_ALARM) ||
   9617	    (qsfp_interrupt_status[5] & QSFP_LOW_BIAS_WARNING))
   9618		dd_dev_err(dd, "%s: Cable TX channel 1/2 bias too low\n",
   9619			   __func__);
   9620
   9621	if ((qsfp_interrupt_status[6] & QSFP_HIGH_BIAS_ALARM) ||
   9622	    (qsfp_interrupt_status[6] & QSFP_HIGH_BIAS_WARNING))
   9623		dd_dev_err(dd, "%s: Cable TX channel 3/4 bias too high\n",
   9624			   __func__);
   9625
   9626	if ((qsfp_interrupt_status[6] & QSFP_LOW_BIAS_ALARM) ||
   9627	    (qsfp_interrupt_status[6] & QSFP_LOW_BIAS_WARNING))
   9628		dd_dev_err(dd, "%s: Cable TX channel 3/4 bias too low\n",
   9629			   __func__);
   9630
   9631	if ((qsfp_interrupt_status[7] & QSFP_HIGH_POWER_ALARM) ||
   9632	    (qsfp_interrupt_status[7] & QSFP_HIGH_POWER_WARNING))
   9633		dd_dev_err(dd, "%s: Cable TX channel 1/2 power too high\n",
   9634			   __func__);
   9635
   9636	if ((qsfp_interrupt_status[7] & QSFP_LOW_POWER_ALARM) ||
   9637	    (qsfp_interrupt_status[7] & QSFP_LOW_POWER_WARNING))
   9638		dd_dev_err(dd, "%s: Cable TX channel 1/2 power too low\n",
   9639			   __func__);
   9640
   9641	if ((qsfp_interrupt_status[8] & QSFP_HIGH_POWER_ALARM) ||
   9642	    (qsfp_interrupt_status[8] & QSFP_HIGH_POWER_WARNING))
   9643		dd_dev_err(dd, "%s: Cable TX channel 3/4 power too high\n",
   9644			   __func__);
   9645
   9646	if ((qsfp_interrupt_status[8] & QSFP_LOW_POWER_ALARM) ||
   9647	    (qsfp_interrupt_status[8] & QSFP_LOW_POWER_WARNING))
   9648		dd_dev_err(dd, "%s: Cable TX channel 3/4 power too low\n",
   9649			   __func__);
   9650
   9651	/* Bytes 9-10 and 11-12 are reserved */
   9652	/* Bytes 13-15 are vendor specific */
   9653
   9654	return 0;
   9655}
   9656
   9657/* This routine will only be scheduled if the QSFP module present is asserted */
   9658void qsfp_event(struct work_struct *work)
   9659{
   9660	struct qsfp_data *qd;
   9661	struct hfi1_pportdata *ppd;
   9662	struct hfi1_devdata *dd;
   9663
   9664	qd = container_of(work, struct qsfp_data, qsfp_work);
   9665	ppd = qd->ppd;
   9666	dd = ppd->dd;
   9667
   9668	/* Sanity check */
   9669	if (!qsfp_mod_present(ppd))
   9670		return;
   9671
   9672	if (ppd->host_link_state == HLS_DN_DISABLE) {
   9673		dd_dev_info(ppd->dd,
   9674			    "%s: stopping link start because link is disabled\n",
   9675			    __func__);
   9676		return;
   9677	}
   9678
   9679	/*
   9680	 * Turn DC back on after cable has been re-inserted. Up until
   9681	 * now, the DC has been in reset to save power.
   9682	 */
   9683	dc_start(dd);
   9684
   9685	if (qd->cache_refresh_required) {
   9686		set_qsfp_int_n(ppd, 0);
   9687
   9688		wait_for_qsfp_init(ppd);
   9689
   9690		/*
   9691		 * Allow INT_N to trigger the QSFP interrupt to watch
   9692		 * for alarms and warnings
   9693		 */
   9694		set_qsfp_int_n(ppd, 1);
   9695
   9696		start_link(ppd);
   9697	}
   9698
   9699	if (qd->check_interrupt_flags) {
   9700		u8 qsfp_interrupt_status[16] = {0,};
   9701
   9702		if (one_qsfp_read(ppd, dd->hfi1_id, 6,
   9703				  &qsfp_interrupt_status[0], 16) != 16) {
   9704			dd_dev_info(dd,
   9705				    "%s: Failed to read status of QSFP module\n",
   9706				    __func__);
   9707		} else {
   9708			unsigned long flags;
   9709
   9710			handle_qsfp_error_conditions(
   9711					ppd, qsfp_interrupt_status);
   9712			spin_lock_irqsave(&ppd->qsfp_info.qsfp_lock, flags);
   9713			ppd->qsfp_info.check_interrupt_flags = 0;
   9714			spin_unlock_irqrestore(&ppd->qsfp_info.qsfp_lock,
   9715					       flags);
   9716		}
   9717	}
   9718}
   9719
   9720void init_qsfp_int(struct hfi1_devdata *dd)
   9721{
   9722	struct hfi1_pportdata *ppd = dd->pport;
   9723	u64 qsfp_mask;
   9724
   9725	qsfp_mask = (u64)(QSFP_HFI0_INT_N | QSFP_HFI0_MODPRST_N);
   9726	/* Clear current status to avoid spurious interrupts */
   9727	write_csr(dd, dd->hfi1_id ? ASIC_QSFP2_CLEAR : ASIC_QSFP1_CLEAR,
   9728		  qsfp_mask);
   9729	write_csr(dd, dd->hfi1_id ? ASIC_QSFP2_MASK : ASIC_QSFP1_MASK,
   9730		  qsfp_mask);
   9731
   9732	set_qsfp_int_n(ppd, 0);
   9733
   9734	/* Handle active low nature of INT_N and MODPRST_N pins */
   9735	if (qsfp_mod_present(ppd))
   9736		qsfp_mask &= ~(u64)QSFP_HFI0_MODPRST_N;
   9737	write_csr(dd,
   9738		  dd->hfi1_id ? ASIC_QSFP2_INVERT : ASIC_QSFP1_INVERT,
   9739		  qsfp_mask);
   9740
   9741	/* Enable the appropriate QSFP IRQ source */
   9742	if (!dd->hfi1_id)
   9743		set_intr_bits(dd, QSFP1_INT, QSFP1_INT, true);
   9744	else
   9745		set_intr_bits(dd, QSFP2_INT, QSFP2_INT, true);
   9746}
   9747
   9748/*
   9749 * Do a one-time initialize of the LCB block.
   9750 */
   9751static void init_lcb(struct hfi1_devdata *dd)
   9752{
   9753	/* simulator does not correctly handle LCB cclk loopback, skip */
   9754	if (dd->icode == ICODE_FUNCTIONAL_SIMULATOR)
   9755		return;
   9756
   9757	/* the DC has been reset earlier in the driver load */
   9758
   9759	/* set LCB for cclk loopback on the port */
   9760	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET, 0x01);
   9761	write_csr(dd, DC_LCB_CFG_LANE_WIDTH, 0x00);
   9762	write_csr(dd, DC_LCB_CFG_REINIT_AS_SLAVE, 0x00);
   9763	write_csr(dd, DC_LCB_CFG_CNT_FOR_SKIP_STALL, 0x110);
   9764	write_csr(dd, DC_LCB_CFG_CLK_CNTR, 0x08);
   9765	write_csr(dd, DC_LCB_CFG_LOOPBACK, 0x02);
   9766	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET, 0x00);
   9767}
   9768
   9769/*
   9770 * Perform a test read on the QSFP.  Return 0 on success, -ERRNO
   9771 * on error.
   9772 */
   9773static int test_qsfp_read(struct hfi1_pportdata *ppd)
   9774{
   9775	int ret;
   9776	u8 status;
   9777
   9778	/*
   9779	 * Report success if not a QSFP or, if it is a QSFP, but the cable is
   9780	 * not present
   9781	 */
   9782	if (ppd->port_type != PORT_TYPE_QSFP || !qsfp_mod_present(ppd))
   9783		return 0;
   9784
   9785	/* read byte 2, the status byte */
   9786	ret = one_qsfp_read(ppd, ppd->dd->hfi1_id, 2, &status, 1);
   9787	if (ret < 0)
   9788		return ret;
   9789	if (ret != 1)
   9790		return -EIO;
   9791
   9792	return 0; /* success */
   9793}
   9794
   9795/*
   9796 * Values for QSFP retry.
   9797 *
   9798 * Give up after 10s (20 x 500ms).  The overall timeout was empirically
   9799 * arrived at from experience on a large cluster.
   9800 */
   9801#define MAX_QSFP_RETRIES 20
   9802#define QSFP_RETRY_WAIT 500 /* msec */
   9803
   9804/*
   9805 * Try a QSFP read.  If it fails, schedule a retry for later.
   9806 * Called on first link activation after driver load.
   9807 */
   9808static void try_start_link(struct hfi1_pportdata *ppd)
   9809{
   9810	if (test_qsfp_read(ppd)) {
   9811		/* read failed */
   9812		if (ppd->qsfp_retry_count >= MAX_QSFP_RETRIES) {
   9813			dd_dev_err(ppd->dd, "QSFP not responding, giving up\n");
   9814			return;
   9815		}
   9816		dd_dev_info(ppd->dd,
   9817			    "QSFP not responding, waiting and retrying %d\n",
   9818			    (int)ppd->qsfp_retry_count);
   9819		ppd->qsfp_retry_count++;
   9820		queue_delayed_work(ppd->link_wq, &ppd->start_link_work,
   9821				   msecs_to_jiffies(QSFP_RETRY_WAIT));
   9822		return;
   9823	}
   9824	ppd->qsfp_retry_count = 0;
   9825
   9826	start_link(ppd);
   9827}
   9828
   9829/*
   9830 * Workqueue function to start the link after a delay.
   9831 */
   9832void handle_start_link(struct work_struct *work)
   9833{
   9834	struct hfi1_pportdata *ppd = container_of(work, struct hfi1_pportdata,
   9835						  start_link_work.work);
   9836	try_start_link(ppd);
   9837}
   9838
   9839int bringup_serdes(struct hfi1_pportdata *ppd)
   9840{
   9841	struct hfi1_devdata *dd = ppd->dd;
   9842	u64 guid;
   9843	int ret;
   9844
   9845	if (HFI1_CAP_IS_KSET(EXTENDED_PSN))
   9846		add_rcvctrl(dd, RCV_CTRL_RCV_EXTENDED_PSN_ENABLE_SMASK);
   9847
   9848	guid = ppd->guids[HFI1_PORT_GUID_INDEX];
   9849	if (!guid) {
   9850		if (dd->base_guid)
   9851			guid = dd->base_guid + ppd->port - 1;
   9852		ppd->guids[HFI1_PORT_GUID_INDEX] = guid;
   9853	}
   9854
   9855	/* Set linkinit_reason on power up per OPA spec */
   9856	ppd->linkinit_reason = OPA_LINKINIT_REASON_LINKUP;
   9857
   9858	/* one-time init of the LCB */
   9859	init_lcb(dd);
   9860
   9861	if (loopback) {
   9862		ret = init_loopback(dd);
   9863		if (ret < 0)
   9864			return ret;
   9865	}
   9866
   9867	get_port_type(ppd);
   9868	if (ppd->port_type == PORT_TYPE_QSFP) {
   9869		set_qsfp_int_n(ppd, 0);
   9870		wait_for_qsfp_init(ppd);
   9871		set_qsfp_int_n(ppd, 1);
   9872	}
   9873
   9874	try_start_link(ppd);
   9875	return 0;
   9876}
   9877
   9878void hfi1_quiet_serdes(struct hfi1_pportdata *ppd)
   9879{
   9880	struct hfi1_devdata *dd = ppd->dd;
   9881
   9882	/*
   9883	 * Shut down the link and keep it down.   First turn off that the
   9884	 * driver wants to allow the link to be up (driver_link_ready).
   9885	 * Then make sure the link is not automatically restarted
   9886	 * (link_enabled).  Cancel any pending restart.  And finally
   9887	 * go offline.
   9888	 */
   9889	ppd->driver_link_ready = 0;
   9890	ppd->link_enabled = 0;
   9891
   9892	ppd->qsfp_retry_count = MAX_QSFP_RETRIES; /* prevent more retries */
   9893	flush_delayed_work(&ppd->start_link_work);
   9894	cancel_delayed_work_sync(&ppd->start_link_work);
   9895
   9896	ppd->offline_disabled_reason =
   9897			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_REBOOT);
   9898	set_link_down_reason(ppd, OPA_LINKDOWN_REASON_REBOOT, 0,
   9899			     OPA_LINKDOWN_REASON_REBOOT);
   9900	set_link_state(ppd, HLS_DN_OFFLINE);
   9901
   9902	/* disable the port */
   9903	clear_rcvctrl(dd, RCV_CTRL_RCV_PORT_ENABLE_SMASK);
   9904	cancel_work_sync(&ppd->freeze_work);
   9905}
   9906
   9907static inline int init_cpu_counters(struct hfi1_devdata *dd)
   9908{
   9909	struct hfi1_pportdata *ppd;
   9910	int i;
   9911
   9912	ppd = (struct hfi1_pportdata *)(dd + 1);
   9913	for (i = 0; i < dd->num_pports; i++, ppd++) {
   9914		ppd->ibport_data.rvp.rc_acks = NULL;
   9915		ppd->ibport_data.rvp.rc_qacks = NULL;
   9916		ppd->ibport_data.rvp.rc_acks = alloc_percpu(u64);
   9917		ppd->ibport_data.rvp.rc_qacks = alloc_percpu(u64);
   9918		ppd->ibport_data.rvp.rc_delayed_comp = alloc_percpu(u64);
   9919		if (!ppd->ibport_data.rvp.rc_acks ||
   9920		    !ppd->ibport_data.rvp.rc_delayed_comp ||
   9921		    !ppd->ibport_data.rvp.rc_qacks)
   9922			return -ENOMEM;
   9923	}
   9924
   9925	return 0;
   9926}
   9927
   9928/*
   9929 * index is the index into the receive array
   9930 */
   9931void hfi1_put_tid(struct hfi1_devdata *dd, u32 index,
   9932		  u32 type, unsigned long pa, u16 order)
   9933{
   9934	u64 reg;
   9935
   9936	if (!(dd->flags & HFI1_PRESENT))
   9937		goto done;
   9938
   9939	if (type == PT_INVALID || type == PT_INVALID_FLUSH) {
   9940		pa = 0;
   9941		order = 0;
   9942	} else if (type > PT_INVALID) {
   9943		dd_dev_err(dd,
   9944			   "unexpected receive array type %u for index %u, not handled\n",
   9945			   type, index);
   9946		goto done;
   9947	}
   9948	trace_hfi1_put_tid(dd, index, type, pa, order);
   9949
   9950#define RT_ADDR_SHIFT 12	/* 4KB kernel address boundary */
   9951	reg = RCV_ARRAY_RT_WRITE_ENABLE_SMASK
   9952		| (u64)order << RCV_ARRAY_RT_BUF_SIZE_SHIFT
   9953		| ((pa >> RT_ADDR_SHIFT) & RCV_ARRAY_RT_ADDR_MASK)
   9954					<< RCV_ARRAY_RT_ADDR_SHIFT;
   9955	trace_hfi1_write_rcvarray(dd->rcvarray_wc + (index * 8), reg);
   9956	writeq(reg, dd->rcvarray_wc + (index * 8));
   9957
   9958	if (type == PT_EAGER || type == PT_INVALID_FLUSH || (index & 3) == 3)
   9959		/*
   9960		 * Eager entries are written and flushed
   9961		 *
   9962		 * Expected entries are flushed every 4 writes
   9963		 */
   9964		flush_wc();
   9965done:
   9966	return;
   9967}
   9968
   9969void hfi1_clear_tids(struct hfi1_ctxtdata *rcd)
   9970{
   9971	struct hfi1_devdata *dd = rcd->dd;
   9972	u32 i;
   9973
   9974	/* this could be optimized */
   9975	for (i = rcd->eager_base; i < rcd->eager_base +
   9976		     rcd->egrbufs.alloced; i++)
   9977		hfi1_put_tid(dd, i, PT_INVALID, 0, 0);
   9978
   9979	for (i = rcd->expected_base;
   9980			i < rcd->expected_base + rcd->expected_count; i++)
   9981		hfi1_put_tid(dd, i, PT_INVALID, 0, 0);
   9982}
   9983
   9984static const char * const ib_cfg_name_strings[] = {
   9985	"HFI1_IB_CFG_LIDLMC",
   9986	"HFI1_IB_CFG_LWID_DG_ENB",
   9987	"HFI1_IB_CFG_LWID_ENB",
   9988	"HFI1_IB_CFG_LWID",
   9989	"HFI1_IB_CFG_SPD_ENB",
   9990	"HFI1_IB_CFG_SPD",
   9991	"HFI1_IB_CFG_RXPOL_ENB",
   9992	"HFI1_IB_CFG_LREV_ENB",
   9993	"HFI1_IB_CFG_LINKLATENCY",
   9994	"HFI1_IB_CFG_HRTBT",
   9995	"HFI1_IB_CFG_OP_VLS",
   9996	"HFI1_IB_CFG_VL_HIGH_CAP",
   9997	"HFI1_IB_CFG_VL_LOW_CAP",
   9998	"HFI1_IB_CFG_OVERRUN_THRESH",
   9999	"HFI1_IB_CFG_PHYERR_THRESH",
  10000	"HFI1_IB_CFG_LINKDEFAULT",
  10001	"HFI1_IB_CFG_PKEYS",
  10002	"HFI1_IB_CFG_MTU",
  10003	"HFI1_IB_CFG_LSTATE",
  10004	"HFI1_IB_CFG_VL_HIGH_LIMIT",
  10005	"HFI1_IB_CFG_PMA_TICKS",
  10006	"HFI1_IB_CFG_PORT"
  10007};
  10008
  10009static const char *ib_cfg_name(int which)
  10010{
  10011	if (which < 0 || which >= ARRAY_SIZE(ib_cfg_name_strings))
  10012		return "invalid";
  10013	return ib_cfg_name_strings[which];
  10014}
  10015
  10016int hfi1_get_ib_cfg(struct hfi1_pportdata *ppd, int which)
  10017{
  10018	struct hfi1_devdata *dd = ppd->dd;
  10019	int val = 0;
  10020
  10021	switch (which) {
  10022	case HFI1_IB_CFG_LWID_ENB: /* allowed Link-width */
  10023		val = ppd->link_width_enabled;
  10024		break;
  10025	case HFI1_IB_CFG_LWID: /* currently active Link-width */
  10026		val = ppd->link_width_active;
  10027		break;
  10028	case HFI1_IB_CFG_SPD_ENB: /* allowed Link speeds */
  10029		val = ppd->link_speed_enabled;
  10030		break;
  10031	case HFI1_IB_CFG_SPD: /* current Link speed */
  10032		val = ppd->link_speed_active;
  10033		break;
  10034
  10035	case HFI1_IB_CFG_RXPOL_ENB: /* Auto-RX-polarity enable */
  10036	case HFI1_IB_CFG_LREV_ENB: /* Auto-Lane-reversal enable */
  10037	case HFI1_IB_CFG_LINKLATENCY:
  10038		goto unimplemented;
  10039
  10040	case HFI1_IB_CFG_OP_VLS:
  10041		val = ppd->actual_vls_operational;
  10042		break;
  10043	case HFI1_IB_CFG_VL_HIGH_CAP: /* VL arb high priority table size */
  10044		val = VL_ARB_HIGH_PRIO_TABLE_SIZE;
  10045		break;
  10046	case HFI1_IB_CFG_VL_LOW_CAP: /* VL arb low priority table size */
  10047		val = VL_ARB_LOW_PRIO_TABLE_SIZE;
  10048		break;
  10049	case HFI1_IB_CFG_OVERRUN_THRESH: /* IB overrun threshold */
  10050		val = ppd->overrun_threshold;
  10051		break;
  10052	case HFI1_IB_CFG_PHYERR_THRESH: /* IB PHY error threshold */
  10053		val = ppd->phy_error_threshold;
  10054		break;
  10055	case HFI1_IB_CFG_LINKDEFAULT: /* IB link default (sleep/poll) */
  10056		val = HLS_DEFAULT;
  10057		break;
  10058
  10059	case HFI1_IB_CFG_HRTBT: /* Heartbeat off/enable/auto */
  10060	case HFI1_IB_CFG_PMA_TICKS:
  10061	default:
  10062unimplemented:
  10063		if (HFI1_CAP_IS_KSET(PRINT_UNIMPL))
  10064			dd_dev_info(
  10065				dd,
  10066				"%s: which %s: not implemented\n",
  10067				__func__,
  10068				ib_cfg_name(which));
  10069		break;
  10070	}
  10071
  10072	return val;
  10073}
  10074
  10075/*
  10076 * The largest MAD packet size.
  10077 */
  10078#define MAX_MAD_PACKET 2048
  10079
  10080/*
  10081 * Return the maximum header bytes that can go on the _wire_
  10082 * for this device. This count includes the ICRC which is
  10083 * not part of the packet held in memory but it is appended
  10084 * by the HW.
  10085 * This is dependent on the device's receive header entry size.
  10086 * HFI allows this to be set per-receive context, but the
  10087 * driver presently enforces a global value.
  10088 */
  10089u32 lrh_max_header_bytes(struct hfi1_devdata *dd)
  10090{
  10091	/*
  10092	 * The maximum non-payload (MTU) bytes in LRH.PktLen are
  10093	 * the Receive Header Entry Size minus the PBC (or RHF) size
  10094	 * plus one DW for the ICRC appended by HW.
  10095	 *
  10096	 * dd->rcd[0].rcvhdrqentsize is in DW.
  10097	 * We use rcd[0] as all context will have the same value. Also,
  10098	 * the first kernel context would have been allocated by now so
  10099	 * we are guaranteed a valid value.
  10100	 */
  10101	return (get_hdrqentsize(dd->rcd[0]) - 2/*PBC/RHF*/ + 1/*ICRC*/) << 2;
  10102}
  10103
  10104/*
  10105 * Set Send Length
  10106 * @ppd: per port data
  10107 *
  10108 * Set the MTU by limiting how many DWs may be sent.  The SendLenCheck*
  10109 * registers compare against LRH.PktLen, so use the max bytes included
  10110 * in the LRH.
  10111 *
  10112 * This routine changes all VL values except VL15, which it maintains at
  10113 * the same value.
  10114 */
  10115static void set_send_length(struct hfi1_pportdata *ppd)
  10116{
  10117	struct hfi1_devdata *dd = ppd->dd;
  10118	u32 max_hb = lrh_max_header_bytes(dd), dcmtu;
  10119	u32 maxvlmtu = dd->vld[15].mtu;
  10120	u64 len1 = 0, len2 = (((dd->vld[15].mtu + max_hb) >> 2)
  10121			      & SEND_LEN_CHECK1_LEN_VL15_MASK) <<
  10122		SEND_LEN_CHECK1_LEN_VL15_SHIFT;
  10123	int i, j;
  10124	u32 thres;
  10125
  10126	for (i = 0; i < ppd->vls_supported; i++) {
  10127		if (dd->vld[i].mtu > maxvlmtu)
  10128			maxvlmtu = dd->vld[i].mtu;
  10129		if (i <= 3)
  10130			len1 |= (((dd->vld[i].mtu + max_hb) >> 2)
  10131				 & SEND_LEN_CHECK0_LEN_VL0_MASK) <<
  10132				((i % 4) * SEND_LEN_CHECK0_LEN_VL1_SHIFT);
  10133		else
  10134			len2 |= (((dd->vld[i].mtu + max_hb) >> 2)
  10135				 & SEND_LEN_CHECK1_LEN_VL4_MASK) <<
  10136				((i % 4) * SEND_LEN_CHECK1_LEN_VL5_SHIFT);
  10137	}
  10138	write_csr(dd, SEND_LEN_CHECK0, len1);
  10139	write_csr(dd, SEND_LEN_CHECK1, len2);
  10140	/* adjust kernel credit return thresholds based on new MTUs */
  10141	/* all kernel receive contexts have the same hdrqentsize */
  10142	for (i = 0; i < ppd->vls_supported; i++) {
  10143		thres = min(sc_percent_to_threshold(dd->vld[i].sc, 50),
  10144			    sc_mtu_to_threshold(dd->vld[i].sc,
  10145						dd->vld[i].mtu,
  10146						get_hdrqentsize(dd->rcd[0])));
  10147		for (j = 0; j < INIT_SC_PER_VL; j++)
  10148			sc_set_cr_threshold(
  10149					pio_select_send_context_vl(dd, j, i),
  10150					    thres);
  10151	}
  10152	thres = min(sc_percent_to_threshold(dd->vld[15].sc, 50),
  10153		    sc_mtu_to_threshold(dd->vld[15].sc,
  10154					dd->vld[15].mtu,
  10155					dd->rcd[0]->rcvhdrqentsize));
  10156	sc_set_cr_threshold(dd->vld[15].sc, thres);
  10157
  10158	/* Adjust maximum MTU for the port in DC */
  10159	dcmtu = maxvlmtu == 10240 ? DCC_CFG_PORT_MTU_CAP_10240 :
  10160		(ilog2(maxvlmtu >> 8) + 1);
  10161	len1 = read_csr(ppd->dd, DCC_CFG_PORT_CONFIG);
  10162	len1 &= ~DCC_CFG_PORT_CONFIG_MTU_CAP_SMASK;
  10163	len1 |= ((u64)dcmtu & DCC_CFG_PORT_CONFIG_MTU_CAP_MASK) <<
  10164		DCC_CFG_PORT_CONFIG_MTU_CAP_SHIFT;
  10165	write_csr(ppd->dd, DCC_CFG_PORT_CONFIG, len1);
  10166}
  10167
  10168static void set_lidlmc(struct hfi1_pportdata *ppd)
  10169{
  10170	int i;
  10171	u64 sreg = 0;
  10172	struct hfi1_devdata *dd = ppd->dd;
  10173	u32 mask = ~((1U << ppd->lmc) - 1);
  10174	u64 c1 = read_csr(ppd->dd, DCC_CFG_PORT_CONFIG1);
  10175	u32 lid;
  10176
  10177	/*
  10178	 * Program 0 in CSR if port lid is extended. This prevents
  10179	 * 9B packets being sent out for large lids.
  10180	 */
  10181	lid = (ppd->lid >= be16_to_cpu(IB_MULTICAST_LID_BASE)) ? 0 : ppd->lid;
  10182	c1 &= ~(DCC_CFG_PORT_CONFIG1_TARGET_DLID_SMASK
  10183		| DCC_CFG_PORT_CONFIG1_DLID_MASK_SMASK);
  10184	c1 |= ((lid & DCC_CFG_PORT_CONFIG1_TARGET_DLID_MASK)
  10185			<< DCC_CFG_PORT_CONFIG1_TARGET_DLID_SHIFT) |
  10186	      ((mask & DCC_CFG_PORT_CONFIG1_DLID_MASK_MASK)
  10187			<< DCC_CFG_PORT_CONFIG1_DLID_MASK_SHIFT);
  10188	write_csr(ppd->dd, DCC_CFG_PORT_CONFIG1, c1);
  10189
  10190	/*
  10191	 * Iterate over all the send contexts and set their SLID check
  10192	 */
  10193	sreg = ((mask & SEND_CTXT_CHECK_SLID_MASK_MASK) <<
  10194			SEND_CTXT_CHECK_SLID_MASK_SHIFT) |
  10195	       (((lid & mask) & SEND_CTXT_CHECK_SLID_VALUE_MASK) <<
  10196			SEND_CTXT_CHECK_SLID_VALUE_SHIFT);
  10197
  10198	for (i = 0; i < chip_send_contexts(dd); i++) {
  10199		hfi1_cdbg(LINKVERB, "SendContext[%d].SLID_CHECK = 0x%x",
  10200			  i, (u32)sreg);
  10201		write_kctxt_csr(dd, i, SEND_CTXT_CHECK_SLID, sreg);
  10202	}
  10203
  10204	/* Now we have to do the same thing for the sdma engines */
  10205	sdma_update_lmc(dd, mask, lid);
  10206}
  10207
  10208static const char *state_completed_string(u32 completed)
  10209{
  10210	static const char * const state_completed[] = {
  10211		"EstablishComm",
  10212		"OptimizeEQ",
  10213		"VerifyCap"
  10214	};
  10215
  10216	if (completed < ARRAY_SIZE(state_completed))
  10217		return state_completed[completed];
  10218
  10219	return "unknown";
  10220}
  10221
  10222static const char all_lanes_dead_timeout_expired[] =
  10223	"All lanes were inactive – was the interconnect media removed?";
  10224static const char tx_out_of_policy[] =
  10225	"Passing lanes on local port do not meet the local link width policy";
  10226static const char no_state_complete[] =
  10227	"State timeout occurred before link partner completed the state";
  10228static const char * const state_complete_reasons[] = {
  10229	[0x00] = "Reason unknown",
  10230	[0x01] = "Link was halted by driver, refer to LinkDownReason",
  10231	[0x02] = "Link partner reported failure",
  10232	[0x10] = "Unable to achieve frame sync on any lane",
  10233	[0x11] =
  10234	  "Unable to find a common bit rate with the link partner",
  10235	[0x12] =
  10236	  "Unable to achieve frame sync on sufficient lanes to meet the local link width policy",
  10237	[0x13] =
  10238	  "Unable to identify preset equalization on sufficient lanes to meet the local link width policy",
  10239	[0x14] = no_state_complete,
  10240	[0x15] =
  10241	  "State timeout occurred before link partner identified equalization presets",
  10242	[0x16] =
  10243	  "Link partner completed the EstablishComm state, but the passing lanes do not meet the local link width policy",
  10244	[0x17] = tx_out_of_policy,
  10245	[0x20] = all_lanes_dead_timeout_expired,
  10246	[0x21] =
  10247	  "Unable to achieve acceptable BER on sufficient lanes to meet the local link width policy",
  10248	[0x22] = no_state_complete,
  10249	[0x23] =
  10250	  "Link partner completed the OptimizeEq state, but the passing lanes do not meet the local link width policy",
  10251	[0x24] = tx_out_of_policy,
  10252	[0x30] = all_lanes_dead_timeout_expired,
  10253	[0x31] =
  10254	  "State timeout occurred waiting for host to process received frames",
  10255	[0x32] = no_state_complete,
  10256	[0x33] =
  10257	  "Link partner completed the VerifyCap state, but the passing lanes do not meet the local link width policy",
  10258	[0x34] = tx_out_of_policy,
  10259	[0x35] = "Negotiated link width is mutually exclusive",
  10260	[0x36] =
  10261	  "Timed out before receiving verifycap frames in VerifyCap.Exchange",
  10262	[0x37] = "Unable to resolve secure data exchange",
  10263};
  10264
  10265static const char *state_complete_reason_code_string(struct hfi1_pportdata *ppd,
  10266						     u32 code)
  10267{
  10268	const char *str = NULL;
  10269
  10270	if (code < ARRAY_SIZE(state_complete_reasons))
  10271		str = state_complete_reasons[code];
  10272
  10273	if (str)
  10274		return str;
  10275	return "Reserved";
  10276}
  10277
  10278/* describe the given last state complete frame */
  10279static void decode_state_complete(struct hfi1_pportdata *ppd, u32 frame,
  10280				  const char *prefix)
  10281{
  10282	struct hfi1_devdata *dd = ppd->dd;
  10283	u32 success;
  10284	u32 state;
  10285	u32 reason;
  10286	u32 lanes;
  10287
  10288	/*
  10289	 * Decode frame:
  10290	 *  [ 0: 0] - success
  10291	 *  [ 3: 1] - state
  10292	 *  [ 7: 4] - next state timeout
  10293	 *  [15: 8] - reason code
  10294	 *  [31:16] - lanes
  10295	 */
  10296	success = frame & 0x1;
  10297	state = (frame >> 1) & 0x7;
  10298	reason = (frame >> 8) & 0xff;
  10299	lanes = (frame >> 16) & 0xffff;
  10300
  10301	dd_dev_err(dd, "Last %s LNI state complete frame 0x%08x:\n",
  10302		   prefix, frame);
  10303	dd_dev_err(dd, "    last reported state state: %s (0x%x)\n",
  10304		   state_completed_string(state), state);
  10305	dd_dev_err(dd, "    state successfully completed: %s\n",
  10306		   success ? "yes" : "no");
  10307	dd_dev_err(dd, "    fail reason 0x%x: %s\n",
  10308		   reason, state_complete_reason_code_string(ppd, reason));
  10309	dd_dev_err(dd, "    passing lane mask: 0x%x", lanes);
  10310}
  10311
  10312/*
  10313 * Read the last state complete frames and explain them.  This routine
  10314 * expects to be called if the link went down during link negotiation
  10315 * and initialization (LNI).  That is, anywhere between polling and link up.
  10316 */
  10317static void check_lni_states(struct hfi1_pportdata *ppd)
  10318{
  10319	u32 last_local_state;
  10320	u32 last_remote_state;
  10321
  10322	read_last_local_state(ppd->dd, &last_local_state);
  10323	read_last_remote_state(ppd->dd, &last_remote_state);
  10324
  10325	/*
  10326	 * Don't report anything if there is nothing to report.  A value of
  10327	 * 0 means the link was taken down while polling and there was no
  10328	 * training in-process.
  10329	 */
  10330	if (last_local_state == 0 && last_remote_state == 0)
  10331		return;
  10332
  10333	decode_state_complete(ppd, last_local_state, "transmitted");
  10334	decode_state_complete(ppd, last_remote_state, "received");
  10335}
  10336
  10337/* wait for wait_ms for LINK_TRANSFER_ACTIVE to go to 1 */
  10338static int wait_link_transfer_active(struct hfi1_devdata *dd, int wait_ms)
  10339{
  10340	u64 reg;
  10341	unsigned long timeout;
  10342
  10343	/* watch LCB_STS_LINK_TRANSFER_ACTIVE */
  10344	timeout = jiffies + msecs_to_jiffies(wait_ms);
  10345	while (1) {
  10346		reg = read_csr(dd, DC_LCB_STS_LINK_TRANSFER_ACTIVE);
  10347		if (reg)
  10348			break;
  10349		if (time_after(jiffies, timeout)) {
  10350			dd_dev_err(dd,
  10351				   "timeout waiting for LINK_TRANSFER_ACTIVE\n");
  10352			return -ETIMEDOUT;
  10353		}
  10354		udelay(2);
  10355	}
  10356	return 0;
  10357}
  10358
  10359/* called when the logical link state is not down as it should be */
  10360static void force_logical_link_state_down(struct hfi1_pportdata *ppd)
  10361{
  10362	struct hfi1_devdata *dd = ppd->dd;
  10363
  10364	/*
  10365	 * Bring link up in LCB loopback
  10366	 */
  10367	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET, 1);
  10368	write_csr(dd, DC_LCB_CFG_IGNORE_LOST_RCLK,
  10369		  DC_LCB_CFG_IGNORE_LOST_RCLK_EN_SMASK);
  10370
  10371	write_csr(dd, DC_LCB_CFG_LANE_WIDTH, 0);
  10372	write_csr(dd, DC_LCB_CFG_REINIT_AS_SLAVE, 0);
  10373	write_csr(dd, DC_LCB_CFG_CNT_FOR_SKIP_STALL, 0x110);
  10374	write_csr(dd, DC_LCB_CFG_LOOPBACK, 0x2);
  10375
  10376	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET, 0);
  10377	(void)read_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET);
  10378	udelay(3);
  10379	write_csr(dd, DC_LCB_CFG_ALLOW_LINK_UP, 1);
  10380	write_csr(dd, DC_LCB_CFG_RUN, 1ull << DC_LCB_CFG_RUN_EN_SHIFT);
  10381
  10382	wait_link_transfer_active(dd, 100);
  10383
  10384	/*
  10385	 * Bring the link down again.
  10386	 */
  10387	write_csr(dd, DC_LCB_CFG_TX_FIFOS_RESET, 1);
  10388	write_csr(dd, DC_LCB_CFG_ALLOW_LINK_UP, 0);
  10389	write_csr(dd, DC_LCB_CFG_IGNORE_LOST_RCLK, 0);
  10390
  10391	dd_dev_info(ppd->dd, "logical state forced to LINK_DOWN\n");
  10392}
  10393
  10394/*
  10395 * Helper for set_link_state().  Do not call except from that routine.
  10396 * Expects ppd->hls_mutex to be held.
  10397 *
  10398 * @rem_reason value to be sent to the neighbor
  10399 *
  10400 * LinkDownReasons only set if transition succeeds.
  10401 */
  10402static int goto_offline(struct hfi1_pportdata *ppd, u8 rem_reason)
  10403{
  10404	struct hfi1_devdata *dd = ppd->dd;
  10405	u32 previous_state;
  10406	int offline_state_ret;
  10407	int ret;
  10408
  10409	update_lcb_cache(dd);
  10410
  10411	previous_state = ppd->host_link_state;
  10412	ppd->host_link_state = HLS_GOING_OFFLINE;
  10413
  10414	/* start offline transition */
  10415	ret = set_physical_link_state(dd, (rem_reason << 8) | PLS_OFFLINE);
  10416
  10417	if (ret != HCMD_SUCCESS) {
  10418		dd_dev_err(dd,
  10419			   "Failed to transition to Offline link state, return %d\n",
  10420			   ret);
  10421		return -EINVAL;
  10422	}
  10423	if (ppd->offline_disabled_reason ==
  10424			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE))
  10425		ppd->offline_disabled_reason =
  10426		HFI1_ODR_MASK(OPA_LINKDOWN_REASON_TRANSIENT);
  10427
  10428	offline_state_ret = wait_phys_link_offline_substates(ppd, 10000);
  10429	if (offline_state_ret < 0)
  10430		return offline_state_ret;
  10431
  10432	/* Disabling AOC transmitters */
  10433	if (ppd->port_type == PORT_TYPE_QSFP &&
  10434	    ppd->qsfp_info.limiting_active &&
  10435	    qsfp_mod_present(ppd)) {
  10436		int ret;
  10437
  10438		ret = acquire_chip_resource(dd, qsfp_resource(dd), QSFP_WAIT);
  10439		if (ret == 0) {
  10440			set_qsfp_tx(ppd, 0);
  10441			release_chip_resource(dd, qsfp_resource(dd));
  10442		} else {
  10443			/* not fatal, but should warn */
  10444			dd_dev_err(dd,
  10445				   "Unable to acquire lock to turn off QSFP TX\n");
  10446		}
  10447	}
  10448
  10449	/*
  10450	 * Wait for the offline.Quiet transition if it hasn't happened yet. It
  10451	 * can take a while for the link to go down.
  10452	 */
  10453	if (offline_state_ret != PLS_OFFLINE_QUIET) {
  10454		ret = wait_physical_linkstate(ppd, PLS_OFFLINE, 30000);
  10455		if (ret < 0)
  10456			return ret;
  10457	}
  10458
  10459	/*
  10460	 * Now in charge of LCB - must be after the physical state is
  10461	 * offline.quiet and before host_link_state is changed.
  10462	 */
  10463	set_host_lcb_access(dd);
  10464	write_csr(dd, DC_LCB_ERR_EN, ~0ull); /* watch LCB errors */
  10465
  10466	/* make sure the logical state is also down */
  10467	ret = wait_logical_linkstate(ppd, IB_PORT_DOWN, 1000);
  10468	if (ret)
  10469		force_logical_link_state_down(ppd);
  10470
  10471	ppd->host_link_state = HLS_LINK_COOLDOWN; /* LCB access allowed */
  10472	update_statusp(ppd, IB_PORT_DOWN);
  10473
  10474	/*
  10475	 * The LNI has a mandatory wait time after the physical state
  10476	 * moves to Offline.Quiet.  The wait time may be different
  10477	 * depending on how the link went down.  The 8051 firmware
  10478	 * will observe the needed wait time and only move to ready
  10479	 * when that is completed.  The largest of the quiet timeouts
  10480	 * is 6s, so wait that long and then at least 0.5s more for
  10481	 * other transitions, and another 0.5s for a buffer.
  10482	 */
  10483	ret = wait_fm_ready(dd, 7000);
  10484	if (ret) {
  10485		dd_dev_err(dd,
  10486			   "After going offline, timed out waiting for the 8051 to become ready to accept host requests\n");
  10487		/* state is really offline, so make it so */
  10488		ppd->host_link_state = HLS_DN_OFFLINE;
  10489		return ret;
  10490	}
  10491
  10492	/*
  10493	 * The state is now offline and the 8051 is ready to accept host
  10494	 * requests.
  10495	 *	- change our state
  10496	 *	- notify others if we were previously in a linkup state
  10497	 */
  10498	ppd->host_link_state = HLS_DN_OFFLINE;
  10499	if (previous_state & HLS_UP) {
  10500		/* went down while link was up */
  10501		handle_linkup_change(dd, 0);
  10502	} else if (previous_state
  10503			& (HLS_DN_POLL | HLS_VERIFY_CAP | HLS_GOING_UP)) {
  10504		/* went down while attempting link up */
  10505		check_lni_states(ppd);
  10506
  10507		/* The QSFP doesn't need to be reset on LNI failure */
  10508		ppd->qsfp_info.reset_needed = 0;
  10509	}
  10510
  10511	/* the active link width (downgrade) is 0 on link down */
  10512	ppd->link_width_active = 0;
  10513	ppd->link_width_downgrade_tx_active = 0;
  10514	ppd->link_width_downgrade_rx_active = 0;
  10515	ppd->current_egress_rate = 0;
  10516	return 0;
  10517}
  10518
  10519/* return the link state name */
  10520static const char *link_state_name(u32 state)
  10521{
  10522	const char *name;
  10523	int n = ilog2(state);
  10524	static const char * const names[] = {
  10525		[__HLS_UP_INIT_BP]	 = "INIT",
  10526		[__HLS_UP_ARMED_BP]	 = "ARMED",
  10527		[__HLS_UP_ACTIVE_BP]	 = "ACTIVE",
  10528		[__HLS_DN_DOWNDEF_BP]	 = "DOWNDEF",
  10529		[__HLS_DN_POLL_BP]	 = "POLL",
  10530		[__HLS_DN_DISABLE_BP]	 = "DISABLE",
  10531		[__HLS_DN_OFFLINE_BP]	 = "OFFLINE",
  10532		[__HLS_VERIFY_CAP_BP]	 = "VERIFY_CAP",
  10533		[__HLS_GOING_UP_BP]	 = "GOING_UP",
  10534		[__HLS_GOING_OFFLINE_BP] = "GOING_OFFLINE",
  10535		[__HLS_LINK_COOLDOWN_BP] = "LINK_COOLDOWN"
  10536	};
  10537
  10538	name = n < ARRAY_SIZE(names) ? names[n] : NULL;
  10539	return name ? name : "unknown";
  10540}
  10541
  10542/* return the link state reason name */
  10543static const char *link_state_reason_name(struct hfi1_pportdata *ppd, u32 state)
  10544{
  10545	if (state == HLS_UP_INIT) {
  10546		switch (ppd->linkinit_reason) {
  10547		case OPA_LINKINIT_REASON_LINKUP:
  10548			return "(LINKUP)";
  10549		case OPA_LINKINIT_REASON_FLAPPING:
  10550			return "(FLAPPING)";
  10551		case OPA_LINKINIT_OUTSIDE_POLICY:
  10552			return "(OUTSIDE_POLICY)";
  10553		case OPA_LINKINIT_QUARANTINED:
  10554			return "(QUARANTINED)";
  10555		case OPA_LINKINIT_INSUFIC_CAPABILITY:
  10556			return "(INSUFIC_CAPABILITY)";
  10557		default:
  10558			break;
  10559		}
  10560	}
  10561	return "";
  10562}
  10563
  10564/*
  10565 * driver_pstate - convert the driver's notion of a port's
  10566 * state (an HLS_*) into a physical state (a {IB,OPA}_PORTPHYSSTATE_*).
  10567 * Return -1 (converted to a u32) to indicate error.
  10568 */
  10569u32 driver_pstate(struct hfi1_pportdata *ppd)
  10570{
  10571	switch (ppd->host_link_state) {
  10572	case HLS_UP_INIT:
  10573	case HLS_UP_ARMED:
  10574	case HLS_UP_ACTIVE:
  10575		return IB_PORTPHYSSTATE_LINKUP;
  10576	case HLS_DN_POLL:
  10577		return IB_PORTPHYSSTATE_POLLING;
  10578	case HLS_DN_DISABLE:
  10579		return IB_PORTPHYSSTATE_DISABLED;
  10580	case HLS_DN_OFFLINE:
  10581		return OPA_PORTPHYSSTATE_OFFLINE;
  10582	case HLS_VERIFY_CAP:
  10583		return IB_PORTPHYSSTATE_TRAINING;
  10584	case HLS_GOING_UP:
  10585		return IB_PORTPHYSSTATE_TRAINING;
  10586	case HLS_GOING_OFFLINE:
  10587		return OPA_PORTPHYSSTATE_OFFLINE;
  10588	case HLS_LINK_COOLDOWN:
  10589		return OPA_PORTPHYSSTATE_OFFLINE;
  10590	case HLS_DN_DOWNDEF:
  10591	default:
  10592		dd_dev_err(ppd->dd, "invalid host_link_state 0x%x\n",
  10593			   ppd->host_link_state);
  10594		return  -1;
  10595	}
  10596}
  10597
  10598/*
  10599 * driver_lstate - convert the driver's notion of a port's
  10600 * state (an HLS_*) into a logical state (a IB_PORT_*). Return -1
  10601 * (converted to a u32) to indicate error.
  10602 */
  10603u32 driver_lstate(struct hfi1_pportdata *ppd)
  10604{
  10605	if (ppd->host_link_state && (ppd->host_link_state & HLS_DOWN))
  10606		return IB_PORT_DOWN;
  10607
  10608	switch (ppd->host_link_state & HLS_UP) {
  10609	case HLS_UP_INIT:
  10610		return IB_PORT_INIT;
  10611	case HLS_UP_ARMED:
  10612		return IB_PORT_ARMED;
  10613	case HLS_UP_ACTIVE:
  10614		return IB_PORT_ACTIVE;
  10615	default:
  10616		dd_dev_err(ppd->dd, "invalid host_link_state 0x%x\n",
  10617			   ppd->host_link_state);
  10618	return -1;
  10619	}
  10620}
  10621
  10622void set_link_down_reason(struct hfi1_pportdata *ppd, u8 lcl_reason,
  10623			  u8 neigh_reason, u8 rem_reason)
  10624{
  10625	if (ppd->local_link_down_reason.latest == 0 &&
  10626	    ppd->neigh_link_down_reason.latest == 0) {
  10627		ppd->local_link_down_reason.latest = lcl_reason;
  10628		ppd->neigh_link_down_reason.latest = neigh_reason;
  10629		ppd->remote_link_down_reason = rem_reason;
  10630	}
  10631}
  10632
  10633/**
  10634 * data_vls_operational() - Verify if data VL BCT credits and MTU
  10635 *			    are both set.
  10636 * @ppd: pointer to hfi1_pportdata structure
  10637 *
  10638 * Return: true - Ok, false -otherwise.
  10639 */
  10640static inline bool data_vls_operational(struct hfi1_pportdata *ppd)
  10641{
  10642	int i;
  10643	u64 reg;
  10644
  10645	if (!ppd->actual_vls_operational)
  10646		return false;
  10647
  10648	for (i = 0; i < ppd->vls_supported; i++) {
  10649		reg = read_csr(ppd->dd, SEND_CM_CREDIT_VL + (8 * i));
  10650		if ((reg && !ppd->dd->vld[i].mtu) ||
  10651		    (!reg && ppd->dd->vld[i].mtu))
  10652			return false;
  10653	}
  10654
  10655	return true;
  10656}
  10657
  10658/*
  10659 * Change the physical and/or logical link state.
  10660 *
  10661 * Do not call this routine while inside an interrupt.  It contains
  10662 * calls to routines that can take multiple seconds to finish.
  10663 *
  10664 * Returns 0 on success, -errno on failure.
  10665 */
  10666int set_link_state(struct hfi1_pportdata *ppd, u32 state)
  10667{
  10668	struct hfi1_devdata *dd = ppd->dd;
  10669	struct ib_event event = {.device = NULL};
  10670	int ret1, ret = 0;
  10671	int orig_new_state, poll_bounce;
  10672
  10673	mutex_lock(&ppd->hls_lock);
  10674
  10675	orig_new_state = state;
  10676	if (state == HLS_DN_DOWNDEF)
  10677		state = HLS_DEFAULT;
  10678
  10679	/* interpret poll -> poll as a link bounce */
  10680	poll_bounce = ppd->host_link_state == HLS_DN_POLL &&
  10681		      state == HLS_DN_POLL;
  10682
  10683	dd_dev_info(dd, "%s: current %s, new %s %s%s\n", __func__,
  10684		    link_state_name(ppd->host_link_state),
  10685		    link_state_name(orig_new_state),
  10686		    poll_bounce ? "(bounce) " : "",
  10687		    link_state_reason_name(ppd, state));
  10688
  10689	/*
  10690	 * If we're going to a (HLS_*) link state that implies the logical
  10691	 * link state is neither of (IB_PORT_ARMED, IB_PORT_ACTIVE), then
  10692	 * reset is_sm_config_started to 0.
  10693	 */
  10694	if (!(state & (HLS_UP_ARMED | HLS_UP_ACTIVE)))
  10695		ppd->is_sm_config_started = 0;
  10696
  10697	/*
  10698	 * Do nothing if the states match.  Let a poll to poll link bounce
  10699	 * go through.
  10700	 */
  10701	if (ppd->host_link_state == state && !poll_bounce)
  10702		goto done;
  10703
  10704	switch (state) {
  10705	case HLS_UP_INIT:
  10706		if (ppd->host_link_state == HLS_DN_POLL &&
  10707		    (quick_linkup || dd->icode == ICODE_FUNCTIONAL_SIMULATOR)) {
  10708			/*
  10709			 * Quick link up jumps from polling to here.
  10710			 *
  10711			 * Whether in normal or loopback mode, the
  10712			 * simulator jumps from polling to link up.
  10713			 * Accept that here.
  10714			 */
  10715			/* OK */
  10716		} else if (ppd->host_link_state != HLS_GOING_UP) {
  10717			goto unexpected;
  10718		}
  10719
  10720		/*
  10721		 * Wait for Link_Up physical state.
  10722		 * Physical and Logical states should already be
  10723		 * be transitioned to LinkUp and LinkInit respectively.
  10724		 */
  10725		ret = wait_physical_linkstate(ppd, PLS_LINKUP, 1000);
  10726		if (ret) {
  10727			dd_dev_err(dd,
  10728				   "%s: physical state did not change to LINK-UP\n",
  10729				   __func__);
  10730			break;
  10731		}
  10732
  10733		ret = wait_logical_linkstate(ppd, IB_PORT_INIT, 1000);
  10734		if (ret) {
  10735			dd_dev_err(dd,
  10736				   "%s: logical state did not change to INIT\n",
  10737				   __func__);
  10738			break;
  10739		}
  10740
  10741		/* clear old transient LINKINIT_REASON code */
  10742		if (ppd->linkinit_reason >= OPA_LINKINIT_REASON_CLEAR)
  10743			ppd->linkinit_reason =
  10744				OPA_LINKINIT_REASON_LINKUP;
  10745
  10746		/* enable the port */
  10747		add_rcvctrl(dd, RCV_CTRL_RCV_PORT_ENABLE_SMASK);
  10748
  10749		handle_linkup_change(dd, 1);
  10750		pio_kernel_linkup(dd);
  10751
  10752		/*
  10753		 * After link up, a new link width will have been set.
  10754		 * Update the xmit counters with regards to the new
  10755		 * link width.
  10756		 */
  10757		update_xmit_counters(ppd, ppd->link_width_active);
  10758
  10759		ppd->host_link_state = HLS_UP_INIT;
  10760		update_statusp(ppd, IB_PORT_INIT);
  10761		break;
  10762	case HLS_UP_ARMED:
  10763		if (ppd->host_link_state != HLS_UP_INIT)
  10764			goto unexpected;
  10765
  10766		if (!data_vls_operational(ppd)) {
  10767			dd_dev_err(dd,
  10768				   "%s: Invalid data VL credits or mtu\n",
  10769				   __func__);
  10770			ret = -EINVAL;
  10771			break;
  10772		}
  10773
  10774		set_logical_state(dd, LSTATE_ARMED);
  10775		ret = wait_logical_linkstate(ppd, IB_PORT_ARMED, 1000);
  10776		if (ret) {
  10777			dd_dev_err(dd,
  10778				   "%s: logical state did not change to ARMED\n",
  10779				   __func__);
  10780			break;
  10781		}
  10782		ppd->host_link_state = HLS_UP_ARMED;
  10783		update_statusp(ppd, IB_PORT_ARMED);
  10784		/*
  10785		 * The simulator does not currently implement SMA messages,
  10786		 * so neighbor_normal is not set.  Set it here when we first
  10787		 * move to Armed.
  10788		 */
  10789		if (dd->icode == ICODE_FUNCTIONAL_SIMULATOR)
  10790			ppd->neighbor_normal = 1;
  10791		break;
  10792	case HLS_UP_ACTIVE:
  10793		if (ppd->host_link_state != HLS_UP_ARMED)
  10794			goto unexpected;
  10795
  10796		set_logical_state(dd, LSTATE_ACTIVE);
  10797		ret = wait_logical_linkstate(ppd, IB_PORT_ACTIVE, 1000);
  10798		if (ret) {
  10799			dd_dev_err(dd,
  10800				   "%s: logical state did not change to ACTIVE\n",
  10801				   __func__);
  10802		} else {
  10803			/* tell all engines to go running */
  10804			sdma_all_running(dd);
  10805			ppd->host_link_state = HLS_UP_ACTIVE;
  10806			update_statusp(ppd, IB_PORT_ACTIVE);
  10807
  10808			/* Signal the IB layer that the port has went active */
  10809			event.device = &dd->verbs_dev.rdi.ibdev;
  10810			event.element.port_num = ppd->port;
  10811			event.event = IB_EVENT_PORT_ACTIVE;
  10812		}
  10813		break;
  10814	case HLS_DN_POLL:
  10815		if ((ppd->host_link_state == HLS_DN_DISABLE ||
  10816		     ppd->host_link_state == HLS_DN_OFFLINE) &&
  10817		    dd->dc_shutdown)
  10818			dc_start(dd);
  10819		/* Hand LED control to the DC */
  10820		write_csr(dd, DCC_CFG_LED_CNTRL, 0);
  10821
  10822		if (ppd->host_link_state != HLS_DN_OFFLINE) {
  10823			u8 tmp = ppd->link_enabled;
  10824
  10825			ret = goto_offline(ppd, ppd->remote_link_down_reason);
  10826			if (ret) {
  10827				ppd->link_enabled = tmp;
  10828				break;
  10829			}
  10830			ppd->remote_link_down_reason = 0;
  10831
  10832			if (ppd->driver_link_ready)
  10833				ppd->link_enabled = 1;
  10834		}
  10835
  10836		set_all_slowpath(ppd->dd);
  10837		ret = set_local_link_attributes(ppd);
  10838		if (ret)
  10839			break;
  10840
  10841		ppd->port_error_action = 0;
  10842
  10843		if (quick_linkup) {
  10844			/* quick linkup does not go into polling */
  10845			ret = do_quick_linkup(dd);
  10846		} else {
  10847			ret1 = set_physical_link_state(dd, PLS_POLLING);
  10848			if (!ret1)
  10849				ret1 = wait_phys_link_out_of_offline(ppd,
  10850								     3000);
  10851			if (ret1 != HCMD_SUCCESS) {
  10852				dd_dev_err(dd,
  10853					   "Failed to transition to Polling link state, return 0x%x\n",
  10854					   ret1);
  10855				ret = -EINVAL;
  10856			}
  10857		}
  10858
  10859		/*
  10860		 * Change the host link state after requesting DC8051 to
  10861		 * change its physical state so that we can ignore any
  10862		 * interrupt with stale LNI(XX) error, which will not be
  10863		 * cleared until DC8051 transitions to Polling state.
  10864		 */
  10865		ppd->host_link_state = HLS_DN_POLL;
  10866		ppd->offline_disabled_reason =
  10867			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE);
  10868		/*
  10869		 * If an error occurred above, go back to offline.  The
  10870		 * caller may reschedule another attempt.
  10871		 */
  10872		if (ret)
  10873			goto_offline(ppd, 0);
  10874		else
  10875			log_physical_state(ppd, PLS_POLLING);
  10876		break;
  10877	case HLS_DN_DISABLE:
  10878		/* link is disabled */
  10879		ppd->link_enabled = 0;
  10880
  10881		/* allow any state to transition to disabled */
  10882
  10883		/* must transition to offline first */
  10884		if (ppd->host_link_state != HLS_DN_OFFLINE) {
  10885			ret = goto_offline(ppd, ppd->remote_link_down_reason);
  10886			if (ret)
  10887				break;
  10888			ppd->remote_link_down_reason = 0;
  10889		}
  10890
  10891		if (!dd->dc_shutdown) {
  10892			ret1 = set_physical_link_state(dd, PLS_DISABLED);
  10893			if (ret1 != HCMD_SUCCESS) {
  10894				dd_dev_err(dd,
  10895					   "Failed to transition to Disabled link state, return 0x%x\n",
  10896					   ret1);
  10897				ret = -EINVAL;
  10898				break;
  10899			}
  10900			ret = wait_physical_linkstate(ppd, PLS_DISABLED, 10000);
  10901			if (ret) {
  10902				dd_dev_err(dd,
  10903					   "%s: physical state did not change to DISABLED\n",
  10904					   __func__);
  10905				break;
  10906			}
  10907			dc_shutdown(dd);
  10908		}
  10909		ppd->host_link_state = HLS_DN_DISABLE;
  10910		break;
  10911	case HLS_DN_OFFLINE:
  10912		if (ppd->host_link_state == HLS_DN_DISABLE)
  10913			dc_start(dd);
  10914
  10915		/* allow any state to transition to offline */
  10916		ret = goto_offline(ppd, ppd->remote_link_down_reason);
  10917		if (!ret)
  10918			ppd->remote_link_down_reason = 0;
  10919		break;
  10920	case HLS_VERIFY_CAP:
  10921		if (ppd->host_link_state != HLS_DN_POLL)
  10922			goto unexpected;
  10923		ppd->host_link_state = HLS_VERIFY_CAP;
  10924		log_physical_state(ppd, PLS_CONFIGPHY_VERIFYCAP);
  10925		break;
  10926	case HLS_GOING_UP:
  10927		if (ppd->host_link_state != HLS_VERIFY_CAP)
  10928			goto unexpected;
  10929
  10930		ret1 = set_physical_link_state(dd, PLS_LINKUP);
  10931		if (ret1 != HCMD_SUCCESS) {
  10932			dd_dev_err(dd,
  10933				   "Failed to transition to link up state, return 0x%x\n",
  10934				   ret1);
  10935			ret = -EINVAL;
  10936			break;
  10937		}
  10938		ppd->host_link_state = HLS_GOING_UP;
  10939		break;
  10940
  10941	case HLS_GOING_OFFLINE:		/* transient within goto_offline() */
  10942	case HLS_LINK_COOLDOWN:		/* transient within goto_offline() */
  10943	default:
  10944		dd_dev_info(dd, "%s: state 0x%x: not supported\n",
  10945			    __func__, state);
  10946		ret = -EINVAL;
  10947		break;
  10948	}
  10949
  10950	goto done;
  10951
  10952unexpected:
  10953	dd_dev_err(dd, "%s: unexpected state transition from %s to %s\n",
  10954		   __func__, link_state_name(ppd->host_link_state),
  10955		   link_state_name(state));
  10956	ret = -EINVAL;
  10957
  10958done:
  10959	mutex_unlock(&ppd->hls_lock);
  10960
  10961	if (event.device)
  10962		ib_dispatch_event(&event);
  10963
  10964	return ret;
  10965}
  10966
  10967int hfi1_set_ib_cfg(struct hfi1_pportdata *ppd, int which, u32 val)
  10968{
  10969	u64 reg;
  10970	int ret = 0;
  10971
  10972	switch (which) {
  10973	case HFI1_IB_CFG_LIDLMC:
  10974		set_lidlmc(ppd);
  10975		break;
  10976	case HFI1_IB_CFG_VL_HIGH_LIMIT:
  10977		/*
  10978		 * The VL Arbitrator high limit is sent in units of 4k
  10979		 * bytes, while HFI stores it in units of 64 bytes.
  10980		 */
  10981		val *= 4096 / 64;
  10982		reg = ((u64)val & SEND_HIGH_PRIORITY_LIMIT_LIMIT_MASK)
  10983			<< SEND_HIGH_PRIORITY_LIMIT_LIMIT_SHIFT;
  10984		write_csr(ppd->dd, SEND_HIGH_PRIORITY_LIMIT, reg);
  10985		break;
  10986	case HFI1_IB_CFG_LINKDEFAULT: /* IB link default (sleep/poll) */
  10987		/* HFI only supports POLL as the default link down state */
  10988		if (val != HLS_DN_POLL)
  10989			ret = -EINVAL;
  10990		break;
  10991	case HFI1_IB_CFG_OP_VLS:
  10992		if (ppd->vls_operational != val) {
  10993			ppd->vls_operational = val;
  10994			if (!ppd->port)
  10995				ret = -EINVAL;
  10996		}
  10997		break;
  10998	/*
  10999	 * For link width, link width downgrade, and speed enable, always AND
  11000	 * the setting with what is actually supported.  This has two benefits.
  11001	 * First, enabled can't have unsupported values, no matter what the
  11002	 * SM or FM might want.  Second, the ALL_SUPPORTED wildcards that mean
  11003	 * "fill in with your supported value" have all the bits in the
  11004	 * field set, so simply ANDing with supported has the desired result.
  11005	 */
  11006	case HFI1_IB_CFG_LWID_ENB: /* set allowed Link-width */
  11007		ppd->link_width_enabled = val & ppd->link_width_supported;
  11008		break;
  11009	case HFI1_IB_CFG_LWID_DG_ENB: /* set allowed link width downgrade */
  11010		ppd->link_width_downgrade_enabled =
  11011				val & ppd->link_width_downgrade_supported;
  11012		break;
  11013	case HFI1_IB_CFG_SPD_ENB: /* allowed Link speeds */
  11014		ppd->link_speed_enabled = val & ppd->link_speed_supported;
  11015		break;
  11016	case HFI1_IB_CFG_OVERRUN_THRESH: /* IB overrun threshold */
  11017		/*
  11018		 * HFI does not follow IB specs, save this value
  11019		 * so we can report it, if asked.
  11020		 */
  11021		ppd->overrun_threshold = val;
  11022		break;
  11023	case HFI1_IB_CFG_PHYERR_THRESH: /* IB PHY error threshold */
  11024		/*
  11025		 * HFI does not follow IB specs, save this value
  11026		 * so we can report it, if asked.
  11027		 */
  11028		ppd->phy_error_threshold = val;
  11029		break;
  11030
  11031	case HFI1_IB_CFG_MTU:
  11032		set_send_length(ppd);
  11033		break;
  11034
  11035	case HFI1_IB_CFG_PKEYS:
  11036		if (HFI1_CAP_IS_KSET(PKEY_CHECK))
  11037			set_partition_keys(ppd);
  11038		break;
  11039
  11040	default:
  11041		if (HFI1_CAP_IS_KSET(PRINT_UNIMPL))
  11042			dd_dev_info(ppd->dd,
  11043				    "%s: which %s, val 0x%x: not implemented\n",
  11044				    __func__, ib_cfg_name(which), val);
  11045		break;
  11046	}
  11047	return ret;
  11048}
  11049
  11050/* begin functions related to vl arbitration table caching */
  11051static void init_vl_arb_caches(struct hfi1_pportdata *ppd)
  11052{
  11053	int i;
  11054
  11055	BUILD_BUG_ON(VL_ARB_TABLE_SIZE !=
  11056			VL_ARB_LOW_PRIO_TABLE_SIZE);
  11057	BUILD_BUG_ON(VL_ARB_TABLE_SIZE !=
  11058			VL_ARB_HIGH_PRIO_TABLE_SIZE);
  11059
  11060	/*
  11061	 * Note that we always return values directly from the
  11062	 * 'vl_arb_cache' (and do no CSR reads) in response to a
  11063	 * 'Get(VLArbTable)'. This is obviously correct after a
  11064	 * 'Set(VLArbTable)', since the cache will then be up to
  11065	 * date. But it's also correct prior to any 'Set(VLArbTable)'
  11066	 * since then both the cache, and the relevant h/w registers
  11067	 * will be zeroed.
  11068	 */
  11069
  11070	for (i = 0; i < MAX_PRIO_TABLE; i++)
  11071		spin_lock_init(&ppd->vl_arb_cache[i].lock);
  11072}
  11073
  11074/*
  11075 * vl_arb_lock_cache
  11076 *
  11077 * All other vl_arb_* functions should be called only after locking
  11078 * the cache.
  11079 */
  11080static inline struct vl_arb_cache *
  11081vl_arb_lock_cache(struct hfi1_pportdata *ppd, int idx)
  11082{
  11083	if (idx != LO_PRIO_TABLE && idx != HI_PRIO_TABLE)
  11084		return NULL;
  11085	spin_lock(&ppd->vl_arb_cache[idx].lock);
  11086	return &ppd->vl_arb_cache[idx];
  11087}
  11088
  11089static inline void vl_arb_unlock_cache(struct hfi1_pportdata *ppd, int idx)
  11090{
  11091	spin_unlock(&ppd->vl_arb_cache[idx].lock);
  11092}
  11093
  11094static void vl_arb_get_cache(struct vl_arb_cache *cache,
  11095			     struct ib_vl_weight_elem *vl)
  11096{
  11097	memcpy(vl, cache->table, VL_ARB_TABLE_SIZE * sizeof(*vl));
  11098}
  11099
  11100static void vl_arb_set_cache(struct vl_arb_cache *cache,
  11101			     struct ib_vl_weight_elem *vl)
  11102{
  11103	memcpy(cache->table, vl, VL_ARB_TABLE_SIZE * sizeof(*vl));
  11104}
  11105
  11106static int vl_arb_match_cache(struct vl_arb_cache *cache,
  11107			      struct ib_vl_weight_elem *vl)
  11108{
  11109	return !memcmp(cache->table, vl, VL_ARB_TABLE_SIZE * sizeof(*vl));
  11110}
  11111
  11112/* end functions related to vl arbitration table caching */
  11113
  11114static int set_vl_weights(struct hfi1_pportdata *ppd, u32 target,
  11115			  u32 size, struct ib_vl_weight_elem *vl)
  11116{
  11117	struct hfi1_devdata *dd = ppd->dd;
  11118	u64 reg;
  11119	unsigned int i, is_up = 0;
  11120	int drain, ret = 0;
  11121
  11122	mutex_lock(&ppd->hls_lock);
  11123
  11124	if (ppd->host_link_state & HLS_UP)
  11125		is_up = 1;
  11126
  11127	drain = !is_ax(dd) && is_up;
  11128
  11129	if (drain)
  11130		/*
  11131		 * Before adjusting VL arbitration weights, empty per-VL
  11132		 * FIFOs, otherwise a packet whose VL weight is being
  11133		 * set to 0 could get stuck in a FIFO with no chance to
  11134		 * egress.
  11135		 */
  11136		ret = stop_drain_data_vls(dd);
  11137
  11138	if (ret) {
  11139		dd_dev_err(
  11140			dd,
  11141			"%s: cannot stop/drain VLs - refusing to change VL arbitration weights\n",
  11142			__func__);
  11143		goto err;
  11144	}
  11145
  11146	for (i = 0; i < size; i++, vl++) {
  11147		/*
  11148		 * NOTE: The low priority shift and mask are used here, but
  11149		 * they are the same for both the low and high registers.
  11150		 */
  11151		reg = (((u64)vl->vl & SEND_LOW_PRIORITY_LIST_VL_MASK)
  11152				<< SEND_LOW_PRIORITY_LIST_VL_SHIFT)
  11153		      | (((u64)vl->weight
  11154				& SEND_LOW_PRIORITY_LIST_WEIGHT_MASK)
  11155				<< SEND_LOW_PRIORITY_LIST_WEIGHT_SHIFT);
  11156		write_csr(dd, target + (i * 8), reg);
  11157	}
  11158	pio_send_control(dd, PSC_GLOBAL_VLARB_ENABLE);
  11159
  11160	if (drain)
  11161		open_fill_data_vls(dd); /* reopen all VLs */
  11162
  11163err:
  11164	mutex_unlock(&ppd->hls_lock);
  11165
  11166	return ret;
  11167}
  11168
  11169/*
  11170 * Read one credit merge VL register.
  11171 */
  11172static void read_one_cm_vl(struct hfi1_devdata *dd, u32 csr,
  11173			   struct vl_limit *vll)
  11174{
  11175	u64 reg = read_csr(dd, csr);
  11176
  11177	vll->dedicated = cpu_to_be16(
  11178		(reg >> SEND_CM_CREDIT_VL_DEDICATED_LIMIT_VL_SHIFT)
  11179		& SEND_CM_CREDIT_VL_DEDICATED_LIMIT_VL_MASK);
  11180	vll->shared = cpu_to_be16(
  11181		(reg >> SEND_CM_CREDIT_VL_SHARED_LIMIT_VL_SHIFT)
  11182		& SEND_CM_CREDIT_VL_SHARED_LIMIT_VL_MASK);
  11183}
  11184
  11185/*
  11186 * Read the current credit merge limits.
  11187 */
  11188static int get_buffer_control(struct hfi1_devdata *dd,
  11189			      struct buffer_control *bc, u16 *overall_limit)
  11190{
  11191	u64 reg;
  11192	int i;
  11193
  11194	/* not all entries are filled in */
  11195	memset(bc, 0, sizeof(*bc));
  11196
  11197	/* OPA and HFI have a 1-1 mapping */
  11198	for (i = 0; i < TXE_NUM_DATA_VL; i++)
  11199		read_one_cm_vl(dd, SEND_CM_CREDIT_VL + (8 * i), &bc->vl[i]);
  11200
  11201	/* NOTE: assumes that VL* and VL15 CSRs are bit-wise identical */
  11202	read_one_cm_vl(dd, SEND_CM_CREDIT_VL15, &bc->vl[15]);
  11203
  11204	reg = read_csr(dd, SEND_CM_GLOBAL_CREDIT);
  11205	bc->overall_shared_limit = cpu_to_be16(
  11206		(reg >> SEND_CM_GLOBAL_CREDIT_SHARED_LIMIT_SHIFT)
  11207		& SEND_CM_GLOBAL_CREDIT_SHARED_LIMIT_MASK);
  11208	if (overall_limit)
  11209		*overall_limit = (reg
  11210			>> SEND_CM_GLOBAL_CREDIT_TOTAL_CREDIT_LIMIT_SHIFT)
  11211			& SEND_CM_GLOBAL_CREDIT_TOTAL_CREDIT_LIMIT_MASK;
  11212	return sizeof(struct buffer_control);
  11213}
  11214
  11215static int get_sc2vlnt(struct hfi1_devdata *dd, struct sc2vlnt *dp)
  11216{
  11217	u64 reg;
  11218	int i;
  11219
  11220	/* each register contains 16 SC->VLnt mappings, 4 bits each */
  11221	reg = read_csr(dd, DCC_CFG_SC_VL_TABLE_15_0);
  11222	for (i = 0; i < sizeof(u64); i++) {
  11223		u8 byte = *(((u8 *)&reg) + i);
  11224
  11225		dp->vlnt[2 * i] = byte & 0xf;
  11226		dp->vlnt[(2 * i) + 1] = (byte & 0xf0) >> 4;
  11227	}
  11228
  11229	reg = read_csr(dd, DCC_CFG_SC_VL_TABLE_31_16);
  11230	for (i = 0; i < sizeof(u64); i++) {
  11231		u8 byte = *(((u8 *)&reg) + i);
  11232
  11233		dp->vlnt[16 + (2 * i)] = byte & 0xf;
  11234		dp->vlnt[16 + (2 * i) + 1] = (byte & 0xf0) >> 4;
  11235	}
  11236	return sizeof(struct sc2vlnt);
  11237}
  11238
  11239static void get_vlarb_preempt(struct hfi1_devdata *dd, u32 nelems,
  11240			      struct ib_vl_weight_elem *vl)
  11241{
  11242	unsigned int i;
  11243
  11244	for (i = 0; i < nelems; i++, vl++) {
  11245		vl->vl = 0xf;
  11246		vl->weight = 0;
  11247	}
  11248}
  11249
  11250static void set_sc2vlnt(struct hfi1_devdata *dd, struct sc2vlnt *dp)
  11251{
  11252	write_csr(dd, DCC_CFG_SC_VL_TABLE_15_0,
  11253		  DC_SC_VL_VAL(15_0,
  11254			       0, dp->vlnt[0] & 0xf,
  11255			       1, dp->vlnt[1] & 0xf,
  11256			       2, dp->vlnt[2] & 0xf,
  11257			       3, dp->vlnt[3] & 0xf,
  11258			       4, dp->vlnt[4] & 0xf,
  11259			       5, dp->vlnt[5] & 0xf,
  11260			       6, dp->vlnt[6] & 0xf,
  11261			       7, dp->vlnt[7] & 0xf,
  11262			       8, dp->vlnt[8] & 0xf,
  11263			       9, dp->vlnt[9] & 0xf,
  11264			       10, dp->vlnt[10] & 0xf,
  11265			       11, dp->vlnt[11] & 0xf,
  11266			       12, dp->vlnt[12] & 0xf,
  11267			       13, dp->vlnt[13] & 0xf,
  11268			       14, dp->vlnt[14] & 0xf,
  11269			       15, dp->vlnt[15] & 0xf));
  11270	write_csr(dd, DCC_CFG_SC_VL_TABLE_31_16,
  11271		  DC_SC_VL_VAL(31_16,
  11272			       16, dp->vlnt[16] & 0xf,
  11273			       17, dp->vlnt[17] & 0xf,
  11274			       18, dp->vlnt[18] & 0xf,
  11275			       19, dp->vlnt[19] & 0xf,
  11276			       20, dp->vlnt[20] & 0xf,
  11277			       21, dp->vlnt[21] & 0xf,
  11278			       22, dp->vlnt[22] & 0xf,
  11279			       23, dp->vlnt[23] & 0xf,
  11280			       24, dp->vlnt[24] & 0xf,
  11281			       25, dp->vlnt[25] & 0xf,
  11282			       26, dp->vlnt[26] & 0xf,
  11283			       27, dp->vlnt[27] & 0xf,
  11284			       28, dp->vlnt[28] & 0xf,
  11285			       29, dp->vlnt[29] & 0xf,
  11286			       30, dp->vlnt[30] & 0xf,
  11287			       31, dp->vlnt[31] & 0xf));
  11288}
  11289
  11290static void nonzero_msg(struct hfi1_devdata *dd, int idx, const char *what,
  11291			u16 limit)
  11292{
  11293	if (limit != 0)
  11294		dd_dev_info(dd, "Invalid %s limit %d on VL %d, ignoring\n",
  11295			    what, (int)limit, idx);
  11296}
  11297
  11298/* change only the shared limit portion of SendCmGLobalCredit */
  11299static void set_global_shared(struct hfi1_devdata *dd, u16 limit)
  11300{
  11301	u64 reg;
  11302
  11303	reg = read_csr(dd, SEND_CM_GLOBAL_CREDIT);
  11304	reg &= ~SEND_CM_GLOBAL_CREDIT_SHARED_LIMIT_SMASK;
  11305	reg |= (u64)limit << SEND_CM_GLOBAL_CREDIT_SHARED_LIMIT_SHIFT;
  11306	write_csr(dd, SEND_CM_GLOBAL_CREDIT, reg);
  11307}
  11308
  11309/* change only the total credit limit portion of SendCmGLobalCredit */
  11310static void set_global_limit(struct hfi1_devdata *dd, u16 limit)
  11311{
  11312	u64 reg;
  11313
  11314	reg = read_csr(dd, SEND_CM_GLOBAL_CREDIT);
  11315	reg &= ~SEND_CM_GLOBAL_CREDIT_TOTAL_CREDIT_LIMIT_SMASK;
  11316	reg |= (u64)limit << SEND_CM_GLOBAL_CREDIT_TOTAL_CREDIT_LIMIT_SHIFT;
  11317	write_csr(dd, SEND_CM_GLOBAL_CREDIT, reg);
  11318}
  11319
  11320/* set the given per-VL shared limit */
  11321static void set_vl_shared(struct hfi1_devdata *dd, int vl, u16 limit)
  11322{
  11323	u64 reg;
  11324	u32 addr;
  11325
  11326	if (vl < TXE_NUM_DATA_VL)
  11327		addr = SEND_CM_CREDIT_VL + (8 * vl);
  11328	else
  11329		addr = SEND_CM_CREDIT_VL15;
  11330
  11331	reg = read_csr(dd, addr);
  11332	reg &= ~SEND_CM_CREDIT_VL_SHARED_LIMIT_VL_SMASK;
  11333	reg |= (u64)limit << SEND_CM_CREDIT_VL_SHARED_LIMIT_VL_SHIFT;
  11334	write_csr(dd, addr, reg);
  11335}
  11336
  11337/* set the given per-VL dedicated limit */
  11338static void set_vl_dedicated(struct hfi1_devdata *dd, int vl, u16 limit)
  11339{
  11340	u64 reg;
  11341	u32 addr;
  11342
  11343	if (vl < TXE_NUM_DATA_VL)
  11344		addr = SEND_CM_CREDIT_VL + (8 * vl);
  11345	else
  11346		addr = SEND_CM_CREDIT_VL15;
  11347
  11348	reg = read_csr(dd, addr);
  11349	reg &= ~SEND_CM_CREDIT_VL_DEDICATED_LIMIT_VL_SMASK;
  11350	reg |= (u64)limit << SEND_CM_CREDIT_VL_DEDICATED_LIMIT_VL_SHIFT;
  11351	write_csr(dd, addr, reg);
  11352}
  11353
  11354/* spin until the given per-VL status mask bits clear */
  11355static void wait_for_vl_status_clear(struct hfi1_devdata *dd, u64 mask,
  11356				     const char *which)
  11357{
  11358	unsigned long timeout;
  11359	u64 reg;
  11360
  11361	timeout = jiffies + msecs_to_jiffies(VL_STATUS_CLEAR_TIMEOUT);
  11362	while (1) {
  11363		reg = read_csr(dd, SEND_CM_CREDIT_USED_STATUS) & mask;
  11364
  11365		if (reg == 0)
  11366			return;	/* success */
  11367		if (time_after(jiffies, timeout))
  11368			break;		/* timed out */
  11369		udelay(1);
  11370	}
  11371
  11372	dd_dev_err(dd,
  11373		   "%s credit change status not clearing after %dms, mask 0x%llx, not clear 0x%llx\n",
  11374		   which, VL_STATUS_CLEAR_TIMEOUT, mask, reg);
  11375	/*
  11376	 * If this occurs, it is likely there was a credit loss on the link.
  11377	 * The only recovery from that is a link bounce.
  11378	 */
  11379	dd_dev_err(dd,
  11380		   "Continuing anyway.  A credit loss may occur.  Suggest a link bounce\n");
  11381}
  11382
  11383/*
  11384 * The number of credits on the VLs may be changed while everything
  11385 * is "live", but the following algorithm must be followed due to
  11386 * how the hardware is actually implemented.  In particular,
  11387 * Return_Credit_Status[] is the only correct status check.
  11388 *
  11389 * if (reducing Global_Shared_Credit_Limit or any shared limit changing)
  11390 *     set Global_Shared_Credit_Limit = 0
  11391 *     use_all_vl = 1
  11392 * mask0 = all VLs that are changing either dedicated or shared limits
  11393 * set Shared_Limit[mask0] = 0
  11394 * spin until Return_Credit_Status[use_all_vl ? all VL : mask0] == 0
  11395 * if (changing any dedicated limit)
  11396 *     mask1 = all VLs that are lowering dedicated limits
  11397 *     lower Dedicated_Limit[mask1]
  11398 *     spin until Return_Credit_Status[mask1] == 0
  11399 *     raise Dedicated_Limits
  11400 * raise Shared_Limits
  11401 * raise Global_Shared_Credit_Limit
  11402 *
  11403 * lower = if the new limit is lower, set the limit to the new value
  11404 * raise = if the new limit is higher than the current value (may be changed
  11405 *	earlier in the algorithm), set the new limit to the new value
  11406 */
  11407int set_buffer_control(struct hfi1_pportdata *ppd,
  11408		       struct buffer_control *new_bc)
  11409{
  11410	struct hfi1_devdata *dd = ppd->dd;
  11411	u64 changing_mask, ld_mask, stat_mask;
  11412	int change_count;
  11413	int i, use_all_mask;
  11414	int this_shared_changing;
  11415	int vl_count = 0, ret;
  11416	/*
  11417	 * A0: add the variable any_shared_limit_changing below and in the
  11418	 * algorithm above.  If removing A0 support, it can be removed.
  11419	 */
  11420	int any_shared_limit_changing;
  11421	struct buffer_control cur_bc;
  11422	u8 changing[OPA_MAX_VLS];
  11423	u8 lowering_dedicated[OPA_MAX_VLS];
  11424	u16 cur_total;
  11425	u32 new_total = 0;
  11426	const u64 all_mask =
  11427	SEND_CM_CREDIT_USED_STATUS_VL0_RETURN_CREDIT_STATUS_SMASK
  11428	 | SEND_CM_CREDIT_USED_STATUS_VL1_RETURN_CREDIT_STATUS_SMASK
  11429	 | SEND_CM_CREDIT_USED_STATUS_VL2_RETURN_CREDIT_STATUS_SMASK
  11430	 | SEND_CM_CREDIT_USED_STATUS_VL3_RETURN_CREDIT_STATUS_SMASK
  11431	 | SEND_CM_CREDIT_USED_STATUS_VL4_RETURN_CREDIT_STATUS_SMASK
  11432	 | SEND_CM_CREDIT_USED_STATUS_VL5_RETURN_CREDIT_STATUS_SMASK
  11433	 | SEND_CM_CREDIT_USED_STATUS_VL6_RETURN_CREDIT_STATUS_SMASK
  11434	 | SEND_CM_CREDIT_USED_STATUS_VL7_RETURN_CREDIT_STATUS_SMASK
  11435	 | SEND_CM_CREDIT_USED_STATUS_VL15_RETURN_CREDIT_STATUS_SMASK;
  11436
  11437#define valid_vl(idx) ((idx) < TXE_NUM_DATA_VL || (idx) == 15)
  11438#define NUM_USABLE_VLS 16	/* look at VL15 and less */
  11439
  11440	/* find the new total credits, do sanity check on unused VLs */
  11441	for (i = 0; i < OPA_MAX_VLS; i++) {
  11442		if (valid_vl(i)) {
  11443			new_total += be16_to_cpu(new_bc->vl[i].dedicated);
  11444			continue;
  11445		}
  11446		nonzero_msg(dd, i, "dedicated",
  11447			    be16_to_cpu(new_bc->vl[i].dedicated));
  11448		nonzero_msg(dd, i, "shared",
  11449			    be16_to_cpu(new_bc->vl[i].shared));
  11450		new_bc->vl[i].dedicated = 0;
  11451		new_bc->vl[i].shared = 0;
  11452	}
  11453	new_total += be16_to_cpu(new_bc->overall_shared_limit);
  11454
  11455	/* fetch the current values */
  11456	get_buffer_control(dd, &cur_bc, &cur_total);
  11457
  11458	/*
  11459	 * Create the masks we will use.
  11460	 */
  11461	memset(changing, 0, sizeof(changing));
  11462	memset(lowering_dedicated, 0, sizeof(lowering_dedicated));
  11463	/*
  11464	 * NOTE: Assumes that the individual VL bits are adjacent and in
  11465	 * increasing order
  11466	 */
  11467	stat_mask =
  11468		SEND_CM_CREDIT_USED_STATUS_VL0_RETURN_CREDIT_STATUS_SMASK;
  11469	changing_mask = 0;
  11470	ld_mask = 0;
  11471	change_count = 0;
  11472	any_shared_limit_changing = 0;
  11473	for (i = 0; i < NUM_USABLE_VLS; i++, stat_mask <<= 1) {
  11474		if (!valid_vl(i))
  11475			continue;
  11476		this_shared_changing = new_bc->vl[i].shared
  11477						!= cur_bc.vl[i].shared;
  11478		if (this_shared_changing)
  11479			any_shared_limit_changing = 1;
  11480		if (new_bc->vl[i].dedicated != cur_bc.vl[i].dedicated ||
  11481		    this_shared_changing) {
  11482			changing[i] = 1;
  11483			changing_mask |= stat_mask;
  11484			change_count++;
  11485		}
  11486		if (be16_to_cpu(new_bc->vl[i].dedicated) <
  11487					be16_to_cpu(cur_bc.vl[i].dedicated)) {
  11488			lowering_dedicated[i] = 1;
  11489			ld_mask |= stat_mask;
  11490		}
  11491	}
  11492
  11493	/* bracket the credit change with a total adjustment */
  11494	if (new_total > cur_total)
  11495		set_global_limit(dd, new_total);
  11496
  11497	/*
  11498	 * Start the credit change algorithm.
  11499	 */
  11500	use_all_mask = 0;
  11501	if ((be16_to_cpu(new_bc->overall_shared_limit) <
  11502	     be16_to_cpu(cur_bc.overall_shared_limit)) ||
  11503	    (is_ax(dd) && any_shared_limit_changing)) {
  11504		set_global_shared(dd, 0);
  11505		cur_bc.overall_shared_limit = 0;
  11506		use_all_mask = 1;
  11507	}
  11508
  11509	for (i = 0; i < NUM_USABLE_VLS; i++) {
  11510		if (!valid_vl(i))
  11511			continue;
  11512
  11513		if (changing[i]) {
  11514			set_vl_shared(dd, i, 0);
  11515			cur_bc.vl[i].shared = 0;
  11516		}
  11517	}
  11518
  11519	wait_for_vl_status_clear(dd, use_all_mask ? all_mask : changing_mask,
  11520				 "shared");
  11521
  11522	if (change_count > 0) {
  11523		for (i = 0; i < NUM_USABLE_VLS; i++) {
  11524			if (!valid_vl(i))
  11525				continue;
  11526
  11527			if (lowering_dedicated[i]) {
  11528				set_vl_dedicated(dd, i,
  11529						 be16_to_cpu(new_bc->
  11530							     vl[i].dedicated));
  11531				cur_bc.vl[i].dedicated =
  11532						new_bc->vl[i].dedicated;
  11533			}
  11534		}
  11535
  11536		wait_for_vl_status_clear(dd, ld_mask, "dedicated");
  11537
  11538		/* now raise all dedicated that are going up */
  11539		for (i = 0; i < NUM_USABLE_VLS; i++) {
  11540			if (!valid_vl(i))
  11541				continue;
  11542
  11543			if (be16_to_cpu(new_bc->vl[i].dedicated) >
  11544					be16_to_cpu(cur_bc.vl[i].dedicated))
  11545				set_vl_dedicated(dd, i,
  11546						 be16_to_cpu(new_bc->
  11547							     vl[i].dedicated));
  11548		}
  11549	}
  11550
  11551	/* next raise all shared that are going up */
  11552	for (i = 0; i < NUM_USABLE_VLS; i++) {
  11553		if (!valid_vl(i))
  11554			continue;
  11555
  11556		if (be16_to_cpu(new_bc->vl[i].shared) >
  11557				be16_to_cpu(cur_bc.vl[i].shared))
  11558			set_vl_shared(dd, i, be16_to_cpu(new_bc->vl[i].shared));
  11559	}
  11560
  11561	/* finally raise the global shared */
  11562	if (be16_to_cpu(new_bc->overall_shared_limit) >
  11563	    be16_to_cpu(cur_bc.overall_shared_limit))
  11564		set_global_shared(dd,
  11565				  be16_to_cpu(new_bc->overall_shared_limit));
  11566
  11567	/* bracket the credit change with a total adjustment */
  11568	if (new_total < cur_total)
  11569		set_global_limit(dd, new_total);
  11570
  11571	/*
  11572	 * Determine the actual number of operational VLS using the number of
  11573	 * dedicated and shared credits for each VL.
  11574	 */
  11575	if (change_count > 0) {
  11576		for (i = 0; i < TXE_NUM_DATA_VL; i++)
  11577			if (be16_to_cpu(new_bc->vl[i].dedicated) > 0 ||
  11578			    be16_to_cpu(new_bc->vl[i].shared) > 0)
  11579				vl_count++;
  11580		ppd->actual_vls_operational = vl_count;
  11581		ret = sdma_map_init(dd, ppd->port - 1, vl_count ?
  11582				    ppd->actual_vls_operational :
  11583				    ppd->vls_operational,
  11584				    NULL);
  11585		if (ret == 0)
  11586			ret = pio_map_init(dd, ppd->port - 1, vl_count ?
  11587					   ppd->actual_vls_operational :
  11588					   ppd->vls_operational, NULL);
  11589		if (ret)
  11590			return ret;
  11591	}
  11592	return 0;
  11593}
  11594
  11595/*
  11596 * Read the given fabric manager table. Return the size of the
  11597 * table (in bytes) on success, and a negative error code on
  11598 * failure.
  11599 */
  11600int fm_get_table(struct hfi1_pportdata *ppd, int which, void *t)
  11601
  11602{
  11603	int size;
  11604	struct vl_arb_cache *vlc;
  11605
  11606	switch (which) {
  11607	case FM_TBL_VL_HIGH_ARB:
  11608		size = 256;
  11609		/*
  11610		 * OPA specifies 128 elements (of 2 bytes each), though
  11611		 * HFI supports only 16 elements in h/w.
  11612		 */
  11613		vlc = vl_arb_lock_cache(ppd, HI_PRIO_TABLE);
  11614		vl_arb_get_cache(vlc, t);
  11615		vl_arb_unlock_cache(ppd, HI_PRIO_TABLE);
  11616		break;
  11617	case FM_TBL_VL_LOW_ARB:
  11618		size = 256;
  11619		/*
  11620		 * OPA specifies 128 elements (of 2 bytes each), though
  11621		 * HFI supports only 16 elements in h/w.
  11622		 */
  11623		vlc = vl_arb_lock_cache(ppd, LO_PRIO_TABLE);
  11624		vl_arb_get_cache(vlc, t);
  11625		vl_arb_unlock_cache(ppd, LO_PRIO_TABLE);
  11626		break;
  11627	case FM_TBL_BUFFER_CONTROL:
  11628		size = get_buffer_control(ppd->dd, t, NULL);
  11629		break;
  11630	case FM_TBL_SC2VLNT:
  11631		size = get_sc2vlnt(ppd->dd, t);
  11632		break;
  11633	case FM_TBL_VL_PREEMPT_ELEMS:
  11634		size = 256;
  11635		/* OPA specifies 128 elements, of 2 bytes each */
  11636		get_vlarb_preempt(ppd->dd, OPA_MAX_VLS, t);
  11637		break;
  11638	case FM_TBL_VL_PREEMPT_MATRIX:
  11639		size = 256;
  11640		/*
  11641		 * OPA specifies that this is the same size as the VL
  11642		 * arbitration tables (i.e., 256 bytes).
  11643		 */
  11644		break;
  11645	default:
  11646		return -EINVAL;
  11647	}
  11648	return size;
  11649}
  11650
  11651/*
  11652 * Write the given fabric manager table.
  11653 */
  11654int fm_set_table(struct hfi1_pportdata *ppd, int which, void *t)
  11655{
  11656	int ret = 0;
  11657	struct vl_arb_cache *vlc;
  11658
  11659	switch (which) {
  11660	case FM_TBL_VL_HIGH_ARB:
  11661		vlc = vl_arb_lock_cache(ppd, HI_PRIO_TABLE);
  11662		if (vl_arb_match_cache(vlc, t)) {
  11663			vl_arb_unlock_cache(ppd, HI_PRIO_TABLE);
  11664			break;
  11665		}
  11666		vl_arb_set_cache(vlc, t);
  11667		vl_arb_unlock_cache(ppd, HI_PRIO_TABLE);
  11668		ret = set_vl_weights(ppd, SEND_HIGH_PRIORITY_LIST,
  11669				     VL_ARB_HIGH_PRIO_TABLE_SIZE, t);
  11670		break;
  11671	case FM_TBL_VL_LOW_ARB:
  11672		vlc = vl_arb_lock_cache(ppd, LO_PRIO_TABLE);
  11673		if (vl_arb_match_cache(vlc, t)) {
  11674			vl_arb_unlock_cache(ppd, LO_PRIO_TABLE);
  11675			break;
  11676		}
  11677		vl_arb_set_cache(vlc, t);
  11678		vl_arb_unlock_cache(ppd, LO_PRIO_TABLE);
  11679		ret = set_vl_weights(ppd, SEND_LOW_PRIORITY_LIST,
  11680				     VL_ARB_LOW_PRIO_TABLE_SIZE, t);
  11681		break;
  11682	case FM_TBL_BUFFER_CONTROL:
  11683		ret = set_buffer_control(ppd, t);
  11684		break;
  11685	case FM_TBL_SC2VLNT:
  11686		set_sc2vlnt(ppd->dd, t);
  11687		break;
  11688	default:
  11689		ret = -EINVAL;
  11690	}
  11691	return ret;
  11692}
  11693
  11694/*
  11695 * Disable all data VLs.
  11696 *
  11697 * Return 0 if disabled, non-zero if the VLs cannot be disabled.
  11698 */
  11699static int disable_data_vls(struct hfi1_devdata *dd)
  11700{
  11701	if (is_ax(dd))
  11702		return 1;
  11703
  11704	pio_send_control(dd, PSC_DATA_VL_DISABLE);
  11705
  11706	return 0;
  11707}
  11708
  11709/*
  11710 * open_fill_data_vls() - the counterpart to stop_drain_data_vls().
  11711 * Just re-enables all data VLs (the "fill" part happens
  11712 * automatically - the name was chosen for symmetry with
  11713 * stop_drain_data_vls()).
  11714 *
  11715 * Return 0 if successful, non-zero if the VLs cannot be enabled.
  11716 */
  11717int open_fill_data_vls(struct hfi1_devdata *dd)
  11718{
  11719	if (is_ax(dd))
  11720		return 1;
  11721
  11722	pio_send_control(dd, PSC_DATA_VL_ENABLE);
  11723
  11724	return 0;
  11725}
  11726
  11727/*
  11728 * drain_data_vls() - assumes that disable_data_vls() has been called,
  11729 * wait for occupancy (of per-VL FIFOs) for all contexts, and SDMA
  11730 * engines to drop to 0.
  11731 */
  11732static void drain_data_vls(struct hfi1_devdata *dd)
  11733{
  11734	sc_wait(dd);
  11735	sdma_wait(dd);
  11736	pause_for_credit_return(dd);
  11737}
  11738
  11739/*
  11740 * stop_drain_data_vls() - disable, then drain all per-VL fifos.
  11741 *
  11742 * Use open_fill_data_vls() to resume using data VLs.  This pair is
  11743 * meant to be used like this:
  11744 *
  11745 * stop_drain_data_vls(dd);
  11746 * // do things with per-VL resources
  11747 * open_fill_data_vls(dd);
  11748 */
  11749int stop_drain_data_vls(struct hfi1_devdata *dd)
  11750{
  11751	int ret;
  11752
  11753	ret = disable_data_vls(dd);
  11754	if (ret == 0)
  11755		drain_data_vls(dd);
  11756
  11757	return ret;
  11758}
  11759
  11760/*
  11761 * Convert a nanosecond time to a cclock count.  No matter how slow
  11762 * the cclock, a non-zero ns will always have a non-zero result.
  11763 */
  11764u32 ns_to_cclock(struct hfi1_devdata *dd, u32 ns)
  11765{
  11766	u32 cclocks;
  11767
  11768	if (dd->icode == ICODE_FPGA_EMULATION)
  11769		cclocks = (ns * 1000) / FPGA_CCLOCK_PS;
  11770	else  /* simulation pretends to be ASIC */
  11771		cclocks = (ns * 1000) / ASIC_CCLOCK_PS;
  11772	if (ns && !cclocks)	/* if ns nonzero, must be at least 1 */
  11773		cclocks = 1;
  11774	return cclocks;
  11775}
  11776
  11777/*
  11778 * Convert a cclock count to nanoseconds. Not matter how slow
  11779 * the cclock, a non-zero cclocks will always have a non-zero result.
  11780 */
  11781u32 cclock_to_ns(struct hfi1_devdata *dd, u32 cclocks)
  11782{
  11783	u32 ns;
  11784
  11785	if (dd->icode == ICODE_FPGA_EMULATION)
  11786		ns = (cclocks * FPGA_CCLOCK_PS) / 1000;
  11787	else  /* simulation pretends to be ASIC */
  11788		ns = (cclocks * ASIC_CCLOCK_PS) / 1000;
  11789	if (cclocks && !ns)
  11790		ns = 1;
  11791	return ns;
  11792}
  11793
  11794/*
  11795 * Dynamically adjust the receive interrupt timeout for a context based on
  11796 * incoming packet rate.
  11797 *
  11798 * NOTE: Dynamic adjustment does not allow rcv_intr_count to be zero.
  11799 */
  11800static void adjust_rcv_timeout(struct hfi1_ctxtdata *rcd, u32 npkts)
  11801{
  11802	struct hfi1_devdata *dd = rcd->dd;
  11803	u32 timeout = rcd->rcvavail_timeout;
  11804
  11805	/*
  11806	 * This algorithm doubles or halves the timeout depending on whether
  11807	 * the number of packets received in this interrupt were less than or
  11808	 * greater equal the interrupt count.
  11809	 *
  11810	 * The calculations below do not allow a steady state to be achieved.
  11811	 * Only at the endpoints it is possible to have an unchanging
  11812	 * timeout.
  11813	 */
  11814	if (npkts < rcv_intr_count) {
  11815		/*
  11816		 * Not enough packets arrived before the timeout, adjust
  11817		 * timeout downward.
  11818		 */
  11819		if (timeout < 2) /* already at minimum? */
  11820			return;
  11821		timeout >>= 1;
  11822	} else {
  11823		/*
  11824		 * More than enough packets arrived before the timeout, adjust
  11825		 * timeout upward.
  11826		 */
  11827		if (timeout >= dd->rcv_intr_timeout_csr) /* already at max? */
  11828			return;
  11829		timeout = min(timeout << 1, dd->rcv_intr_timeout_csr);
  11830	}
  11831
  11832	rcd->rcvavail_timeout = timeout;
  11833	/*
  11834	 * timeout cannot be larger than rcv_intr_timeout_csr which has already
  11835	 * been verified to be in range
  11836	 */
  11837	write_kctxt_csr(dd, rcd->ctxt, RCV_AVAIL_TIME_OUT,
  11838			(u64)timeout <<
  11839			RCV_AVAIL_TIME_OUT_TIME_OUT_RELOAD_SHIFT);
  11840}
  11841
  11842void update_usrhead(struct hfi1_ctxtdata *rcd, u32 hd, u32 updegr, u32 egrhd,
  11843		    u32 intr_adjust, u32 npkts)
  11844{
  11845	struct hfi1_devdata *dd = rcd->dd;
  11846	u64 reg;
  11847	u32 ctxt = rcd->ctxt;
  11848
  11849	/*
  11850	 * Need to write timeout register before updating RcvHdrHead to ensure
  11851	 * that a new value is used when the HW decides to restart counting.
  11852	 */
  11853	if (intr_adjust)
  11854		adjust_rcv_timeout(rcd, npkts);
  11855	if (updegr) {
  11856		reg = (egrhd & RCV_EGR_INDEX_HEAD_HEAD_MASK)
  11857			<< RCV_EGR_INDEX_HEAD_HEAD_SHIFT;
  11858		write_uctxt_csr(dd, ctxt, RCV_EGR_INDEX_HEAD, reg);
  11859	}
  11860	reg = ((u64)rcv_intr_count << RCV_HDR_HEAD_COUNTER_SHIFT) |
  11861		(((u64)hd & RCV_HDR_HEAD_HEAD_MASK)
  11862			<< RCV_HDR_HEAD_HEAD_SHIFT);
  11863	write_uctxt_csr(dd, ctxt, RCV_HDR_HEAD, reg);
  11864}
  11865
  11866u32 hdrqempty(struct hfi1_ctxtdata *rcd)
  11867{
  11868	u32 head, tail;
  11869
  11870	head = (read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_HEAD)
  11871		& RCV_HDR_HEAD_HEAD_SMASK) >> RCV_HDR_HEAD_HEAD_SHIFT;
  11872
  11873	if (hfi1_rcvhdrtail_kvaddr(rcd))
  11874		tail = get_rcvhdrtail(rcd);
  11875	else
  11876		tail = read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_TAIL);
  11877
  11878	return head == tail;
  11879}
  11880
  11881/*
  11882 * Context Control and Receive Array encoding for buffer size:
  11883 *	0x0 invalid
  11884 *	0x1   4 KB
  11885 *	0x2   8 KB
  11886 *	0x3  16 KB
  11887 *	0x4  32 KB
  11888 *	0x5  64 KB
  11889 *	0x6 128 KB
  11890 *	0x7 256 KB
  11891 *	0x8 512 KB (Receive Array only)
  11892 *	0x9   1 MB (Receive Array only)
  11893 *	0xa   2 MB (Receive Array only)
  11894 *
  11895 *	0xB-0xF - reserved (Receive Array only)
  11896 *
  11897 *
  11898 * This routine assumes that the value has already been sanity checked.
  11899 */
  11900static u32 encoded_size(u32 size)
  11901{
  11902	switch (size) {
  11903	case   4 * 1024: return 0x1;
  11904	case   8 * 1024: return 0x2;
  11905	case  16 * 1024: return 0x3;
  11906	case  32 * 1024: return 0x4;
  11907	case  64 * 1024: return 0x5;
  11908	case 128 * 1024: return 0x6;
  11909	case 256 * 1024: return 0x7;
  11910	case 512 * 1024: return 0x8;
  11911	case   1 * 1024 * 1024: return 0x9;
  11912	case   2 * 1024 * 1024: return 0xa;
  11913	}
  11914	return 0x1;	/* if invalid, go with the minimum size */
  11915}
  11916
  11917/**
  11918 * encode_rcv_header_entry_size - return chip specific encoding for size
  11919 * @size: size in dwords
  11920 *
  11921 * Convert a receive header entry size that to the encoding used in the CSR.
  11922 *
  11923 * Return a zero if the given size is invalid, otherwise the encoding.
  11924 */
  11925u8 encode_rcv_header_entry_size(u8 size)
  11926{
  11927	/* there are only 3 valid receive header entry sizes */
  11928	if (size == 2)
  11929		return 1;
  11930	if (size == 16)
  11931		return 2;
  11932	if (size == 32)
  11933		return 4;
  11934	return 0; /* invalid */
  11935}
  11936
  11937/**
  11938 * hfi1_validate_rcvhdrcnt - validate hdrcnt
  11939 * @dd: the device data
  11940 * @thecnt: the header count
  11941 */
  11942int hfi1_validate_rcvhdrcnt(struct hfi1_devdata *dd, uint thecnt)
  11943{
  11944	if (thecnt <= HFI1_MIN_HDRQ_EGRBUF_CNT) {
  11945		dd_dev_err(dd, "Receive header queue count too small\n");
  11946		return -EINVAL;
  11947	}
  11948
  11949	if (thecnt > HFI1_MAX_HDRQ_EGRBUF_CNT) {
  11950		dd_dev_err(dd,
  11951			   "Receive header queue count cannot be greater than %u\n",
  11952			   HFI1_MAX_HDRQ_EGRBUF_CNT);
  11953		return -EINVAL;
  11954	}
  11955
  11956	if (thecnt % HDRQ_INCREMENT) {
  11957		dd_dev_err(dd, "Receive header queue count %d must be divisible by %lu\n",
  11958			   thecnt, HDRQ_INCREMENT);
  11959		return -EINVAL;
  11960	}
  11961
  11962	return 0;
  11963}
  11964
  11965/**
  11966 * set_hdrq_regs - set header queue registers for context
  11967 * @dd: the device data
  11968 * @ctxt: the context
  11969 * @entsize: the dword entry size
  11970 * @hdrcnt: the number of header entries
  11971 */
  11972void set_hdrq_regs(struct hfi1_devdata *dd, u8 ctxt, u8 entsize, u16 hdrcnt)
  11973{
  11974	u64 reg;
  11975
  11976	reg = (((u64)hdrcnt >> HDRQ_SIZE_SHIFT) & RCV_HDR_CNT_CNT_MASK) <<
  11977	      RCV_HDR_CNT_CNT_SHIFT;
  11978	write_kctxt_csr(dd, ctxt, RCV_HDR_CNT, reg);
  11979	reg = ((u64)encode_rcv_header_entry_size(entsize) &
  11980	       RCV_HDR_ENT_SIZE_ENT_SIZE_MASK) <<
  11981	      RCV_HDR_ENT_SIZE_ENT_SIZE_SHIFT;
  11982	write_kctxt_csr(dd, ctxt, RCV_HDR_ENT_SIZE, reg);
  11983	reg = ((u64)DEFAULT_RCVHDRSIZE & RCV_HDR_SIZE_HDR_SIZE_MASK) <<
  11984	      RCV_HDR_SIZE_HDR_SIZE_SHIFT;
  11985	write_kctxt_csr(dd, ctxt, RCV_HDR_SIZE, reg);
  11986
  11987	/*
  11988	 * Program dummy tail address for every receive context
  11989	 * before enabling any receive context
  11990	 */
  11991	write_kctxt_csr(dd, ctxt, RCV_HDR_TAIL_ADDR,
  11992			dd->rcvhdrtail_dummy_dma);
  11993}
  11994
  11995void hfi1_rcvctrl(struct hfi1_devdata *dd, unsigned int op,
  11996		  struct hfi1_ctxtdata *rcd)
  11997{
  11998	u64 rcvctrl, reg;
  11999	int did_enable = 0;
  12000	u16 ctxt;
  12001
  12002	if (!rcd)
  12003		return;
  12004
  12005	ctxt = rcd->ctxt;
  12006
  12007	hfi1_cdbg(RCVCTRL, "ctxt %d op 0x%x", ctxt, op);
  12008
  12009	rcvctrl = read_kctxt_csr(dd, ctxt, RCV_CTXT_CTRL);
  12010	/* if the context already enabled, don't do the extra steps */
  12011	if ((op & HFI1_RCVCTRL_CTXT_ENB) &&
  12012	    !(rcvctrl & RCV_CTXT_CTRL_ENABLE_SMASK)) {
  12013		/* reset the tail and hdr addresses, and sequence count */
  12014		write_kctxt_csr(dd, ctxt, RCV_HDR_ADDR,
  12015				rcd->rcvhdrq_dma);
  12016		if (hfi1_rcvhdrtail_kvaddr(rcd))
  12017			write_kctxt_csr(dd, ctxt, RCV_HDR_TAIL_ADDR,
  12018					rcd->rcvhdrqtailaddr_dma);
  12019		hfi1_set_seq_cnt(rcd, 1);
  12020
  12021		/* reset the cached receive header queue head value */
  12022		hfi1_set_rcd_head(rcd, 0);
  12023
  12024		/*
  12025		 * Zero the receive header queue so we don't get false
  12026		 * positives when checking the sequence number.  The
  12027		 * sequence numbers could land exactly on the same spot.
  12028		 * E.g. a rcd restart before the receive header wrapped.
  12029		 */
  12030		memset(rcd->rcvhdrq, 0, rcvhdrq_size(rcd));
  12031
  12032		/* starting timeout */
  12033		rcd->rcvavail_timeout = dd->rcv_intr_timeout_csr;
  12034
  12035		/* enable the context */
  12036		rcvctrl |= RCV_CTXT_CTRL_ENABLE_SMASK;
  12037
  12038		/* clean the egr buffer size first */
  12039		rcvctrl &= ~RCV_CTXT_CTRL_EGR_BUF_SIZE_SMASK;
  12040		rcvctrl |= ((u64)encoded_size(rcd->egrbufs.rcvtid_size)
  12041				& RCV_CTXT_CTRL_EGR_BUF_SIZE_MASK)
  12042					<< RCV_CTXT_CTRL_EGR_BUF_SIZE_SHIFT;
  12043
  12044		/* zero RcvHdrHead - set RcvHdrHead.Counter after enable */
  12045		write_uctxt_csr(dd, ctxt, RCV_HDR_HEAD, 0);
  12046		did_enable = 1;
  12047
  12048		/* zero RcvEgrIndexHead */
  12049		write_uctxt_csr(dd, ctxt, RCV_EGR_INDEX_HEAD, 0);
  12050
  12051		/* set eager count and base index */
  12052		reg = (((u64)(rcd->egrbufs.alloced >> RCV_SHIFT)
  12053			& RCV_EGR_CTRL_EGR_CNT_MASK)
  12054		       << RCV_EGR_CTRL_EGR_CNT_SHIFT) |
  12055			(((rcd->eager_base >> RCV_SHIFT)
  12056			  & RCV_EGR_CTRL_EGR_BASE_INDEX_MASK)
  12057			 << RCV_EGR_CTRL_EGR_BASE_INDEX_SHIFT);
  12058		write_kctxt_csr(dd, ctxt, RCV_EGR_CTRL, reg);
  12059
  12060		/*
  12061		 * Set TID (expected) count and base index.
  12062		 * rcd->expected_count is set to individual RcvArray entries,
  12063		 * not pairs, and the CSR takes a pair-count in groups of
  12064		 * four, so divide by 8.
  12065		 */
  12066		reg = (((rcd->expected_count >> RCV_SHIFT)
  12067					& RCV_TID_CTRL_TID_PAIR_CNT_MASK)
  12068				<< RCV_TID_CTRL_TID_PAIR_CNT_SHIFT) |
  12069		      (((rcd->expected_base >> RCV_SHIFT)
  12070					& RCV_TID_CTRL_TID_BASE_INDEX_MASK)
  12071				<< RCV_TID_CTRL_TID_BASE_INDEX_SHIFT);
  12072		write_kctxt_csr(dd, ctxt, RCV_TID_CTRL, reg);
  12073		if (ctxt == HFI1_CTRL_CTXT)
  12074			write_csr(dd, RCV_VL15, HFI1_CTRL_CTXT);
  12075	}
  12076	if (op & HFI1_RCVCTRL_CTXT_DIS) {
  12077		write_csr(dd, RCV_VL15, 0);
  12078		/*
  12079		 * When receive context is being disabled turn on tail
  12080		 * update with a dummy tail address and then disable
  12081		 * receive context.
  12082		 */
  12083		if (dd->rcvhdrtail_dummy_dma) {
  12084			write_kctxt_csr(dd, ctxt, RCV_HDR_TAIL_ADDR,
  12085					dd->rcvhdrtail_dummy_dma);
  12086			/* Enabling RcvCtxtCtrl.TailUpd is intentional. */
  12087			rcvctrl |= RCV_CTXT_CTRL_TAIL_UPD_SMASK;
  12088		}
  12089
  12090		rcvctrl &= ~RCV_CTXT_CTRL_ENABLE_SMASK;
  12091	}
  12092	if (op & HFI1_RCVCTRL_INTRAVAIL_ENB) {
  12093		set_intr_bits(dd, IS_RCVAVAIL_START + rcd->ctxt,
  12094			      IS_RCVAVAIL_START + rcd->ctxt, true);
  12095		rcvctrl |= RCV_CTXT_CTRL_INTR_AVAIL_SMASK;
  12096	}
  12097	if (op & HFI1_RCVCTRL_INTRAVAIL_DIS) {
  12098		set_intr_bits(dd, IS_RCVAVAIL_START + rcd->ctxt,
  12099			      IS_RCVAVAIL_START + rcd->ctxt, false);
  12100		rcvctrl &= ~RCV_CTXT_CTRL_INTR_AVAIL_SMASK;
  12101	}
  12102	if ((op & HFI1_RCVCTRL_TAILUPD_ENB) && hfi1_rcvhdrtail_kvaddr(rcd))
  12103		rcvctrl |= RCV_CTXT_CTRL_TAIL_UPD_SMASK;
  12104	if (op & HFI1_RCVCTRL_TAILUPD_DIS) {
  12105		/* See comment on RcvCtxtCtrl.TailUpd above */
  12106		if (!(op & HFI1_RCVCTRL_CTXT_DIS))
  12107			rcvctrl &= ~RCV_CTXT_CTRL_TAIL_UPD_SMASK;
  12108	}
  12109	if (op & HFI1_RCVCTRL_TIDFLOW_ENB)
  12110		rcvctrl |= RCV_CTXT_CTRL_TID_FLOW_ENABLE_SMASK;
  12111	if (op & HFI1_RCVCTRL_TIDFLOW_DIS)
  12112		rcvctrl &= ~RCV_CTXT_CTRL_TID_FLOW_ENABLE_SMASK;
  12113	if (op & HFI1_RCVCTRL_ONE_PKT_EGR_ENB) {
  12114		/*
  12115		 * In one-packet-per-eager mode, the size comes from
  12116		 * the RcvArray entry.
  12117		 */
  12118		rcvctrl &= ~RCV_CTXT_CTRL_EGR_BUF_SIZE_SMASK;
  12119		rcvctrl |= RCV_CTXT_CTRL_ONE_PACKET_PER_EGR_BUFFER_SMASK;
  12120	}
  12121	if (op & HFI1_RCVCTRL_ONE_PKT_EGR_DIS)
  12122		rcvctrl &= ~RCV_CTXT_CTRL_ONE_PACKET_PER_EGR_BUFFER_SMASK;
  12123	if (op & HFI1_RCVCTRL_NO_RHQ_DROP_ENB)
  12124		rcvctrl |= RCV_CTXT_CTRL_DONT_DROP_RHQ_FULL_SMASK;
  12125	if (op & HFI1_RCVCTRL_NO_RHQ_DROP_DIS)
  12126		rcvctrl &= ~RCV_CTXT_CTRL_DONT_DROP_RHQ_FULL_SMASK;
  12127	if (op & HFI1_RCVCTRL_NO_EGR_DROP_ENB)
  12128		rcvctrl |= RCV_CTXT_CTRL_DONT_DROP_EGR_FULL_SMASK;
  12129	if (op & HFI1_RCVCTRL_NO_EGR_DROP_DIS)
  12130		rcvctrl &= ~RCV_CTXT_CTRL_DONT_DROP_EGR_FULL_SMASK;
  12131	if (op & HFI1_RCVCTRL_URGENT_ENB)
  12132		set_intr_bits(dd, IS_RCVURGENT_START + rcd->ctxt,
  12133			      IS_RCVURGENT_START + rcd->ctxt, true);
  12134	if (op & HFI1_RCVCTRL_URGENT_DIS)
  12135		set_intr_bits(dd, IS_RCVURGENT_START + rcd->ctxt,
  12136			      IS_RCVURGENT_START + rcd->ctxt, false);
  12137
  12138	hfi1_cdbg(RCVCTRL, "ctxt %d rcvctrl 0x%llx\n", ctxt, rcvctrl);
  12139	write_kctxt_csr(dd, ctxt, RCV_CTXT_CTRL, rcvctrl);
  12140
  12141	/* work around sticky RcvCtxtStatus.BlockedRHQFull */
  12142	if (did_enable &&
  12143	    (rcvctrl & RCV_CTXT_CTRL_DONT_DROP_RHQ_FULL_SMASK)) {
  12144		reg = read_kctxt_csr(dd, ctxt, RCV_CTXT_STATUS);
  12145		if (reg != 0) {
  12146			dd_dev_info(dd, "ctxt %d status %lld (blocked)\n",
  12147				    ctxt, reg);
  12148			read_uctxt_csr(dd, ctxt, RCV_HDR_HEAD);
  12149			write_uctxt_csr(dd, ctxt, RCV_HDR_HEAD, 0x10);
  12150			write_uctxt_csr(dd, ctxt, RCV_HDR_HEAD, 0x00);
  12151			read_uctxt_csr(dd, ctxt, RCV_HDR_HEAD);
  12152			reg = read_kctxt_csr(dd, ctxt, RCV_CTXT_STATUS);
  12153			dd_dev_info(dd, "ctxt %d status %lld (%s blocked)\n",
  12154				    ctxt, reg, reg == 0 ? "not" : "still");
  12155		}
  12156	}
  12157
  12158	if (did_enable) {
  12159		/*
  12160		 * The interrupt timeout and count must be set after
  12161		 * the context is enabled to take effect.
  12162		 */
  12163		/* set interrupt timeout */
  12164		write_kctxt_csr(dd, ctxt, RCV_AVAIL_TIME_OUT,
  12165				(u64)rcd->rcvavail_timeout <<
  12166				RCV_AVAIL_TIME_OUT_TIME_OUT_RELOAD_SHIFT);
  12167
  12168		/* set RcvHdrHead.Counter, zero RcvHdrHead.Head (again) */
  12169		reg = (u64)rcv_intr_count << RCV_HDR_HEAD_COUNTER_SHIFT;
  12170		write_uctxt_csr(dd, ctxt, RCV_HDR_HEAD, reg);
  12171	}
  12172
  12173	if (op & (HFI1_RCVCTRL_TAILUPD_DIS | HFI1_RCVCTRL_CTXT_DIS))
  12174		/*
  12175		 * If the context has been disabled and the Tail Update has
  12176		 * been cleared, set the RCV_HDR_TAIL_ADDR CSR to dummy address
  12177		 * so it doesn't contain an address that is invalid.
  12178		 */
  12179		write_kctxt_csr(dd, ctxt, RCV_HDR_TAIL_ADDR,
  12180				dd->rcvhdrtail_dummy_dma);
  12181}
  12182
  12183u32 hfi1_read_cntrs(struct hfi1_devdata *dd, char **namep, u64 **cntrp)
  12184{
  12185	int ret;
  12186	u64 val = 0;
  12187
  12188	if (namep) {
  12189		ret = dd->cntrnameslen;
  12190		*namep = dd->cntrnames;
  12191	} else {
  12192		const struct cntr_entry *entry;
  12193		int i, j;
  12194
  12195		ret = (dd->ndevcntrs) * sizeof(u64);
  12196
  12197		/* Get the start of the block of counters */
  12198		*cntrp = dd->cntrs;
  12199
  12200		/*
  12201		 * Now go and fill in each counter in the block.
  12202		 */
  12203		for (i = 0; i < DEV_CNTR_LAST; i++) {
  12204			entry = &dev_cntrs[i];
  12205			hfi1_cdbg(CNTR, "reading %s", entry->name);
  12206			if (entry->flags & CNTR_DISABLED) {
  12207				/* Nothing */
  12208				hfi1_cdbg(CNTR, "\tDisabled\n");
  12209			} else {
  12210				if (entry->flags & CNTR_VL) {
  12211					hfi1_cdbg(CNTR, "\tPer VL\n");
  12212					for (j = 0; j < C_VL_COUNT; j++) {
  12213						val = entry->rw_cntr(entry,
  12214								  dd, j,
  12215								  CNTR_MODE_R,
  12216								  0);
  12217						hfi1_cdbg(
  12218						   CNTR,
  12219						   "\t\tRead 0x%llx for %d\n",
  12220						   val, j);
  12221						dd->cntrs[entry->offset + j] =
  12222									    val;
  12223					}
  12224				} else if (entry->flags & CNTR_SDMA) {
  12225					hfi1_cdbg(CNTR,
  12226						  "\t Per SDMA Engine\n");
  12227					for (j = 0; j < chip_sdma_engines(dd);
  12228					     j++) {
  12229						val =
  12230						entry->rw_cntr(entry, dd, j,
  12231							       CNTR_MODE_R, 0);
  12232						hfi1_cdbg(CNTR,
  12233							  "\t\tRead 0x%llx for %d\n",
  12234							  val, j);
  12235						dd->cntrs[entry->offset + j] =
  12236									val;
  12237					}
  12238				} else {
  12239					val = entry->rw_cntr(entry, dd,
  12240							CNTR_INVALID_VL,
  12241							CNTR_MODE_R, 0);
  12242					dd->cntrs[entry->offset] = val;
  12243					hfi1_cdbg(CNTR, "\tRead 0x%llx", val);
  12244				}
  12245			}
  12246		}
  12247	}
  12248	return ret;
  12249}
  12250
  12251/*
  12252 * Used by sysfs to create files for hfi stats to read
  12253 */
  12254u32 hfi1_read_portcntrs(struct hfi1_pportdata *ppd, char **namep, u64 **cntrp)
  12255{
  12256	int ret;
  12257	u64 val = 0;
  12258
  12259	if (namep) {
  12260		ret = ppd->dd->portcntrnameslen;
  12261		*namep = ppd->dd->portcntrnames;
  12262	} else {
  12263		const struct cntr_entry *entry;
  12264		int i, j;
  12265
  12266		ret = ppd->dd->nportcntrs * sizeof(u64);
  12267		*cntrp = ppd->cntrs;
  12268
  12269		for (i = 0; i < PORT_CNTR_LAST; i++) {
  12270			entry = &port_cntrs[i];
  12271			hfi1_cdbg(CNTR, "reading %s", entry->name);
  12272			if (entry->flags & CNTR_DISABLED) {
  12273				/* Nothing */
  12274				hfi1_cdbg(CNTR, "\tDisabled\n");
  12275				continue;
  12276			}
  12277
  12278			if (entry->flags & CNTR_VL) {
  12279				hfi1_cdbg(CNTR, "\tPer VL");
  12280				for (j = 0; j < C_VL_COUNT; j++) {
  12281					val = entry->rw_cntr(entry, ppd, j,
  12282							       CNTR_MODE_R,
  12283							       0);
  12284					hfi1_cdbg(
  12285					   CNTR,
  12286					   "\t\tRead 0x%llx for %d",
  12287					   val, j);
  12288					ppd->cntrs[entry->offset + j] = val;
  12289				}
  12290			} else {
  12291				val = entry->rw_cntr(entry, ppd,
  12292						       CNTR_INVALID_VL,
  12293						       CNTR_MODE_R,
  12294						       0);
  12295				ppd->cntrs[entry->offset] = val;
  12296				hfi1_cdbg(CNTR, "\tRead 0x%llx", val);
  12297			}
  12298		}
  12299	}
  12300	return ret;
  12301}
  12302
  12303static void free_cntrs(struct hfi1_devdata *dd)
  12304{
  12305	struct hfi1_pportdata *ppd;
  12306	int i;
  12307
  12308	if (dd->synth_stats_timer.function)
  12309		del_timer_sync(&dd->synth_stats_timer);
  12310	ppd = (struct hfi1_pportdata *)(dd + 1);
  12311	for (i = 0; i < dd->num_pports; i++, ppd++) {
  12312		kfree(ppd->cntrs);
  12313		kfree(ppd->scntrs);
  12314		free_percpu(ppd->ibport_data.rvp.rc_acks);
  12315		free_percpu(ppd->ibport_data.rvp.rc_qacks);
  12316		free_percpu(ppd->ibport_data.rvp.rc_delayed_comp);
  12317		ppd->cntrs = NULL;
  12318		ppd->scntrs = NULL;
  12319		ppd->ibport_data.rvp.rc_acks = NULL;
  12320		ppd->ibport_data.rvp.rc_qacks = NULL;
  12321		ppd->ibport_data.rvp.rc_delayed_comp = NULL;
  12322	}
  12323	kfree(dd->portcntrnames);
  12324	dd->portcntrnames = NULL;
  12325	kfree(dd->cntrs);
  12326	dd->cntrs = NULL;
  12327	kfree(dd->scntrs);
  12328	dd->scntrs = NULL;
  12329	kfree(dd->cntrnames);
  12330	dd->cntrnames = NULL;
  12331	if (dd->update_cntr_wq) {
  12332		destroy_workqueue(dd->update_cntr_wq);
  12333		dd->update_cntr_wq = NULL;
  12334	}
  12335}
  12336
  12337static u64 read_dev_port_cntr(struct hfi1_devdata *dd, struct cntr_entry *entry,
  12338			      u64 *psval, void *context, int vl)
  12339{
  12340	u64 val;
  12341	u64 sval = *psval;
  12342
  12343	if (entry->flags & CNTR_DISABLED) {
  12344		dd_dev_err(dd, "Counter %s not enabled", entry->name);
  12345		return 0;
  12346	}
  12347
  12348	hfi1_cdbg(CNTR, "cntr: %s vl %d psval 0x%llx", entry->name, vl, *psval);
  12349
  12350	val = entry->rw_cntr(entry, context, vl, CNTR_MODE_R, 0);
  12351
  12352	/* If its a synthetic counter there is more work we need to do */
  12353	if (entry->flags & CNTR_SYNTH) {
  12354		if (sval == CNTR_MAX) {
  12355			/* No need to read already saturated */
  12356			return CNTR_MAX;
  12357		}
  12358
  12359		if (entry->flags & CNTR_32BIT) {
  12360			/* 32bit counters can wrap multiple times */
  12361			u64 upper = sval >> 32;
  12362			u64 lower = (sval << 32) >> 32;
  12363
  12364			if (lower > val) { /* hw wrapped */
  12365				if (upper == CNTR_32BIT_MAX)
  12366					val = CNTR_MAX;
  12367				else
  12368					upper++;
  12369			}
  12370
  12371			if (val != CNTR_MAX)
  12372				val = (upper << 32) | val;
  12373
  12374		} else {
  12375			/* If we rolled we are saturated */
  12376			if ((val < sval) || (val > CNTR_MAX))
  12377				val = CNTR_MAX;
  12378		}
  12379	}
  12380
  12381	*psval = val;
  12382
  12383	hfi1_cdbg(CNTR, "\tNew val=0x%llx", val);
  12384
  12385	return val;
  12386}
  12387
  12388static u64 write_dev_port_cntr(struct hfi1_devdata *dd,
  12389			       struct cntr_entry *entry,
  12390			       u64 *psval, void *context, int vl, u64 data)
  12391{
  12392	u64 val;
  12393
  12394	if (entry->flags & CNTR_DISABLED) {
  12395		dd_dev_err(dd, "Counter %s not enabled", entry->name);
  12396		return 0;
  12397	}
  12398
  12399	hfi1_cdbg(CNTR, "cntr: %s vl %d psval 0x%llx", entry->name, vl, *psval);
  12400
  12401	if (entry->flags & CNTR_SYNTH) {
  12402		*psval = data;
  12403		if (entry->flags & CNTR_32BIT) {
  12404			val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W,
  12405					     (data << 32) >> 32);
  12406			val = data; /* return the full 64bit value */
  12407		} else {
  12408			val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W,
  12409					     data);
  12410		}
  12411	} else {
  12412		val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W, data);
  12413	}
  12414
  12415	*psval = val;
  12416
  12417	hfi1_cdbg(CNTR, "\tNew val=0x%llx", val);
  12418
  12419	return val;
  12420}
  12421
  12422u64 read_dev_cntr(struct hfi1_devdata *dd, int index, int vl)
  12423{
  12424	struct cntr_entry *entry;
  12425	u64 *sval;
  12426
  12427	entry = &dev_cntrs[index];
  12428	sval = dd->scntrs + entry->offset;
  12429
  12430	if (vl != CNTR_INVALID_VL)
  12431		sval += vl;
  12432
  12433	return read_dev_port_cntr(dd, entry, sval, dd, vl);
  12434}
  12435
  12436u64 write_dev_cntr(struct hfi1_devdata *dd, int index, int vl, u64 data)
  12437{
  12438	struct cntr_entry *entry;
  12439	u64 *sval;
  12440
  12441	entry = &dev_cntrs[index];
  12442	sval = dd->scntrs + entry->offset;
  12443
  12444	if (vl != CNTR_INVALID_VL)
  12445		sval += vl;
  12446
  12447	return write_dev_port_cntr(dd, entry, sval, dd, vl, data);
  12448}
  12449
  12450u64 read_port_cntr(struct hfi1_pportdata *ppd, int index, int vl)
  12451{
  12452	struct cntr_entry *entry;
  12453	u64 *sval;
  12454
  12455	entry = &port_cntrs[index];
  12456	sval = ppd->scntrs + entry->offset;
  12457
  12458	if (vl != CNTR_INVALID_VL)
  12459		sval += vl;
  12460
  12461	if ((index >= C_RCV_HDR_OVF_FIRST + ppd->dd->num_rcv_contexts) &&
  12462	    (index <= C_RCV_HDR_OVF_LAST)) {
  12463		/* We do not want to bother for disabled contexts */
  12464		return 0;
  12465	}
  12466
  12467	return read_dev_port_cntr(ppd->dd, entry, sval, ppd, vl);
  12468}
  12469
  12470u64 write_port_cntr(struct hfi1_pportdata *ppd, int index, int vl, u64 data)
  12471{
  12472	struct cntr_entry *entry;
  12473	u64 *sval;
  12474
  12475	entry = &port_cntrs[index];
  12476	sval = ppd->scntrs + entry->offset;
  12477
  12478	if (vl != CNTR_INVALID_VL)
  12479		sval += vl;
  12480
  12481	if ((index >= C_RCV_HDR_OVF_FIRST + ppd->dd->num_rcv_contexts) &&
  12482	    (index <= C_RCV_HDR_OVF_LAST)) {
  12483		/* We do not want to bother for disabled contexts */
  12484		return 0;
  12485	}
  12486
  12487	return write_dev_port_cntr(ppd->dd, entry, sval, ppd, vl, data);
  12488}
  12489
  12490static void do_update_synth_timer(struct work_struct *work)
  12491{
  12492	u64 cur_tx;
  12493	u64 cur_rx;
  12494	u64 total_flits;
  12495	u8 update = 0;
  12496	int i, j, vl;
  12497	struct hfi1_pportdata *ppd;
  12498	struct cntr_entry *entry;
  12499	struct hfi1_devdata *dd = container_of(work, struct hfi1_devdata,
  12500					       update_cntr_work);
  12501
  12502	/*
  12503	 * Rather than keep beating on the CSRs pick a minimal set that we can
  12504	 * check to watch for potential roll over. We can do this by looking at
  12505	 * the number of flits sent/recv. If the total flits exceeds 32bits then
  12506	 * we have to iterate all the counters and update.
  12507	 */
  12508	entry = &dev_cntrs[C_DC_RCV_FLITS];
  12509	cur_rx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0);
  12510
  12511	entry = &dev_cntrs[C_DC_XMIT_FLITS];
  12512	cur_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0);
  12513
  12514	hfi1_cdbg(
  12515	    CNTR,
  12516	    "[%d] curr tx=0x%llx rx=0x%llx :: last tx=0x%llx rx=0x%llx\n",
  12517	    dd->unit, cur_tx, cur_rx, dd->last_tx, dd->last_rx);
  12518
  12519	if ((cur_tx < dd->last_tx) || (cur_rx < dd->last_rx)) {
  12520		/*
  12521		 * May not be strictly necessary to update but it won't hurt and
  12522		 * simplifies the logic here.
  12523		 */
  12524		update = 1;
  12525		hfi1_cdbg(CNTR, "[%d] Tripwire counter rolled, updating",
  12526			  dd->unit);
  12527	} else {
  12528		total_flits = (cur_tx - dd->last_tx) + (cur_rx - dd->last_rx);
  12529		hfi1_cdbg(CNTR,
  12530			  "[%d] total flits 0x%llx limit 0x%llx\n", dd->unit,
  12531			  total_flits, (u64)CNTR_32BIT_MAX);
  12532		if (total_flits >= CNTR_32BIT_MAX) {
  12533			hfi1_cdbg(CNTR, "[%d] 32bit limit hit, updating",
  12534				  dd->unit);
  12535			update = 1;
  12536		}
  12537	}
  12538
  12539	if (update) {
  12540		hfi1_cdbg(CNTR, "[%d] Updating dd and ppd counters", dd->unit);
  12541		for (i = 0; i < DEV_CNTR_LAST; i++) {
  12542			entry = &dev_cntrs[i];
  12543			if (entry->flags & CNTR_VL) {
  12544				for (vl = 0; vl < C_VL_COUNT; vl++)
  12545					read_dev_cntr(dd, i, vl);
  12546			} else {
  12547				read_dev_cntr(dd, i, CNTR_INVALID_VL);
  12548			}
  12549		}
  12550		ppd = (struct hfi1_pportdata *)(dd + 1);
  12551		for (i = 0; i < dd->num_pports; i++, ppd++) {
  12552			for (j = 0; j < PORT_CNTR_LAST; j++) {
  12553				entry = &port_cntrs[j];
  12554				if (entry->flags & CNTR_VL) {
  12555					for (vl = 0; vl < C_VL_COUNT; vl++)
  12556						read_port_cntr(ppd, j, vl);
  12557				} else {
  12558					read_port_cntr(ppd, j, CNTR_INVALID_VL);
  12559				}
  12560			}
  12561		}
  12562
  12563		/*
  12564		 * We want the value in the register. The goal is to keep track
  12565		 * of the number of "ticks" not the counter value. In other
  12566		 * words if the register rolls we want to notice it and go ahead
  12567		 * and force an update.
  12568		 */
  12569		entry = &dev_cntrs[C_DC_XMIT_FLITS];
  12570		dd->last_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL,
  12571						CNTR_MODE_R, 0);
  12572
  12573		entry = &dev_cntrs[C_DC_RCV_FLITS];
  12574		dd->last_rx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL,
  12575						CNTR_MODE_R, 0);
  12576
  12577		hfi1_cdbg(CNTR, "[%d] setting last tx/rx to 0x%llx 0x%llx",
  12578			  dd->unit, dd->last_tx, dd->last_rx);
  12579
  12580	} else {
  12581		hfi1_cdbg(CNTR, "[%d] No update necessary", dd->unit);
  12582	}
  12583}
  12584
  12585static void update_synth_timer(struct timer_list *t)
  12586{
  12587	struct hfi1_devdata *dd = from_timer(dd, t, synth_stats_timer);
  12588
  12589	queue_work(dd->update_cntr_wq, &dd->update_cntr_work);
  12590	mod_timer(&dd->synth_stats_timer, jiffies + HZ * SYNTH_CNT_TIME);
  12591}
  12592
  12593#define C_MAX_NAME 16 /* 15 chars + one for /0 */
  12594static int init_cntrs(struct hfi1_devdata *dd)
  12595{
  12596	int i, rcv_ctxts, j;
  12597	size_t sz;
  12598	char *p;
  12599	char name[C_MAX_NAME];
  12600	struct hfi1_pportdata *ppd;
  12601	const char *bit_type_32 = ",32";
  12602	const int bit_type_32_sz = strlen(bit_type_32);
  12603	u32 sdma_engines = chip_sdma_engines(dd);
  12604
  12605	/* set up the stats timer; the add_timer is done at the end */
  12606	timer_setup(&dd->synth_stats_timer, update_synth_timer, 0);
  12607
  12608	/***********************/
  12609	/* per device counters */
  12610	/***********************/
  12611
  12612	/* size names and determine how many we have*/
  12613	dd->ndevcntrs = 0;
  12614	sz = 0;
  12615
  12616	for (i = 0; i < DEV_CNTR_LAST; i++) {
  12617		if (dev_cntrs[i].flags & CNTR_DISABLED) {
  12618			hfi1_dbg_early("\tSkipping %s\n", dev_cntrs[i].name);
  12619			continue;
  12620		}
  12621
  12622		if (dev_cntrs[i].flags & CNTR_VL) {
  12623			dev_cntrs[i].offset = dd->ndevcntrs;
  12624			for (j = 0; j < C_VL_COUNT; j++) {
  12625				snprintf(name, C_MAX_NAME, "%s%d",
  12626					 dev_cntrs[i].name, vl_from_idx(j));
  12627				sz += strlen(name);
  12628				/* Add ",32" for 32-bit counters */
  12629				if (dev_cntrs[i].flags & CNTR_32BIT)
  12630					sz += bit_type_32_sz;
  12631				sz++;
  12632				dd->ndevcntrs++;
  12633			}
  12634		} else if (dev_cntrs[i].flags & CNTR_SDMA) {
  12635			dev_cntrs[i].offset = dd->ndevcntrs;
  12636			for (j = 0; j < sdma_engines; j++) {
  12637				snprintf(name, C_MAX_NAME, "%s%d",
  12638					 dev_cntrs[i].name, j);
  12639				sz += strlen(name);
  12640				/* Add ",32" for 32-bit counters */
  12641				if (dev_cntrs[i].flags & CNTR_32BIT)
  12642					sz += bit_type_32_sz;
  12643				sz++;
  12644				dd->ndevcntrs++;
  12645			}
  12646		} else {
  12647			/* +1 for newline. */
  12648			sz += strlen(dev_cntrs[i].name) + 1;
  12649			/* Add ",32" for 32-bit counters */
  12650			if (dev_cntrs[i].flags & CNTR_32BIT)
  12651				sz += bit_type_32_sz;
  12652			dev_cntrs[i].offset = dd->ndevcntrs;
  12653			dd->ndevcntrs++;
  12654		}
  12655	}
  12656
  12657	/* allocate space for the counter values */
  12658	dd->cntrs = kcalloc(dd->ndevcntrs + num_driver_cntrs, sizeof(u64),
  12659			    GFP_KERNEL);
  12660	if (!dd->cntrs)
  12661		goto bail;
  12662
  12663	dd->scntrs = kcalloc(dd->ndevcntrs, sizeof(u64), GFP_KERNEL);
  12664	if (!dd->scntrs)
  12665		goto bail;
  12666
  12667	/* allocate space for the counter names */
  12668	dd->cntrnameslen = sz;
  12669	dd->cntrnames = kmalloc(sz, GFP_KERNEL);
  12670	if (!dd->cntrnames)
  12671		goto bail;
  12672
  12673	/* fill in the names */
  12674	for (p = dd->cntrnames, i = 0; i < DEV_CNTR_LAST; i++) {
  12675		if (dev_cntrs[i].flags & CNTR_DISABLED) {
  12676			/* Nothing */
  12677		} else if (dev_cntrs[i].flags & CNTR_VL) {
  12678			for (j = 0; j < C_VL_COUNT; j++) {
  12679				snprintf(name, C_MAX_NAME, "%s%d",
  12680					 dev_cntrs[i].name,
  12681					 vl_from_idx(j));
  12682				memcpy(p, name, strlen(name));
  12683				p += strlen(name);
  12684
  12685				/* Counter is 32 bits */
  12686				if (dev_cntrs[i].flags & CNTR_32BIT) {
  12687					memcpy(p, bit_type_32, bit_type_32_sz);
  12688					p += bit_type_32_sz;
  12689				}
  12690
  12691				*p++ = '\n';
  12692			}
  12693		} else if (dev_cntrs[i].flags & CNTR_SDMA) {
  12694			for (j = 0; j < sdma_engines; j++) {
  12695				snprintf(name, C_MAX_NAME, "%s%d",
  12696					 dev_cntrs[i].name, j);
  12697				memcpy(p, name, strlen(name));
  12698				p += strlen(name);
  12699
  12700				/* Counter is 32 bits */
  12701				if (dev_cntrs[i].flags & CNTR_32BIT) {
  12702					memcpy(p, bit_type_32, bit_type_32_sz);
  12703					p += bit_type_32_sz;
  12704				}
  12705
  12706				*p++ = '\n';
  12707			}
  12708		} else {
  12709			memcpy(p, dev_cntrs[i].name, strlen(dev_cntrs[i].name));
  12710			p += strlen(dev_cntrs[i].name);
  12711
  12712			/* Counter is 32 bits */
  12713			if (dev_cntrs[i].flags & CNTR_32BIT) {
  12714				memcpy(p, bit_type_32, bit_type_32_sz);
  12715				p += bit_type_32_sz;
  12716			}
  12717
  12718			*p++ = '\n';
  12719		}
  12720	}
  12721
  12722	/*********************/
  12723	/* per port counters */
  12724	/*********************/
  12725
  12726	/*
  12727	 * Go through the counters for the overflows and disable the ones we
  12728	 * don't need. This varies based on platform so we need to do it
  12729	 * dynamically here.
  12730	 */
  12731	rcv_ctxts = dd->num_rcv_contexts;
  12732	for (i = C_RCV_HDR_OVF_FIRST + rcv_ctxts;
  12733	     i <= C_RCV_HDR_OVF_LAST; i++) {
  12734		port_cntrs[i].flags |= CNTR_DISABLED;
  12735	}
  12736
  12737	/* size port counter names and determine how many we have*/
  12738	sz = 0;
  12739	dd->nportcntrs = 0;
  12740	for (i = 0; i < PORT_CNTR_LAST; i++) {
  12741		if (port_cntrs[i].flags & CNTR_DISABLED) {
  12742			hfi1_dbg_early("\tSkipping %s\n", port_cntrs[i].name);
  12743			continue;
  12744		}
  12745
  12746		if (port_cntrs[i].flags & CNTR_VL) {
  12747			port_cntrs[i].offset = dd->nportcntrs;
  12748			for (j = 0; j < C_VL_COUNT; j++) {
  12749				snprintf(name, C_MAX_NAME, "%s%d",
  12750					 port_cntrs[i].name, vl_from_idx(j));
  12751				sz += strlen(name);
  12752				/* Add ",32" for 32-bit counters */
  12753				if (port_cntrs[i].flags & CNTR_32BIT)
  12754					sz += bit_type_32_sz;
  12755				sz++;
  12756				dd->nportcntrs++;
  12757			}
  12758		} else {
  12759			/* +1 for newline */
  12760			sz += strlen(port_cntrs[i].name) + 1;
  12761			/* Add ",32" for 32-bit counters */
  12762			if (port_cntrs[i].flags & CNTR_32BIT)
  12763				sz += bit_type_32_sz;
  12764			port_cntrs[i].offset = dd->nportcntrs;
  12765			dd->nportcntrs++;
  12766		}
  12767	}
  12768
  12769	/* allocate space for the counter names */
  12770	dd->portcntrnameslen = sz;
  12771	dd->portcntrnames = kmalloc(sz, GFP_KERNEL);
  12772	if (!dd->portcntrnames)
  12773		goto bail;
  12774
  12775	/* fill in port cntr names */
  12776	for (p = dd->portcntrnames, i = 0; i < PORT_CNTR_LAST; i++) {
  12777		if (port_cntrs[i].flags & CNTR_DISABLED)
  12778			continue;
  12779
  12780		if (port_cntrs[i].flags & CNTR_VL) {
  12781			for (j = 0; j < C_VL_COUNT; j++) {
  12782				snprintf(name, C_MAX_NAME, "%s%d",
  12783					 port_cntrs[i].name, vl_from_idx(j));
  12784				memcpy(p, name, strlen(name));
  12785				p += strlen(name);
  12786
  12787				/* Counter is 32 bits */
  12788				if (port_cntrs[i].flags & CNTR_32BIT) {
  12789					memcpy(p, bit_type_32, bit_type_32_sz);
  12790					p += bit_type_32_sz;
  12791				}
  12792
  12793				*p++ = '\n';
  12794			}
  12795		} else {
  12796			memcpy(p, port_cntrs[i].name,
  12797			       strlen(port_cntrs[i].name));
  12798			p += strlen(port_cntrs[i].name);
  12799
  12800			/* Counter is 32 bits */
  12801			if (port_cntrs[i].flags & CNTR_32BIT) {
  12802				memcpy(p, bit_type_32, bit_type_32_sz);
  12803				p += bit_type_32_sz;
  12804			}
  12805
  12806			*p++ = '\n';
  12807		}
  12808	}
  12809
  12810	/* allocate per port storage for counter values */
  12811	ppd = (struct hfi1_pportdata *)(dd + 1);
  12812	for (i = 0; i < dd->num_pports; i++, ppd++) {
  12813		ppd->cntrs = kcalloc(dd->nportcntrs, sizeof(u64), GFP_KERNEL);
  12814		if (!ppd->cntrs)
  12815			goto bail;
  12816
  12817		ppd->scntrs = kcalloc(dd->nportcntrs, sizeof(u64), GFP_KERNEL);
  12818		if (!ppd->scntrs)
  12819			goto bail;
  12820	}
  12821
  12822	/* CPU counters need to be allocated and zeroed */
  12823	if (init_cpu_counters(dd))
  12824		goto bail;
  12825
  12826	dd->update_cntr_wq = alloc_ordered_workqueue("hfi1_update_cntr_%d",
  12827						     WQ_MEM_RECLAIM, dd->unit);
  12828	if (!dd->update_cntr_wq)
  12829		goto bail;
  12830
  12831	INIT_WORK(&dd->update_cntr_work, do_update_synth_timer);
  12832
  12833	mod_timer(&dd->synth_stats_timer, jiffies + HZ * SYNTH_CNT_TIME);
  12834	return 0;
  12835bail:
  12836	free_cntrs(dd);
  12837	return -ENOMEM;
  12838}
  12839
  12840static u32 chip_to_opa_lstate(struct hfi1_devdata *dd, u32 chip_lstate)
  12841{
  12842	switch (chip_lstate) {
  12843	case LSTATE_DOWN:
  12844		return IB_PORT_DOWN;
  12845	case LSTATE_INIT:
  12846		return IB_PORT_INIT;
  12847	case LSTATE_ARMED:
  12848		return IB_PORT_ARMED;
  12849	case LSTATE_ACTIVE:
  12850		return IB_PORT_ACTIVE;
  12851	default:
  12852		dd_dev_err(dd,
  12853			   "Unknown logical state 0x%x, reporting IB_PORT_DOWN\n",
  12854			   chip_lstate);
  12855		return IB_PORT_DOWN;
  12856	}
  12857}
  12858
  12859u32 chip_to_opa_pstate(struct hfi1_devdata *dd, u32 chip_pstate)
  12860{
  12861	/* look at the HFI meta-states only */
  12862	switch (chip_pstate & 0xf0) {
  12863	case PLS_DISABLED:
  12864		return IB_PORTPHYSSTATE_DISABLED;
  12865	case PLS_OFFLINE:
  12866		return OPA_PORTPHYSSTATE_OFFLINE;
  12867	case PLS_POLLING:
  12868		return IB_PORTPHYSSTATE_POLLING;
  12869	case PLS_CONFIGPHY:
  12870		return IB_PORTPHYSSTATE_TRAINING;
  12871	case PLS_LINKUP:
  12872		return IB_PORTPHYSSTATE_LINKUP;
  12873	case PLS_PHYTEST:
  12874		return IB_PORTPHYSSTATE_PHY_TEST;
  12875	default:
  12876		dd_dev_err(dd, "Unexpected chip physical state of 0x%x\n",
  12877			   chip_pstate);
  12878		return IB_PORTPHYSSTATE_DISABLED;
  12879	}
  12880}
  12881
  12882/* return the OPA port logical state name */
  12883const char *opa_lstate_name(u32 lstate)
  12884{
  12885	static const char * const port_logical_names[] = {
  12886		"PORT_NOP",
  12887		"PORT_DOWN",
  12888		"PORT_INIT",
  12889		"PORT_ARMED",
  12890		"PORT_ACTIVE",
  12891		"PORT_ACTIVE_DEFER",
  12892	};
  12893	if (lstate < ARRAY_SIZE(port_logical_names))
  12894		return port_logical_names[lstate];
  12895	return "unknown";
  12896}
  12897
  12898/* return the OPA port physical state name */
  12899const char *opa_pstate_name(u32 pstate)
  12900{
  12901	static const char * const port_physical_names[] = {
  12902		"PHYS_NOP",
  12903		"reserved1",
  12904		"PHYS_POLL",
  12905		"PHYS_DISABLED",
  12906		"PHYS_TRAINING",
  12907		"PHYS_LINKUP",
  12908		"PHYS_LINK_ERR_RECOVER",
  12909		"PHYS_PHY_TEST",
  12910		"reserved8",
  12911		"PHYS_OFFLINE",
  12912		"PHYS_GANGED",
  12913		"PHYS_TEST",
  12914	};
  12915	if (pstate < ARRAY_SIZE(port_physical_names))
  12916		return port_physical_names[pstate];
  12917	return "unknown";
  12918}
  12919
  12920/**
  12921 * update_statusp - Update userspace status flag
  12922 * @ppd: Port data structure
  12923 * @state: port state information
  12924 *
  12925 * Actual port status is determined by the host_link_state value
  12926 * in the ppd.
  12927 *
  12928 * host_link_state MUST be updated before updating the user space
  12929 * statusp.
  12930 */
  12931static void update_statusp(struct hfi1_pportdata *ppd, u32 state)
  12932{
  12933	/*
  12934	 * Set port status flags in the page mapped into userspace
  12935	 * memory. Do it here to ensure a reliable state - this is
  12936	 * the only function called by all state handling code.
  12937	 * Always set the flags due to the fact that the cache value
  12938	 * might have been changed explicitly outside of this
  12939	 * function.
  12940	 */
  12941	if (ppd->statusp) {
  12942		switch (state) {
  12943		case IB_PORT_DOWN:
  12944		case IB_PORT_INIT:
  12945			*ppd->statusp &= ~(HFI1_STATUS_IB_CONF |
  12946					   HFI1_STATUS_IB_READY);
  12947			break;
  12948		case IB_PORT_ARMED:
  12949			*ppd->statusp |= HFI1_STATUS_IB_CONF;
  12950			break;
  12951		case IB_PORT_ACTIVE:
  12952			*ppd->statusp |= HFI1_STATUS_IB_READY;
  12953			break;
  12954		}
  12955	}
  12956	dd_dev_info(ppd->dd, "logical state changed to %s (0x%x)\n",
  12957		    opa_lstate_name(state), state);
  12958}
  12959
  12960/**
  12961 * wait_logical_linkstate - wait for an IB link state change to occur
  12962 * @ppd: port device
  12963 * @state: the state to wait for
  12964 * @msecs: the number of milliseconds to wait
  12965 *
  12966 * Wait up to msecs milliseconds for IB link state change to occur.
  12967 * For now, take the easy polling route.
  12968 * Returns 0 if state reached, otherwise -ETIMEDOUT.
  12969 */
  12970static int wait_logical_linkstate(struct hfi1_pportdata *ppd, u32 state,
  12971				  int msecs)
  12972{
  12973	unsigned long timeout;
  12974	u32 new_state;
  12975
  12976	timeout = jiffies + msecs_to_jiffies(msecs);
  12977	while (1) {
  12978		new_state = chip_to_opa_lstate(ppd->dd,
  12979					       read_logical_state(ppd->dd));
  12980		if (new_state == state)
  12981			break;
  12982		if (time_after(jiffies, timeout)) {
  12983			dd_dev_err(ppd->dd,
  12984				   "timeout waiting for link state 0x%x\n",
  12985				   state);
  12986			return -ETIMEDOUT;
  12987		}
  12988		msleep(20);
  12989	}
  12990
  12991	return 0;
  12992}
  12993
  12994static void log_state_transition(struct hfi1_pportdata *ppd, u32 state)
  12995{
  12996	u32 ib_pstate = chip_to_opa_pstate(ppd->dd, state);
  12997
  12998	dd_dev_info(ppd->dd,
  12999		    "physical state changed to %s (0x%x), phy 0x%x\n",
  13000		    opa_pstate_name(ib_pstate), ib_pstate, state);
  13001}
  13002
  13003/*
  13004 * Read the physical hardware link state and check if it matches host
  13005 * drivers anticipated state.
  13006 */
  13007static void log_physical_state(struct hfi1_pportdata *ppd, u32 state)
  13008{
  13009	u32 read_state = read_physical_state(ppd->dd);
  13010
  13011	if (read_state == state) {
  13012		log_state_transition(ppd, state);
  13013	} else {
  13014		dd_dev_err(ppd->dd,
  13015			   "anticipated phy link state 0x%x, read 0x%x\n",
  13016			   state, read_state);
  13017	}
  13018}
  13019
  13020/*
  13021 * wait_physical_linkstate - wait for an physical link state change to occur
  13022 * @ppd: port device
  13023 * @state: the state to wait for
  13024 * @msecs: the number of milliseconds to wait
  13025 *
  13026 * Wait up to msecs milliseconds for physical link state change to occur.
  13027 * Returns 0 if state reached, otherwise -ETIMEDOUT.
  13028 */
  13029static int wait_physical_linkstate(struct hfi1_pportdata *ppd, u32 state,
  13030				   int msecs)
  13031{
  13032	u32 read_state;
  13033	unsigned long timeout;
  13034
  13035	timeout = jiffies + msecs_to_jiffies(msecs);
  13036	while (1) {
  13037		read_state = read_physical_state(ppd->dd);
  13038		if (read_state == state)
  13039			break;
  13040		if (time_after(jiffies, timeout)) {
  13041			dd_dev_err(ppd->dd,
  13042				   "timeout waiting for phy link state 0x%x\n",
  13043				   state);
  13044			return -ETIMEDOUT;
  13045		}
  13046		usleep_range(1950, 2050); /* sleep 2ms-ish */
  13047	}
  13048
  13049	log_state_transition(ppd, state);
  13050	return 0;
  13051}
  13052
  13053/*
  13054 * wait_phys_link_offline_quiet_substates - wait for any offline substate
  13055 * @ppd: port device
  13056 * @msecs: the number of milliseconds to wait
  13057 *
  13058 * Wait up to msecs milliseconds for any offline physical link
  13059 * state change to occur.
  13060 * Returns 0 if at least one state is reached, otherwise -ETIMEDOUT.
  13061 */
  13062static int wait_phys_link_offline_substates(struct hfi1_pportdata *ppd,
  13063					    int msecs)
  13064{
  13065	u32 read_state;
  13066	unsigned long timeout;
  13067
  13068	timeout = jiffies + msecs_to_jiffies(msecs);
  13069	while (1) {
  13070		read_state = read_physical_state(ppd->dd);
  13071		if ((read_state & 0xF0) == PLS_OFFLINE)
  13072			break;
  13073		if (time_after(jiffies, timeout)) {
  13074			dd_dev_err(ppd->dd,
  13075				   "timeout waiting for phy link offline.quiet substates. Read state 0x%x, %dms\n",
  13076				   read_state, msecs);
  13077			return -ETIMEDOUT;
  13078		}
  13079		usleep_range(1950, 2050); /* sleep 2ms-ish */
  13080	}
  13081
  13082	log_state_transition(ppd, read_state);
  13083	return read_state;
  13084}
  13085
  13086/*
  13087 * wait_phys_link_out_of_offline - wait for any out of offline state
  13088 * @ppd: port device
  13089 * @msecs: the number of milliseconds to wait
  13090 *
  13091 * Wait up to msecs milliseconds for any out of offline physical link
  13092 * state change to occur.
  13093 * Returns 0 if at least one state is reached, otherwise -ETIMEDOUT.
  13094 */
  13095static int wait_phys_link_out_of_offline(struct hfi1_pportdata *ppd,
  13096					 int msecs)
  13097{
  13098	u32 read_state;
  13099	unsigned long timeout;
  13100
  13101	timeout = jiffies + msecs_to_jiffies(msecs);
  13102	while (1) {
  13103		read_state = read_physical_state(ppd->dd);
  13104		if ((read_state & 0xF0) != PLS_OFFLINE)
  13105			break;
  13106		if (time_after(jiffies, timeout)) {
  13107			dd_dev_err(ppd->dd,
  13108				   "timeout waiting for phy link out of offline. Read state 0x%x, %dms\n",
  13109				   read_state, msecs);
  13110			return -ETIMEDOUT;
  13111		}
  13112		usleep_range(1950, 2050); /* sleep 2ms-ish */
  13113	}
  13114
  13115	log_state_transition(ppd, read_state);
  13116	return read_state;
  13117}
  13118
  13119#define CLEAR_STATIC_RATE_CONTROL_SMASK(r) \
  13120(r &= ~SEND_CTXT_CHECK_ENABLE_DISALLOW_PBC_STATIC_RATE_CONTROL_SMASK)
  13121
  13122#define SET_STATIC_RATE_CONTROL_SMASK(r) \
  13123(r |= SEND_CTXT_CHECK_ENABLE_DISALLOW_PBC_STATIC_RATE_CONTROL_SMASK)
  13124
  13125void hfi1_init_ctxt(struct send_context *sc)
  13126{
  13127	if (sc) {
  13128		struct hfi1_devdata *dd = sc->dd;
  13129		u64 reg;
  13130		u8 set = (sc->type == SC_USER ?
  13131			  HFI1_CAP_IS_USET(STATIC_RATE_CTRL) :
  13132			  HFI1_CAP_IS_KSET(STATIC_RATE_CTRL));
  13133		reg = read_kctxt_csr(dd, sc->hw_context,
  13134				     SEND_CTXT_CHECK_ENABLE);
  13135		if (set)
  13136			CLEAR_STATIC_RATE_CONTROL_SMASK(reg);
  13137		else
  13138			SET_STATIC_RATE_CONTROL_SMASK(reg);
  13139		write_kctxt_csr(dd, sc->hw_context,
  13140				SEND_CTXT_CHECK_ENABLE, reg);
  13141	}
  13142}
  13143
  13144int hfi1_tempsense_rd(struct hfi1_devdata *dd, struct hfi1_temp *temp)
  13145{
  13146	int ret = 0;
  13147	u64 reg;
  13148
  13149	if (dd->icode != ICODE_RTL_SILICON) {
  13150		if (HFI1_CAP_IS_KSET(PRINT_UNIMPL))
  13151			dd_dev_info(dd, "%s: tempsense not supported by HW\n",
  13152				    __func__);
  13153		return -EINVAL;
  13154	}
  13155	reg = read_csr(dd, ASIC_STS_THERM);
  13156	temp->curr = ((reg >> ASIC_STS_THERM_CURR_TEMP_SHIFT) &
  13157		      ASIC_STS_THERM_CURR_TEMP_MASK);
  13158	temp->lo_lim = ((reg >> ASIC_STS_THERM_LO_TEMP_SHIFT) &
  13159			ASIC_STS_THERM_LO_TEMP_MASK);
  13160	temp->hi_lim = ((reg >> ASIC_STS_THERM_HI_TEMP_SHIFT) &
  13161			ASIC_STS_THERM_HI_TEMP_MASK);
  13162	temp->crit_lim = ((reg >> ASIC_STS_THERM_CRIT_TEMP_SHIFT) &
  13163			  ASIC_STS_THERM_CRIT_TEMP_MASK);
  13164	/* triggers is a 3-bit value - 1 bit per trigger. */
  13165	temp->triggers = (u8)((reg >> ASIC_STS_THERM_LOW_SHIFT) & 0x7);
  13166
  13167	return ret;
  13168}
  13169
  13170/* ========================================================================= */
  13171
  13172/**
  13173 * read_mod_write() - Calculate the IRQ register index and set/clear the bits
  13174 * @dd: valid devdata
  13175 * @src: IRQ source to determine register index from
  13176 * @bits: the bits to set or clear
  13177 * @set: true == set the bits, false == clear the bits
  13178 *
  13179 */
  13180static void read_mod_write(struct hfi1_devdata *dd, u16 src, u64 bits,
  13181			   bool set)
  13182{
  13183	u64 reg;
  13184	u16 idx = src / BITS_PER_REGISTER;
  13185
  13186	spin_lock(&dd->irq_src_lock);
  13187	reg = read_csr(dd, CCE_INT_MASK + (8 * idx));
  13188	if (set)
  13189		reg |= bits;
  13190	else
  13191		reg &= ~bits;
  13192	write_csr(dd, CCE_INT_MASK + (8 * idx), reg);
  13193	spin_unlock(&dd->irq_src_lock);
  13194}
  13195
  13196/**
  13197 * set_intr_bits() - Enable/disable a range (one or more) IRQ sources
  13198 * @dd: valid devdata
  13199 * @first: first IRQ source to set/clear
  13200 * @last: last IRQ source (inclusive) to set/clear
  13201 * @set: true == set the bits, false == clear the bits
  13202 *
  13203 * If first == last, set the exact source.
  13204 */
  13205int set_intr_bits(struct hfi1_devdata *dd, u16 first, u16 last, bool set)
  13206{
  13207	u64 bits = 0;
  13208	u64 bit;
  13209	u16 src;
  13210
  13211	if (first > NUM_INTERRUPT_SOURCES || last > NUM_INTERRUPT_SOURCES)
  13212		return -EINVAL;
  13213
  13214	if (last < first)
  13215		return -ERANGE;
  13216
  13217	for (src = first; src <= last; src++) {
  13218		bit = src % BITS_PER_REGISTER;
  13219		/* wrapped to next register? */
  13220		if (!bit && bits) {
  13221			read_mod_write(dd, src - 1, bits, set);
  13222			bits = 0;
  13223		}
  13224		bits |= BIT_ULL(bit);
  13225	}
  13226	read_mod_write(dd, last, bits, set);
  13227
  13228	return 0;
  13229}
  13230
  13231/*
  13232 * Clear all interrupt sources on the chip.
  13233 */
  13234void clear_all_interrupts(struct hfi1_devdata *dd)
  13235{
  13236	int i;
  13237
  13238	for (i = 0; i < CCE_NUM_INT_CSRS; i++)
  13239		write_csr(dd, CCE_INT_CLEAR + (8 * i), ~(u64)0);
  13240
  13241	write_csr(dd, CCE_ERR_CLEAR, ~(u64)0);
  13242	write_csr(dd, MISC_ERR_CLEAR, ~(u64)0);
  13243	write_csr(dd, RCV_ERR_CLEAR, ~(u64)0);
  13244	write_csr(dd, SEND_ERR_CLEAR, ~(u64)0);
  13245	write_csr(dd, SEND_PIO_ERR_CLEAR, ~(u64)0);
  13246	write_csr(dd, SEND_DMA_ERR_CLEAR, ~(u64)0);
  13247	write_csr(dd, SEND_EGRESS_ERR_CLEAR, ~(u64)0);
  13248	for (i = 0; i < chip_send_contexts(dd); i++)
  13249		write_kctxt_csr(dd, i, SEND_CTXT_ERR_CLEAR, ~(u64)0);
  13250	for (i = 0; i < chip_sdma_engines(dd); i++)
  13251		write_kctxt_csr(dd, i, SEND_DMA_ENG_ERR_CLEAR, ~(u64)0);
  13252
  13253	write_csr(dd, DCC_ERR_FLG_CLR, ~(u64)0);
  13254	write_csr(dd, DC_LCB_ERR_CLR, ~(u64)0);
  13255	write_csr(dd, DC_DC8051_ERR_CLR, ~(u64)0);
  13256}
  13257
  13258/*
  13259 * Remap the interrupt source from the general handler to the given MSI-X
  13260 * interrupt.
  13261 */
  13262void remap_intr(struct hfi1_devdata *dd, int isrc, int msix_intr)
  13263{
  13264	u64 reg;
  13265	int m, n;
  13266
  13267	/* clear from the handled mask of the general interrupt */
  13268	m = isrc / 64;
  13269	n = isrc % 64;
  13270	if (likely(m < CCE_NUM_INT_CSRS)) {
  13271		dd->gi_mask[m] &= ~((u64)1 << n);
  13272	} else {
  13273		dd_dev_err(dd, "remap interrupt err\n");
  13274		return;
  13275	}
  13276
  13277	/* direct the chip source to the given MSI-X interrupt */
  13278	m = isrc / 8;
  13279	n = isrc % 8;
  13280	reg = read_csr(dd, CCE_INT_MAP + (8 * m));
  13281	reg &= ~((u64)0xff << (8 * n));
  13282	reg |= ((u64)msix_intr & 0xff) << (8 * n);
  13283	write_csr(dd, CCE_INT_MAP + (8 * m), reg);
  13284}
  13285
  13286void remap_sdma_interrupts(struct hfi1_devdata *dd, int engine, int msix_intr)
  13287{
  13288	/*
  13289	 * SDMA engine interrupt sources grouped by type, rather than
  13290	 * engine.  Per-engine interrupts are as follows:
  13291	 *	SDMA
  13292	 *	SDMAProgress
  13293	 *	SDMAIdle
  13294	 */
  13295	remap_intr(dd, IS_SDMA_START + engine, msix_intr);
  13296	remap_intr(dd, IS_SDMA_PROGRESS_START + engine, msix_intr);
  13297	remap_intr(dd, IS_SDMA_IDLE_START + engine, msix_intr);
  13298}
  13299
  13300/*
  13301 * Set the general handler to accept all interrupts, remap all
  13302 * chip interrupts back to MSI-X 0.
  13303 */
  13304void reset_interrupts(struct hfi1_devdata *dd)
  13305{
  13306	int i;
  13307
  13308	/* all interrupts handled by the general handler */
  13309	for (i = 0; i < CCE_NUM_INT_CSRS; i++)
  13310		dd->gi_mask[i] = ~(u64)0;
  13311
  13312	/* all chip interrupts map to MSI-X 0 */
  13313	for (i = 0; i < CCE_NUM_INT_MAP_CSRS; i++)
  13314		write_csr(dd, CCE_INT_MAP + (8 * i), 0);
  13315}
  13316
  13317/**
  13318 * set_up_interrupts() - Initialize the IRQ resources and state
  13319 * @dd: valid devdata
  13320 *
  13321 */
  13322static int set_up_interrupts(struct hfi1_devdata *dd)
  13323{
  13324	int ret;
  13325
  13326	/* mask all interrupts */
  13327	set_intr_bits(dd, IS_FIRST_SOURCE, IS_LAST_SOURCE, false);
  13328
  13329	/* clear all pending interrupts */
  13330	clear_all_interrupts(dd);
  13331
  13332	/* reset general handler mask, chip MSI-X mappings */
  13333	reset_interrupts(dd);
  13334
  13335	/* ask for MSI-X interrupts */
  13336	ret = msix_initialize(dd);
  13337	if (ret)
  13338		return ret;
  13339
  13340	ret = msix_request_irqs(dd);
  13341	if (ret)
  13342		msix_clean_up_interrupts(dd);
  13343
  13344	return ret;
  13345}
  13346
  13347/*
  13348 * Set up context values in dd.  Sets:
  13349 *
  13350 *	num_rcv_contexts - number of contexts being used
  13351 *	n_krcv_queues - number of kernel contexts
  13352 *	first_dyn_alloc_ctxt - first dynamically allocated context
  13353 *                             in array of contexts
  13354 *	freectxts  - number of free user contexts
  13355 *	num_send_contexts - number of PIO send contexts being used
  13356 *	num_netdev_contexts - number of contexts reserved for netdev
  13357 */
  13358static int set_up_context_variables(struct hfi1_devdata *dd)
  13359{
  13360	unsigned long num_kernel_contexts;
  13361	u16 num_netdev_contexts;
  13362	int ret;
  13363	unsigned ngroups;
  13364	int rmt_count;
  13365	int user_rmt_reduced;
  13366	u32 n_usr_ctxts;
  13367	u32 send_contexts = chip_send_contexts(dd);
  13368	u32 rcv_contexts = chip_rcv_contexts(dd);
  13369
  13370	/*
  13371	 * Kernel receive contexts:
  13372	 * - Context 0 - control context (VL15/multicast/error)
  13373	 * - Context 1 - first kernel context
  13374	 * - Context 2 - second kernel context
  13375	 * ...
  13376	 */
  13377	if (n_krcvqs)
  13378		/*
  13379		 * n_krcvqs is the sum of module parameter kernel receive
  13380		 * contexts, krcvqs[].  It does not include the control
  13381		 * context, so add that.
  13382		 */
  13383		num_kernel_contexts = n_krcvqs + 1;
  13384	else
  13385		num_kernel_contexts = DEFAULT_KRCVQS + 1;
  13386	/*
  13387	 * Every kernel receive context needs an ACK send context.
  13388	 * one send context is allocated for each VL{0-7} and VL15
  13389	 */
  13390	if (num_kernel_contexts > (send_contexts - num_vls - 1)) {
  13391		dd_dev_err(dd,
  13392			   "Reducing # kernel rcv contexts to: %d, from %lu\n",
  13393			   send_contexts - num_vls - 1,
  13394			   num_kernel_contexts);
  13395		num_kernel_contexts = send_contexts - num_vls - 1;
  13396	}
  13397
  13398	/*
  13399	 * User contexts:
  13400	 *	- default to 1 user context per real (non-HT) CPU core if
  13401	 *	  num_user_contexts is negative
  13402	 */
  13403	if (num_user_contexts < 0)
  13404		n_usr_ctxts = cpumask_weight(&node_affinity.real_cpu_mask);
  13405	else
  13406		n_usr_ctxts = num_user_contexts;
  13407	/*
  13408	 * Adjust the counts given a global max.
  13409	 */
  13410	if (num_kernel_contexts + n_usr_ctxts > rcv_contexts) {
  13411		dd_dev_err(dd,
  13412			   "Reducing # user receive contexts to: %u, from %u\n",
  13413			   (u32)(rcv_contexts - num_kernel_contexts),
  13414			   n_usr_ctxts);
  13415		/* recalculate */
  13416		n_usr_ctxts = rcv_contexts - num_kernel_contexts;
  13417	}
  13418
  13419	num_netdev_contexts =
  13420		hfi1_num_netdev_contexts(dd, rcv_contexts -
  13421					 (num_kernel_contexts + n_usr_ctxts),
  13422					 &node_affinity.real_cpu_mask);
  13423	/*
  13424	 * The RMT entries are currently allocated as shown below:
  13425	 * 1. QOS (0 to 128 entries);
  13426	 * 2. FECN (num_kernel_context - 1 + num_user_contexts +
  13427	 *    num_netdev_contexts);
  13428	 * 3. netdev (num_netdev_contexts).
  13429	 * It should be noted that FECN oversubscribe num_netdev_contexts
  13430	 * entries of RMT because both netdev and PSM could allocate any receive
  13431	 * context between dd->first_dyn_alloc_text and dd->num_rcv_contexts,
  13432	 * and PSM FECN must reserve an RMT entry for each possible PSM receive
  13433	 * context.
  13434	 */
  13435	rmt_count = qos_rmt_entries(dd, NULL, NULL) + (num_netdev_contexts * 2);
  13436	if (HFI1_CAP_IS_KSET(TID_RDMA))
  13437		rmt_count += num_kernel_contexts - 1;
  13438	if (rmt_count + n_usr_ctxts > NUM_MAP_ENTRIES) {
  13439		user_rmt_reduced = NUM_MAP_ENTRIES - rmt_count;
  13440		dd_dev_err(dd,
  13441			   "RMT size is reducing the number of user receive contexts from %u to %d\n",
  13442			   n_usr_ctxts,
  13443			   user_rmt_reduced);
  13444		/* recalculate */
  13445		n_usr_ctxts = user_rmt_reduced;
  13446	}
  13447
  13448	/* the first N are kernel contexts, the rest are user/netdev contexts */
  13449	dd->num_rcv_contexts =
  13450		num_kernel_contexts + n_usr_ctxts + num_netdev_contexts;
  13451	dd->n_krcv_queues = num_kernel_contexts;
  13452	dd->first_dyn_alloc_ctxt = num_kernel_contexts;
  13453	dd->num_netdev_contexts = num_netdev_contexts;
  13454	dd->num_user_contexts = n_usr_ctxts;
  13455	dd->freectxts = n_usr_ctxts;
  13456	dd_dev_info(dd,
  13457		    "rcv contexts: chip %d, used %d (kernel %d, netdev %u, user %u)\n",
  13458		    rcv_contexts,
  13459		    (int)dd->num_rcv_contexts,
  13460		    (int)dd->n_krcv_queues,
  13461		    dd->num_netdev_contexts,
  13462		    dd->num_user_contexts);
  13463
  13464	/*
  13465	 * Receive array allocation:
  13466	 *   All RcvArray entries are divided into groups of 8. This
  13467	 *   is required by the hardware and will speed up writes to
  13468	 *   consecutive entries by using write-combining of the entire
  13469	 *   cacheline.
  13470	 *
  13471	 *   The number of groups are evenly divided among all contexts.
  13472	 *   any left over groups will be given to the first N user
  13473	 *   contexts.
  13474	 */
  13475	dd->rcv_entries.group_size = RCV_INCREMENT;
  13476	ngroups = chip_rcv_array_count(dd) / dd->rcv_entries.group_size;
  13477	dd->rcv_entries.ngroups = ngroups / dd->num_rcv_contexts;
  13478	dd->rcv_entries.nctxt_extra = ngroups -
  13479		(dd->num_rcv_contexts * dd->rcv_entries.ngroups);
  13480	dd_dev_info(dd, "RcvArray groups %u, ctxts extra %u\n",
  13481		    dd->rcv_entries.ngroups,
  13482		    dd->rcv_entries.nctxt_extra);
  13483	if (dd->rcv_entries.ngroups * dd->rcv_entries.group_size >
  13484	    MAX_EAGER_ENTRIES * 2) {
  13485		dd->rcv_entries.ngroups = (MAX_EAGER_ENTRIES * 2) /
  13486			dd->rcv_entries.group_size;
  13487		dd_dev_info(dd,
  13488			    "RcvArray group count too high, change to %u\n",
  13489			    dd->rcv_entries.ngroups);
  13490		dd->rcv_entries.nctxt_extra = 0;
  13491	}
  13492	/*
  13493	 * PIO send contexts
  13494	 */
  13495	ret = init_sc_pools_and_sizes(dd);
  13496	if (ret >= 0) {	/* success */
  13497		dd->num_send_contexts = ret;
  13498		dd_dev_info(
  13499			dd,
  13500			"send contexts: chip %d, used %d (kernel %d, ack %d, user %d, vl15 %d)\n",
  13501			send_contexts,
  13502			dd->num_send_contexts,
  13503			dd->sc_sizes[SC_KERNEL].count,
  13504			dd->sc_sizes[SC_ACK].count,
  13505			dd->sc_sizes[SC_USER].count,
  13506			dd->sc_sizes[SC_VL15].count);
  13507		ret = 0;	/* success */
  13508	}
  13509
  13510	return ret;
  13511}
  13512
  13513/*
  13514 * Set the device/port partition key table. The MAD code
  13515 * will ensure that, at least, the partial management
  13516 * partition key is present in the table.
  13517 */
  13518static void set_partition_keys(struct hfi1_pportdata *ppd)
  13519{
  13520	struct hfi1_devdata *dd = ppd->dd;
  13521	u64 reg = 0;
  13522	int i;
  13523
  13524	dd_dev_info(dd, "Setting partition keys\n");
  13525	for (i = 0; i < hfi1_get_npkeys(dd); i++) {
  13526		reg |= (ppd->pkeys[i] &
  13527			RCV_PARTITION_KEY_PARTITION_KEY_A_MASK) <<
  13528			((i % 4) *
  13529			 RCV_PARTITION_KEY_PARTITION_KEY_B_SHIFT);
  13530		/* Each register holds 4 PKey values. */
  13531		if ((i % 4) == 3) {
  13532			write_csr(dd, RCV_PARTITION_KEY +
  13533				  ((i - 3) * 2), reg);
  13534			reg = 0;
  13535		}
  13536	}
  13537
  13538	/* Always enable HW pkeys check when pkeys table is set */
  13539	add_rcvctrl(dd, RCV_CTRL_RCV_PARTITION_KEY_ENABLE_SMASK);
  13540}
  13541
  13542/*
  13543 * These CSRs and memories are uninitialized on reset and must be
  13544 * written before reading to set the ECC/parity bits.
  13545 *
  13546 * NOTE: All user context CSRs that are not mmaped write-only
  13547 * (e.g. the TID flows) must be initialized even if the driver never
  13548 * reads them.
  13549 */
  13550static void write_uninitialized_csrs_and_memories(struct hfi1_devdata *dd)
  13551{
  13552	int i, j;
  13553
  13554	/* CceIntMap */
  13555	for (i = 0; i < CCE_NUM_INT_MAP_CSRS; i++)
  13556		write_csr(dd, CCE_INT_MAP + (8 * i), 0);
  13557
  13558	/* SendCtxtCreditReturnAddr */
  13559	for (i = 0; i < chip_send_contexts(dd); i++)
  13560		write_kctxt_csr(dd, i, SEND_CTXT_CREDIT_RETURN_ADDR, 0);
  13561
  13562	/* PIO Send buffers */
  13563	/* SDMA Send buffers */
  13564	/*
  13565	 * These are not normally read, and (presently) have no method
  13566	 * to be read, so are not pre-initialized
  13567	 */
  13568
  13569	/* RcvHdrAddr */
  13570	/* RcvHdrTailAddr */
  13571	/* RcvTidFlowTable */
  13572	for (i = 0; i < chip_rcv_contexts(dd); i++) {
  13573		write_kctxt_csr(dd, i, RCV_HDR_ADDR, 0);
  13574		write_kctxt_csr(dd, i, RCV_HDR_TAIL_ADDR, 0);
  13575		for (j = 0; j < RXE_NUM_TID_FLOWS; j++)
  13576			write_uctxt_csr(dd, i, RCV_TID_FLOW_TABLE + (8 * j), 0);
  13577	}
  13578
  13579	/* RcvArray */
  13580	for (i = 0; i < chip_rcv_array_count(dd); i++)
  13581		hfi1_put_tid(dd, i, PT_INVALID_FLUSH, 0, 0);
  13582
  13583	/* RcvQPMapTable */
  13584	for (i = 0; i < 32; i++)
  13585		write_csr(dd, RCV_QP_MAP_TABLE + (8 * i), 0);
  13586}
  13587
  13588/*
  13589 * Use the ctrl_bits in CceCtrl to clear the status_bits in CceStatus.
  13590 */
  13591static void clear_cce_status(struct hfi1_devdata *dd, u64 status_bits,
  13592			     u64 ctrl_bits)
  13593{
  13594	unsigned long timeout;
  13595	u64 reg;
  13596
  13597	/* is the condition present? */
  13598	reg = read_csr(dd, CCE_STATUS);
  13599	if ((reg & status_bits) == 0)
  13600		return;
  13601
  13602	/* clear the condition */
  13603	write_csr(dd, CCE_CTRL, ctrl_bits);
  13604
  13605	/* wait for the condition to clear */
  13606	timeout = jiffies + msecs_to_jiffies(CCE_STATUS_TIMEOUT);
  13607	while (1) {
  13608		reg = read_csr(dd, CCE_STATUS);
  13609		if ((reg & status_bits) == 0)
  13610			return;
  13611		if (time_after(jiffies, timeout)) {
  13612			dd_dev_err(dd,
  13613				   "Timeout waiting for CceStatus to clear bits 0x%llx, remaining 0x%llx\n",
  13614				   status_bits, reg & status_bits);
  13615			return;
  13616		}
  13617		udelay(1);
  13618	}
  13619}
  13620
  13621/* set CCE CSRs to chip reset defaults */
  13622static void reset_cce_csrs(struct hfi1_devdata *dd)
  13623{
  13624	int i;
  13625
  13626	/* CCE_REVISION read-only */
  13627	/* CCE_REVISION2 read-only */
  13628	/* CCE_CTRL - bits clear automatically */
  13629	/* CCE_STATUS read-only, use CceCtrl to clear */
  13630	clear_cce_status(dd, ALL_FROZE, CCE_CTRL_SPC_UNFREEZE_SMASK);
  13631	clear_cce_status(dd, ALL_TXE_PAUSE, CCE_CTRL_TXE_RESUME_SMASK);
  13632	clear_cce_status(dd, ALL_RXE_PAUSE, CCE_CTRL_RXE_RESUME_SMASK);
  13633	for (i = 0; i < CCE_NUM_SCRATCH; i++)
  13634		write_csr(dd, CCE_SCRATCH + (8 * i), 0);
  13635	/* CCE_ERR_STATUS read-only */
  13636	write_csr(dd, CCE_ERR_MASK, 0);
  13637	write_csr(dd, CCE_ERR_CLEAR, ~0ull);
  13638	/* CCE_ERR_FORCE leave alone */
  13639	for (i = 0; i < CCE_NUM_32_BIT_COUNTERS; i++)
  13640		write_csr(dd, CCE_COUNTER_ARRAY32 + (8 * i), 0);
  13641	write_csr(dd, CCE_DC_CTRL, CCE_DC_CTRL_RESETCSR);
  13642	/* CCE_PCIE_CTRL leave alone */
  13643	for (i = 0; i < CCE_NUM_MSIX_VECTORS; i++) {
  13644		write_csr(dd, CCE_MSIX_TABLE_LOWER + (8 * i), 0);
  13645		write_csr(dd, CCE_MSIX_TABLE_UPPER + (8 * i),
  13646			  CCE_MSIX_TABLE_UPPER_RESETCSR);
  13647	}
  13648	for (i = 0; i < CCE_NUM_MSIX_PBAS; i++) {
  13649		/* CCE_MSIX_PBA read-only */
  13650		write_csr(dd, CCE_MSIX_INT_GRANTED, ~0ull);
  13651		write_csr(dd, CCE_MSIX_VEC_CLR_WITHOUT_INT, ~0ull);
  13652	}
  13653	for (i = 0; i < CCE_NUM_INT_MAP_CSRS; i++)
  13654		write_csr(dd, CCE_INT_MAP, 0);
  13655	for (i = 0; i < CCE_NUM_INT_CSRS; i++) {
  13656		/* CCE_INT_STATUS read-only */
  13657		write_csr(dd, CCE_INT_MASK + (8 * i), 0);
  13658		write_csr(dd, CCE_INT_CLEAR + (8 * i), ~0ull);
  13659		/* CCE_INT_FORCE leave alone */
  13660		/* CCE_INT_BLOCKED read-only */
  13661	}
  13662	for (i = 0; i < CCE_NUM_32_BIT_INT_COUNTERS; i++)
  13663		write_csr(dd, CCE_INT_COUNTER_ARRAY32 + (8 * i), 0);
  13664}
  13665
  13666/* set MISC CSRs to chip reset defaults */
  13667static void reset_misc_csrs(struct hfi1_devdata *dd)
  13668{
  13669	int i;
  13670
  13671	for (i = 0; i < 32; i++) {
  13672		write_csr(dd, MISC_CFG_RSA_R2 + (8 * i), 0);
  13673		write_csr(dd, MISC_CFG_RSA_SIGNATURE + (8 * i), 0);
  13674		write_csr(dd, MISC_CFG_RSA_MODULUS + (8 * i), 0);
  13675	}
  13676	/*
  13677	 * MISC_CFG_SHA_PRELOAD leave alone - always reads 0 and can
  13678	 * only be written 128-byte chunks
  13679	 */
  13680	/* init RSA engine to clear lingering errors */
  13681	write_csr(dd, MISC_CFG_RSA_CMD, 1);
  13682	write_csr(dd, MISC_CFG_RSA_MU, 0);
  13683	write_csr(dd, MISC_CFG_FW_CTRL, 0);
  13684	/* MISC_STS_8051_DIGEST read-only */
  13685	/* MISC_STS_SBM_DIGEST read-only */
  13686	/* MISC_STS_PCIE_DIGEST read-only */
  13687	/* MISC_STS_FAB_DIGEST read-only */
  13688	/* MISC_ERR_STATUS read-only */
  13689	write_csr(dd, MISC_ERR_MASK, 0);
  13690	write_csr(dd, MISC_ERR_CLEAR, ~0ull);
  13691	/* MISC_ERR_FORCE leave alone */
  13692}
  13693
  13694/* set TXE CSRs to chip reset defaults */
  13695static void reset_txe_csrs(struct hfi1_devdata *dd)
  13696{
  13697	int i;
  13698
  13699	/*
  13700	 * TXE Kernel CSRs
  13701	 */
  13702	write_csr(dd, SEND_CTRL, 0);
  13703	__cm_reset(dd, 0);	/* reset CM internal state */
  13704	/* SEND_CONTEXTS read-only */
  13705	/* SEND_DMA_ENGINES read-only */
  13706	/* SEND_PIO_MEM_SIZE read-only */
  13707	/* SEND_DMA_MEM_SIZE read-only */
  13708	write_csr(dd, SEND_HIGH_PRIORITY_LIMIT, 0);
  13709	pio_reset_all(dd);	/* SEND_PIO_INIT_CTXT */
  13710	/* SEND_PIO_ERR_STATUS read-only */
  13711	write_csr(dd, SEND_PIO_ERR_MASK, 0);
  13712	write_csr(dd, SEND_PIO_ERR_CLEAR, ~0ull);
  13713	/* SEND_PIO_ERR_FORCE leave alone */
  13714	/* SEND_DMA_ERR_STATUS read-only */
  13715	write_csr(dd, SEND_DMA_ERR_MASK, 0);
  13716	write_csr(dd, SEND_DMA_ERR_CLEAR, ~0ull);
  13717	/* SEND_DMA_ERR_FORCE leave alone */
  13718	/* SEND_EGRESS_ERR_STATUS read-only */
  13719	write_csr(dd, SEND_EGRESS_ERR_MASK, 0);
  13720	write_csr(dd, SEND_EGRESS_ERR_CLEAR, ~0ull);
  13721	/* SEND_EGRESS_ERR_FORCE leave alone */
  13722	write_csr(dd, SEND_BTH_QP, 0);
  13723	write_csr(dd, SEND_STATIC_RATE_CONTROL, 0);
  13724	write_csr(dd, SEND_SC2VLT0, 0);
  13725	write_csr(dd, SEND_SC2VLT1, 0);
  13726	write_csr(dd, SEND_SC2VLT2, 0);
  13727	write_csr(dd, SEND_SC2VLT3, 0);
  13728	write_csr(dd, SEND_LEN_CHECK0, 0);
  13729	write_csr(dd, SEND_LEN_CHECK1, 0);
  13730	/* SEND_ERR_STATUS read-only */
  13731	write_csr(dd, SEND_ERR_MASK, 0);
  13732	write_csr(dd, SEND_ERR_CLEAR, ~0ull);
  13733	/* SEND_ERR_FORCE read-only */
  13734	for (i = 0; i < VL_ARB_LOW_PRIO_TABLE_SIZE; i++)
  13735		write_csr(dd, SEND_LOW_PRIORITY_LIST + (8 * i), 0);
  13736	for (i = 0; i < VL_ARB_HIGH_PRIO_TABLE_SIZE; i++)
  13737		write_csr(dd, SEND_HIGH_PRIORITY_LIST + (8 * i), 0);
  13738	for (i = 0; i < chip_send_contexts(dd) / NUM_CONTEXTS_PER_SET; i++)
  13739		write_csr(dd, SEND_CONTEXT_SET_CTRL + (8 * i), 0);
  13740	for (i = 0; i < TXE_NUM_32_BIT_COUNTER; i++)
  13741		write_csr(dd, SEND_COUNTER_ARRAY32 + (8 * i), 0);
  13742	for (i = 0; i < TXE_NUM_64_BIT_COUNTER; i++)
  13743		write_csr(dd, SEND_COUNTER_ARRAY64 + (8 * i), 0);
  13744	write_csr(dd, SEND_CM_CTRL, SEND_CM_CTRL_RESETCSR);
  13745	write_csr(dd, SEND_CM_GLOBAL_CREDIT, SEND_CM_GLOBAL_CREDIT_RESETCSR);
  13746	/* SEND_CM_CREDIT_USED_STATUS read-only */
  13747	write_csr(dd, SEND_CM_TIMER_CTRL, 0);
  13748	write_csr(dd, SEND_CM_LOCAL_AU_TABLE0_TO3, 0);
  13749	write_csr(dd, SEND_CM_LOCAL_AU_TABLE4_TO7, 0);
  13750	write_csr(dd, SEND_CM_REMOTE_AU_TABLE0_TO3, 0);
  13751	write_csr(dd, SEND_CM_REMOTE_AU_TABLE4_TO7, 0);
  13752	for (i = 0; i < TXE_NUM_DATA_VL; i++)
  13753		write_csr(dd, SEND_CM_CREDIT_VL + (8 * i), 0);
  13754	write_csr(dd, SEND_CM_CREDIT_VL15, 0);
  13755	/* SEND_CM_CREDIT_USED_VL read-only */
  13756	/* SEND_CM_CREDIT_USED_VL15 read-only */
  13757	/* SEND_EGRESS_CTXT_STATUS read-only */
  13758	/* SEND_EGRESS_SEND_DMA_STATUS read-only */
  13759	write_csr(dd, SEND_EGRESS_ERR_INFO, ~0ull);
  13760	/* SEND_EGRESS_ERR_INFO read-only */
  13761	/* SEND_EGRESS_ERR_SOURCE read-only */
  13762
  13763	/*
  13764	 * TXE Per-Context CSRs
  13765	 */
  13766	for (i = 0; i < chip_send_contexts(dd); i++) {
  13767		write_kctxt_csr(dd, i, SEND_CTXT_CTRL, 0);
  13768		write_kctxt_csr(dd, i, SEND_CTXT_CREDIT_CTRL, 0);
  13769		write_kctxt_csr(dd, i, SEND_CTXT_CREDIT_RETURN_ADDR, 0);
  13770		write_kctxt_csr(dd, i, SEND_CTXT_CREDIT_FORCE, 0);
  13771		write_kctxt_csr(dd, i, SEND_CTXT_ERR_MASK, 0);
  13772		write_kctxt_csr(dd, i, SEND_CTXT_ERR_CLEAR, ~0ull);
  13773		write_kctxt_csr(dd, i, SEND_CTXT_CHECK_ENABLE, 0);
  13774		write_kctxt_csr(dd, i, SEND_CTXT_CHECK_VL, 0);
  13775		write_kctxt_csr(dd, i, SEND_CTXT_CHECK_JOB_KEY, 0);
  13776		write_kctxt_csr(dd, i, SEND_CTXT_CHECK_PARTITION_KEY, 0);
  13777		write_kctxt_csr(dd, i, SEND_CTXT_CHECK_SLID, 0);
  13778		write_kctxt_csr(dd, i, SEND_CTXT_CHECK_OPCODE, 0);
  13779	}
  13780
  13781	/*
  13782	 * TXE Per-SDMA CSRs
  13783	 */
  13784	for (i = 0; i < chip_sdma_engines(dd); i++) {
  13785		write_kctxt_csr(dd, i, SEND_DMA_CTRL, 0);
  13786		/* SEND_DMA_STATUS read-only */
  13787		write_kctxt_csr(dd, i, SEND_DMA_BASE_ADDR, 0);
  13788		write_kctxt_csr(dd, i, SEND_DMA_LEN_GEN, 0);
  13789		write_kctxt_csr(dd, i, SEND_DMA_TAIL, 0);
  13790		/* SEND_DMA_HEAD read-only */
  13791		write_kctxt_csr(dd, i, SEND_DMA_HEAD_ADDR, 0);
  13792		write_kctxt_csr(dd, i, SEND_DMA_PRIORITY_THLD, 0);
  13793		/* SEND_DMA_IDLE_CNT read-only */
  13794		write_kctxt_csr(dd, i, SEND_DMA_RELOAD_CNT, 0);
  13795		write_kctxt_csr(dd, i, SEND_DMA_DESC_CNT, 0);
  13796		/* SEND_DMA_DESC_FETCHED_CNT read-only */
  13797		/* SEND_DMA_ENG_ERR_STATUS read-only */
  13798		write_kctxt_csr(dd, i, SEND_DMA_ENG_ERR_MASK, 0);
  13799		write_kctxt_csr(dd, i, SEND_DMA_ENG_ERR_CLEAR, ~0ull);
  13800		/* SEND_DMA_ENG_ERR_FORCE leave alone */
  13801		write_kctxt_csr(dd, i, SEND_DMA_CHECK_ENABLE, 0);
  13802		write_kctxt_csr(dd, i, SEND_DMA_CHECK_VL, 0);
  13803		write_kctxt_csr(dd, i, SEND_DMA_CHECK_JOB_KEY, 0);
  13804		write_kctxt_csr(dd, i, SEND_DMA_CHECK_PARTITION_KEY, 0);
  13805		write_kctxt_csr(dd, i, SEND_DMA_CHECK_SLID, 0);
  13806		write_kctxt_csr(dd, i, SEND_DMA_CHECK_OPCODE, 0);
  13807		write_kctxt_csr(dd, i, SEND_DMA_MEMORY, 0);
  13808	}
  13809}
  13810
  13811/*
  13812 * Expect on entry:
  13813 * o Packet ingress is disabled, i.e. RcvCtrl.RcvPortEnable == 0
  13814 */
  13815static void init_rbufs(struct hfi1_devdata *dd)
  13816{
  13817	u64 reg;
  13818	int count;
  13819
  13820	/*
  13821	 * Wait for DMA to stop: RxRbufPktPending and RxPktInProgress are
  13822	 * clear.
  13823	 */
  13824	count = 0;
  13825	while (1) {
  13826		reg = read_csr(dd, RCV_STATUS);
  13827		if ((reg & (RCV_STATUS_RX_RBUF_PKT_PENDING_SMASK
  13828			    | RCV_STATUS_RX_PKT_IN_PROGRESS_SMASK)) == 0)
  13829			break;
  13830		/*
  13831		 * Give up after 1ms - maximum wait time.
  13832		 *
  13833		 * RBuf size is 136KiB.  Slowest possible is PCIe Gen1 x1 at
  13834		 * 250MB/s bandwidth.  Lower rate to 66% for overhead to get:
  13835		 *	136 KB / (66% * 250MB/s) = 844us
  13836		 */
  13837		if (count++ > 500) {
  13838			dd_dev_err(dd,
  13839				   "%s: in-progress DMA not clearing: RcvStatus 0x%llx, continuing\n",
  13840				   __func__, reg);
  13841			break;
  13842		}
  13843		udelay(2); /* do not busy-wait the CSR */
  13844	}
  13845
  13846	/* start the init - expect RcvCtrl to be 0 */
  13847	write_csr(dd, RCV_CTRL, RCV_CTRL_RX_RBUF_INIT_SMASK);
  13848
  13849	/*
  13850	 * Read to force the write of Rcvtrl.RxRbufInit.  There is a brief
  13851	 * period after the write before RcvStatus.RxRbufInitDone is valid.
  13852	 * The delay in the first run through the loop below is sufficient and
  13853	 * required before the first read of RcvStatus.RxRbufInintDone.
  13854	 */
  13855	read_csr(dd, RCV_CTRL);
  13856
  13857	/* wait for the init to finish */
  13858	count = 0;
  13859	while (1) {
  13860		/* delay is required first time through - see above */
  13861		udelay(2); /* do not busy-wait the CSR */
  13862		reg = read_csr(dd, RCV_STATUS);
  13863		if (reg & (RCV_STATUS_RX_RBUF_INIT_DONE_SMASK))
  13864			break;
  13865
  13866		/* give up after 100us - slowest possible at 33MHz is 73us */
  13867		if (count++ > 50) {
  13868			dd_dev_err(dd,
  13869				   "%s: RcvStatus.RxRbufInit not set, continuing\n",
  13870				   __func__);
  13871			break;
  13872		}
  13873	}
  13874}
  13875
  13876/* set RXE CSRs to chip reset defaults */
  13877static void reset_rxe_csrs(struct hfi1_devdata *dd)
  13878{
  13879	int i, j;
  13880
  13881	/*
  13882	 * RXE Kernel CSRs
  13883	 */
  13884	write_csr(dd, RCV_CTRL, 0);
  13885	init_rbufs(dd);
  13886	/* RCV_STATUS read-only */
  13887	/* RCV_CONTEXTS read-only */
  13888	/* RCV_ARRAY_CNT read-only */
  13889	/* RCV_BUF_SIZE read-only */
  13890	write_csr(dd, RCV_BTH_QP, 0);
  13891	write_csr(dd, RCV_MULTICAST, 0);
  13892	write_csr(dd, RCV_BYPASS, 0);
  13893	write_csr(dd, RCV_VL15, 0);
  13894	/* this is a clear-down */
  13895	write_csr(dd, RCV_ERR_INFO,
  13896		  RCV_ERR_INFO_RCV_EXCESS_BUFFER_OVERRUN_SMASK);
  13897	/* RCV_ERR_STATUS read-only */
  13898	write_csr(dd, RCV_ERR_MASK, 0);
  13899	write_csr(dd, RCV_ERR_CLEAR, ~0ull);
  13900	/* RCV_ERR_FORCE leave alone */
  13901	for (i = 0; i < 32; i++)
  13902		write_csr(dd, RCV_QP_MAP_TABLE + (8 * i), 0);
  13903	for (i = 0; i < 4; i++)
  13904		write_csr(dd, RCV_PARTITION_KEY + (8 * i), 0);
  13905	for (i = 0; i < RXE_NUM_32_BIT_COUNTERS; i++)
  13906		write_csr(dd, RCV_COUNTER_ARRAY32 + (8 * i), 0);
  13907	for (i = 0; i < RXE_NUM_64_BIT_COUNTERS; i++)
  13908		write_csr(dd, RCV_COUNTER_ARRAY64 + (8 * i), 0);
  13909	for (i = 0; i < RXE_NUM_RSM_INSTANCES; i++)
  13910		clear_rsm_rule(dd, i);
  13911	for (i = 0; i < 32; i++)
  13912		write_csr(dd, RCV_RSM_MAP_TABLE + (8 * i), 0);
  13913
  13914	/*
  13915	 * RXE Kernel and User Per-Context CSRs
  13916	 */
  13917	for (i = 0; i < chip_rcv_contexts(dd); i++) {
  13918		/* kernel */
  13919		write_kctxt_csr(dd, i, RCV_CTXT_CTRL, 0);
  13920		/* RCV_CTXT_STATUS read-only */
  13921		write_kctxt_csr(dd, i, RCV_EGR_CTRL, 0);
  13922		write_kctxt_csr(dd, i, RCV_TID_CTRL, 0);
  13923		write_kctxt_csr(dd, i, RCV_KEY_CTRL, 0);
  13924		write_kctxt_csr(dd, i, RCV_HDR_ADDR, 0);
  13925		write_kctxt_csr(dd, i, RCV_HDR_CNT, 0);
  13926		write_kctxt_csr(dd, i, RCV_HDR_ENT_SIZE, 0);
  13927		write_kctxt_csr(dd, i, RCV_HDR_SIZE, 0);
  13928		write_kctxt_csr(dd, i, RCV_HDR_TAIL_ADDR, 0);
  13929		write_kctxt_csr(dd, i, RCV_AVAIL_TIME_OUT, 0);
  13930		write_kctxt_csr(dd, i, RCV_HDR_OVFL_CNT, 0);
  13931
  13932		/* user */
  13933		/* RCV_HDR_TAIL read-only */
  13934		write_uctxt_csr(dd, i, RCV_HDR_HEAD, 0);
  13935		/* RCV_EGR_INDEX_TAIL read-only */
  13936		write_uctxt_csr(dd, i, RCV_EGR_INDEX_HEAD, 0);
  13937		/* RCV_EGR_OFFSET_TAIL read-only */
  13938		for (j = 0; j < RXE_NUM_TID_FLOWS; j++) {
  13939			write_uctxt_csr(dd, i,
  13940					RCV_TID_FLOW_TABLE + (8 * j), 0);
  13941		}
  13942	}
  13943}
  13944
  13945/*
  13946 * Set sc2vl tables.
  13947 *
  13948 * They power on to zeros, so to avoid send context errors
  13949 * they need to be set:
  13950 *
  13951 * SC 0-7 -> VL 0-7 (respectively)
  13952 * SC 15  -> VL 15
  13953 * otherwise
  13954 *        -> VL 0
  13955 */
  13956static void init_sc2vl_tables(struct hfi1_devdata *dd)
  13957{
  13958	int i;
  13959	/* init per architecture spec, constrained by hardware capability */
  13960
  13961	/* HFI maps sent packets */
  13962	write_csr(dd, SEND_SC2VLT0, SC2VL_VAL(
  13963		0,
  13964		0, 0, 1, 1,
  13965		2, 2, 3, 3,
  13966		4, 4, 5, 5,
  13967		6, 6, 7, 7));
  13968	write_csr(dd, SEND_SC2VLT1, SC2VL_VAL(
  13969		1,
  13970		8, 0, 9, 0,
  13971		10, 0, 11, 0,
  13972		12, 0, 13, 0,
  13973		14, 0, 15, 15));
  13974	write_csr(dd, SEND_SC2VLT2, SC2VL_VAL(
  13975		2,
  13976		16, 0, 17, 0,
  13977		18, 0, 19, 0,
  13978		20, 0, 21, 0,
  13979		22, 0, 23, 0));
  13980	write_csr(dd, SEND_SC2VLT3, SC2VL_VAL(
  13981		3,
  13982		24, 0, 25, 0,
  13983		26, 0, 27, 0,
  13984		28, 0, 29, 0,
  13985		30, 0, 31, 0));
  13986
  13987	/* DC maps received packets */
  13988	write_csr(dd, DCC_CFG_SC_VL_TABLE_15_0, DC_SC_VL_VAL(
  13989		15_0,
  13990		0, 0, 1, 1,  2, 2,  3, 3,  4, 4,  5, 5,  6, 6,  7,  7,
  13991		8, 0, 9, 0, 10, 0, 11, 0, 12, 0, 13, 0, 14, 0, 15, 15));
  13992	write_csr(dd, DCC_CFG_SC_VL_TABLE_31_16, DC_SC_VL_VAL(
  13993		31_16,
  13994		16, 0, 17, 0, 18, 0, 19, 0, 20, 0, 21, 0, 22, 0, 23, 0,
  13995		24, 0, 25, 0, 26, 0, 27, 0, 28, 0, 29, 0, 30, 0, 31, 0));
  13996
  13997	/* initialize the cached sc2vl values consistently with h/w */
  13998	for (i = 0; i < 32; i++) {
  13999		if (i < 8 || i == 15)
  14000			*((u8 *)(dd->sc2vl) + i) = (u8)i;
  14001		else
  14002			*((u8 *)(dd->sc2vl) + i) = 0;
  14003	}
  14004}
  14005
  14006/*
  14007 * Read chip sizes and then reset parts to sane, disabled, values.  We cannot
  14008 * depend on the chip going through a power-on reset - a driver may be loaded
  14009 * and unloaded many times.
  14010 *
  14011 * Do not write any CSR values to the chip in this routine - there may be
  14012 * a reset following the (possible) FLR in this routine.
  14013 *
  14014 */
  14015static int init_chip(struct hfi1_devdata *dd)
  14016{
  14017	int i;
  14018	int ret = 0;
  14019
  14020	/*
  14021	 * Put the HFI CSRs in a known state.
  14022	 * Combine this with a DC reset.
  14023	 *
  14024	 * Stop the device from doing anything while we do a
  14025	 * reset.  We know there are no other active users of
  14026	 * the device since we are now in charge.  Turn off
  14027	 * off all outbound and inbound traffic and make sure
  14028	 * the device does not generate any interrupts.
  14029	 */
  14030
  14031	/* disable send contexts and SDMA engines */
  14032	write_csr(dd, SEND_CTRL, 0);
  14033	for (i = 0; i < chip_send_contexts(dd); i++)
  14034		write_kctxt_csr(dd, i, SEND_CTXT_CTRL, 0);
  14035	for (i = 0; i < chip_sdma_engines(dd); i++)
  14036		write_kctxt_csr(dd, i, SEND_DMA_CTRL, 0);
  14037	/* disable port (turn off RXE inbound traffic) and contexts */
  14038	write_csr(dd, RCV_CTRL, 0);
  14039	for (i = 0; i < chip_rcv_contexts(dd); i++)
  14040		write_csr(dd, RCV_CTXT_CTRL, 0);
  14041	/* mask all interrupt sources */
  14042	for (i = 0; i < CCE_NUM_INT_CSRS; i++)
  14043		write_csr(dd, CCE_INT_MASK + (8 * i), 0ull);
  14044
  14045	/*
  14046	 * DC Reset: do a full DC reset before the register clear.
  14047	 * A recommended length of time to hold is one CSR read,
  14048	 * so reread the CceDcCtrl.  Then, hold the DC in reset
  14049	 * across the clear.
  14050	 */
  14051	write_csr(dd, CCE_DC_CTRL, CCE_DC_CTRL_DC_RESET_SMASK);
  14052	(void)read_csr(dd, CCE_DC_CTRL);
  14053
  14054	if (use_flr) {
  14055		/*
  14056		 * A FLR will reset the SPC core and part of the PCIe.
  14057		 * The parts that need to be restored have already been
  14058		 * saved.
  14059		 */
  14060		dd_dev_info(dd, "Resetting CSRs with FLR\n");
  14061
  14062		/* do the FLR, the DC reset will remain */
  14063		pcie_flr(dd->pcidev);
  14064
  14065		/* restore command and BARs */
  14066		ret = restore_pci_variables(dd);
  14067		if (ret) {
  14068			dd_dev_err(dd, "%s: Could not restore PCI variables\n",
  14069				   __func__);
  14070			return ret;
  14071		}
  14072
  14073		if (is_ax(dd)) {
  14074			dd_dev_info(dd, "Resetting CSRs with FLR\n");
  14075			pcie_flr(dd->pcidev);
  14076			ret = restore_pci_variables(dd);
  14077			if (ret) {
  14078				dd_dev_err(dd, "%s: Could not restore PCI variables\n",
  14079					   __func__);
  14080				return ret;
  14081			}
  14082		}
  14083	} else {
  14084		dd_dev_info(dd, "Resetting CSRs with writes\n");
  14085		reset_cce_csrs(dd);
  14086		reset_txe_csrs(dd);
  14087		reset_rxe_csrs(dd);
  14088		reset_misc_csrs(dd);
  14089	}
  14090	/* clear the DC reset */
  14091	write_csr(dd, CCE_DC_CTRL, 0);
  14092
  14093	/* Set the LED off */
  14094	setextled(dd, 0);
  14095
  14096	/*
  14097	 * Clear the QSFP reset.
  14098	 * An FLR enforces a 0 on all out pins. The driver does not touch
  14099	 * ASIC_QSFPn_OUT otherwise.  This leaves RESET_N low and
  14100	 * anything plugged constantly in reset, if it pays attention
  14101	 * to RESET_N.
  14102	 * Prime examples of this are optical cables. Set all pins high.
  14103	 * I2CCLK and I2CDAT will change per direction, and INT_N and
  14104	 * MODPRS_N are input only and their value is ignored.
  14105	 */
  14106	write_csr(dd, ASIC_QSFP1_OUT, 0x1f);
  14107	write_csr(dd, ASIC_QSFP2_OUT, 0x1f);
  14108	init_chip_resources(dd);
  14109	return ret;
  14110}
  14111
  14112static void init_early_variables(struct hfi1_devdata *dd)
  14113{
  14114	int i;
  14115
  14116	/* assign link credit variables */
  14117	dd->vau = CM_VAU;
  14118	dd->link_credits = CM_GLOBAL_CREDITS;
  14119	if (is_ax(dd))
  14120		dd->link_credits--;
  14121	dd->vcu = cu_to_vcu(hfi1_cu);
  14122	/* enough room for 8 MAD packets plus header - 17K */
  14123	dd->vl15_init = (8 * (2048 + 128)) / vau_to_au(dd->vau);
  14124	if (dd->vl15_init > dd->link_credits)
  14125		dd->vl15_init = dd->link_credits;
  14126
  14127	write_uninitialized_csrs_and_memories(dd);
  14128
  14129	if (HFI1_CAP_IS_KSET(PKEY_CHECK))
  14130		for (i = 0; i < dd->num_pports; i++) {
  14131			struct hfi1_pportdata *ppd = &dd->pport[i];
  14132
  14133			set_partition_keys(ppd);
  14134		}
  14135	init_sc2vl_tables(dd);
  14136}
  14137
  14138static void init_kdeth_qp(struct hfi1_devdata *dd)
  14139{
  14140	write_csr(dd, SEND_BTH_QP,
  14141		  (RVT_KDETH_QP_PREFIX & SEND_BTH_QP_KDETH_QP_MASK) <<
  14142		  SEND_BTH_QP_KDETH_QP_SHIFT);
  14143
  14144	write_csr(dd, RCV_BTH_QP,
  14145		  (RVT_KDETH_QP_PREFIX & RCV_BTH_QP_KDETH_QP_MASK) <<
  14146		  RCV_BTH_QP_KDETH_QP_SHIFT);
  14147}
  14148
  14149/**
  14150 * hfi1_get_qp_map - get qp map
  14151 * @dd: device data
  14152 * @idx: index to read
  14153 */
  14154u8 hfi1_get_qp_map(struct hfi1_devdata *dd, u8 idx)
  14155{
  14156	u64 reg = read_csr(dd, RCV_QP_MAP_TABLE + (idx / 8) * 8);
  14157
  14158	reg >>= (idx % 8) * 8;
  14159	return reg;
  14160}
  14161
  14162/**
  14163 * init_qpmap_table - init qp map
  14164 * @dd: device data
  14165 * @first_ctxt: first context
  14166 * @last_ctxt: first context
  14167 *
  14168 * This return sets the qpn mapping table that
  14169 * is indexed by qpn[8:1].
  14170 *
  14171 * The routine will round robin the 256 settings
  14172 * from first_ctxt to last_ctxt.
  14173 *
  14174 * The first/last looks ahead to having specialized
  14175 * receive contexts for mgmt and bypass.  Normal
  14176 * verbs traffic will assumed to be on a range
  14177 * of receive contexts.
  14178 */
  14179static void init_qpmap_table(struct hfi1_devdata *dd,
  14180			     u32 first_ctxt,
  14181			     u32 last_ctxt)
  14182{
  14183	u64 reg = 0;
  14184	u64 regno = RCV_QP_MAP_TABLE;
  14185	int i;
  14186	u64 ctxt = first_ctxt;
  14187
  14188	for (i = 0; i < 256; i++) {
  14189		reg |= ctxt << (8 * (i % 8));
  14190		ctxt++;
  14191		if (ctxt > last_ctxt)
  14192			ctxt = first_ctxt;
  14193		if (i % 8 == 7) {
  14194			write_csr(dd, regno, reg);
  14195			reg = 0;
  14196			regno += 8;
  14197		}
  14198	}
  14199
  14200	add_rcvctrl(dd, RCV_CTRL_RCV_QP_MAP_ENABLE_SMASK
  14201			| RCV_CTRL_RCV_BYPASS_ENABLE_SMASK);
  14202}
  14203
  14204struct rsm_map_table {
  14205	u64 map[NUM_MAP_REGS];
  14206	unsigned int used;
  14207};
  14208
  14209struct rsm_rule_data {
  14210	u8 offset;
  14211	u8 pkt_type;
  14212	u32 field1_off;
  14213	u32 field2_off;
  14214	u32 index1_off;
  14215	u32 index1_width;
  14216	u32 index2_off;
  14217	u32 index2_width;
  14218	u32 mask1;
  14219	u32 value1;
  14220	u32 mask2;
  14221	u32 value2;
  14222};
  14223
  14224/*
  14225 * Return an initialized RMT map table for users to fill in.  OK if it
  14226 * returns NULL, indicating no table.
  14227 */
  14228static struct rsm_map_table *alloc_rsm_map_table(struct hfi1_devdata *dd)
  14229{
  14230	struct rsm_map_table *rmt;
  14231	u8 rxcontext = is_ax(dd) ? 0 : 0xff;  /* 0 is default if a0 ver. */
  14232
  14233	rmt = kmalloc(sizeof(*rmt), GFP_KERNEL);
  14234	if (rmt) {
  14235		memset(rmt->map, rxcontext, sizeof(rmt->map));
  14236		rmt->used = 0;
  14237	}
  14238
  14239	return rmt;
  14240}
  14241
  14242/*
  14243 * Write the final RMT map table to the chip and free the table.  OK if
  14244 * table is NULL.
  14245 */
  14246static void complete_rsm_map_table(struct hfi1_devdata *dd,
  14247				   struct rsm_map_table *rmt)
  14248{
  14249	int i;
  14250
  14251	if (rmt) {
  14252		/* write table to chip */
  14253		for (i = 0; i < NUM_MAP_REGS; i++)
  14254			write_csr(dd, RCV_RSM_MAP_TABLE + (8 * i), rmt->map[i]);
  14255
  14256		/* enable RSM */
  14257		add_rcvctrl(dd, RCV_CTRL_RCV_RSM_ENABLE_SMASK);
  14258	}
  14259}
  14260
  14261/* Is a receive side mapping rule */
  14262static bool has_rsm_rule(struct hfi1_devdata *dd, u8 rule_index)
  14263{
  14264	return read_csr(dd, RCV_RSM_CFG + (8 * rule_index)) != 0;
  14265}
  14266
  14267/*
  14268 * Add a receive side mapping rule.
  14269 */
  14270static void add_rsm_rule(struct hfi1_devdata *dd, u8 rule_index,
  14271			 struct rsm_rule_data *rrd)
  14272{
  14273	write_csr(dd, RCV_RSM_CFG + (8 * rule_index),
  14274		  (u64)rrd->offset << RCV_RSM_CFG_OFFSET_SHIFT |
  14275		  1ull << rule_index | /* enable bit */
  14276		  (u64)rrd->pkt_type << RCV_RSM_CFG_PACKET_TYPE_SHIFT);
  14277	write_csr(dd, RCV_RSM_SELECT + (8 * rule_index),
  14278		  (u64)rrd->field1_off << RCV_RSM_SELECT_FIELD1_OFFSET_SHIFT |
  14279		  (u64)rrd->field2_off << RCV_RSM_SELECT_FIELD2_OFFSET_SHIFT |
  14280		  (u64)rrd->index1_off << RCV_RSM_SELECT_INDEX1_OFFSET_SHIFT |
  14281		  (u64)rrd->index1_width << RCV_RSM_SELECT_INDEX1_WIDTH_SHIFT |
  14282		  (u64)rrd->index2_off << RCV_RSM_SELECT_INDEX2_OFFSET_SHIFT |
  14283		  (u64)rrd->index2_width << RCV_RSM_SELECT_INDEX2_WIDTH_SHIFT);
  14284	write_csr(dd, RCV_RSM_MATCH + (8 * rule_index),
  14285		  (u64)rrd->mask1 << RCV_RSM_MATCH_MASK1_SHIFT |
  14286		  (u64)rrd->value1 << RCV_RSM_MATCH_VALUE1_SHIFT |
  14287		  (u64)rrd->mask2 << RCV_RSM_MATCH_MASK2_SHIFT |
  14288		  (u64)rrd->value2 << RCV_RSM_MATCH_VALUE2_SHIFT);
  14289}
  14290
  14291/*
  14292 * Clear a receive side mapping rule.
  14293 */
  14294static void clear_rsm_rule(struct hfi1_devdata *dd, u8 rule_index)
  14295{
  14296	write_csr(dd, RCV_RSM_CFG + (8 * rule_index), 0);
  14297	write_csr(dd, RCV_RSM_SELECT + (8 * rule_index), 0);
  14298	write_csr(dd, RCV_RSM_MATCH + (8 * rule_index), 0);
  14299}
  14300
  14301/* return the number of RSM map table entries that will be used for QOS */
  14302static int qos_rmt_entries(struct hfi1_devdata *dd, unsigned int *mp,
  14303			   unsigned int *np)
  14304{
  14305	int i;
  14306	unsigned int m, n;
  14307	u8 max_by_vl = 0;
  14308
  14309	/* is QOS active at all? */
  14310	if (dd->n_krcv_queues <= MIN_KERNEL_KCTXTS ||
  14311	    num_vls == 1 ||
  14312	    krcvqsset <= 1)
  14313		goto no_qos;
  14314
  14315	/* determine bits for qpn */
  14316	for (i = 0; i < min_t(unsigned int, num_vls, krcvqsset); i++)
  14317		if (krcvqs[i] > max_by_vl)
  14318			max_by_vl = krcvqs[i];
  14319	if (max_by_vl > 32)
  14320		goto no_qos;
  14321	m = ilog2(__roundup_pow_of_two(max_by_vl));
  14322
  14323	/* determine bits for vl */
  14324	n = ilog2(__roundup_pow_of_two(num_vls));
  14325
  14326	/* reject if too much is used */
  14327	if ((m + n) > 7)
  14328		goto no_qos;
  14329
  14330	if (mp)
  14331		*mp = m;
  14332	if (np)
  14333		*np = n;
  14334
  14335	return 1 << (m + n);
  14336
  14337no_qos:
  14338	if (mp)
  14339		*mp = 0;
  14340	if (np)
  14341		*np = 0;
  14342	return 0;
  14343}
  14344
  14345/**
  14346 * init_qos - init RX qos
  14347 * @dd: device data
  14348 * @rmt: RSM map table
  14349 *
  14350 * This routine initializes Rule 0 and the RSM map table to implement
  14351 * quality of service (qos).
  14352 *
  14353 * If all of the limit tests succeed, qos is applied based on the array
  14354 * interpretation of krcvqs where entry 0 is VL0.
  14355 *
  14356 * The number of vl bits (n) and the number of qpn bits (m) are computed to
  14357 * feed both the RSM map table and the single rule.
  14358 */
  14359static void init_qos(struct hfi1_devdata *dd, struct rsm_map_table *rmt)
  14360{
  14361	struct rsm_rule_data rrd;
  14362	unsigned qpns_per_vl, ctxt, i, qpn, n = 1, m;
  14363	unsigned int rmt_entries;
  14364	u64 reg;
  14365
  14366	if (!rmt)
  14367		goto bail;
  14368	rmt_entries = qos_rmt_entries(dd, &m, &n);
  14369	if (rmt_entries == 0)
  14370		goto bail;
  14371	qpns_per_vl = 1 << m;
  14372
  14373	/* enough room in the map table? */
  14374	rmt_entries = 1 << (m + n);
  14375	if (rmt->used + rmt_entries >= NUM_MAP_ENTRIES)
  14376		goto bail;
  14377
  14378	/* add qos entries to the RSM map table */
  14379	for (i = 0, ctxt = FIRST_KERNEL_KCTXT; i < num_vls; i++) {
  14380		unsigned tctxt;
  14381
  14382		for (qpn = 0, tctxt = ctxt;
  14383		     krcvqs[i] && qpn < qpns_per_vl; qpn++) {
  14384			unsigned idx, regoff, regidx;
  14385
  14386			/* generate the index the hardware will produce */
  14387			idx = rmt->used + ((qpn << n) ^ i);
  14388			regoff = (idx % 8) * 8;
  14389			regidx = idx / 8;
  14390			/* replace default with context number */
  14391			reg = rmt->map[regidx];
  14392			reg &= ~(RCV_RSM_MAP_TABLE_RCV_CONTEXT_A_MASK
  14393				<< regoff);
  14394			reg |= (u64)(tctxt++) << regoff;
  14395			rmt->map[regidx] = reg;
  14396			if (tctxt == ctxt + krcvqs[i])
  14397				tctxt = ctxt;
  14398		}
  14399		ctxt += krcvqs[i];
  14400	}
  14401
  14402	rrd.offset = rmt->used;
  14403	rrd.pkt_type = 2;
  14404	rrd.field1_off = LRH_BTH_MATCH_OFFSET;
  14405	rrd.field2_off = LRH_SC_MATCH_OFFSET;
  14406	rrd.index1_off = LRH_SC_SELECT_OFFSET;
  14407	rrd.index1_width = n;
  14408	rrd.index2_off = QPN_SELECT_OFFSET;
  14409	rrd.index2_width = m + n;
  14410	rrd.mask1 = LRH_BTH_MASK;
  14411	rrd.value1 = LRH_BTH_VALUE;
  14412	rrd.mask2 = LRH_SC_MASK;
  14413	rrd.value2 = LRH_SC_VALUE;
  14414
  14415	/* add rule 0 */
  14416	add_rsm_rule(dd, RSM_INS_VERBS, &rrd);
  14417
  14418	/* mark RSM map entries as used */
  14419	rmt->used += rmt_entries;
  14420	/* map everything else to the mcast/err/vl15 context */
  14421	init_qpmap_table(dd, HFI1_CTRL_CTXT, HFI1_CTRL_CTXT);
  14422	dd->qos_shift = n + 1;
  14423	return;
  14424bail:
  14425	dd->qos_shift = 1;
  14426	init_qpmap_table(dd, FIRST_KERNEL_KCTXT, dd->n_krcv_queues - 1);
  14427}
  14428
  14429static void init_fecn_handling(struct hfi1_devdata *dd,
  14430			       struct rsm_map_table *rmt)
  14431{
  14432	struct rsm_rule_data rrd;
  14433	u64 reg;
  14434	int i, idx, regoff, regidx, start;
  14435	u8 offset;
  14436	u32 total_cnt;
  14437
  14438	if (HFI1_CAP_IS_KSET(TID_RDMA))
  14439		/* Exclude context 0 */
  14440		start = 1;
  14441	else
  14442		start = dd->first_dyn_alloc_ctxt;
  14443
  14444	total_cnt = dd->num_rcv_contexts - start;
  14445
  14446	/* there needs to be enough room in the map table */
  14447	if (rmt->used + total_cnt >= NUM_MAP_ENTRIES) {
  14448		dd_dev_err(dd, "FECN handling disabled - too many contexts allocated\n");
  14449		return;
  14450	}
  14451
  14452	/*
  14453	 * RSM will extract the destination context as an index into the
  14454	 * map table.  The destination contexts are a sequential block
  14455	 * in the range start...num_rcv_contexts-1 (inclusive).
  14456	 * Map entries are accessed as offset + extracted value.  Adjust
  14457	 * the added offset so this sequence can be placed anywhere in
  14458	 * the table - as long as the entries themselves do not wrap.
  14459	 * There are only enough bits in offset for the table size, so
  14460	 * start with that to allow for a "negative" offset.
  14461	 */
  14462	offset = (u8)(NUM_MAP_ENTRIES + rmt->used - start);
  14463
  14464	for (i = start, idx = rmt->used; i < dd->num_rcv_contexts;
  14465	     i++, idx++) {
  14466		/* replace with identity mapping */
  14467		regoff = (idx % 8) * 8;
  14468		regidx = idx / 8;
  14469		reg = rmt->map[regidx];
  14470		reg &= ~(RCV_RSM_MAP_TABLE_RCV_CONTEXT_A_MASK << regoff);
  14471		reg |= (u64)i << regoff;
  14472		rmt->map[regidx] = reg;
  14473	}
  14474
  14475	/*
  14476	 * For RSM intercept of Expected FECN packets:
  14477	 * o packet type 0 - expected
  14478	 * o match on F (bit 95), using select/match 1, and
  14479	 * o match on SH (bit 133), using select/match 2.
  14480	 *
  14481	 * Use index 1 to extract the 8-bit receive context from DestQP
  14482	 * (start at bit 64).  Use that as the RSM map table index.
  14483	 */
  14484	rrd.offset = offset;
  14485	rrd.pkt_type = 0;
  14486	rrd.field1_off = 95;
  14487	rrd.field2_off = 133;
  14488	rrd.index1_off = 64;
  14489	rrd.index1_width = 8;
  14490	rrd.index2_off = 0;
  14491	rrd.index2_width = 0;
  14492	rrd.mask1 = 1;
  14493	rrd.value1 = 1;
  14494	rrd.mask2 = 1;
  14495	rrd.value2 = 1;
  14496
  14497	/* add rule 1 */
  14498	add_rsm_rule(dd, RSM_INS_FECN, &rrd);
  14499
  14500	rmt->used += total_cnt;
  14501}
  14502
  14503static inline bool hfi1_is_rmt_full(int start, int spare)
  14504{
  14505	return (start + spare) > NUM_MAP_ENTRIES;
  14506}
  14507
  14508static bool hfi1_netdev_update_rmt(struct hfi1_devdata *dd)
  14509{
  14510	u8 i, j;
  14511	u8 ctx_id = 0;
  14512	u64 reg;
  14513	u32 regoff;
  14514	int rmt_start = hfi1_netdev_get_free_rmt_idx(dd);
  14515	int ctxt_count = hfi1_netdev_ctxt_count(dd);
  14516
  14517	/* We already have contexts mapped in RMT */
  14518	if (has_rsm_rule(dd, RSM_INS_VNIC) || has_rsm_rule(dd, RSM_INS_AIP)) {
  14519		dd_dev_info(dd, "Contexts are already mapped in RMT\n");
  14520		return true;
  14521	}
  14522
  14523	if (hfi1_is_rmt_full(rmt_start, NUM_NETDEV_MAP_ENTRIES)) {
  14524		dd_dev_err(dd, "Not enough RMT entries used = %d\n",
  14525			   rmt_start);
  14526		return false;
  14527	}
  14528
  14529	dev_dbg(&(dd)->pcidev->dev, "RMT start = %d, end %d\n",
  14530		rmt_start,
  14531		rmt_start + NUM_NETDEV_MAP_ENTRIES);
  14532
  14533	/* Update RSM mapping table, 32 regs, 256 entries - 1 ctx per byte */
  14534	regoff = RCV_RSM_MAP_TABLE + (rmt_start / 8) * 8;
  14535	reg = read_csr(dd, regoff);
  14536	for (i = 0; i < NUM_NETDEV_MAP_ENTRIES; i++) {
  14537		/* Update map register with netdev context */
  14538		j = (rmt_start + i) % 8;
  14539		reg &= ~(0xffllu << (j * 8));
  14540		reg |= (u64)hfi1_netdev_get_ctxt(dd, ctx_id++)->ctxt << (j * 8);
  14541		/* Wrap up netdev ctx index */
  14542		ctx_id %= ctxt_count;
  14543		/* Write back map register */
  14544		if (j == 7 || ((i + 1) == NUM_NETDEV_MAP_ENTRIES)) {
  14545			dev_dbg(&(dd)->pcidev->dev,
  14546				"RMT[%d] =0x%llx\n",
  14547				regoff - RCV_RSM_MAP_TABLE, reg);
  14548
  14549			write_csr(dd, regoff, reg);
  14550			regoff += 8;
  14551			if (i < (NUM_NETDEV_MAP_ENTRIES - 1))
  14552				reg = read_csr(dd, regoff);
  14553		}
  14554	}
  14555
  14556	return true;
  14557}
  14558
  14559static void hfi1_enable_rsm_rule(struct hfi1_devdata *dd,
  14560				 int rule, struct rsm_rule_data *rrd)
  14561{
  14562	if (!hfi1_netdev_update_rmt(dd)) {
  14563		dd_dev_err(dd, "Failed to update RMT for RSM%d rule\n", rule);
  14564		return;
  14565	}
  14566
  14567	add_rsm_rule(dd, rule, rrd);
  14568	add_rcvctrl(dd, RCV_CTRL_RCV_RSM_ENABLE_SMASK);
  14569}
  14570
  14571void hfi1_init_aip_rsm(struct hfi1_devdata *dd)
  14572{
  14573	/*
  14574	 * go through with the initialisation only if this rule actually doesn't
  14575	 * exist yet
  14576	 */
  14577	if (atomic_fetch_inc(&dd->ipoib_rsm_usr_num) == 0) {
  14578		int rmt_start = hfi1_netdev_get_free_rmt_idx(dd);
  14579		struct rsm_rule_data rrd = {
  14580			.offset = rmt_start,
  14581			.pkt_type = IB_PACKET_TYPE,
  14582			.field1_off = LRH_BTH_MATCH_OFFSET,
  14583			.mask1 = LRH_BTH_MASK,
  14584			.value1 = LRH_BTH_VALUE,
  14585			.field2_off = BTH_DESTQP_MATCH_OFFSET,
  14586			.mask2 = BTH_DESTQP_MASK,
  14587			.value2 = BTH_DESTQP_VALUE,
  14588			.index1_off = DETH_AIP_SQPN_SELECT_OFFSET +
  14589					ilog2(NUM_NETDEV_MAP_ENTRIES),
  14590			.index1_width = ilog2(NUM_NETDEV_MAP_ENTRIES),
  14591			.index2_off = DETH_AIP_SQPN_SELECT_OFFSET,
  14592			.index2_width = ilog2(NUM_NETDEV_MAP_ENTRIES)
  14593		};
  14594
  14595		hfi1_enable_rsm_rule(dd, RSM_INS_AIP, &rrd);
  14596	}
  14597}
  14598
  14599/* Initialize RSM for VNIC */
  14600void hfi1_init_vnic_rsm(struct hfi1_devdata *dd)
  14601{
  14602	int rmt_start = hfi1_netdev_get_free_rmt_idx(dd);
  14603	struct rsm_rule_data rrd = {
  14604		/* Add rule for vnic */
  14605		.offset = rmt_start,
  14606		.pkt_type = 4,
  14607		/* Match 16B packets */
  14608		.field1_off = L2_TYPE_MATCH_OFFSET,
  14609		.mask1 = L2_TYPE_MASK,
  14610		.value1 = L2_16B_VALUE,
  14611		/* Match ETH L4 packets */
  14612		.field2_off = L4_TYPE_MATCH_OFFSET,
  14613		.mask2 = L4_16B_TYPE_MASK,
  14614		.value2 = L4_16B_ETH_VALUE,
  14615		/* Calc context from veswid and entropy */
  14616		.index1_off = L4_16B_HDR_VESWID_OFFSET,
  14617		.index1_width = ilog2(NUM_NETDEV_MAP_ENTRIES),
  14618		.index2_off = L2_16B_ENTROPY_OFFSET,
  14619		.index2_width = ilog2(NUM_NETDEV_MAP_ENTRIES)
  14620	};
  14621
  14622	hfi1_enable_rsm_rule(dd, RSM_INS_VNIC, &rrd);
  14623}
  14624
  14625void hfi1_deinit_vnic_rsm(struct hfi1_devdata *dd)
  14626{
  14627	clear_rsm_rule(dd, RSM_INS_VNIC);
  14628}
  14629
  14630void hfi1_deinit_aip_rsm(struct hfi1_devdata *dd)
  14631{
  14632	/* only actually clear the rule if it's the last user asking to do so */
  14633	if (atomic_fetch_add_unless(&dd->ipoib_rsm_usr_num, -1, 0) == 1)
  14634		clear_rsm_rule(dd, RSM_INS_AIP);
  14635}
  14636
  14637static int init_rxe(struct hfi1_devdata *dd)
  14638{
  14639	struct rsm_map_table *rmt;
  14640	u64 val;
  14641
  14642	/* enable all receive errors */
  14643	write_csr(dd, RCV_ERR_MASK, ~0ull);
  14644
  14645	rmt = alloc_rsm_map_table(dd);
  14646	if (!rmt)
  14647		return -ENOMEM;
  14648
  14649	/* set up QOS, including the QPN map table */
  14650	init_qos(dd, rmt);
  14651	init_fecn_handling(dd, rmt);
  14652	complete_rsm_map_table(dd, rmt);
  14653	/* record number of used rsm map entries for netdev */
  14654	hfi1_netdev_set_free_rmt_idx(dd, rmt->used);
  14655	kfree(rmt);
  14656
  14657	/*
  14658	 * make sure RcvCtrl.RcvWcb <= PCIe Device Control
  14659	 * Register Max_Payload_Size (PCI_EXP_DEVCTL in Linux PCIe config
  14660	 * space, PciCfgCap2.MaxPayloadSize in HFI).  There is only one
  14661	 * invalid configuration: RcvCtrl.RcvWcb set to its max of 256 and
  14662	 * Max_PayLoad_Size set to its minimum of 128.
  14663	 *
  14664	 * Presently, RcvCtrl.RcvWcb is not modified from its default of 0
  14665	 * (64 bytes).  Max_Payload_Size is possibly modified upward in
  14666	 * tune_pcie_caps() which is called after this routine.
  14667	 */
  14668
  14669	/* Have 16 bytes (4DW) of bypass header available in header queue */
  14670	val = read_csr(dd, RCV_BYPASS);
  14671	val &= ~RCV_BYPASS_HDR_SIZE_SMASK;
  14672	val |= ((4ull & RCV_BYPASS_HDR_SIZE_MASK) <<
  14673		RCV_BYPASS_HDR_SIZE_SHIFT);
  14674	write_csr(dd, RCV_BYPASS, val);
  14675	return 0;
  14676}
  14677
  14678static void init_other(struct hfi1_devdata *dd)
  14679{
  14680	/* enable all CCE errors */
  14681	write_csr(dd, CCE_ERR_MASK, ~0ull);
  14682	/* enable *some* Misc errors */
  14683	write_csr(dd, MISC_ERR_MASK, DRIVER_MISC_MASK);
  14684	/* enable all DC errors, except LCB */
  14685	write_csr(dd, DCC_ERR_FLG_EN, ~0ull);
  14686	write_csr(dd, DC_DC8051_ERR_EN, ~0ull);
  14687}
  14688
  14689/*
  14690 * Fill out the given AU table using the given CU.  A CU is defined in terms
  14691 * AUs.  The table is a an encoding: given the index, how many AUs does that
  14692 * represent?
  14693 *
  14694 * NOTE: Assumes that the register layout is the same for the
  14695 * local and remote tables.
  14696 */
  14697static void assign_cm_au_table(struct hfi1_devdata *dd, u32 cu,
  14698			       u32 csr0to3, u32 csr4to7)
  14699{
  14700	write_csr(dd, csr0to3,
  14701		  0ull << SEND_CM_LOCAL_AU_TABLE0_TO3_LOCAL_AU_TABLE0_SHIFT |
  14702		  1ull << SEND_CM_LOCAL_AU_TABLE0_TO3_LOCAL_AU_TABLE1_SHIFT |
  14703		  2ull * cu <<
  14704		  SEND_CM_LOCAL_AU_TABLE0_TO3_LOCAL_AU_TABLE2_SHIFT |
  14705		  4ull * cu <<
  14706		  SEND_CM_LOCAL_AU_TABLE0_TO3_LOCAL_AU_TABLE3_SHIFT);
  14707	write_csr(dd, csr4to7,
  14708		  8ull * cu <<
  14709		  SEND_CM_LOCAL_AU_TABLE4_TO7_LOCAL_AU_TABLE4_SHIFT |
  14710		  16ull * cu <<
  14711		  SEND_CM_LOCAL_AU_TABLE4_TO7_LOCAL_AU_TABLE5_SHIFT |
  14712		  32ull * cu <<
  14713		  SEND_CM_LOCAL_AU_TABLE4_TO7_LOCAL_AU_TABLE6_SHIFT |
  14714		  64ull * cu <<
  14715		  SEND_CM_LOCAL_AU_TABLE4_TO7_LOCAL_AU_TABLE7_SHIFT);
  14716}
  14717
  14718static void assign_local_cm_au_table(struct hfi1_devdata *dd, u8 vcu)
  14719{
  14720	assign_cm_au_table(dd, vcu_to_cu(vcu), SEND_CM_LOCAL_AU_TABLE0_TO3,
  14721			   SEND_CM_LOCAL_AU_TABLE4_TO7);
  14722}
  14723
  14724void assign_remote_cm_au_table(struct hfi1_devdata *dd, u8 vcu)
  14725{
  14726	assign_cm_au_table(dd, vcu_to_cu(vcu), SEND_CM_REMOTE_AU_TABLE0_TO3,
  14727			   SEND_CM_REMOTE_AU_TABLE4_TO7);
  14728}
  14729
  14730static void init_txe(struct hfi1_devdata *dd)
  14731{
  14732	int i;
  14733
  14734	/* enable all PIO, SDMA, general, and Egress errors */
  14735	write_csr(dd, SEND_PIO_ERR_MASK, ~0ull);
  14736	write_csr(dd, SEND_DMA_ERR_MASK, ~0ull);
  14737	write_csr(dd, SEND_ERR_MASK, ~0ull);
  14738	write_csr(dd, SEND_EGRESS_ERR_MASK, ~0ull);
  14739
  14740	/* enable all per-context and per-SDMA engine errors */
  14741	for (i = 0; i < chip_send_contexts(dd); i++)
  14742		write_kctxt_csr(dd, i, SEND_CTXT_ERR_MASK, ~0ull);
  14743	for (i = 0; i < chip_sdma_engines(dd); i++)
  14744		write_kctxt_csr(dd, i, SEND_DMA_ENG_ERR_MASK, ~0ull);
  14745
  14746	/* set the local CU to AU mapping */
  14747	assign_local_cm_au_table(dd, dd->vcu);
  14748
  14749	/*
  14750	 * Set reasonable default for Credit Return Timer
  14751	 * Don't set on Simulator - causes it to choke.
  14752	 */
  14753	if (dd->icode != ICODE_FUNCTIONAL_SIMULATOR)
  14754		write_csr(dd, SEND_CM_TIMER_CTRL, HFI1_CREDIT_RETURN_RATE);
  14755}
  14756
  14757int hfi1_set_ctxt_jkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd,
  14758		       u16 jkey)
  14759{
  14760	u8 hw_ctxt;
  14761	u64 reg;
  14762
  14763	if (!rcd || !rcd->sc)
  14764		return -EINVAL;
  14765
  14766	hw_ctxt = rcd->sc->hw_context;
  14767	reg = SEND_CTXT_CHECK_JOB_KEY_MASK_SMASK | /* mask is always 1's */
  14768		((jkey & SEND_CTXT_CHECK_JOB_KEY_VALUE_MASK) <<
  14769		 SEND_CTXT_CHECK_JOB_KEY_VALUE_SHIFT);
  14770	/* JOB_KEY_ALLOW_PERMISSIVE is not allowed by default */
  14771	if (HFI1_CAP_KGET_MASK(rcd->flags, ALLOW_PERM_JKEY))
  14772		reg |= SEND_CTXT_CHECK_JOB_KEY_ALLOW_PERMISSIVE_SMASK;
  14773	write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_JOB_KEY, reg);
  14774	/*
  14775	 * Enable send-side J_KEY integrity check, unless this is A0 h/w
  14776	 */
  14777	if (!is_ax(dd)) {
  14778		reg = read_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE);
  14779		reg |= SEND_CTXT_CHECK_ENABLE_CHECK_JOB_KEY_SMASK;
  14780		write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE, reg);
  14781	}
  14782
  14783	/* Enable J_KEY check on receive context. */
  14784	reg = RCV_KEY_CTRL_JOB_KEY_ENABLE_SMASK |
  14785		((jkey & RCV_KEY_CTRL_JOB_KEY_VALUE_MASK) <<
  14786		 RCV_KEY_CTRL_JOB_KEY_VALUE_SHIFT);
  14787	write_kctxt_csr(dd, rcd->ctxt, RCV_KEY_CTRL, reg);
  14788
  14789	return 0;
  14790}
  14791
  14792int hfi1_clear_ctxt_jkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd)
  14793{
  14794	u8 hw_ctxt;
  14795	u64 reg;
  14796
  14797	if (!rcd || !rcd->sc)
  14798		return -EINVAL;
  14799
  14800	hw_ctxt = rcd->sc->hw_context;
  14801	write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_JOB_KEY, 0);
  14802	/*
  14803	 * Disable send-side J_KEY integrity check, unless this is A0 h/w.
  14804	 * This check would not have been enabled for A0 h/w, see
  14805	 * set_ctxt_jkey().
  14806	 */
  14807	if (!is_ax(dd)) {
  14808		reg = read_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE);
  14809		reg &= ~SEND_CTXT_CHECK_ENABLE_CHECK_JOB_KEY_SMASK;
  14810		write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE, reg);
  14811	}
  14812	/* Turn off the J_KEY on the receive side */
  14813	write_kctxt_csr(dd, rcd->ctxt, RCV_KEY_CTRL, 0);
  14814
  14815	return 0;
  14816}
  14817
  14818int hfi1_set_ctxt_pkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd,
  14819		       u16 pkey)
  14820{
  14821	u8 hw_ctxt;
  14822	u64 reg;
  14823
  14824	if (!rcd || !rcd->sc)
  14825		return -EINVAL;
  14826
  14827	hw_ctxt = rcd->sc->hw_context;
  14828	reg = ((u64)pkey & SEND_CTXT_CHECK_PARTITION_KEY_VALUE_MASK) <<
  14829		SEND_CTXT_CHECK_PARTITION_KEY_VALUE_SHIFT;
  14830	write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_PARTITION_KEY, reg);
  14831	reg = read_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE);
  14832	reg |= SEND_CTXT_CHECK_ENABLE_CHECK_PARTITION_KEY_SMASK;
  14833	reg &= ~SEND_CTXT_CHECK_ENABLE_DISALLOW_KDETH_PACKETS_SMASK;
  14834	write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE, reg);
  14835
  14836	return 0;
  14837}
  14838
  14839int hfi1_clear_ctxt_pkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *ctxt)
  14840{
  14841	u8 hw_ctxt;
  14842	u64 reg;
  14843
  14844	if (!ctxt || !ctxt->sc)
  14845		return -EINVAL;
  14846
  14847	hw_ctxt = ctxt->sc->hw_context;
  14848	reg = read_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE);
  14849	reg &= ~SEND_CTXT_CHECK_ENABLE_CHECK_PARTITION_KEY_SMASK;
  14850	write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_ENABLE, reg);
  14851	write_kctxt_csr(dd, hw_ctxt, SEND_CTXT_CHECK_PARTITION_KEY, 0);
  14852
  14853	return 0;
  14854}
  14855
  14856/*
  14857 * Start doing the clean up the chip. Our clean up happens in multiple
  14858 * stages and this is just the first.
  14859 */
  14860void hfi1_start_cleanup(struct hfi1_devdata *dd)
  14861{
  14862	aspm_exit(dd);
  14863	free_cntrs(dd);
  14864	free_rcverr(dd);
  14865	finish_chip_resources(dd);
  14866}
  14867
  14868#define HFI_BASE_GUID(dev) \
  14869	((dev)->base_guid & ~(1ULL << GUID_HFI_INDEX_SHIFT))
  14870
  14871/*
  14872 * Information can be shared between the two HFIs on the same ASIC
  14873 * in the same OS.  This function finds the peer device and sets
  14874 * up a shared structure.
  14875 */
  14876static int init_asic_data(struct hfi1_devdata *dd)
  14877{
  14878	unsigned long index;
  14879	struct hfi1_devdata *peer;
  14880	struct hfi1_asic_data *asic_data;
  14881	int ret = 0;
  14882
  14883	/* pre-allocate the asic structure in case we are the first device */
  14884	asic_data = kzalloc(sizeof(*dd->asic_data), GFP_KERNEL);
  14885	if (!asic_data)
  14886		return -ENOMEM;
  14887
  14888	xa_lock_irq(&hfi1_dev_table);
  14889	/* Find our peer device */
  14890	xa_for_each(&hfi1_dev_table, index, peer) {
  14891		if ((HFI_BASE_GUID(dd) == HFI_BASE_GUID(peer)) &&
  14892		    dd->unit != peer->unit)
  14893			break;
  14894	}
  14895
  14896	if (peer) {
  14897		/* use already allocated structure */
  14898		dd->asic_data = peer->asic_data;
  14899		kfree(asic_data);
  14900	} else {
  14901		dd->asic_data = asic_data;
  14902		mutex_init(&dd->asic_data->asic_resource_mutex);
  14903	}
  14904	dd->asic_data->dds[dd->hfi1_id] = dd; /* self back-pointer */
  14905	xa_unlock_irq(&hfi1_dev_table);
  14906
  14907	/* first one through - set up i2c devices */
  14908	if (!peer)
  14909		ret = set_up_i2c(dd, dd->asic_data);
  14910
  14911	return ret;
  14912}
  14913
  14914/*
  14915 * Set dd->boardname.  Use a generic name if a name is not returned from
  14916 * EFI variable space.
  14917 *
  14918 * Return 0 on success, -ENOMEM if space could not be allocated.
  14919 */
  14920static int obtain_boardname(struct hfi1_devdata *dd)
  14921{
  14922	/* generic board description */
  14923	const char generic[] =
  14924		"Cornelis Omni-Path Host Fabric Interface Adapter 100 Series";
  14925	unsigned long size;
  14926	int ret;
  14927
  14928	ret = read_hfi1_efi_var(dd, "description", &size,
  14929				(void **)&dd->boardname);
  14930	if (ret) {
  14931		dd_dev_info(dd, "Board description not found\n");
  14932		/* use generic description */
  14933		dd->boardname = kstrdup(generic, GFP_KERNEL);
  14934		if (!dd->boardname)
  14935			return -ENOMEM;
  14936	}
  14937	return 0;
  14938}
  14939
  14940/*
  14941 * Check the interrupt registers to make sure that they are mapped correctly.
  14942 * It is intended to help user identify any mismapping by VMM when the driver
  14943 * is running in a VM. This function should only be called before interrupt
  14944 * is set up properly.
  14945 *
  14946 * Return 0 on success, -EINVAL on failure.
  14947 */
  14948static int check_int_registers(struct hfi1_devdata *dd)
  14949{
  14950	u64 reg;
  14951	u64 all_bits = ~(u64)0;
  14952	u64 mask;
  14953
  14954	/* Clear CceIntMask[0] to avoid raising any interrupts */
  14955	mask = read_csr(dd, CCE_INT_MASK);
  14956	write_csr(dd, CCE_INT_MASK, 0ull);
  14957	reg = read_csr(dd, CCE_INT_MASK);
  14958	if (reg)
  14959		goto err_exit;
  14960
  14961	/* Clear all interrupt status bits */
  14962	write_csr(dd, CCE_INT_CLEAR, all_bits);
  14963	reg = read_csr(dd, CCE_INT_STATUS);
  14964	if (reg)
  14965		goto err_exit;
  14966
  14967	/* Set all interrupt status bits */
  14968	write_csr(dd, CCE_INT_FORCE, all_bits);
  14969	reg = read_csr(dd, CCE_INT_STATUS);
  14970	if (reg != all_bits)
  14971		goto err_exit;
  14972
  14973	/* Restore the interrupt mask */
  14974	write_csr(dd, CCE_INT_CLEAR, all_bits);
  14975	write_csr(dd, CCE_INT_MASK, mask);
  14976
  14977	return 0;
  14978err_exit:
  14979	write_csr(dd, CCE_INT_MASK, mask);
  14980	dd_dev_err(dd, "Interrupt registers not properly mapped by VMM\n");
  14981	return -EINVAL;
  14982}
  14983
  14984/**
  14985 * hfi1_init_dd() - Initialize most of the dd structure.
  14986 * @dd: the dd device
  14987 *
  14988 * This is global, and is called directly at init to set up the
  14989 * chip-specific function pointers for later use.
  14990 */
  14991int hfi1_init_dd(struct hfi1_devdata *dd)
  14992{
  14993	struct pci_dev *pdev = dd->pcidev;
  14994	struct hfi1_pportdata *ppd;
  14995	u64 reg;
  14996	int i, ret;
  14997	static const char * const inames[] = { /* implementation names */
  14998		"RTL silicon",
  14999		"RTL VCS simulation",
  15000		"RTL FPGA emulation",
  15001		"Functional simulator"
  15002	};
  15003	struct pci_dev *parent = pdev->bus->self;
  15004	u32 sdma_engines = chip_sdma_engines(dd);
  15005
  15006	ppd = dd->pport;
  15007	for (i = 0; i < dd->num_pports; i++, ppd++) {
  15008		int vl;
  15009		/* init common fields */
  15010		hfi1_init_pportdata(pdev, ppd, dd, 0, 1);
  15011		/* DC supports 4 link widths */
  15012		ppd->link_width_supported =
  15013			OPA_LINK_WIDTH_1X | OPA_LINK_WIDTH_2X |
  15014			OPA_LINK_WIDTH_3X | OPA_LINK_WIDTH_4X;
  15015		ppd->link_width_downgrade_supported =
  15016			ppd->link_width_supported;
  15017		/* start out enabling only 4X */
  15018		ppd->link_width_enabled = OPA_LINK_WIDTH_4X;
  15019		ppd->link_width_downgrade_enabled =
  15020					ppd->link_width_downgrade_supported;
  15021		/* link width active is 0 when link is down */
  15022		/* link width downgrade active is 0 when link is down */
  15023
  15024		if (num_vls < HFI1_MIN_VLS_SUPPORTED ||
  15025		    num_vls > HFI1_MAX_VLS_SUPPORTED) {
  15026			dd_dev_err(dd, "Invalid num_vls %u, using %u VLs\n",
  15027				   num_vls, HFI1_MAX_VLS_SUPPORTED);
  15028			num_vls = HFI1_MAX_VLS_SUPPORTED;
  15029		}
  15030		ppd->vls_supported = num_vls;
  15031		ppd->vls_operational = ppd->vls_supported;
  15032		/* Set the default MTU. */
  15033		for (vl = 0; vl < num_vls; vl++)
  15034			dd->vld[vl].mtu = hfi1_max_mtu;
  15035		dd->vld[15].mtu = MAX_MAD_PACKET;
  15036		/*
  15037		 * Set the initial values to reasonable default, will be set
  15038		 * for real when link is up.
  15039		 */
  15040		ppd->overrun_threshold = 0x4;
  15041		ppd->phy_error_threshold = 0xf;
  15042		ppd->port_crc_mode_enabled = link_crc_mask;
  15043		/* initialize supported LTP CRC mode */
  15044		ppd->port_ltp_crc_mode = cap_to_port_ltp(link_crc_mask) << 8;
  15045		/* initialize enabled LTP CRC mode */
  15046		ppd->port_ltp_crc_mode |= cap_to_port_ltp(link_crc_mask) << 4;
  15047		/* start in offline */
  15048		ppd->host_link_state = HLS_DN_OFFLINE;
  15049		init_vl_arb_caches(ppd);
  15050	}
  15051
  15052	/*
  15053	 * Do remaining PCIe setup and save PCIe values in dd.
  15054	 * Any error printing is already done by the init code.
  15055	 * On return, we have the chip mapped.
  15056	 */
  15057	ret = hfi1_pcie_ddinit(dd, pdev);
  15058	if (ret < 0)
  15059		goto bail_free;
  15060
  15061	/* Save PCI space registers to rewrite after device reset */
  15062	ret = save_pci_variables(dd);
  15063	if (ret < 0)
  15064		goto bail_cleanup;
  15065
  15066	dd->majrev = (dd->revision >> CCE_REVISION_CHIP_REV_MAJOR_SHIFT)
  15067			& CCE_REVISION_CHIP_REV_MAJOR_MASK;
  15068	dd->minrev = (dd->revision >> CCE_REVISION_CHIP_REV_MINOR_SHIFT)
  15069			& CCE_REVISION_CHIP_REV_MINOR_MASK;
  15070
  15071	/*
  15072	 * Check interrupt registers mapping if the driver has no access to
  15073	 * the upstream component. In this case, it is likely that the driver
  15074	 * is running in a VM.
  15075	 */
  15076	if (!parent) {
  15077		ret = check_int_registers(dd);
  15078		if (ret)
  15079			goto bail_cleanup;
  15080	}
  15081
  15082	/*
  15083	 * obtain the hardware ID - NOT related to unit, which is a
  15084	 * software enumeration
  15085	 */
  15086	reg = read_csr(dd, CCE_REVISION2);
  15087	dd->hfi1_id = (reg >> CCE_REVISION2_HFI_ID_SHIFT)
  15088					& CCE_REVISION2_HFI_ID_MASK;
  15089	/* the variable size will remove unwanted bits */
  15090	dd->icode = reg >> CCE_REVISION2_IMPL_CODE_SHIFT;
  15091	dd->irev = reg >> CCE_REVISION2_IMPL_REVISION_SHIFT;
  15092	dd_dev_info(dd, "Implementation: %s, revision 0x%x\n",
  15093		    dd->icode < ARRAY_SIZE(inames) ?
  15094		    inames[dd->icode] : "unknown", (int)dd->irev);
  15095
  15096	/* speeds the hardware can support */
  15097	dd->pport->link_speed_supported = OPA_LINK_SPEED_25G;
  15098	/* speeds allowed to run at */
  15099	dd->pport->link_speed_enabled = dd->pport->link_speed_supported;
  15100	/* give a reasonable active value, will be set on link up */
  15101	dd->pport->link_speed_active = OPA_LINK_SPEED_25G;
  15102
  15103	/* fix up link widths for emulation _p */
  15104	ppd = dd->pport;
  15105	if (dd->icode == ICODE_FPGA_EMULATION && is_emulator_p(dd)) {
  15106		ppd->link_width_supported =
  15107			ppd->link_width_enabled =
  15108			ppd->link_width_downgrade_supported =
  15109			ppd->link_width_downgrade_enabled =
  15110				OPA_LINK_WIDTH_1X;
  15111	}
  15112	/* insure num_vls isn't larger than number of sdma engines */
  15113	if (HFI1_CAP_IS_KSET(SDMA) && num_vls > sdma_engines) {
  15114		dd_dev_err(dd, "num_vls %u too large, using %u VLs\n",
  15115			   num_vls, sdma_engines);
  15116		num_vls = sdma_engines;
  15117		ppd->vls_supported = sdma_engines;
  15118		ppd->vls_operational = ppd->vls_supported;
  15119	}
  15120
  15121	/*
  15122	 * Convert the ns parameter to the 64 * cclocks used in the CSR.
  15123	 * Limit the max if larger than the field holds.  If timeout is
  15124	 * non-zero, then the calculated field will be at least 1.
  15125	 *
  15126	 * Must be after icode is set up - the cclock rate depends
  15127	 * on knowing the hardware being used.
  15128	 */
  15129	dd->rcv_intr_timeout_csr = ns_to_cclock(dd, rcv_intr_timeout) / 64;
  15130	if (dd->rcv_intr_timeout_csr >
  15131			RCV_AVAIL_TIME_OUT_TIME_OUT_RELOAD_MASK)
  15132		dd->rcv_intr_timeout_csr =
  15133			RCV_AVAIL_TIME_OUT_TIME_OUT_RELOAD_MASK;
  15134	else if (dd->rcv_intr_timeout_csr == 0 && rcv_intr_timeout)
  15135		dd->rcv_intr_timeout_csr = 1;
  15136
  15137	/* needs to be done before we look for the peer device */
  15138	read_guid(dd);
  15139
  15140	/* set up shared ASIC data with peer device */
  15141	ret = init_asic_data(dd);
  15142	if (ret)
  15143		goto bail_cleanup;
  15144
  15145	/* obtain chip sizes, reset chip CSRs */
  15146	ret = init_chip(dd);
  15147	if (ret)
  15148		goto bail_cleanup;
  15149
  15150	/* read in the PCIe link speed information */
  15151	ret = pcie_speeds(dd);
  15152	if (ret)
  15153		goto bail_cleanup;
  15154
  15155	/* call before get_platform_config(), after init_chip_resources() */
  15156	ret = eprom_init(dd);
  15157	if (ret)
  15158		goto bail_free_rcverr;
  15159
  15160	/* Needs to be called before hfi1_firmware_init */
  15161	get_platform_config(dd);
  15162
  15163	/* read in firmware */
  15164	ret = hfi1_firmware_init(dd);
  15165	if (ret)
  15166		goto bail_cleanup;
  15167
  15168	/*
  15169	 * In general, the PCIe Gen3 transition must occur after the
  15170	 * chip has been idled (so it won't initiate any PCIe transactions
  15171	 * e.g. an interrupt) and before the driver changes any registers
  15172	 * (the transition will reset the registers).
  15173	 *
  15174	 * In particular, place this call after:
  15175	 * - init_chip()     - the chip will not initiate any PCIe transactions
  15176	 * - pcie_speeds()   - reads the current link speed
  15177	 * - hfi1_firmware_init() - the needed firmware is ready to be
  15178	 *			    downloaded
  15179	 */
  15180	ret = do_pcie_gen3_transition(dd);
  15181	if (ret)
  15182		goto bail_cleanup;
  15183
  15184	/*
  15185	 * This should probably occur in hfi1_pcie_init(), but historically
  15186	 * occurs after the do_pcie_gen3_transition() code.
  15187	 */
  15188	tune_pcie_caps(dd);
  15189
  15190	/* start setting dd values and adjusting CSRs */
  15191	init_early_variables(dd);
  15192
  15193	parse_platform_config(dd);
  15194
  15195	ret = obtain_boardname(dd);
  15196	if (ret)
  15197		goto bail_cleanup;
  15198
  15199	snprintf(dd->boardversion, BOARD_VERS_MAX,
  15200		 "ChipABI %u.%u, ChipRev %u.%u, SW Compat %llu\n",
  15201		 HFI1_CHIP_VERS_MAJ, HFI1_CHIP_VERS_MIN,
  15202		 (u32)dd->majrev,
  15203		 (u32)dd->minrev,
  15204		 (dd->revision >> CCE_REVISION_SW_SHIFT)
  15205		    & CCE_REVISION_SW_MASK);
  15206
  15207	/* alloc VNIC/AIP rx data */
  15208	ret = hfi1_alloc_rx(dd);
  15209	if (ret)
  15210		goto bail_cleanup;
  15211
  15212	ret = set_up_context_variables(dd);
  15213	if (ret)
  15214		goto bail_cleanup;
  15215
  15216	/* set initial RXE CSRs */
  15217	ret = init_rxe(dd);
  15218	if (ret)
  15219		goto bail_cleanup;
  15220
  15221	/* set initial TXE CSRs */
  15222	init_txe(dd);
  15223	/* set initial non-RXE, non-TXE CSRs */
  15224	init_other(dd);
  15225	/* set up KDETH QP prefix in both RX and TX CSRs */
  15226	init_kdeth_qp(dd);
  15227
  15228	ret = hfi1_dev_affinity_init(dd);
  15229	if (ret)
  15230		goto bail_cleanup;
  15231
  15232	/* send contexts must be set up before receive contexts */
  15233	ret = init_send_contexts(dd);
  15234	if (ret)
  15235		goto bail_cleanup;
  15236
  15237	ret = hfi1_create_kctxts(dd);
  15238	if (ret)
  15239		goto bail_cleanup;
  15240
  15241	/*
  15242	 * Initialize aspm, to be done after gen3 transition and setting up
  15243	 * contexts and before enabling interrupts
  15244	 */
  15245	aspm_init(dd);
  15246
  15247	ret = init_pervl_scs(dd);
  15248	if (ret)
  15249		goto bail_cleanup;
  15250
  15251	/* sdma init */
  15252	for (i = 0; i < dd->num_pports; ++i) {
  15253		ret = sdma_init(dd, i);
  15254		if (ret)
  15255			goto bail_cleanup;
  15256	}
  15257
  15258	/* use contexts created by hfi1_create_kctxts */
  15259	ret = set_up_interrupts(dd);
  15260	if (ret)
  15261		goto bail_cleanup;
  15262
  15263	ret = hfi1_comp_vectors_set_up(dd);
  15264	if (ret)
  15265		goto bail_clear_intr;
  15266
  15267	/* set up LCB access - must be after set_up_interrupts() */
  15268	init_lcb_access(dd);
  15269
  15270	/*
  15271	 * Serial number is created from the base guid:
  15272	 * [27:24] = base guid [38:35]
  15273	 * [23: 0] = base guid [23: 0]
  15274	 */
  15275	snprintf(dd->serial, SERIAL_MAX, "0x%08llx\n",
  15276		 (dd->base_guid & 0xFFFFFF) |
  15277		     ((dd->base_guid >> 11) & 0xF000000));
  15278
  15279	dd->oui1 = dd->base_guid >> 56 & 0xFF;
  15280	dd->oui2 = dd->base_guid >> 48 & 0xFF;
  15281	dd->oui3 = dd->base_guid >> 40 & 0xFF;
  15282
  15283	ret = load_firmware(dd); /* asymmetric with dispose_firmware() */
  15284	if (ret)
  15285		goto bail_clear_intr;
  15286
  15287	thermal_init(dd);
  15288
  15289	ret = init_cntrs(dd);
  15290	if (ret)
  15291		goto bail_clear_intr;
  15292
  15293	ret = init_rcverr(dd);
  15294	if (ret)
  15295		goto bail_free_cntrs;
  15296
  15297	init_completion(&dd->user_comp);
  15298
  15299	/* The user refcount starts with one to inidicate an active device */
  15300	refcount_set(&dd->user_refcount, 1);
  15301
  15302	goto bail;
  15303
  15304bail_free_rcverr:
  15305	free_rcverr(dd);
  15306bail_free_cntrs:
  15307	free_cntrs(dd);
  15308bail_clear_intr:
  15309	hfi1_comp_vectors_clean_up(dd);
  15310	msix_clean_up_interrupts(dd);
  15311bail_cleanup:
  15312	hfi1_free_rx(dd);
  15313	hfi1_pcie_ddcleanup(dd);
  15314bail_free:
  15315	hfi1_free_devdata(dd);
  15316bail:
  15317	return ret;
  15318}
  15319
  15320static u16 delay_cycles(struct hfi1_pportdata *ppd, u32 desired_egress_rate,
  15321			u32 dw_len)
  15322{
  15323	u32 delta_cycles;
  15324	u32 current_egress_rate = ppd->current_egress_rate;
  15325	/* rates here are in units of 10^6 bits/sec */
  15326
  15327	if (desired_egress_rate == -1)
  15328		return 0; /* shouldn't happen */
  15329
  15330	if (desired_egress_rate >= current_egress_rate)
  15331		return 0; /* we can't help go faster, only slower */
  15332
  15333	delta_cycles = egress_cycles(dw_len * 4, desired_egress_rate) -
  15334			egress_cycles(dw_len * 4, current_egress_rate);
  15335
  15336	return (u16)delta_cycles;
  15337}
  15338
  15339/**
  15340 * create_pbc - build a pbc for transmission
  15341 * @ppd: info of physical Hfi port
  15342 * @flags: special case flags or-ed in built pbc
  15343 * @srate_mbs: static rate
  15344 * @vl: vl
  15345 * @dw_len: dword length (header words + data words + pbc words)
  15346 *
  15347 * Create a PBC with the given flags, rate, VL, and length.
  15348 *
  15349 * NOTE: The PBC created will not insert any HCRC - all callers but one are
  15350 * for verbs, which does not use this PSM feature.  The lone other caller
  15351 * is for the diagnostic interface which calls this if the user does not
  15352 * supply their own PBC.
  15353 */
  15354u64 create_pbc(struct hfi1_pportdata *ppd, u64 flags, int srate_mbs, u32 vl,
  15355	       u32 dw_len)
  15356{
  15357	u64 pbc, delay = 0;
  15358
  15359	if (unlikely(srate_mbs))
  15360		delay = delay_cycles(ppd, srate_mbs, dw_len);
  15361
  15362	pbc = flags
  15363		| (delay << PBC_STATIC_RATE_CONTROL_COUNT_SHIFT)
  15364		| ((u64)PBC_IHCRC_NONE << PBC_INSERT_HCRC_SHIFT)
  15365		| (vl & PBC_VL_MASK) << PBC_VL_SHIFT
  15366		| (dw_len & PBC_LENGTH_DWS_MASK)
  15367			<< PBC_LENGTH_DWS_SHIFT;
  15368
  15369	return pbc;
  15370}
  15371
  15372#define SBUS_THERMAL    0x4f
  15373#define SBUS_THERM_MONITOR_MODE 0x1
  15374
  15375#define THERM_FAILURE(dev, ret, reason) \
  15376	dd_dev_err((dd),						\
  15377		   "Thermal sensor initialization failed: %s (%d)\n",	\
  15378		   (reason), (ret))
  15379
  15380/*
  15381 * Initialize the thermal sensor.
  15382 *
  15383 * After initialization, enable polling of thermal sensor through
  15384 * SBus interface. In order for this to work, the SBus Master
  15385 * firmware has to be loaded due to the fact that the HW polling
  15386 * logic uses SBus interrupts, which are not supported with
  15387 * default firmware. Otherwise, no data will be returned through
  15388 * the ASIC_STS_THERM CSR.
  15389 */
  15390static int thermal_init(struct hfi1_devdata *dd)
  15391{
  15392	int ret = 0;
  15393
  15394	if (dd->icode != ICODE_RTL_SILICON ||
  15395	    check_chip_resource(dd, CR_THERM_INIT, NULL))
  15396		return ret;
  15397
  15398	ret = acquire_chip_resource(dd, CR_SBUS, SBUS_TIMEOUT);
  15399	if (ret) {
  15400		THERM_FAILURE(dd, ret, "Acquire SBus");
  15401		return ret;
  15402	}
  15403
  15404	dd_dev_info(dd, "Initializing thermal sensor\n");
  15405	/* Disable polling of thermal readings */
  15406	write_csr(dd, ASIC_CFG_THERM_POLL_EN, 0x0);
  15407	msleep(100);
  15408	/* Thermal Sensor Initialization */
  15409	/*    Step 1: Reset the Thermal SBus Receiver */
  15410	ret = sbus_request_slow(dd, SBUS_THERMAL, 0x0,
  15411				RESET_SBUS_RECEIVER, 0);
  15412	if (ret) {
  15413		THERM_FAILURE(dd, ret, "Bus Reset");
  15414		goto done;
  15415	}
  15416	/*    Step 2: Set Reset bit in Thermal block */
  15417	ret = sbus_request_slow(dd, SBUS_THERMAL, 0x0,
  15418				WRITE_SBUS_RECEIVER, 0x1);
  15419	if (ret) {
  15420		THERM_FAILURE(dd, ret, "Therm Block Reset");
  15421		goto done;
  15422	}
  15423	/*    Step 3: Write clock divider value (100MHz -> 2MHz) */
  15424	ret = sbus_request_slow(dd, SBUS_THERMAL, 0x1,
  15425				WRITE_SBUS_RECEIVER, 0x32);
  15426	if (ret) {
  15427		THERM_FAILURE(dd, ret, "Write Clock Div");
  15428		goto done;
  15429	}
  15430	/*    Step 4: Select temperature mode */
  15431	ret = sbus_request_slow(dd, SBUS_THERMAL, 0x3,
  15432				WRITE_SBUS_RECEIVER,
  15433				SBUS_THERM_MONITOR_MODE);
  15434	if (ret) {
  15435		THERM_FAILURE(dd, ret, "Write Mode Sel");
  15436		goto done;
  15437	}
  15438	/*    Step 5: De-assert block reset and start conversion */
  15439	ret = sbus_request_slow(dd, SBUS_THERMAL, 0x0,
  15440				WRITE_SBUS_RECEIVER, 0x2);
  15441	if (ret) {
  15442		THERM_FAILURE(dd, ret, "Write Reset Deassert");
  15443		goto done;
  15444	}
  15445	/*    Step 5.1: Wait for first conversion (21.5ms per spec) */
  15446	msleep(22);
  15447
  15448	/* Enable polling of thermal readings */
  15449	write_csr(dd, ASIC_CFG_THERM_POLL_EN, 0x1);
  15450
  15451	/* Set initialized flag */
  15452	ret = acquire_chip_resource(dd, CR_THERM_INIT, 0);
  15453	if (ret)
  15454		THERM_FAILURE(dd, ret, "Unable to set thermal init flag");
  15455
  15456done:
  15457	release_chip_resource(dd, CR_SBUS);
  15458	return ret;
  15459}
  15460
  15461static void handle_temp_err(struct hfi1_devdata *dd)
  15462{
  15463	struct hfi1_pportdata *ppd = &dd->pport[0];
  15464	/*
  15465	 * Thermal Critical Interrupt
  15466	 * Put the device into forced freeze mode, take link down to
  15467	 * offline, and put DC into reset.
  15468	 */
  15469	dd_dev_emerg(dd,
  15470		     "Critical temperature reached! Forcing device into freeze mode!\n");
  15471	dd->flags |= HFI1_FORCED_FREEZE;
  15472	start_freeze_handling(ppd, FREEZE_SELF | FREEZE_ABORT);
  15473	/*
  15474	 * Shut DC down as much and as quickly as possible.
  15475	 *
  15476	 * Step 1: Take the link down to OFFLINE. This will cause the
  15477	 *         8051 to put the Serdes in reset. However, we don't want to
  15478	 *         go through the entire link state machine since we want to
  15479	 *         shutdown ASAP. Furthermore, this is not a graceful shutdown
  15480	 *         but rather an attempt to save the chip.
  15481	 *         Code below is almost the same as quiet_serdes() but avoids
  15482	 *         all the extra work and the sleeps.
  15483	 */
  15484	ppd->driver_link_ready = 0;
  15485	ppd->link_enabled = 0;
  15486	set_physical_link_state(dd, (OPA_LINKDOWN_REASON_SMA_DISABLED << 8) |
  15487				PLS_OFFLINE);
  15488	/*
  15489	 * Step 2: Shutdown LCB and 8051
  15490	 *         After shutdown, do not restore DC_CFG_RESET value.
  15491	 */
  15492	dc_shutdown(dd);
  15493}