cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

ixgbe_dcb_82599.c (9584B)


      1// SPDX-License-Identifier: GPL-2.0
      2/* Copyright(c) 1999 - 2018 Intel Corporation. */
      3
      4#include "ixgbe.h"
      5#include "ixgbe_type.h"
      6#include "ixgbe_dcb.h"
      7#include "ixgbe_dcb_82599.h"
      8
      9/**
     10 * ixgbe_dcb_config_rx_arbiter_82599 - Config Rx Data arbiter
     11 * @hw: pointer to hardware structure
     12 * @refill: refill credits index by traffic class
     13 * @max: max credits index by traffic class
     14 * @bwg_id: bandwidth grouping indexed by traffic class
     15 * @prio_type: priority type indexed by traffic class
     16 * @prio_tc: priority to tc assignments indexed by priority
     17 *
     18 * Configure Rx Packet Arbiter and credits for each traffic class.
     19 */
     20s32 ixgbe_dcb_config_rx_arbiter_82599(struct ixgbe_hw *hw,
     21				      u16 *refill,
     22				      u16 *max,
     23				      u8 *bwg_id,
     24				      u8 *prio_type,
     25				      u8 *prio_tc)
     26{
     27	u32    reg           = 0;
     28	u32    credit_refill = 0;
     29	u32    credit_max    = 0;
     30	u8     i             = 0;
     31
     32	/*
     33	 * Disable the arbiter before changing parameters
     34	 * (always enable recycle mode; WSP)
     35	 */
     36	reg = IXGBE_RTRPCS_RRM | IXGBE_RTRPCS_RAC | IXGBE_RTRPCS_ARBDIS;
     37	IXGBE_WRITE_REG(hw, IXGBE_RTRPCS, reg);
     38
     39	/* Map all traffic classes to their UP */
     40	reg = 0;
     41	for (i = 0; i < MAX_USER_PRIORITY; i++)
     42		reg |= (prio_tc[i] << (i * IXGBE_RTRUP2TC_UP_SHIFT));
     43	IXGBE_WRITE_REG(hw, IXGBE_RTRUP2TC, reg);
     44
     45	/* Configure traffic class credits and priority */
     46	for (i = 0; i < MAX_TRAFFIC_CLASS; i++) {
     47		credit_refill = refill[i];
     48		credit_max    = max[i];
     49		reg = credit_refill | (credit_max << IXGBE_RTRPT4C_MCL_SHIFT);
     50
     51		reg |= (u32)(bwg_id[i]) << IXGBE_RTRPT4C_BWG_SHIFT;
     52
     53		if (prio_type[i] == prio_link)
     54			reg |= IXGBE_RTRPT4C_LSP;
     55
     56		IXGBE_WRITE_REG(hw, IXGBE_RTRPT4C(i), reg);
     57	}
     58
     59	/*
     60	 * Configure Rx packet plane (recycle mode; WSP) and
     61	 * enable arbiter
     62	 */
     63	reg = IXGBE_RTRPCS_RRM | IXGBE_RTRPCS_RAC;
     64	IXGBE_WRITE_REG(hw, IXGBE_RTRPCS, reg);
     65
     66	return 0;
     67}
     68
     69/**
     70 * ixgbe_dcb_config_tx_desc_arbiter_82599 - Config Tx Desc. arbiter
     71 * @hw: pointer to hardware structure
     72 * @refill: refill credits index by traffic class
     73 * @max: max credits index by traffic class
     74 * @bwg_id: bandwidth grouping indexed by traffic class
     75 * @prio_type: priority type indexed by traffic class
     76 *
     77 * Configure Tx Descriptor Arbiter and credits for each traffic class.
     78 */
     79s32 ixgbe_dcb_config_tx_desc_arbiter_82599(struct ixgbe_hw *hw,
     80					   u16 *refill,
     81					   u16 *max,
     82					   u8 *bwg_id,
     83					   u8 *prio_type)
     84{
     85	u32    reg, max_credits;
     86	u8     i;
     87
     88	/* Clear the per-Tx queue credits; we use per-TC instead */
     89	for (i = 0; i < 128; i++) {
     90		IXGBE_WRITE_REG(hw, IXGBE_RTTDQSEL, i);
     91		IXGBE_WRITE_REG(hw, IXGBE_RTTDT1C, 0);
     92	}
     93
     94	/* Configure traffic class credits and priority */
     95	for (i = 0; i < MAX_TRAFFIC_CLASS; i++) {
     96		max_credits = max[i];
     97		reg = max_credits << IXGBE_RTTDT2C_MCL_SHIFT;
     98		reg |= refill[i];
     99		reg |= (u32)(bwg_id[i]) << IXGBE_RTTDT2C_BWG_SHIFT;
    100
    101		if (prio_type[i] == prio_group)
    102			reg |= IXGBE_RTTDT2C_GSP;
    103
    104		if (prio_type[i] == prio_link)
    105			reg |= IXGBE_RTTDT2C_LSP;
    106
    107		IXGBE_WRITE_REG(hw, IXGBE_RTTDT2C(i), reg);
    108	}
    109
    110	/*
    111	 * Configure Tx descriptor plane (recycle mode; WSP) and
    112	 * enable arbiter
    113	 */
    114	reg = IXGBE_RTTDCS_TDPAC | IXGBE_RTTDCS_TDRM;
    115	IXGBE_WRITE_REG(hw, IXGBE_RTTDCS, reg);
    116
    117	return 0;
    118}
    119
    120/**
    121 * ixgbe_dcb_config_tx_data_arbiter_82599 - Config Tx Data arbiter
    122 * @hw: pointer to hardware structure
    123 * @refill: refill credits index by traffic class
    124 * @max: max credits index by traffic class
    125 * @bwg_id: bandwidth grouping indexed by traffic class
    126 * @prio_type: priority type indexed by traffic class
    127 * @prio_tc: priority to tc assignments indexed by priority
    128 *
    129 * Configure Tx Packet Arbiter and credits for each traffic class.
    130 */
    131s32 ixgbe_dcb_config_tx_data_arbiter_82599(struct ixgbe_hw *hw,
    132					   u16 *refill,
    133					   u16 *max,
    134					   u8 *bwg_id,
    135					   u8 *prio_type,
    136					   u8 *prio_tc)
    137{
    138	u32 reg;
    139	u8 i;
    140
    141	/*
    142	 * Disable the arbiter before changing parameters
    143	 * (always enable recycle mode; SP; arb delay)
    144	 */
    145	reg = IXGBE_RTTPCS_TPPAC | IXGBE_RTTPCS_TPRM |
    146	      (IXGBE_RTTPCS_ARBD_DCB << IXGBE_RTTPCS_ARBD_SHIFT) |
    147	      IXGBE_RTTPCS_ARBDIS;
    148	IXGBE_WRITE_REG(hw, IXGBE_RTTPCS, reg);
    149
    150	/* Map all traffic classes to their UP */
    151	reg = 0;
    152	for (i = 0; i < MAX_USER_PRIORITY; i++)
    153		reg |= (prio_tc[i] << (i * IXGBE_RTTUP2TC_UP_SHIFT));
    154	IXGBE_WRITE_REG(hw, IXGBE_RTTUP2TC, reg);
    155
    156	/* Configure traffic class credits and priority */
    157	for (i = 0; i < MAX_TRAFFIC_CLASS; i++) {
    158		reg = refill[i];
    159		reg |= (u32)(max[i]) << IXGBE_RTTPT2C_MCL_SHIFT;
    160		reg |= (u32)(bwg_id[i]) << IXGBE_RTTPT2C_BWG_SHIFT;
    161
    162		if (prio_type[i] == prio_group)
    163			reg |= IXGBE_RTTPT2C_GSP;
    164
    165		if (prio_type[i] == prio_link)
    166			reg |= IXGBE_RTTPT2C_LSP;
    167
    168		IXGBE_WRITE_REG(hw, IXGBE_RTTPT2C(i), reg);
    169	}
    170
    171	/*
    172	 * Configure Tx packet plane (recycle mode; SP; arb delay) and
    173	 * enable arbiter
    174	 */
    175	reg = IXGBE_RTTPCS_TPPAC | IXGBE_RTTPCS_TPRM |
    176	      (IXGBE_RTTPCS_ARBD_DCB << IXGBE_RTTPCS_ARBD_SHIFT);
    177	IXGBE_WRITE_REG(hw, IXGBE_RTTPCS, reg);
    178
    179	return 0;
    180}
    181
    182/**
    183 * ixgbe_dcb_config_pfc_82599 - Configure priority flow control
    184 * @hw: pointer to hardware structure
    185 * @pfc_en: enabled pfc bitmask
    186 * @prio_tc: priority to tc assignments indexed by priority
    187 *
    188 * Configure Priority Flow Control (PFC) for each traffic class.
    189 */
    190s32 ixgbe_dcb_config_pfc_82599(struct ixgbe_hw *hw, u8 pfc_en, u8 *prio_tc)
    191{
    192	u32 i, j, fcrtl, reg;
    193	u8 max_tc = 0;
    194
    195	/* Enable Transmit Priority Flow Control */
    196	IXGBE_WRITE_REG(hw, IXGBE_FCCFG, IXGBE_FCCFG_TFCE_PRIORITY);
    197
    198	/* Enable Receive Priority Flow Control */
    199	reg = IXGBE_READ_REG(hw, IXGBE_MFLCN);
    200	reg |= IXGBE_MFLCN_DPF;
    201
    202	/*
    203	 * X540 & X550 supports per TC Rx priority flow control.
    204	 * So clear all TCs and only enable those that should be
    205	 * enabled.
    206	 */
    207	reg &= ~(IXGBE_MFLCN_RPFCE_MASK | IXGBE_MFLCN_RFCE);
    208
    209	if (hw->mac.type >= ixgbe_mac_X540)
    210		reg |= pfc_en << IXGBE_MFLCN_RPFCE_SHIFT;
    211
    212	if (pfc_en)
    213		reg |= IXGBE_MFLCN_RPFCE;
    214
    215	IXGBE_WRITE_REG(hw, IXGBE_MFLCN, reg);
    216
    217	for (i = 0; i < MAX_USER_PRIORITY; i++) {
    218		if (prio_tc[i] > max_tc)
    219			max_tc = prio_tc[i];
    220	}
    221
    222
    223	/* Configure PFC Tx thresholds per TC */
    224	for (i = 0; i <= max_tc; i++) {
    225		int enabled = 0;
    226
    227		for (j = 0; j < MAX_USER_PRIORITY; j++) {
    228			if ((prio_tc[j] == i) && (pfc_en & BIT(j))) {
    229				enabled = 1;
    230				break;
    231			}
    232		}
    233
    234		if (enabled) {
    235			reg = (hw->fc.high_water[i] << 10) | IXGBE_FCRTH_FCEN;
    236			fcrtl = (hw->fc.low_water[i] << 10) | IXGBE_FCRTL_XONE;
    237			IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(i), fcrtl);
    238		} else {
    239			/* In order to prevent Tx hangs when the internal Tx
    240			 * switch is enabled we must set the high water mark
    241			 * to the Rx packet buffer size - 24KB.  This allows
    242			 * the Tx switch to function even under heavy Rx
    243			 * workloads.
    244			 */
    245			reg = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(i)) - 24576;
    246			IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(i), 0);
    247		}
    248
    249		IXGBE_WRITE_REG(hw, IXGBE_FCRTH_82599(i), reg);
    250	}
    251
    252	for (; i < MAX_TRAFFIC_CLASS; i++) {
    253		IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(i), 0);
    254		IXGBE_WRITE_REG(hw, IXGBE_FCRTH_82599(i), 0);
    255	}
    256
    257	/* Configure pause time (2 TCs per register) */
    258	reg = hw->fc.pause_time * 0x00010001;
    259	for (i = 0; i < (MAX_TRAFFIC_CLASS / 2); i++)
    260		IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
    261
    262	/* Configure flow control refresh threshold value */
    263	IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
    264
    265	return 0;
    266}
    267
    268/**
    269 * ixgbe_dcb_config_tc_stats_82599 - Config traffic class statistics
    270 * @hw: pointer to hardware structure
    271 *
    272 * Configure queue statistics registers, all queues belonging to same traffic
    273 * class uses a single set of queue statistics counters.
    274 */
    275static s32 ixgbe_dcb_config_tc_stats_82599(struct ixgbe_hw *hw)
    276{
    277	u32 reg = 0;
    278	u8  i   = 0;
    279
    280	/*
    281	 * Receive Queues stats setting
    282	 * 32 RQSMR registers, each configuring 4 queues.
    283	 * Set all 16 queues of each TC to the same stat
    284	 * with TC 'n' going to stat 'n'.
    285	 */
    286	for (i = 0; i < 32; i++) {
    287		reg = 0x01010101 * (i / 4);
    288		IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), reg);
    289	}
    290	/*
    291	 * Transmit Queues stats setting
    292	 * 32 TQSM registers, each controlling 4 queues.
    293	 * Set all queues of each TC to the same stat
    294	 * with TC 'n' going to stat 'n'.
    295	 * Tx queues are allocated non-uniformly to TCs:
    296	 * 32, 32, 16, 16, 8, 8, 8, 8.
    297	 */
    298	for (i = 0; i < 32; i++) {
    299		if (i < 8)
    300			reg = 0x00000000;
    301		else if (i < 16)
    302			reg = 0x01010101;
    303		else if (i < 20)
    304			reg = 0x02020202;
    305		else if (i < 24)
    306			reg = 0x03030303;
    307		else if (i < 26)
    308			reg = 0x04040404;
    309		else if (i < 28)
    310			reg = 0x05050505;
    311		else if (i < 30)
    312			reg = 0x06060606;
    313		else
    314			reg = 0x07070707;
    315		IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), reg);
    316	}
    317
    318	return 0;
    319}
    320
    321/**
    322 * ixgbe_dcb_hw_config_82599 - Configure and enable DCB
    323 * @hw: pointer to hardware structure
    324 * @pfc_en: enabled pfc bitmask
    325 * @refill: refill credits index by traffic class
    326 * @max: max credits index by traffic class
    327 * @bwg_id: bandwidth grouping indexed by traffic class
    328 * @prio_type: priority type indexed by traffic class
    329 * @prio_tc: priority to tc assignments indexed by priority
    330 *
    331 * Configure dcb settings and enable dcb mode.
    332 */
    333s32 ixgbe_dcb_hw_config_82599(struct ixgbe_hw *hw, u8 pfc_en, u16 *refill,
    334			      u16 *max, u8 *bwg_id, u8 *prio_type, u8 *prio_tc)
    335{
    336	ixgbe_dcb_config_rx_arbiter_82599(hw, refill, max, bwg_id,
    337					  prio_type, prio_tc);
    338	ixgbe_dcb_config_tx_desc_arbiter_82599(hw, refill, max,
    339					       bwg_id, prio_type);
    340	ixgbe_dcb_config_tx_data_arbiter_82599(hw, refill, max,
    341					       bwg_id, prio_type, prio_tc);
    342	ixgbe_dcb_config_pfc_82599(hw, pfc_en, prio_tc);
    343	ixgbe_dcb_config_tc_stats_82599(hw);
    344
    345	return 0;
    346}
    347