cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

platform.c (27091B)


      1// SPDX-License-Identifier: GPL-2.0 or BSD-3-Clause
      2/*
      3 * Copyright(c) 2015, 2016 Intel Corporation.
      4 */
      5
      6#include <linux/firmware.h>
      7
      8#include "hfi.h"
      9#include "efivar.h"
     10#include "eprom.h"
     11
     12#define DEFAULT_PLATFORM_CONFIG_NAME "hfi1_platform.dat"
     13
     14static int validate_scratch_checksum(struct hfi1_devdata *dd)
     15{
     16	u64 checksum = 0, temp_scratch = 0;
     17	int i, j, version;
     18
     19	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH);
     20	version = (temp_scratch & BITMAP_VERSION_SMASK) >> BITMAP_VERSION_SHIFT;
     21
     22	/* Prevent power on default of all zeroes from passing checksum */
     23	if (!version) {
     24		dd_dev_err(dd, "%s: Config bitmap uninitialized\n", __func__);
     25		dd_dev_err(dd,
     26			   "%s: Please update your BIOS to support active channels\n",
     27			   __func__);
     28		return 0;
     29	}
     30
     31	/*
     32	 * ASIC scratch 0 only contains the checksum and bitmap version as
     33	 * fields of interest, both of which are handled separately from the
     34	 * loop below, so skip it
     35	 */
     36	checksum += version;
     37	for (i = 1; i < ASIC_NUM_SCRATCH; i++) {
     38		temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH + (8 * i));
     39		for (j = sizeof(u64); j != 0; j -= 2) {
     40			checksum += (temp_scratch & 0xFFFF);
     41			temp_scratch >>= 16;
     42		}
     43	}
     44
     45	while (checksum >> 16)
     46		checksum = (checksum & CHECKSUM_MASK) + (checksum >> 16);
     47
     48	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH);
     49	temp_scratch &= CHECKSUM_SMASK;
     50	temp_scratch >>= CHECKSUM_SHIFT;
     51
     52	if (checksum + temp_scratch == 0xFFFF)
     53		return 1;
     54
     55	dd_dev_err(dd, "%s: Configuration bitmap corrupted\n", __func__);
     56	return 0;
     57}
     58
     59static void save_platform_config_fields(struct hfi1_devdata *dd)
     60{
     61	struct hfi1_pportdata *ppd = dd->pport;
     62	u64 temp_scratch = 0, temp_dest = 0;
     63
     64	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH_1);
     65
     66	temp_dest = temp_scratch &
     67		    (dd->hfi1_id ? PORT1_PORT_TYPE_SMASK :
     68		     PORT0_PORT_TYPE_SMASK);
     69	ppd->port_type = temp_dest >>
     70			 (dd->hfi1_id ? PORT1_PORT_TYPE_SHIFT :
     71			  PORT0_PORT_TYPE_SHIFT);
     72
     73	temp_dest = temp_scratch &
     74		    (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SMASK :
     75		     PORT0_LOCAL_ATTEN_SMASK);
     76	ppd->local_atten = temp_dest >>
     77			   (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SHIFT :
     78			    PORT0_LOCAL_ATTEN_SHIFT);
     79
     80	temp_dest = temp_scratch &
     81		    (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SMASK :
     82		     PORT0_REMOTE_ATTEN_SMASK);
     83	ppd->remote_atten = temp_dest >>
     84			    (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SHIFT :
     85			     PORT0_REMOTE_ATTEN_SHIFT);
     86
     87	temp_dest = temp_scratch &
     88		    (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SMASK :
     89		     PORT0_DEFAULT_ATTEN_SMASK);
     90	ppd->default_atten = temp_dest >>
     91			     (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SHIFT :
     92			      PORT0_DEFAULT_ATTEN_SHIFT);
     93
     94	temp_scratch = read_csr(dd, dd->hfi1_id ? ASIC_CFG_SCRATCH_3 :
     95				ASIC_CFG_SCRATCH_2);
     96
     97	ppd->tx_preset_eq = (temp_scratch & TX_EQ_SMASK) >> TX_EQ_SHIFT;
     98	ppd->tx_preset_noeq = (temp_scratch & TX_NO_EQ_SMASK) >> TX_NO_EQ_SHIFT;
     99	ppd->rx_preset = (temp_scratch & RX_SMASK) >> RX_SHIFT;
    100
    101	ppd->max_power_class = (temp_scratch & QSFP_MAX_POWER_SMASK) >>
    102				QSFP_MAX_POWER_SHIFT;
    103
    104	ppd->config_from_scratch = true;
    105}
    106
    107void get_platform_config(struct hfi1_devdata *dd)
    108{
    109	int ret = 0;
    110	u8 *temp_platform_config = NULL;
    111	u32 esize;
    112	const struct firmware *platform_config_file = NULL;
    113
    114	if (is_integrated(dd)) {
    115		if (validate_scratch_checksum(dd)) {
    116			save_platform_config_fields(dd);
    117			return;
    118		}
    119	} else {
    120		ret = eprom_read_platform_config(dd,
    121						 (void **)&temp_platform_config,
    122						 &esize);
    123		if (!ret) {
    124			/* success */
    125			dd->platform_config.data = temp_platform_config;
    126			dd->platform_config.size = esize;
    127			return;
    128		}
    129	}
    130	dd_dev_err(dd,
    131		   "%s: Failed to get platform config, falling back to sub-optimal default file\n",
    132		   __func__);
    133
    134	ret = request_firmware(&platform_config_file,
    135			       DEFAULT_PLATFORM_CONFIG_NAME,
    136			       &dd->pcidev->dev);
    137	if (ret) {
    138		dd_dev_err(dd,
    139			   "%s: No default platform config file found\n",
    140			   __func__);
    141		return;
    142	}
    143
    144	/*
    145	 * Allocate separate memory block to store data and free firmware
    146	 * structure. This allows free_platform_config to treat EPROM and
    147	 * fallback configs in the same manner.
    148	 */
    149	dd->platform_config.data = kmemdup(platform_config_file->data,
    150					   platform_config_file->size,
    151					   GFP_KERNEL);
    152	dd->platform_config.size = platform_config_file->size;
    153	release_firmware(platform_config_file);
    154}
    155
    156void free_platform_config(struct hfi1_devdata *dd)
    157{
    158	/* Release memory allocated for eprom or fallback file read. */
    159	kfree(dd->platform_config.data);
    160	dd->platform_config.data = NULL;
    161}
    162
    163void get_port_type(struct hfi1_pportdata *ppd)
    164{
    165	int ret;
    166	u32 temp;
    167
    168	ret = get_platform_config_field(ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    169					PORT_TABLE_PORT_TYPE, &temp,
    170					4);
    171	if (ret) {
    172		ppd->port_type = PORT_TYPE_UNKNOWN;
    173		return;
    174	}
    175	ppd->port_type = temp;
    176}
    177
    178int set_qsfp_tx(struct hfi1_pportdata *ppd, int on)
    179{
    180	u8 tx_ctrl_byte = on ? 0x0 : 0xF;
    181	int ret = 0;
    182
    183	ret = qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_TX_CTRL_BYTE_OFFS,
    184			 &tx_ctrl_byte, 1);
    185	/* we expected 1, so consider 0 an error */
    186	if (ret == 0)
    187		ret = -EIO;
    188	else if (ret == 1)
    189		ret = 0;
    190	return ret;
    191}
    192
    193static int qual_power(struct hfi1_pportdata *ppd)
    194{
    195	u32 cable_power_class = 0, power_class_max = 0;
    196	u8 *cache = ppd->qsfp_info.cache;
    197	int ret = 0;
    198
    199	ret = get_platform_config_field(
    200		ppd->dd, PLATFORM_CONFIG_SYSTEM_TABLE, 0,
    201		SYSTEM_TABLE_QSFP_POWER_CLASS_MAX, &power_class_max, 4);
    202	if (ret)
    203		return ret;
    204
    205	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
    206
    207	if (cable_power_class > power_class_max)
    208		ppd->offline_disabled_reason =
    209			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY);
    210
    211	if (ppd->offline_disabled_reason ==
    212			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY)) {
    213		dd_dev_err(
    214			ppd->dd,
    215			"%s: Port disabled due to system power restrictions\n",
    216			__func__);
    217		ret = -EPERM;
    218	}
    219	return ret;
    220}
    221
    222static int qual_bitrate(struct hfi1_pportdata *ppd)
    223{
    224	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
    225	u8 *cache = ppd->qsfp_info.cache;
    226
    227	if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G) &&
    228	    cache[QSFP_NOM_BIT_RATE_250_OFFS] < 0x64)
    229		ppd->offline_disabled_reason =
    230			   HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY);
    231
    232	if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G) &&
    233	    cache[QSFP_NOM_BIT_RATE_100_OFFS] < 0x7D)
    234		ppd->offline_disabled_reason =
    235			   HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY);
    236
    237	if (ppd->offline_disabled_reason ==
    238			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY)) {
    239		dd_dev_err(
    240			ppd->dd,
    241			"%s: Cable failed bitrate check, disabling port\n",
    242			__func__);
    243		return -EPERM;
    244	}
    245	return 0;
    246}
    247
    248static int set_qsfp_high_power(struct hfi1_pportdata *ppd)
    249{
    250	u8 cable_power_class = 0, power_ctrl_byte = 0;
    251	u8 *cache = ppd->qsfp_info.cache;
    252	int ret;
    253
    254	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
    255
    256	if (cable_power_class > QSFP_POWER_CLASS_1) {
    257		power_ctrl_byte = cache[QSFP_PWR_CTRL_BYTE_OFFS];
    258
    259		power_ctrl_byte |= 1;
    260		power_ctrl_byte &= ~(0x2);
    261
    262		ret = qsfp_write(ppd, ppd->dd->hfi1_id,
    263				 QSFP_PWR_CTRL_BYTE_OFFS,
    264				 &power_ctrl_byte, 1);
    265		if (ret != 1)
    266			return -EIO;
    267
    268		if (cable_power_class > QSFP_POWER_CLASS_4) {
    269			power_ctrl_byte |= (1 << 2);
    270			ret = qsfp_write(ppd, ppd->dd->hfi1_id,
    271					 QSFP_PWR_CTRL_BYTE_OFFS,
    272					 &power_ctrl_byte, 1);
    273			if (ret != 1)
    274				return -EIO;
    275		}
    276
    277		/* SFF 8679 rev 1.7 LPMode Deassert time */
    278		msleep(300);
    279	}
    280	return 0;
    281}
    282
    283static void apply_rx_cdr(struct hfi1_pportdata *ppd,
    284			 u32 rx_preset_index,
    285			 u8 *cdr_ctrl_byte)
    286{
    287	u32 rx_preset;
    288	u8 *cache = ppd->qsfp_info.cache;
    289	int cable_power_class;
    290
    291	if (!((cache[QSFP_MOD_PWR_OFFS] & 0x4) &&
    292	      (cache[QSFP_CDR_INFO_OFFS] & 0x40)))
    293		return;
    294
    295	/* RX CDR present, bypass supported */
    296	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
    297
    298	if (cable_power_class <= QSFP_POWER_CLASS_3) {
    299		/* Power class <= 3, ignore config & turn RX CDR on */
    300		*cdr_ctrl_byte |= 0xF;
    301		return;
    302	}
    303
    304	get_platform_config_field(
    305		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
    306		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR_APPLY,
    307		&rx_preset, 4);
    308
    309	if (!rx_preset) {
    310		dd_dev_info(
    311			ppd->dd,
    312			"%s: RX_CDR_APPLY is set to disabled\n",
    313			__func__);
    314		return;
    315	}
    316	get_platform_config_field(
    317		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
    318		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR,
    319		&rx_preset, 4);
    320
    321	/* Expand cdr setting to all 4 lanes */
    322	rx_preset = (rx_preset | (rx_preset << 1) |
    323			(rx_preset << 2) | (rx_preset << 3));
    324
    325	if (rx_preset) {
    326		*cdr_ctrl_byte |= rx_preset;
    327	} else {
    328		*cdr_ctrl_byte &= rx_preset;
    329		/* Preserve current TX CDR status */
    330		*cdr_ctrl_byte |= (cache[QSFP_CDR_CTRL_BYTE_OFFS] & 0xF0);
    331	}
    332}
    333
    334static void apply_tx_cdr(struct hfi1_pportdata *ppd,
    335			 u32 tx_preset_index,
    336			 u8 *cdr_ctrl_byte)
    337{
    338	u32 tx_preset;
    339	u8 *cache = ppd->qsfp_info.cache;
    340	int cable_power_class;
    341
    342	if (!((cache[QSFP_MOD_PWR_OFFS] & 0x8) &&
    343	      (cache[QSFP_CDR_INFO_OFFS] & 0x80)))
    344		return;
    345
    346	/* TX CDR present, bypass supported */
    347	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
    348
    349	if (cable_power_class <= QSFP_POWER_CLASS_3) {
    350		/* Power class <= 3, ignore config & turn TX CDR on */
    351		*cdr_ctrl_byte |= 0xF0;
    352		return;
    353	}
    354
    355	get_platform_config_field(
    356		ppd->dd,
    357		PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index,
    358		TX_PRESET_TABLE_QSFP_TX_CDR_APPLY, &tx_preset, 4);
    359
    360	if (!tx_preset) {
    361		dd_dev_info(
    362			ppd->dd,
    363			"%s: TX_CDR_APPLY is set to disabled\n",
    364			__func__);
    365		return;
    366	}
    367	get_platform_config_field(
    368		ppd->dd,
    369		PLATFORM_CONFIG_TX_PRESET_TABLE,
    370		tx_preset_index,
    371		TX_PRESET_TABLE_QSFP_TX_CDR, &tx_preset, 4);
    372
    373	/* Expand cdr setting to all 4 lanes */
    374	tx_preset = (tx_preset | (tx_preset << 1) |
    375			(tx_preset << 2) | (tx_preset << 3));
    376
    377	if (tx_preset)
    378		*cdr_ctrl_byte |= (tx_preset << 4);
    379	else
    380		/* Preserve current/determined RX CDR status */
    381		*cdr_ctrl_byte &= ((tx_preset << 4) | 0xF);
    382}
    383
    384static void apply_cdr_settings(
    385		struct hfi1_pportdata *ppd, u32 rx_preset_index,
    386		u32 tx_preset_index)
    387{
    388	u8 *cache = ppd->qsfp_info.cache;
    389	u8 cdr_ctrl_byte = cache[QSFP_CDR_CTRL_BYTE_OFFS];
    390
    391	apply_rx_cdr(ppd, rx_preset_index, &cdr_ctrl_byte);
    392
    393	apply_tx_cdr(ppd, tx_preset_index, &cdr_ctrl_byte);
    394
    395	qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_CDR_CTRL_BYTE_OFFS,
    396		   &cdr_ctrl_byte, 1);
    397}
    398
    399static void apply_tx_eq_auto(struct hfi1_pportdata *ppd)
    400{
    401	u8 *cache = ppd->qsfp_info.cache;
    402	u8 tx_eq;
    403
    404	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x8))
    405		return;
    406	/* Disable adaptive TX EQ if present */
    407	tx_eq = cache[(128 * 3) + 241];
    408	tx_eq &= 0xF0;
    409	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 241, &tx_eq, 1);
    410}
    411
    412static void apply_tx_eq_prog(struct hfi1_pportdata *ppd, u32 tx_preset_index)
    413{
    414	u8 *cache = ppd->qsfp_info.cache;
    415	u32 tx_preset;
    416	u8 tx_eq;
    417
    418	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x4))
    419		return;
    420
    421	get_platform_config_field(
    422		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
    423		tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ_APPLY,
    424		&tx_preset, 4);
    425	if (!tx_preset) {
    426		dd_dev_info(
    427			ppd->dd,
    428			"%s: TX_EQ_APPLY is set to disabled\n",
    429			__func__);
    430		return;
    431	}
    432	get_platform_config_field(
    433			ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
    434			tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ,
    435			&tx_preset, 4);
    436
    437	if (((cache[(128 * 3) + 224] & 0xF0) >> 4) < tx_preset) {
    438		dd_dev_info(
    439			ppd->dd,
    440			"%s: TX EQ %x unsupported\n",
    441			__func__, tx_preset);
    442
    443		dd_dev_info(
    444			ppd->dd,
    445			"%s: Applying EQ %x\n",
    446			__func__, cache[608] & 0xF0);
    447
    448		tx_preset = (cache[608] & 0xF0) >> 4;
    449	}
    450
    451	tx_eq = tx_preset | (tx_preset << 4);
    452	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 234, &tx_eq, 1);
    453	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 235, &tx_eq, 1);
    454}
    455
    456static void apply_rx_eq_emp(struct hfi1_pportdata *ppd, u32 rx_preset_index)
    457{
    458	u32 rx_preset;
    459	u8 rx_eq, *cache = ppd->qsfp_info.cache;
    460
    461	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x2))
    462		return;
    463	get_platform_config_field(
    464			ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
    465			rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP_APPLY,
    466			&rx_preset, 4);
    467
    468	if (!rx_preset) {
    469		dd_dev_info(
    470			ppd->dd,
    471			"%s: RX_EMP_APPLY is set to disabled\n",
    472			__func__);
    473		return;
    474	}
    475	get_platform_config_field(
    476		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
    477		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP,
    478		&rx_preset, 4);
    479
    480	if ((cache[(128 * 3) + 224] & 0xF) < rx_preset) {
    481		dd_dev_info(
    482			ppd->dd,
    483			"%s: Requested RX EMP %x\n",
    484			__func__, rx_preset);
    485
    486		dd_dev_info(
    487			ppd->dd,
    488			"%s: Applying supported EMP %x\n",
    489			__func__, cache[608] & 0xF);
    490
    491		rx_preset = cache[608] & 0xF;
    492	}
    493
    494	rx_eq = rx_preset | (rx_preset << 4);
    495
    496	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 236, &rx_eq, 1);
    497	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 237, &rx_eq, 1);
    498}
    499
    500static void apply_eq_settings(struct hfi1_pportdata *ppd,
    501			      u32 rx_preset_index, u32 tx_preset_index)
    502{
    503	u8 *cache = ppd->qsfp_info.cache;
    504
    505	/* no point going on w/o a page 3 */
    506	if (cache[2] & 4) {
    507		dd_dev_info(ppd->dd,
    508			    "%s: Upper page 03 not present\n",
    509			    __func__);
    510		return;
    511	}
    512
    513	apply_tx_eq_auto(ppd);
    514
    515	apply_tx_eq_prog(ppd, tx_preset_index);
    516
    517	apply_rx_eq_emp(ppd, rx_preset_index);
    518}
    519
    520static void apply_rx_amplitude_settings(
    521		struct hfi1_pportdata *ppd, u32 rx_preset_index,
    522		u32 tx_preset_index)
    523{
    524	u32 rx_preset;
    525	u8 rx_amp = 0, i = 0, preferred = 0, *cache = ppd->qsfp_info.cache;
    526
    527	/* no point going on w/o a page 3 */
    528	if (cache[2] & 4) {
    529		dd_dev_info(ppd->dd,
    530			    "%s: Upper page 03 not present\n",
    531			    __func__);
    532		return;
    533	}
    534	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x1)) {
    535		dd_dev_info(ppd->dd,
    536			    "%s: RX_AMP_APPLY is set to disabled\n",
    537			    __func__);
    538		return;
    539	}
    540
    541	get_platform_config_field(ppd->dd,
    542				  PLATFORM_CONFIG_RX_PRESET_TABLE,
    543				  rx_preset_index,
    544				  RX_PRESET_TABLE_QSFP_RX_AMP_APPLY,
    545				  &rx_preset, 4);
    546
    547	if (!rx_preset) {
    548		dd_dev_info(ppd->dd,
    549			    "%s: RX_AMP_APPLY is set to disabled\n",
    550			    __func__);
    551		return;
    552	}
    553	get_platform_config_field(ppd->dd,
    554				  PLATFORM_CONFIG_RX_PRESET_TABLE,
    555				  rx_preset_index,
    556				  RX_PRESET_TABLE_QSFP_RX_AMP,
    557				  &rx_preset, 4);
    558
    559	dd_dev_info(ppd->dd,
    560		    "%s: Requested RX AMP %x\n",
    561		    __func__,
    562		    rx_preset);
    563
    564	for (i = 0; i < 4; i++) {
    565		if (cache[(128 * 3) + 225] & (1 << i)) {
    566			preferred = i;
    567			if (preferred == rx_preset)
    568				break;
    569		}
    570	}
    571
    572	/*
    573	 * Verify that preferred RX amplitude is not just a
    574	 * fall through of the default
    575	 */
    576	if (!preferred && !(cache[(128 * 3) + 225] & 0x1)) {
    577		dd_dev_info(ppd->dd, "No supported RX AMP, not applying\n");
    578		return;
    579	}
    580
    581	dd_dev_info(ppd->dd,
    582		    "%s: Applying RX AMP %x\n", __func__, preferred);
    583
    584	rx_amp = preferred | (preferred << 4);
    585	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 238, &rx_amp, 1);
    586	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 239, &rx_amp, 1);
    587}
    588
    589#define OPA_INVALID_INDEX 0xFFF
    590
    591static void apply_tx_lanes(struct hfi1_pportdata *ppd, u8 field_id,
    592			   u32 config_data, const char *message)
    593{
    594	u8 i;
    595	int ret;
    596
    597	for (i = 0; i < 4; i++) {
    598		ret = load_8051_config(ppd->dd, field_id, i, config_data);
    599		if (ret != HCMD_SUCCESS) {
    600			dd_dev_err(
    601				ppd->dd,
    602				"%s: %s for lane %u failed\n",
    603				message, __func__, i);
    604		}
    605	}
    606}
    607
    608/*
    609 * Return a special SerDes setting for low power AOC cables.  The power class
    610 * threshold and setting being used were all found by empirical testing.
    611 *
    612 * Summary of the logic:
    613 *
    614 * if (QSFP and QSFP_TYPE == AOC and QSFP_POWER_CLASS < 4)
    615 *     return 0xe
    616 * return 0; // leave at default
    617 */
    618static u8 aoc_low_power_setting(struct hfi1_pportdata *ppd)
    619{
    620	u8 *cache = ppd->qsfp_info.cache;
    621	int power_class;
    622
    623	/* QSFP only */
    624	if (ppd->port_type != PORT_TYPE_QSFP)
    625		return 0; /* leave at default */
    626
    627	/* active optical cables only */
    628	switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) {
    629	case 0x0 ... 0x9: fallthrough;
    630	case 0xC: fallthrough;
    631	case 0xE:
    632		/* active AOC */
    633		power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
    634		if (power_class < QSFP_POWER_CLASS_4)
    635			return 0xe;
    636	}
    637	return 0; /* leave at default */
    638}
    639
    640static void apply_tunings(
    641		struct hfi1_pportdata *ppd, u32 tx_preset_index,
    642		u8 tuning_method, u32 total_atten, u8 limiting_active)
    643{
    644	int ret = 0;
    645	u32 config_data = 0, tx_preset = 0;
    646	u8 precur = 0, attn = 0, postcur = 0, external_device_config = 0;
    647	u8 *cache = ppd->qsfp_info.cache;
    648
    649	/* Pass tuning method to 8051 */
    650	read_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG,
    651			 &config_data);
    652	config_data &= ~(0xff << TUNING_METHOD_SHIFT);
    653	config_data |= ((u32)tuning_method << TUNING_METHOD_SHIFT);
    654	ret = load_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG,
    655			       config_data);
    656	if (ret != HCMD_SUCCESS)
    657		dd_dev_err(ppd->dd, "%s: Failed to set tuning method\n",
    658			   __func__);
    659
    660	/* Set same channel loss for both TX and RX */
    661	config_data = 0 | (total_atten << 16) | (total_atten << 24);
    662	apply_tx_lanes(ppd, CHANNEL_LOSS_SETTINGS, config_data,
    663		       "Setting channel loss");
    664
    665	/* Inform 8051 of cable capabilities */
    666	if (ppd->qsfp_info.cache_valid) {
    667		external_device_config =
    668			((cache[QSFP_MOD_PWR_OFFS] & 0x4) << 3) |
    669			((cache[QSFP_MOD_PWR_OFFS] & 0x8) << 2) |
    670			((cache[QSFP_EQ_INFO_OFFS] & 0x2) << 1) |
    671			(cache[QSFP_EQ_INFO_OFFS] & 0x4);
    672		ret = read_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS,
    673				       GENERAL_CONFIG, &config_data);
    674		/* Clear, then set the external device config field */
    675		config_data &= ~(u32)0xFF;
    676		config_data |= external_device_config;
    677		ret = load_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS,
    678				       GENERAL_CONFIG, config_data);
    679		if (ret != HCMD_SUCCESS)
    680			dd_dev_err(ppd->dd,
    681				   "%s: Failed set ext device config params\n",
    682				   __func__);
    683	}
    684
    685	if (tx_preset_index == OPA_INVALID_INDEX) {
    686		if (ppd->port_type == PORT_TYPE_QSFP && limiting_active)
    687			dd_dev_err(ppd->dd, "%s: Invalid Tx preset index\n",
    688				   __func__);
    689		return;
    690	}
    691
    692	/* Following for limiting active channels only */
    693	get_platform_config_field(
    694		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index,
    695		TX_PRESET_TABLE_PRECUR, &tx_preset, 4);
    696	precur = tx_preset;
    697
    698	get_platform_config_field(
    699		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
    700		tx_preset_index, TX_PRESET_TABLE_ATTN, &tx_preset, 4);
    701	attn = tx_preset;
    702
    703	get_platform_config_field(
    704		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
    705		tx_preset_index, TX_PRESET_TABLE_POSTCUR, &tx_preset, 4);
    706	postcur = tx_preset;
    707
    708	/*
    709	 * NOTES:
    710	 * o The aoc_low_power_setting is applied to all lanes even
    711	 *   though only lane 0's value is examined by the firmware.
    712	 * o A lingering low power setting after a cable swap does
    713	 *   not occur.  On cable unplug the 8051 is reset and
    714	 *   restarted on cable insert.  This resets all settings to
    715	 *   their default, erasing any previous low power setting.
    716	 */
    717	config_data = precur | (attn << 8) | (postcur << 16) |
    718			(aoc_low_power_setting(ppd) << 24);
    719
    720	apply_tx_lanes(ppd, TX_EQ_SETTINGS, config_data,
    721		       "Applying TX settings");
    722}
    723
    724/* Must be holding the QSFP i2c resource */
    725static int tune_active_qsfp(struct hfi1_pportdata *ppd, u32 *ptr_tx_preset,
    726			    u32 *ptr_rx_preset, u32 *ptr_total_atten)
    727{
    728	int ret;
    729	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
    730	u8 *cache = ppd->qsfp_info.cache;
    731
    732	ppd->qsfp_info.limiting_active = 1;
    733
    734	ret = set_qsfp_tx(ppd, 0);
    735	if (ret)
    736		return ret;
    737
    738	ret = qual_power(ppd);
    739	if (ret)
    740		return ret;
    741
    742	ret = qual_bitrate(ppd);
    743	if (ret)
    744		return ret;
    745
    746	/*
    747	 * We'll change the QSFP memory contents from here on out, thus we set a
    748	 * flag here to remind ourselves to reset the QSFP module. This prevents
    749	 * reuse of stale settings established in our previous pass through.
    750	 */
    751	if (ppd->qsfp_info.reset_needed) {
    752		ret = reset_qsfp(ppd);
    753		if (ret)
    754			return ret;
    755		refresh_qsfp_cache(ppd, &ppd->qsfp_info);
    756	} else {
    757		ppd->qsfp_info.reset_needed = 1;
    758	}
    759
    760	ret = set_qsfp_high_power(ppd);
    761	if (ret)
    762		return ret;
    763
    764	if (cache[QSFP_EQ_INFO_OFFS] & 0x4) {
    765		ret = get_platform_config_field(
    766			ppd->dd,
    767			PLATFORM_CONFIG_PORT_TABLE, 0,
    768			PORT_TABLE_TX_PRESET_IDX_ACTIVE_EQ,
    769			ptr_tx_preset, 4);
    770		if (ret) {
    771			*ptr_tx_preset = OPA_INVALID_INDEX;
    772			return ret;
    773		}
    774	} else {
    775		ret = get_platform_config_field(
    776			ppd->dd,
    777			PLATFORM_CONFIG_PORT_TABLE, 0,
    778			PORT_TABLE_TX_PRESET_IDX_ACTIVE_NO_EQ,
    779			ptr_tx_preset, 4);
    780		if (ret) {
    781			*ptr_tx_preset = OPA_INVALID_INDEX;
    782			return ret;
    783		}
    784	}
    785
    786	ret = get_platform_config_field(
    787		ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    788		PORT_TABLE_RX_PRESET_IDX, ptr_rx_preset, 4);
    789	if (ret) {
    790		*ptr_rx_preset = OPA_INVALID_INDEX;
    791		return ret;
    792	}
    793
    794	if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G))
    795		get_platform_config_field(
    796			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    797			PORT_TABLE_LOCAL_ATTEN_25G, ptr_total_atten, 4);
    798	else if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G))
    799		get_platform_config_field(
    800			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    801			PORT_TABLE_LOCAL_ATTEN_12G, ptr_total_atten, 4);
    802
    803	apply_cdr_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
    804
    805	apply_eq_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
    806
    807	apply_rx_amplitude_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
    808
    809	ret = set_qsfp_tx(ppd, 1);
    810
    811	return ret;
    812}
    813
    814static int tune_qsfp(struct hfi1_pportdata *ppd,
    815		     u32 *ptr_tx_preset, u32 *ptr_rx_preset,
    816		     u8 *ptr_tuning_method, u32 *ptr_total_atten)
    817{
    818	u32 cable_atten = 0, remote_atten = 0, platform_atten = 0;
    819	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
    820	int ret = 0;
    821	u8 *cache = ppd->qsfp_info.cache;
    822
    823	switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) {
    824	case 0xA ... 0xB:
    825		ret = get_platform_config_field(
    826			ppd->dd,
    827			PLATFORM_CONFIG_PORT_TABLE, 0,
    828			PORT_TABLE_LOCAL_ATTEN_25G,
    829			&platform_atten, 4);
    830		if (ret)
    831			return ret;
    832
    833		if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G))
    834			cable_atten = cache[QSFP_CU_ATTEN_12G_OFFS];
    835		else if ((lss & OPA_LINK_SPEED_12_5G) &&
    836			 (lse & OPA_LINK_SPEED_12_5G))
    837			cable_atten = cache[QSFP_CU_ATTEN_7G_OFFS];
    838
    839		/* Fallback to configured attenuation if cable memory is bad */
    840		if (cable_atten == 0 || cable_atten > 36) {
    841			ret = get_platform_config_field(
    842				ppd->dd,
    843				PLATFORM_CONFIG_SYSTEM_TABLE, 0,
    844				SYSTEM_TABLE_QSFP_ATTENUATION_DEFAULT_25G,
    845				&cable_atten, 4);
    846			if (ret)
    847				return ret;
    848		}
    849
    850		ret = get_platform_config_field(
    851			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    852			PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4);
    853		if (ret)
    854			return ret;
    855
    856		*ptr_total_atten = platform_atten + cable_atten + remote_atten;
    857
    858		*ptr_tuning_method = OPA_PASSIVE_TUNING;
    859		break;
    860	case 0x0 ... 0x9: fallthrough;
    861	case 0xC: fallthrough;
    862	case 0xE:
    863		ret = tune_active_qsfp(ppd, ptr_tx_preset, ptr_rx_preset,
    864				       ptr_total_atten);
    865		if (ret)
    866			return ret;
    867
    868		*ptr_tuning_method = OPA_ACTIVE_TUNING;
    869		break;
    870	case 0xD: fallthrough;
    871	case 0xF:
    872	default:
    873		dd_dev_warn(ppd->dd, "%s: Unknown/unsupported cable\n",
    874			    __func__);
    875		break;
    876	}
    877	return ret;
    878}
    879
    880/*
    881 * This function communicates its success or failure via ppd->driver_link_ready
    882 * Thus, it depends on its association with start_link(...) which checks
    883 * driver_link_ready before proceeding with the link negotiation and
    884 * initialization process.
    885 */
    886void tune_serdes(struct hfi1_pportdata *ppd)
    887{
    888	int ret = 0;
    889	u32 total_atten = 0;
    890	u32 remote_atten = 0, platform_atten = 0;
    891	u32 rx_preset_index, tx_preset_index;
    892	u8 tuning_method = 0, limiting_active = 0;
    893	struct hfi1_devdata *dd = ppd->dd;
    894
    895	rx_preset_index = OPA_INVALID_INDEX;
    896	tx_preset_index = OPA_INVALID_INDEX;
    897
    898	/* the link defaults to enabled */
    899	ppd->link_enabled = 1;
    900	/* the driver link ready state defaults to not ready */
    901	ppd->driver_link_ready = 0;
    902	ppd->offline_disabled_reason = HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE);
    903
    904	/* Skip the tuning for testing (loopback != none) and simulations */
    905	if (loopback != LOOPBACK_NONE ||
    906	    ppd->dd->icode == ICODE_FUNCTIONAL_SIMULATOR) {
    907		ppd->driver_link_ready = 1;
    908
    909		if (qsfp_mod_present(ppd)) {
    910			ret = acquire_chip_resource(ppd->dd,
    911						    qsfp_resource(ppd->dd),
    912						    QSFP_WAIT);
    913			if (ret) {
    914				dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n",
    915					   __func__, (int)ppd->dd->hfi1_id);
    916				goto bail;
    917			}
    918
    919			refresh_qsfp_cache(ppd, &ppd->qsfp_info);
    920			release_chip_resource(ppd->dd, qsfp_resource(ppd->dd));
    921		}
    922
    923		return;
    924	}
    925
    926	switch (ppd->port_type) {
    927	case PORT_TYPE_DISCONNECTED:
    928		ppd->offline_disabled_reason =
    929			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_DISCONNECTED);
    930		dd_dev_warn(dd, "%s: Port disconnected, disabling port\n",
    931			    __func__);
    932		goto bail;
    933	case PORT_TYPE_FIXED:
    934		/* platform_atten, remote_atten pre-zeroed to catch error */
    935		get_platform_config_field(
    936			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    937			PORT_TABLE_LOCAL_ATTEN_25G, &platform_atten, 4);
    938
    939		get_platform_config_field(
    940			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    941			PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4);
    942
    943		total_atten = platform_atten + remote_atten;
    944
    945		tuning_method = OPA_PASSIVE_TUNING;
    946		break;
    947	case PORT_TYPE_VARIABLE:
    948		if (qsfp_mod_present(ppd)) {
    949			/*
    950			 * platform_atten, remote_atten pre-zeroed to
    951			 * catch error
    952			 */
    953			get_platform_config_field(
    954				ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    955				PORT_TABLE_LOCAL_ATTEN_25G,
    956				&platform_atten, 4);
    957
    958			get_platform_config_field(
    959				ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
    960				PORT_TABLE_REMOTE_ATTEN_25G,
    961				&remote_atten, 4);
    962
    963			total_atten = platform_atten + remote_atten;
    964
    965			tuning_method = OPA_PASSIVE_TUNING;
    966		} else {
    967			ppd->offline_disabled_reason =
    968			     HFI1_ODR_MASK(OPA_LINKDOWN_REASON_CHASSIS_CONFIG);
    969			goto bail;
    970		}
    971		break;
    972	case PORT_TYPE_QSFP:
    973		if (qsfp_mod_present(ppd)) {
    974			ret = acquire_chip_resource(ppd->dd,
    975						    qsfp_resource(ppd->dd),
    976						    QSFP_WAIT);
    977			if (ret) {
    978				dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n",
    979					   __func__, (int)ppd->dd->hfi1_id);
    980				goto bail;
    981			}
    982			refresh_qsfp_cache(ppd, &ppd->qsfp_info);
    983
    984			if (ppd->qsfp_info.cache_valid) {
    985				ret = tune_qsfp(ppd,
    986						&tx_preset_index,
    987						&rx_preset_index,
    988						&tuning_method,
    989						&total_atten);
    990
    991				/*
    992				 * We may have modified the QSFP memory, so
    993				 * update the cache to reflect the changes
    994				 */
    995				refresh_qsfp_cache(ppd, &ppd->qsfp_info);
    996				limiting_active =
    997						ppd->qsfp_info.limiting_active;
    998			} else {
    999				dd_dev_err(dd,
   1000					   "%s: Reading QSFP memory failed\n",
   1001					   __func__);
   1002				ret = -EINVAL; /* a fail indication */
   1003			}
   1004			release_chip_resource(ppd->dd, qsfp_resource(ppd->dd));
   1005			if (ret)
   1006				goto bail;
   1007		} else {
   1008			ppd->offline_disabled_reason =
   1009			   HFI1_ODR_MASK(
   1010				OPA_LINKDOWN_REASON_LOCAL_MEDIA_NOT_INSTALLED);
   1011			goto bail;
   1012		}
   1013		break;
   1014	default:
   1015		dd_dev_warn(ppd->dd, "%s: Unknown port type\n", __func__);
   1016		ppd->port_type = PORT_TYPE_UNKNOWN;
   1017		tuning_method = OPA_UNKNOWN_TUNING;
   1018		total_atten = 0;
   1019		limiting_active = 0;
   1020		tx_preset_index = OPA_INVALID_INDEX;
   1021		break;
   1022	}
   1023
   1024	if (ppd->offline_disabled_reason ==
   1025			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE))
   1026		apply_tunings(ppd, tx_preset_index, tuning_method,
   1027			      total_atten, limiting_active);
   1028
   1029	if (!ret)
   1030		ppd->driver_link_ready = 1;
   1031
   1032	return;
   1033bail:
   1034	ppd->driver_link_ready = 0;
   1035}