cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

qed_ptp.c (12199B)


      1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
      2/* QLogic qed NIC Driver
      3 * Copyright (c) 2015-2017  QLogic Corporation
      4 * Copyright (c) 2019-2020 Marvell International Ltd.
      5 */
      6
      7#include <linux/types.h>
      8#include "qed.h"
      9#include "qed_dev_api.h"
     10#include "qed_hw.h"
     11#include "qed_l2.h"
     12#include "qed_mcp.h"
     13#include "qed_ptp.h"
     14#include "qed_reg_addr.h"
     15
     16/* 16 nano second time quantas to wait before making a Drift adjustment */
     17#define QED_DRIFT_CNTR_TIME_QUANTA_SHIFT	0
     18/* Nano seconds to add/subtract when making a Drift adjustment */
     19#define QED_DRIFT_CNTR_ADJUSTMENT_SHIFT		28
     20/* Add/subtract the Adjustment_Value when making a Drift adjustment */
     21#define QED_DRIFT_CNTR_DIRECTION_SHIFT		31
     22#define QED_TIMESTAMP_MASK			BIT(16)
     23/* Param mask for Hardware to detect/timestamp the L2/L4 unicast PTP packets */
     24#define QED_PTP_UCAST_PARAM_MASK              0x70F
     25
     26static enum qed_resc_lock qed_ptcdev_to_resc(struct qed_hwfn *p_hwfn)
     27{
     28	switch (MFW_PORT(p_hwfn)) {
     29	case 0:
     30		return QED_RESC_LOCK_PTP_PORT0;
     31	case 1:
     32		return QED_RESC_LOCK_PTP_PORT1;
     33	case 2:
     34		return QED_RESC_LOCK_PTP_PORT2;
     35	case 3:
     36		return QED_RESC_LOCK_PTP_PORT3;
     37	default:
     38		return QED_RESC_LOCK_RESC_INVALID;
     39	}
     40}
     41
     42static int qed_ptp_res_lock(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt)
     43{
     44	struct qed_resc_lock_params params;
     45	enum qed_resc_lock resource;
     46	int rc;
     47
     48	resource = qed_ptcdev_to_resc(p_hwfn);
     49	if (resource == QED_RESC_LOCK_RESC_INVALID)
     50		return -EINVAL;
     51
     52	qed_mcp_resc_lock_default_init(&params, NULL, resource, true);
     53
     54	rc = qed_mcp_resc_lock(p_hwfn, p_ptt, &params);
     55	if (rc && rc != -EINVAL) {
     56		return rc;
     57	} else if (rc == -EINVAL) {
     58		/* MFW doesn't support resource locking, first PF on the port
     59		 * has lock ownership.
     60		 */
     61		if (p_hwfn->abs_pf_id < p_hwfn->cdev->num_ports_in_engine)
     62			return 0;
     63
     64		DP_INFO(p_hwfn, "PF doesn't have lock ownership\n");
     65		return -EBUSY;
     66	} else if (!params.b_granted) {
     67		DP_INFO(p_hwfn, "Failed to acquire ptp resource lock\n");
     68		return -EBUSY;
     69	}
     70
     71	return 0;
     72}
     73
     74static int qed_ptp_res_unlock(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt)
     75{
     76	struct qed_resc_unlock_params params;
     77	enum qed_resc_lock resource;
     78	int rc;
     79
     80	resource = qed_ptcdev_to_resc(p_hwfn);
     81	if (resource == QED_RESC_LOCK_RESC_INVALID)
     82		return -EINVAL;
     83
     84	qed_mcp_resc_lock_default_init(NULL, &params, resource, true);
     85
     86	rc = qed_mcp_resc_unlock(p_hwfn, p_ptt, &params);
     87	if (rc == -EINVAL) {
     88		/* MFW doesn't support locking, first PF has lock ownership */
     89		if (p_hwfn->abs_pf_id < p_hwfn->cdev->num_ports_in_engine) {
     90			rc = 0;
     91		} else {
     92			DP_INFO(p_hwfn, "PF doesn't have lock ownership\n");
     93			return -EINVAL;
     94		}
     95	} else if (rc) {
     96		DP_INFO(p_hwfn, "Failed to release the ptp resource lock\n");
     97	}
     98
     99	return rc;
    100}
    101
    102/* Read Rx timestamp */
    103static int qed_ptp_hw_read_rx_ts(struct qed_dev *cdev, u64 *timestamp)
    104{
    105	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
    106	struct qed_ptt *p_ptt = p_hwfn->p_ptp_ptt;
    107	u32 val;
    108
    109	*timestamp = 0;
    110	val = qed_rd(p_hwfn, p_ptt, NIG_REG_LLH_PTP_HOST_BUF_SEQID);
    111	if (!(val & QED_TIMESTAMP_MASK)) {
    112		DP_INFO(p_hwfn, "Invalid Rx timestamp, buf_seqid = %d\n", val);
    113		return -EINVAL;
    114	}
    115
    116	val = qed_rd(p_hwfn, p_ptt, NIG_REG_LLH_PTP_HOST_BUF_TS_LSB);
    117	*timestamp = qed_rd(p_hwfn, p_ptt, NIG_REG_LLH_PTP_HOST_BUF_TS_MSB);
    118	*timestamp <<= 32;
    119	*timestamp |= val;
    120
    121	/* Reset timestamp register to allow new timestamp */
    122	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_HOST_BUF_SEQID,
    123	       QED_TIMESTAMP_MASK);
    124
    125	return 0;
    126}
    127
    128/* Read Tx timestamp */
    129static int qed_ptp_hw_read_tx_ts(struct qed_dev *cdev, u64 *timestamp)
    130{
    131	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
    132	struct qed_ptt *p_ptt = p_hwfn->p_ptp_ptt;
    133	u32 val;
    134
    135	*timestamp = 0;
    136	val = qed_rd(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_BUF_SEQID);
    137	if (!(val & QED_TIMESTAMP_MASK)) {
    138		DP_VERBOSE(p_hwfn, QED_MSG_DEBUG,
    139			   "Invalid Tx timestamp, buf_seqid = %08x\n", val);
    140		return -EINVAL;
    141	}
    142
    143	val = qed_rd(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_BUF_TS_LSB);
    144	*timestamp = qed_rd(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_BUF_TS_MSB);
    145	*timestamp <<= 32;
    146	*timestamp |= val;
    147
    148	/* Reset timestamp register to allow new timestamp */
    149	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_BUF_SEQID, QED_TIMESTAMP_MASK);
    150
    151	return 0;
    152}
    153
    154/* Read Phy Hardware Clock */
    155static int qed_ptp_hw_read_cc(struct qed_dev *cdev, u64 *phc_cycles)
    156{
    157	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
    158	struct qed_ptt *p_ptt = p_hwfn->p_ptp_ptt;
    159	u32 temp = 0;
    160
    161	temp = qed_rd(p_hwfn, p_ptt, NIG_REG_TSGEN_SYNC_TIME_LSB);
    162	*phc_cycles = qed_rd(p_hwfn, p_ptt, NIG_REG_TSGEN_SYNC_TIME_MSB);
    163	*phc_cycles <<= 32;
    164	*phc_cycles |= temp;
    165
    166	return 0;
    167}
    168
    169/* Filter PTP protocol packets that need to be timestamped */
    170static int qed_ptp_hw_cfg_filters(struct qed_dev *cdev,
    171				  enum qed_ptp_filter_type rx_type,
    172				  enum qed_ptp_hwtstamp_tx_type tx_type)
    173{
    174	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
    175	struct qed_ptt *p_ptt = p_hwfn->p_ptp_ptt;
    176	u32 rule_mask, enable_cfg = 0x0;
    177
    178	switch (rx_type) {
    179	case QED_PTP_FILTER_NONE:
    180		enable_cfg = 0x0;
    181		rule_mask = 0x3FFF;
    182		break;
    183	case QED_PTP_FILTER_ALL:
    184		enable_cfg = 0x7;
    185		rule_mask = 0x3CAA;
    186		break;
    187	case QED_PTP_FILTER_V1_L4_EVENT:
    188		enable_cfg = 0x3;
    189		rule_mask = 0x3FFA;
    190		break;
    191	case QED_PTP_FILTER_V1_L4_GEN:
    192		enable_cfg = 0x3;
    193		rule_mask = 0x3FFE;
    194		break;
    195	case QED_PTP_FILTER_V2_L4_EVENT:
    196		enable_cfg = 0x5;
    197		rule_mask = 0x3FAA;
    198		break;
    199	case QED_PTP_FILTER_V2_L4_GEN:
    200		enable_cfg = 0x5;
    201		rule_mask = 0x3FEE;
    202		break;
    203	case QED_PTP_FILTER_V2_L2_EVENT:
    204		enable_cfg = 0x5;
    205		rule_mask = 0x3CFF;
    206		break;
    207	case QED_PTP_FILTER_V2_L2_GEN:
    208		enable_cfg = 0x5;
    209		rule_mask = 0x3EFF;
    210		break;
    211	case QED_PTP_FILTER_V2_EVENT:
    212		enable_cfg = 0x5;
    213		rule_mask = 0x3CAA;
    214		break;
    215	case QED_PTP_FILTER_V2_GEN:
    216		enable_cfg = 0x5;
    217		rule_mask = 0x3EEE;
    218		break;
    219	default:
    220		DP_INFO(p_hwfn, "Invalid PTP filter type %d\n", rx_type);
    221		return -EINVAL;
    222	}
    223
    224	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_PARAM_MASK,
    225	       QED_PTP_UCAST_PARAM_MASK);
    226	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_RULE_MASK, rule_mask);
    227	qed_wr(p_hwfn, p_ptt, NIG_REG_RX_PTP_EN, enable_cfg);
    228
    229	if (tx_type == QED_PTP_HWTSTAMP_TX_OFF) {
    230		qed_wr(p_hwfn, p_ptt, NIG_REG_TX_PTP_EN, 0x0);
    231		qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_PARAM_MASK, 0x7FF);
    232		qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_RULE_MASK, 0x3FFF);
    233	} else {
    234		qed_wr(p_hwfn, p_ptt, NIG_REG_TX_PTP_EN, enable_cfg);
    235		qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_PARAM_MASK,
    236		       QED_PTP_UCAST_PARAM_MASK);
    237		qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_RULE_MASK, rule_mask);
    238	}
    239
    240	/* Reset possibly old timestamps */
    241	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_HOST_BUF_SEQID,
    242	       QED_TIMESTAMP_MASK);
    243
    244	return 0;
    245}
    246
    247/* Adjust the HW clock by a rate given in parts-per-billion (ppb) units.
    248 * FW/HW accepts the adjustment value in terms of 3 parameters:
    249 *   Drift period - adjustment happens once in certain number of nano seconds.
    250 *   Drift value - time is adjusted by a certain value, for example by 5 ns.
    251 *   Drift direction - add or subtract the adjustment value.
    252 * The routine translates ppb into the adjustment triplet in an optimal manner.
    253 */
    254static int qed_ptp_hw_adjfreq(struct qed_dev *cdev, s32 ppb)
    255{
    256	s64 best_val = 0, val, best_period = 0, period, approx_dev, dif, dif2;
    257	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
    258	struct qed_ptt *p_ptt = p_hwfn->p_ptp_ptt;
    259	u32 drift_ctr_cfg = 0, drift_state;
    260	int drift_dir = 1;
    261
    262	if (ppb < 0) {
    263		ppb = -ppb;
    264		drift_dir = 0;
    265	}
    266
    267	if (ppb > 1) {
    268		s64 best_dif = ppb, best_approx_dev = 1;
    269
    270		/* Adjustment value is up to +/-7ns, find an optimal value in
    271		 * this range.
    272		 */
    273		for (val = 7; val > 0; val--) {
    274			period = div_s64(val * 1000000000, ppb);
    275			period -= 8;
    276			period >>= 4;
    277			if (period < 1)
    278				period = 1;
    279			if (period > 0xFFFFFFE)
    280				period = 0xFFFFFFE;
    281
    282			/* Check both rounding ends for approximate error */
    283			approx_dev = period * 16 + 8;
    284			dif = ppb * approx_dev - val * 1000000000;
    285			dif2 = dif + 16 * ppb;
    286
    287			if (dif < 0)
    288				dif = -dif;
    289			if (dif2 < 0)
    290				dif2 = -dif2;
    291
    292			/* Determine which end gives better approximation */
    293			if (dif * (approx_dev + 16) > dif2 * approx_dev) {
    294				period++;
    295				approx_dev += 16;
    296				dif = dif2;
    297			}
    298
    299			/* Track best approximation found so far */
    300			if (best_dif * approx_dev > dif * best_approx_dev) {
    301				best_dif = dif;
    302				best_val = val;
    303				best_period = period;
    304				best_approx_dev = approx_dev;
    305			}
    306		}
    307	} else if (ppb == 1) {
    308		/* This is a special case as its the only value which wouldn't
    309		 * fit in a s64 variable. In order to prevent castings simple
    310		 * handle it seperately.
    311		 */
    312		best_val = 4;
    313		best_period = 0xee6b27f;
    314	} else {
    315		best_val = 0;
    316		best_period = 0xFFFFFFF;
    317	}
    318
    319	drift_ctr_cfg = (best_period << QED_DRIFT_CNTR_TIME_QUANTA_SHIFT) |
    320			(((int)best_val) << QED_DRIFT_CNTR_ADJUSTMENT_SHIFT) |
    321			(((int)drift_dir) << QED_DRIFT_CNTR_DIRECTION_SHIFT);
    322
    323	qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_RST_DRIFT_CNTR, 0x1);
    324
    325	drift_state = qed_rd(p_hwfn, p_ptt, NIG_REG_TSGEN_RST_DRIFT_CNTR);
    326	if (drift_state & 1) {
    327		qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_DRIFT_CNTR_CONF,
    328		       drift_ctr_cfg);
    329	} else {
    330		DP_INFO(p_hwfn, "Drift counter is not reset\n");
    331		return -EINVAL;
    332	}
    333
    334	qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_RST_DRIFT_CNTR, 0x0);
    335
    336	return 0;
    337}
    338
    339static int qed_ptp_hw_enable(struct qed_dev *cdev)
    340{
    341	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
    342	struct qed_ptt *p_ptt;
    343	int rc;
    344
    345	p_ptt = qed_ptt_acquire(p_hwfn);
    346	if (!p_ptt) {
    347		DP_NOTICE(p_hwfn, "Failed to acquire PTT for PTP\n");
    348		return -EBUSY;
    349	}
    350
    351	p_hwfn->p_ptp_ptt = p_ptt;
    352
    353	rc = qed_ptp_res_lock(p_hwfn, p_ptt);
    354	if (rc) {
    355		DP_INFO(p_hwfn,
    356			"Couldn't acquire the resource lock, skip ptp enable for this PF\n");
    357		qed_ptt_release(p_hwfn, p_ptt);
    358		p_hwfn->p_ptp_ptt = NULL;
    359		return rc;
    360	}
    361
    362	/* Reset PTP event detection rules - will be configured in the IOCTL */
    363	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_PARAM_MASK, 0x7FF);
    364	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_RULE_MASK, 0x3FFF);
    365	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_PARAM_MASK, 0x7FF);
    366	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_RULE_MASK, 0x3FFF);
    367
    368	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_PTP_EN, 7);
    369	qed_wr(p_hwfn, p_ptt, NIG_REG_RX_PTP_EN, 7);
    370
    371	qed_wr(p_hwfn, p_ptt, NIG_REG_TS_OUTPUT_ENABLE_PDA, 0x1);
    372
    373	/* Pause free running counter */
    374	if (QED_IS_BB_B0(p_hwfn->cdev))
    375		qed_wr(p_hwfn, p_ptt, NIG_REG_TIMESYNC_GEN_REG_BB, 2);
    376	if (QED_IS_AH(p_hwfn->cdev))
    377		qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_FREECNT_UPDATE_K2, 2);
    378
    379	qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_FREE_CNT_VALUE_LSB, 0);
    380	qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_FREE_CNT_VALUE_MSB, 0);
    381	/* Resume free running counter */
    382	if (QED_IS_BB_B0(p_hwfn->cdev))
    383		qed_wr(p_hwfn, p_ptt, NIG_REG_TIMESYNC_GEN_REG_BB, 4);
    384	if (QED_IS_AH(p_hwfn->cdev)) {
    385		qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_FREECNT_UPDATE_K2, 4);
    386		qed_wr(p_hwfn, p_ptt, NIG_REG_PTP_LATCH_OSTS_PKT_TIME, 1);
    387	}
    388
    389	/* Disable drift register */
    390	qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_DRIFT_CNTR_CONF, 0x0);
    391	qed_wr(p_hwfn, p_ptt, NIG_REG_TSGEN_RST_DRIFT_CNTR, 0x0);
    392
    393	/* Reset possibly old timestamps */
    394	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_HOST_BUF_SEQID,
    395	       QED_TIMESTAMP_MASK);
    396	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_BUF_SEQID, QED_TIMESTAMP_MASK);
    397
    398	return 0;
    399}
    400
    401static int qed_ptp_hw_disable(struct qed_dev *cdev)
    402{
    403	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
    404	struct qed_ptt *p_ptt = p_hwfn->p_ptp_ptt;
    405
    406	qed_ptp_res_unlock(p_hwfn, p_ptt);
    407
    408	/* Reset PTP event detection rules */
    409	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_PARAM_MASK, 0x7FF);
    410	qed_wr(p_hwfn, p_ptt, NIG_REG_LLH_PTP_RULE_MASK, 0x3FFF);
    411
    412	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_PARAM_MASK, 0x7FF);
    413	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_LLH_PTP_RULE_MASK, 0x3FFF);
    414
    415	/* Disable the PTP feature */
    416	qed_wr(p_hwfn, p_ptt, NIG_REG_RX_PTP_EN, 0x0);
    417	qed_wr(p_hwfn, p_ptt, NIG_REG_TX_PTP_EN, 0x0);
    418
    419	qed_ptt_release(p_hwfn, p_ptt);
    420	p_hwfn->p_ptp_ptt = NULL;
    421
    422	return 0;
    423}
    424
    425const struct qed_eth_ptp_ops qed_ptp_ops_pass = {
    426	.cfg_filters = qed_ptp_hw_cfg_filters,
    427	.read_rx_ts = qed_ptp_hw_read_rx_ts,
    428	.read_tx_ts = qed_ptp_hw_read_tx_ts,
    429	.read_cc = qed_ptp_hw_read_cc,
    430	.adjfreq = qed_ptp_hw_adjfreq,
    431	.disable = qed_ptp_hw_disable,
    432	.enable = qed_ptp_hw_enable,
    433};