cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

ring_mode.c (4194B)


      1// SPDX-License-Identifier: GPL-2.0-only
      2/*******************************************************************************
      3  Specialised functions for managing Ring mode
      4
      5  Copyright(C) 2011  STMicroelectronics Ltd
      6
      7  It defines all the functions used to handle the normal/enhanced
      8  descriptors in case of the DMA is configured to work in chained or
      9  in ring mode.
     10
     11
     12  Author: Giuseppe Cavallaro <peppe.cavallaro@st.com>
     13*******************************************************************************/
     14
     15#include "stmmac.h"
     16
     17static int jumbo_frm(void *p, struct sk_buff *skb, int csum)
     18{
     19	struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)p;
     20	unsigned int nopaged_len = skb_headlen(skb);
     21	struct stmmac_priv *priv = tx_q->priv_data;
     22	unsigned int entry = tx_q->cur_tx;
     23	unsigned int bmax, len, des2;
     24	struct dma_desc *desc;
     25
     26	if (priv->extend_desc)
     27		desc = (struct dma_desc *)(tx_q->dma_etx + entry);
     28	else
     29		desc = tx_q->dma_tx + entry;
     30
     31	if (priv->plat->enh_desc)
     32		bmax = BUF_SIZE_8KiB;
     33	else
     34		bmax = BUF_SIZE_2KiB;
     35
     36	len = nopaged_len - bmax;
     37
     38	if (nopaged_len > BUF_SIZE_8KiB) {
     39
     40		des2 = dma_map_single(priv->device, skb->data, bmax,
     41				      DMA_TO_DEVICE);
     42		desc->des2 = cpu_to_le32(des2);
     43		if (dma_mapping_error(priv->device, des2))
     44			return -1;
     45
     46		tx_q->tx_skbuff_dma[entry].buf = des2;
     47		tx_q->tx_skbuff_dma[entry].len = bmax;
     48		tx_q->tx_skbuff_dma[entry].is_jumbo = true;
     49
     50		desc->des3 = cpu_to_le32(des2 + BUF_SIZE_4KiB);
     51		stmmac_prepare_tx_desc(priv, desc, 1, bmax, csum,
     52				STMMAC_RING_MODE, 0, false, skb->len);
     53		tx_q->tx_skbuff[entry] = NULL;
     54		entry = STMMAC_GET_ENTRY(entry, priv->dma_tx_size);
     55
     56		if (priv->extend_desc)
     57			desc = (struct dma_desc *)(tx_q->dma_etx + entry);
     58		else
     59			desc = tx_q->dma_tx + entry;
     60
     61		des2 = dma_map_single(priv->device, skb->data + bmax, len,
     62				      DMA_TO_DEVICE);
     63		desc->des2 = cpu_to_le32(des2);
     64		if (dma_mapping_error(priv->device, des2))
     65			return -1;
     66		tx_q->tx_skbuff_dma[entry].buf = des2;
     67		tx_q->tx_skbuff_dma[entry].len = len;
     68		tx_q->tx_skbuff_dma[entry].is_jumbo = true;
     69
     70		desc->des3 = cpu_to_le32(des2 + BUF_SIZE_4KiB);
     71		stmmac_prepare_tx_desc(priv, desc, 0, len, csum,
     72				STMMAC_RING_MODE, 1, !skb_is_nonlinear(skb),
     73				skb->len);
     74	} else {
     75		des2 = dma_map_single(priv->device, skb->data,
     76				      nopaged_len, DMA_TO_DEVICE);
     77		desc->des2 = cpu_to_le32(des2);
     78		if (dma_mapping_error(priv->device, des2))
     79			return -1;
     80		tx_q->tx_skbuff_dma[entry].buf = des2;
     81		tx_q->tx_skbuff_dma[entry].len = nopaged_len;
     82		tx_q->tx_skbuff_dma[entry].is_jumbo = true;
     83		desc->des3 = cpu_to_le32(des2 + BUF_SIZE_4KiB);
     84		stmmac_prepare_tx_desc(priv, desc, 1, nopaged_len, csum,
     85				STMMAC_RING_MODE, 0, !skb_is_nonlinear(skb),
     86				skb->len);
     87	}
     88
     89	tx_q->cur_tx = entry;
     90
     91	return entry;
     92}
     93
     94static unsigned int is_jumbo_frm(int len, int enh_desc)
     95{
     96	unsigned int ret = 0;
     97
     98	if (len >= BUF_SIZE_4KiB)
     99		ret = 1;
    100
    101	return ret;
    102}
    103
    104static void refill_desc3(void *priv_ptr, struct dma_desc *p)
    105{
    106	struct stmmac_rx_queue *rx_q = priv_ptr;
    107	struct stmmac_priv *priv = rx_q->priv_data;
    108
    109	/* Fill DES3 in case of RING mode */
    110	if (priv->dma_buf_sz == BUF_SIZE_16KiB)
    111		p->des3 = cpu_to_le32(le32_to_cpu(p->des2) + BUF_SIZE_8KiB);
    112}
    113
    114/* In ring mode we need to fill the desc3 because it is used as buffer */
    115static void init_desc3(struct dma_desc *p)
    116{
    117	p->des3 = cpu_to_le32(le32_to_cpu(p->des2) + BUF_SIZE_8KiB);
    118}
    119
    120static void clean_desc3(void *priv_ptr, struct dma_desc *p)
    121{
    122	struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)priv_ptr;
    123	struct stmmac_priv *priv = tx_q->priv_data;
    124	unsigned int entry = tx_q->dirty_tx;
    125
    126	/* des3 is only used for jumbo frames tx or time stamping */
    127	if (unlikely(tx_q->tx_skbuff_dma[entry].is_jumbo ||
    128		     (tx_q->tx_skbuff_dma[entry].last_segment &&
    129		      !priv->extend_desc && priv->hwts_tx_en)))
    130		p->des3 = 0;
    131}
    132
    133static int set_16kib_bfsize(int mtu)
    134{
    135	int ret = 0;
    136	if (unlikely(mtu > BUF_SIZE_8KiB))
    137		ret = BUF_SIZE_16KiB;
    138	return ret;
    139}
    140
    141const struct stmmac_mode_ops ring_mode_ops = {
    142	.is_jumbo_frm = is_jumbo_frm,
    143	.jumbo_frm = jumbo_frm,
    144	.refill_desc3 = refill_desc3,
    145	.init_desc3 = init_desc3,
    146	.clean_desc3 = clean_desc3,
    147	.set_16kib_bfsize = set_16kib_bfsize,
    148};