cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

ehea_hw.h (5672B)


      1/* SPDX-License-Identifier: GPL-2.0-or-later */
      2/*
      3 *  linux/drivers/net/ethernet/ibm/ehea/ehea_hw.h
      4 *
      5 *  eHEA ethernet device driver for IBM eServer System p
      6 *
      7 *  (C) Copyright IBM Corp. 2006
      8 *
      9 *  Authors:
     10 *       Christoph Raisch <raisch@de.ibm.com>
     11 *       Jan-Bernd Themann <themann@de.ibm.com>
     12 *       Thomas Klein <tklein@de.ibm.com>
     13 */
     14
     15#ifndef __EHEA_HW_H__
     16#define __EHEA_HW_H__
     17
     18#define QPX_SQA_VALUE   EHEA_BMASK_IBM(48, 63)
     19#define QPX_RQ1A_VALUE  EHEA_BMASK_IBM(48, 63)
     20#define QPX_RQ2A_VALUE  EHEA_BMASK_IBM(48, 63)
     21#define QPX_RQ3A_VALUE  EHEA_BMASK_IBM(48, 63)
     22
     23#define QPTEMM_OFFSET(x) offsetof(struct ehea_qptemm, x)
     24
     25struct ehea_qptemm {
     26	u64 qpx_hcr;
     27	u64 qpx_c;
     28	u64 qpx_herr;
     29	u64 qpx_aer;
     30	u64 qpx_sqa;
     31	u64 qpx_sqc;
     32	u64 qpx_rq1a;
     33	u64 qpx_rq1c;
     34	u64 qpx_st;
     35	u64 qpx_aerr;
     36	u64 qpx_tenure;
     37	u64 qpx_reserved1[(0x098 - 0x058) / 8];
     38	u64 qpx_portp;
     39	u64 qpx_reserved2[(0x100 - 0x0A0) / 8];
     40	u64 qpx_t;
     41	u64 qpx_sqhp;
     42	u64 qpx_sqptp;
     43	u64 qpx_reserved3[(0x140 - 0x118) / 8];
     44	u64 qpx_sqwsize;
     45	u64 qpx_reserved4[(0x170 - 0x148) / 8];
     46	u64 qpx_sqsize;
     47	u64 qpx_reserved5[(0x1B0 - 0x178) / 8];
     48	u64 qpx_sigt;
     49	u64 qpx_wqecnt;
     50	u64 qpx_rq1hp;
     51	u64 qpx_rq1ptp;
     52	u64 qpx_rq1size;
     53	u64 qpx_reserved6[(0x220 - 0x1D8) / 8];
     54	u64 qpx_rq1wsize;
     55	u64 qpx_reserved7[(0x240 - 0x228) / 8];
     56	u64 qpx_pd;
     57	u64 qpx_scqn;
     58	u64 qpx_rcqn;
     59	u64 qpx_aeqn;
     60	u64 reserved49;
     61	u64 qpx_ram;
     62	u64 qpx_reserved8[(0x300 - 0x270) / 8];
     63	u64 qpx_rq2a;
     64	u64 qpx_rq2c;
     65	u64 qpx_rq2hp;
     66	u64 qpx_rq2ptp;
     67	u64 qpx_rq2size;
     68	u64 qpx_rq2wsize;
     69	u64 qpx_rq2th;
     70	u64 qpx_rq3a;
     71	u64 qpx_rq3c;
     72	u64 qpx_rq3hp;
     73	u64 qpx_rq3ptp;
     74	u64 qpx_rq3size;
     75	u64 qpx_rq3wsize;
     76	u64 qpx_rq3th;
     77	u64 qpx_lpn;
     78	u64 qpx_reserved9[(0x400 - 0x378) / 8];
     79	u64 reserved_ext[(0x500 - 0x400) / 8];
     80	u64 reserved2[(0x1000 - 0x500) / 8];
     81};
     82
     83#define MRx_HCR_LPARID_VALID EHEA_BMASK_IBM(0, 0)
     84
     85#define MRMWMM_OFFSET(x) offsetof(struct ehea_mrmwmm, x)
     86
     87struct ehea_mrmwmm {
     88	u64 mrx_hcr;
     89	u64 mrx_c;
     90	u64 mrx_herr;
     91	u64 mrx_aer;
     92	u64 mrx_pp;
     93	u64 reserved1;
     94	u64 reserved2;
     95	u64 reserved3;
     96	u64 reserved4[(0x200 - 0x40) / 8];
     97	u64 mrx_ctl[64];
     98};
     99
    100#define QPEDMM_OFFSET(x) offsetof(struct ehea_qpedmm, x)
    101
    102struct ehea_qpedmm {
    103
    104	u64 reserved0[(0x400) / 8];
    105	u64 qpedx_phh;
    106	u64 qpedx_ppsgp;
    107	u64 qpedx_ppsgu;
    108	u64 qpedx_ppdgp;
    109	u64 qpedx_ppdgu;
    110	u64 qpedx_aph;
    111	u64 qpedx_apsgp;
    112	u64 qpedx_apsgu;
    113	u64 qpedx_apdgp;
    114	u64 qpedx_apdgu;
    115	u64 qpedx_apav;
    116	u64 qpedx_apsav;
    117	u64 qpedx_hcr;
    118	u64 reserved1[4];
    119	u64 qpedx_rrl0;
    120	u64 qpedx_rrrkey0;
    121	u64 qpedx_rrva0;
    122	u64 reserved2;
    123	u64 qpedx_rrl1;
    124	u64 qpedx_rrrkey1;
    125	u64 qpedx_rrva1;
    126	u64 reserved3;
    127	u64 qpedx_rrl2;
    128	u64 qpedx_rrrkey2;
    129	u64 qpedx_rrva2;
    130	u64 reserved4;
    131	u64 qpedx_rrl3;
    132	u64 qpedx_rrrkey3;
    133	u64 qpedx_rrva3;
    134};
    135
    136#define CQX_FECADDER EHEA_BMASK_IBM(32, 63)
    137#define CQX_FEC_CQE_CNT EHEA_BMASK_IBM(32, 63)
    138#define CQX_N1_GENERATE_COMP_EVENT EHEA_BMASK_IBM(0, 0)
    139#define CQX_EP_EVENT_PENDING EHEA_BMASK_IBM(0, 0)
    140
    141#define CQTEMM_OFFSET(x) offsetof(struct ehea_cqtemm, x)
    142
    143struct ehea_cqtemm {
    144	u64 cqx_hcr;
    145	u64 cqx_c;
    146	u64 cqx_herr;
    147	u64 cqx_aer;
    148	u64 cqx_ptp;
    149	u64 cqx_tp;
    150	u64 cqx_fec;
    151	u64 cqx_feca;
    152	u64 cqx_ep;
    153	u64 cqx_eq;
    154	u64 reserved1;
    155	u64 cqx_n0;
    156	u64 cqx_n1;
    157	u64 reserved2[(0x1000 - 0x60) / 8];
    158};
    159
    160#define EQTEMM_OFFSET(x) offsetof(struct ehea_eqtemm, x)
    161
    162struct ehea_eqtemm {
    163	u64 eqx_hcr;
    164	u64 eqx_c;
    165	u64 eqx_herr;
    166	u64 eqx_aer;
    167	u64 eqx_ptp;
    168	u64 eqx_tp;
    169	u64 eqx_ssba;
    170	u64 eqx_psba;
    171	u64 eqx_cec;
    172	u64 eqx_meql;
    173	u64 eqx_xisbi;
    174	u64 eqx_xisc;
    175	u64 eqx_it;
    176};
    177
    178/*
    179 * These access functions will be changed when the dissuccsion about
    180 * the new access methods for POWER has settled.
    181 */
    182
    183static inline u64 epa_load(struct h_epa epa, u32 offset)
    184{
    185	return __raw_readq((void __iomem *)(epa.addr + offset));
    186}
    187
    188static inline void epa_store(struct h_epa epa, u32 offset, u64 value)
    189{
    190	__raw_writeq(value, (void __iomem *)(epa.addr + offset));
    191	epa_load(epa, offset);	/* synchronize explicitly to eHEA */
    192}
    193
    194static inline void epa_store_acc(struct h_epa epa, u32 offset, u64 value)
    195{
    196	__raw_writeq(value, (void __iomem *)(epa.addr + offset));
    197}
    198
    199#define epa_store_cq(epa, offset, value)\
    200	epa_store(epa, CQTEMM_OFFSET(offset), value)
    201#define epa_load_cq(epa, offset)\
    202	epa_load(epa, CQTEMM_OFFSET(offset))
    203
    204static inline void ehea_update_sqa(struct ehea_qp *qp, u16 nr_wqes)
    205{
    206	struct h_epa epa = qp->epas.kernel;
    207	epa_store_acc(epa, QPTEMM_OFFSET(qpx_sqa),
    208		      EHEA_BMASK_SET(QPX_SQA_VALUE, nr_wqes));
    209}
    210
    211static inline void ehea_update_rq3a(struct ehea_qp *qp, u16 nr_wqes)
    212{
    213	struct h_epa epa = qp->epas.kernel;
    214	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq3a),
    215		      EHEA_BMASK_SET(QPX_RQ1A_VALUE, nr_wqes));
    216}
    217
    218static inline void ehea_update_rq2a(struct ehea_qp *qp, u16 nr_wqes)
    219{
    220	struct h_epa epa = qp->epas.kernel;
    221	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq2a),
    222		      EHEA_BMASK_SET(QPX_RQ2A_VALUE, nr_wqes));
    223}
    224
    225static inline void ehea_update_rq1a(struct ehea_qp *qp, u16 nr_wqes)
    226{
    227	struct h_epa epa = qp->epas.kernel;
    228	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq1a),
    229		      EHEA_BMASK_SET(QPX_RQ3A_VALUE, nr_wqes));
    230}
    231
    232static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes)
    233{
    234	struct h_epa epa = cq->epas.kernel;
    235	epa_store_acc(epa, CQTEMM_OFFSET(cqx_feca),
    236		      EHEA_BMASK_SET(CQX_FECADDER, nr_cqes));
    237}
    238
    239static inline void ehea_reset_cq_n1(struct ehea_cq *cq)
    240{
    241	struct h_epa epa = cq->epas.kernel;
    242	epa_store_cq(epa, cqx_n1,
    243		     EHEA_BMASK_SET(CQX_N1_GENERATE_COMP_EVENT, 1));
    244}
    245
    246static inline void ehea_reset_cq_ep(struct ehea_cq *my_cq)
    247{
    248	struct h_epa epa = my_cq->epas.kernel;
    249	epa_store_acc(epa, CQTEMM_OFFSET(cqx_ep),
    250		      EHEA_BMASK_SET(CQX_EP_EVENT_PENDING, 0));
    251}
    252
    253#endif	/* __EHEA_HW_H__ */