cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

intel_ring.h (4197B)


      1/* SPDX-License-Identifier: MIT */
      2/*
      3 * Copyright © 2019 Intel Corporation
      4 */
      5
      6#ifndef INTEL_RING_H
      7#define INTEL_RING_H
      8
      9#include "i915_gem.h" /* GEM_BUG_ON */
     10#include "i915_request.h"
     11#include "intel_ring_types.h"
     12
     13struct intel_engine_cs;
     14
     15struct intel_ring *
     16intel_engine_create_ring(struct intel_engine_cs *engine, int size);
     17
     18u32 *intel_ring_begin(struct i915_request *rq, unsigned int num_dwords);
     19int intel_ring_cacheline_align(struct i915_request *rq);
     20
     21unsigned int intel_ring_update_space(struct intel_ring *ring);
     22
     23void __intel_ring_pin(struct intel_ring *ring);
     24int intel_ring_pin(struct intel_ring *ring, struct i915_gem_ww_ctx *ww);
     25void intel_ring_unpin(struct intel_ring *ring);
     26void intel_ring_reset(struct intel_ring *ring, u32 tail);
     27
     28void intel_ring_free(struct kref *ref);
     29
     30static inline struct intel_ring *intel_ring_get(struct intel_ring *ring)
     31{
     32	kref_get(&ring->ref);
     33	return ring;
     34}
     35
     36static inline void intel_ring_put(struct intel_ring *ring)
     37{
     38	kref_put(&ring->ref, intel_ring_free);
     39}
     40
     41static inline void intel_ring_advance(struct i915_request *rq, u32 *cs)
     42{
     43	/* Dummy function.
     44	 *
     45	 * This serves as a placeholder in the code so that the reader
     46	 * can compare against the preceding intel_ring_begin() and
     47	 * check that the number of dwords emitted matches the space
     48	 * reserved for the command packet (i.e. the value passed to
     49	 * intel_ring_begin()).
     50	 */
     51	GEM_BUG_ON((rq->ring->vaddr + rq->ring->emit) != cs);
     52	GEM_BUG_ON(!IS_ALIGNED(rq->ring->emit, 8)); /* RING_TAIL qword align */
     53}
     54
     55static inline u32 intel_ring_wrap(const struct intel_ring *ring, u32 pos)
     56{
     57	return pos & (ring->size - 1);
     58}
     59
     60static inline int intel_ring_direction(const struct intel_ring *ring,
     61				       u32 next, u32 prev)
     62{
     63	typecheck(typeof(ring->size), next);
     64	typecheck(typeof(ring->size), prev);
     65	return (next - prev) << ring->wrap;
     66}
     67
     68static inline bool
     69intel_ring_offset_valid(const struct intel_ring *ring,
     70			unsigned int pos)
     71{
     72	if (pos & -ring->size) /* must be strictly within the ring */
     73		return false;
     74
     75	if (!IS_ALIGNED(pos, 8)) /* must be qword aligned */
     76		return false;
     77
     78	return true;
     79}
     80
     81static inline u32 intel_ring_offset(const struct i915_request *rq, void *addr)
     82{
     83	/* Don't write ring->size (equivalent to 0) as that hangs some GPUs. */
     84	u32 offset = addr - rq->ring->vaddr;
     85
     86	GEM_BUG_ON(offset > rq->ring->size);
     87	return intel_ring_wrap(rq->ring, offset);
     88}
     89
     90static inline void
     91assert_ring_tail_valid(const struct intel_ring *ring, unsigned int tail)
     92{
     93	unsigned int head = READ_ONCE(ring->head);
     94
     95	GEM_BUG_ON(!intel_ring_offset_valid(ring, tail));
     96
     97	/*
     98	 * "Ring Buffer Use"
     99	 *	Gen2 BSpec "1. Programming Environment" / 1.4.4.6
    100	 *	Gen3 BSpec "1c Memory Interface Functions" / 2.3.4.5
    101	 *	Gen4+ BSpec "1c Memory Interface and Command Stream" / 5.3.4.5
    102	 * "If the Ring Buffer Head Pointer and the Tail Pointer are on the
    103	 * same cacheline, the Head Pointer must not be greater than the Tail
    104	 * Pointer."
    105	 *
    106	 * We use ring->head as the last known location of the actual RING_HEAD,
    107	 * it may have advanced but in the worst case it is equally the same
    108	 * as ring->head and so we should never program RING_TAIL to advance
    109	 * into the same cacheline as ring->head.
    110	 */
    111#define cacheline(a) round_down(a, CACHELINE_BYTES)
    112	GEM_BUG_ON(cacheline(tail) == cacheline(head) && tail < head);
    113#undef cacheline
    114}
    115
    116static inline unsigned int
    117intel_ring_set_tail(struct intel_ring *ring, unsigned int tail)
    118{
    119	/* Whilst writes to the tail are strictly order, there is no
    120	 * serialisation between readers and the writers. The tail may be
    121	 * read by i915_request_retire() just as it is being updated
    122	 * by execlists, as although the breadcrumb is complete, the context
    123	 * switch hasn't been seen.
    124	 */
    125	assert_ring_tail_valid(ring, tail);
    126	ring->tail = tail;
    127	return tail;
    128}
    129
    130static inline unsigned int
    131__intel_ring_space(unsigned int head, unsigned int tail, unsigned int size)
    132{
    133	/*
    134	 * "If the Ring Buffer Head Pointer and the Tail Pointer are on the
    135	 * same cacheline, the Head Pointer must not be greater than the Tail
    136	 * Pointer."
    137	 */
    138	GEM_BUG_ON(!is_power_of_2(size));
    139	return (head - tail - CACHELINE_BYTES) & (size - 1);
    140}
    141
    142#endif /* INTEL_RING_H */