cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

amdgpu_res_cursor.h (4405B)


      1// SPDX-License-Identifier: GPL-2.0 OR MIT
      2/*
      3 * Copyright 2020 Advanced Micro Devices, Inc.
      4 *
      5 * Permission is hereby granted, free of charge, to any person obtaining a
      6 * copy of this software and associated documentation files (the "Software"),
      7 * to deal in the Software without restriction, including without limitation
      8 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
      9 * and/or sell copies of the Software, and to permit persons to whom the
     10 * Software is furnished to do so, subject to the following conditions:
     11 *
     12 * The above copyright notice and this permission notice shall be included in
     13 * all copies or substantial portions of the Software.
     14 *
     15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     18 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
     19 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
     20 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
     21 * OTHER DEALINGS IN THE SOFTWARE.
     22 *
     23 * Authors: Christian König
     24 */
     25
     26#ifndef __AMDGPU_RES_CURSOR_H__
     27#define __AMDGPU_RES_CURSOR_H__
     28
     29#include <drm/drm_mm.h>
     30#include <drm/ttm/ttm_resource.h>
     31#include <drm/ttm/ttm_range_manager.h>
     32
     33#include "amdgpu_vram_mgr.h"
     34
     35/* state back for walking over vram_mgr and gtt_mgr allocations */
     36struct amdgpu_res_cursor {
     37	uint64_t		start;
     38	uint64_t		size;
     39	uint64_t		remaining;
     40	void			*node;
     41	uint32_t		mem_type;
     42};
     43
     44/**
     45 * amdgpu_res_first - initialize a amdgpu_res_cursor
     46 *
     47 * @res: TTM resource object to walk
     48 * @start: Start of the range
     49 * @size: Size of the range
     50 * @cur: cursor object to initialize
     51 *
     52 * Start walking over the range of allocations between @start and @size.
     53 */
     54static inline void amdgpu_res_first(struct ttm_resource *res,
     55				    uint64_t start, uint64_t size,
     56				    struct amdgpu_res_cursor *cur)
     57{
     58	struct drm_buddy_block *block;
     59	struct list_head *head, *next;
     60	struct drm_mm_node *node;
     61
     62	if (!res)
     63		goto fallback;
     64
     65	BUG_ON(start + size > res->num_pages << PAGE_SHIFT);
     66
     67	cur->mem_type = res->mem_type;
     68
     69	switch (cur->mem_type) {
     70	case TTM_PL_VRAM:
     71		head = &to_amdgpu_vram_mgr_resource(res)->blocks;
     72
     73		block = list_first_entry_or_null(head,
     74						 struct drm_buddy_block,
     75						 link);
     76		if (!block)
     77			goto fallback;
     78
     79		while (start >= amdgpu_vram_mgr_block_size(block)) {
     80			start -= amdgpu_vram_mgr_block_size(block);
     81
     82			next = block->link.next;
     83			if (next != head)
     84				block = list_entry(next, struct drm_buddy_block, link);
     85		}
     86
     87		cur->start = amdgpu_vram_mgr_block_start(block) + start;
     88		cur->size = min(amdgpu_vram_mgr_block_size(block) - start, size);
     89		cur->remaining = size;
     90		cur->node = block;
     91		break;
     92	case TTM_PL_TT:
     93		node = to_ttm_range_mgr_node(res)->mm_nodes;
     94		while (start >= node->size << PAGE_SHIFT)
     95			start -= node++->size << PAGE_SHIFT;
     96
     97		cur->start = (node->start << PAGE_SHIFT) + start;
     98		cur->size = min((node->size << PAGE_SHIFT) - start, size);
     99		cur->remaining = size;
    100		cur->node = node;
    101		break;
    102	default:
    103		goto fallback;
    104	}
    105
    106	return;
    107
    108fallback:
    109	cur->start = start;
    110	cur->size = size;
    111	cur->remaining = size;
    112	cur->node = NULL;
    113	WARN_ON(res && start + size > res->num_pages << PAGE_SHIFT);
    114	return;
    115}
    116
    117/**
    118 * amdgpu_res_next - advance the cursor
    119 *
    120 * @cur: the cursor to advance
    121 * @size: number of bytes to move forward
    122 *
    123 * Move the cursor @size bytes forwrad, walking to the next node if necessary.
    124 */
    125static inline void amdgpu_res_next(struct amdgpu_res_cursor *cur, uint64_t size)
    126{
    127	struct drm_buddy_block *block;
    128	struct drm_mm_node *node;
    129	struct list_head *next;
    130
    131	BUG_ON(size > cur->remaining);
    132
    133	cur->remaining -= size;
    134	if (!cur->remaining)
    135		return;
    136
    137	cur->size -= size;
    138	if (cur->size) {
    139		cur->start += size;
    140		return;
    141	}
    142
    143	switch (cur->mem_type) {
    144	case TTM_PL_VRAM:
    145		block = cur->node;
    146
    147		next = block->link.next;
    148		block = list_entry(next, struct drm_buddy_block, link);
    149
    150		cur->node = block;
    151		cur->start = amdgpu_vram_mgr_block_start(block);
    152		cur->size = min(amdgpu_vram_mgr_block_size(block), cur->remaining);
    153		break;
    154	case TTM_PL_TT:
    155		node = cur->node;
    156
    157		cur->node = ++node;
    158		cur->start = node->start << PAGE_SHIFT;
    159		cur->size = min(node->size << PAGE_SHIFT, cur->remaining);
    160		break;
    161	default:
    162		return;
    163	}
    164}
    165
    166#endif