cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

gpuobj.c (7307B)


      1/*
      2 * Copyright 2012 Red Hat Inc.
      3 *
      4 * Permission is hereby granted, free of charge, to any person obtaining a
      5 * copy of this software and associated documentation files (the "Software"),
      6 * to deal in the Software without restriction, including without limitation
      7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
      8 * and/or sell copies of the Software, and to permit persons to whom the
      9 * Software is furnished to do so, subject to the following conditions:
     10 *
     11 * The above copyright notice and this permission notice shall be included in
     12 * all copies or substantial portions of the Software.
     13 *
     14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
     18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
     19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
     20 * OTHER DEALINGS IN THE SOFTWARE.
     21 *
     22 * Authors: Ben Skeggs
     23 */
     24#include <core/gpuobj.h>
     25#include <core/engine.h>
     26
     27#include <subdev/instmem.h>
     28#include <subdev/bar.h>
     29#include <subdev/mmu.h>
     30
     31/* fast-path, where backend is able to provide direct pointer to memory */
     32static u32
     33nvkm_gpuobj_rd32_fast(struct nvkm_gpuobj *gpuobj, u32 offset)
     34{
     35	return ioread32_native(gpuobj->map + offset);
     36}
     37
     38static void
     39nvkm_gpuobj_wr32_fast(struct nvkm_gpuobj *gpuobj, u32 offset, u32 data)
     40{
     41	iowrite32_native(data, gpuobj->map + offset);
     42}
     43
     44/* accessor functions for gpuobjs allocated directly from instmem */
     45static int
     46nvkm_gpuobj_heap_map(struct nvkm_gpuobj *gpuobj, u64 offset,
     47		     struct nvkm_vmm *vmm, struct nvkm_vma *vma,
     48		     void *argv, u32 argc)
     49{
     50	return nvkm_memory_map(gpuobj->memory, offset, vmm, vma, argv, argc);
     51}
     52
     53static u32
     54nvkm_gpuobj_heap_rd32(struct nvkm_gpuobj *gpuobj, u32 offset)
     55{
     56	return nvkm_ro32(gpuobj->memory, offset);
     57}
     58
     59static void
     60nvkm_gpuobj_heap_wr32(struct nvkm_gpuobj *gpuobj, u32 offset, u32 data)
     61{
     62	nvkm_wo32(gpuobj->memory, offset, data);
     63}
     64
     65static const struct nvkm_gpuobj_func nvkm_gpuobj_heap;
     66static void
     67nvkm_gpuobj_heap_release(struct nvkm_gpuobj *gpuobj)
     68{
     69	gpuobj->func = &nvkm_gpuobj_heap;
     70	nvkm_done(gpuobj->memory);
     71}
     72
     73static const struct nvkm_gpuobj_func
     74nvkm_gpuobj_heap_fast = {
     75	.release = nvkm_gpuobj_heap_release,
     76	.rd32 = nvkm_gpuobj_rd32_fast,
     77	.wr32 = nvkm_gpuobj_wr32_fast,
     78	.map = nvkm_gpuobj_heap_map,
     79};
     80
     81static const struct nvkm_gpuobj_func
     82nvkm_gpuobj_heap_slow = {
     83	.release = nvkm_gpuobj_heap_release,
     84	.rd32 = nvkm_gpuobj_heap_rd32,
     85	.wr32 = nvkm_gpuobj_heap_wr32,
     86	.map = nvkm_gpuobj_heap_map,
     87};
     88
     89static void *
     90nvkm_gpuobj_heap_acquire(struct nvkm_gpuobj *gpuobj)
     91{
     92	gpuobj->map = nvkm_kmap(gpuobj->memory);
     93	if (likely(gpuobj->map))
     94		gpuobj->func = &nvkm_gpuobj_heap_fast;
     95	else
     96		gpuobj->func = &nvkm_gpuobj_heap_slow;
     97	return gpuobj->map;
     98}
     99
    100static const struct nvkm_gpuobj_func
    101nvkm_gpuobj_heap = {
    102	.acquire = nvkm_gpuobj_heap_acquire,
    103	.map = nvkm_gpuobj_heap_map,
    104};
    105
    106/* accessor functions for gpuobjs sub-allocated from a parent gpuobj */
    107static int
    108nvkm_gpuobj_map(struct nvkm_gpuobj *gpuobj, u64 offset,
    109		struct nvkm_vmm *vmm, struct nvkm_vma *vma,
    110		void *argv, u32 argc)
    111{
    112	return nvkm_memory_map(gpuobj->parent, gpuobj->node->offset + offset,
    113			       vmm, vma, argv, argc);
    114}
    115
    116static u32
    117nvkm_gpuobj_rd32(struct nvkm_gpuobj *gpuobj, u32 offset)
    118{
    119	return nvkm_ro32(gpuobj->parent, gpuobj->node->offset + offset);
    120}
    121
    122static void
    123nvkm_gpuobj_wr32(struct nvkm_gpuobj *gpuobj, u32 offset, u32 data)
    124{
    125	nvkm_wo32(gpuobj->parent, gpuobj->node->offset + offset, data);
    126}
    127
    128static const struct nvkm_gpuobj_func nvkm_gpuobj_func;
    129static void
    130nvkm_gpuobj_release(struct nvkm_gpuobj *gpuobj)
    131{
    132	gpuobj->func = &nvkm_gpuobj_func;
    133	nvkm_done(gpuobj->parent);
    134}
    135
    136static const struct nvkm_gpuobj_func
    137nvkm_gpuobj_fast = {
    138	.release = nvkm_gpuobj_release,
    139	.rd32 = nvkm_gpuobj_rd32_fast,
    140	.wr32 = nvkm_gpuobj_wr32_fast,
    141	.map = nvkm_gpuobj_map,
    142};
    143
    144static const struct nvkm_gpuobj_func
    145nvkm_gpuobj_slow = {
    146	.release = nvkm_gpuobj_release,
    147	.rd32 = nvkm_gpuobj_rd32,
    148	.wr32 = nvkm_gpuobj_wr32,
    149	.map = nvkm_gpuobj_map,
    150};
    151
    152static void *
    153nvkm_gpuobj_acquire(struct nvkm_gpuobj *gpuobj)
    154{
    155	gpuobj->map = nvkm_kmap(gpuobj->parent);
    156	if (likely(gpuobj->map)) {
    157		gpuobj->map  = (u8 *)gpuobj->map + gpuobj->node->offset;
    158		gpuobj->func = &nvkm_gpuobj_fast;
    159	} else {
    160		gpuobj->func = &nvkm_gpuobj_slow;
    161	}
    162	return gpuobj->map;
    163}
    164
    165static const struct nvkm_gpuobj_func
    166nvkm_gpuobj_func = {
    167	.acquire = nvkm_gpuobj_acquire,
    168	.map = nvkm_gpuobj_map,
    169};
    170
    171static int
    172nvkm_gpuobj_ctor(struct nvkm_device *device, u32 size, int align, bool zero,
    173		 struct nvkm_gpuobj *parent, struct nvkm_gpuobj *gpuobj)
    174{
    175	u32 offset;
    176	int ret;
    177
    178	if (parent) {
    179		if (align >= 0) {
    180			ret = nvkm_mm_head(&parent->heap, 0, 1, size, size,
    181					   max(align, 1), &gpuobj->node);
    182		} else {
    183			ret = nvkm_mm_tail(&parent->heap, 0, 1, size, size,
    184					   -align, &gpuobj->node);
    185		}
    186		if (ret)
    187			return ret;
    188
    189		gpuobj->parent = parent;
    190		gpuobj->func = &nvkm_gpuobj_func;
    191		gpuobj->addr = parent->addr + gpuobj->node->offset;
    192		gpuobj->size = gpuobj->node->length;
    193
    194		if (zero) {
    195			nvkm_kmap(gpuobj);
    196			for (offset = 0; offset < gpuobj->size; offset += 4)
    197				nvkm_wo32(gpuobj, offset, 0x00000000);
    198			nvkm_done(gpuobj);
    199		}
    200	} else {
    201		ret = nvkm_memory_new(device, NVKM_MEM_TARGET_INST, size,
    202				      abs(align), zero, &gpuobj->memory);
    203		if (ret)
    204			return ret;
    205
    206		gpuobj->func = &nvkm_gpuobj_heap;
    207		gpuobj->addr = nvkm_memory_addr(gpuobj->memory);
    208		gpuobj->size = nvkm_memory_size(gpuobj->memory);
    209	}
    210
    211	return nvkm_mm_init(&gpuobj->heap, 0, 0, gpuobj->size, 1);
    212}
    213
    214void
    215nvkm_gpuobj_del(struct nvkm_gpuobj **pgpuobj)
    216{
    217	struct nvkm_gpuobj *gpuobj = *pgpuobj;
    218	if (gpuobj) {
    219		if (gpuobj->parent)
    220			nvkm_mm_free(&gpuobj->parent->heap, &gpuobj->node);
    221		nvkm_mm_fini(&gpuobj->heap);
    222		nvkm_memory_unref(&gpuobj->memory);
    223		kfree(*pgpuobj);
    224		*pgpuobj = NULL;
    225	}
    226}
    227
    228int
    229nvkm_gpuobj_new(struct nvkm_device *device, u32 size, int align, bool zero,
    230		struct nvkm_gpuobj *parent, struct nvkm_gpuobj **pgpuobj)
    231{
    232	struct nvkm_gpuobj *gpuobj;
    233	int ret;
    234
    235	if (!(gpuobj = *pgpuobj = kzalloc(sizeof(*gpuobj), GFP_KERNEL)))
    236		return -ENOMEM;
    237
    238	ret = nvkm_gpuobj_ctor(device, size, align, zero, parent, gpuobj);
    239	if (ret)
    240		nvkm_gpuobj_del(pgpuobj);
    241	return ret;
    242}
    243
    244/* the below is basically only here to support sharing the paged dma object
    245 * for PCI(E)GART on <=nv4x chipsets, and should *not* be expected to work
    246 * anywhere else.
    247 */
    248
    249int
    250nvkm_gpuobj_wrap(struct nvkm_memory *memory, struct nvkm_gpuobj **pgpuobj)
    251{
    252	if (!(*pgpuobj = kzalloc(sizeof(**pgpuobj), GFP_KERNEL)))
    253		return -ENOMEM;
    254
    255	(*pgpuobj)->addr = nvkm_memory_addr(memory);
    256	(*pgpuobj)->size = nvkm_memory_size(memory);
    257	return 0;
    258}
    259
    260void
    261nvkm_gpuobj_memcpy_to(struct nvkm_gpuobj *dst, u32 dstoffset, void *src,
    262		      u32 length)
    263{
    264	int i;
    265
    266	for (i = 0; i < length; i += 4)
    267		nvkm_wo32(dst, dstoffset + i, *(u32 *)(src + i));
    268}
    269
    270void
    271nvkm_gpuobj_memcpy_from(void *dst, struct nvkm_gpuobj *src, u32 srcoffset,
    272			u32 length)
    273{
    274	int i;
    275
    276	for (i = 0; i < length; i += 4)
    277		((u32 *)src)[i / 4] = nvkm_ro32(src, srcoffset + i);
    278}