cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

vmwgfx_gmrid_manager.c (6473B)


      1// SPDX-License-Identifier: GPL-2.0 OR MIT
      2/**************************************************************************
      3 *
      4 * Copyright 2007-2010 VMware, Inc., Palo Alto, CA., USA
      5 *
      6 * Permission is hereby granted, free of charge, to any person obtaining a
      7 * copy of this software and associated documentation files (the
      8 * "Software"), to deal in the Software without restriction, including
      9 * without limitation the rights to use, copy, modify, merge, publish,
     10 * distribute, sub license, and/or sell copies of the Software, and to
     11 * permit persons to whom the Software is furnished to do so, subject to
     12 * the following conditions:
     13 *
     14 * The above copyright notice and this permission notice (including the
     15 * next paragraph) shall be included in all copies or substantial portions
     16 * of the Software.
     17 *
     18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     20 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
     21 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
     22 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
     23 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
     24 * USE OR OTHER DEALINGS IN THE SOFTWARE.
     25 *
     26 **************************************************************************/
     27/*
     28 * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com>
     29 */
     30
     31#include "vmwgfx_drv.h"
     32#include <drm/ttm/ttm_bo_driver.h>
     33#include <drm/ttm/ttm_placement.h>
     34#include <linux/idr.h>
     35#include <linux/spinlock.h>
     36#include <linux/kernel.h>
     37
     38struct vmwgfx_gmrid_man {
     39	struct ttm_resource_manager manager;
     40	spinlock_t lock;
     41	struct ida gmr_ida;
     42	uint32_t max_gmr_ids;
     43	uint32_t max_gmr_pages;
     44	uint32_t used_gmr_pages;
     45	uint8_t type;
     46};
     47
     48static struct vmwgfx_gmrid_man *to_gmrid_manager(struct ttm_resource_manager *man)
     49{
     50	return container_of(man, struct vmwgfx_gmrid_man, manager);
     51}
     52
     53static int vmw_gmrid_man_get_node(struct ttm_resource_manager *man,
     54				  struct ttm_buffer_object *bo,
     55				  const struct ttm_place *place,
     56				  struct ttm_resource **res)
     57{
     58	struct vmwgfx_gmrid_man *gman = to_gmrid_manager(man);
     59	int id;
     60
     61	*res = kmalloc(sizeof(**res), GFP_KERNEL);
     62	if (!*res)
     63		return -ENOMEM;
     64
     65	ttm_resource_init(bo, place, *res);
     66
     67	id = ida_alloc_max(&gman->gmr_ida, gman->max_gmr_ids - 1, GFP_KERNEL);
     68	if (id < 0)
     69		return id;
     70
     71	spin_lock(&gman->lock);
     72
     73	if (gman->max_gmr_pages > 0) {
     74		gman->used_gmr_pages += (*res)->num_pages;
     75		/*
     76		 * Because the graphics memory is a soft limit we can try to
     77		 * expand it instead of letting the userspace apps crash.
     78		 * We're just going to have a sane limit (half of RAM)
     79		 * on the number of MOB's that we create and will try to keep
     80		 * the system running until we reach that.
     81		 */
     82		if (unlikely(gman->used_gmr_pages > gman->max_gmr_pages)) {
     83			const unsigned long max_graphics_pages = totalram_pages() / 2;
     84			uint32_t new_max_pages = 0;
     85
     86			DRM_WARN("vmwgfx: mob memory overflow. Consider increasing guest RAM and graphicsMemory.\n");
     87			vmw_host_printf("vmwgfx, warning: mob memory overflow. Consider increasing guest RAM and graphicsMemory.\n");
     88
     89			if (gman->max_gmr_pages > (max_graphics_pages / 2)) {
     90				DRM_WARN("vmwgfx: guest requires more than half of RAM for graphics.\n");
     91				new_max_pages = max_graphics_pages;
     92			} else
     93				new_max_pages = gman->max_gmr_pages * 2;
     94			if (new_max_pages > gman->max_gmr_pages && new_max_pages >= gman->used_gmr_pages) {
     95				DRM_WARN("vmwgfx: increasing guest mob limits to %u kB.\n",
     96					 ((new_max_pages) << (PAGE_SHIFT - 10)));
     97
     98				gman->max_gmr_pages = new_max_pages;
     99			} else {
    100				char buf[256];
    101				snprintf(buf, sizeof(buf),
    102					 "vmwgfx, error: guest graphics is out of memory (mob limit at: %ukB).\n",
    103					 ((gman->max_gmr_pages) << (PAGE_SHIFT - 10)));
    104				vmw_host_printf(buf);
    105				DRM_WARN("%s", buf);
    106				goto nospace;
    107			}
    108		}
    109	}
    110
    111	(*res)->start = id;
    112
    113	spin_unlock(&gman->lock);
    114	return 0;
    115
    116nospace:
    117	gman->used_gmr_pages -= (*res)->num_pages;
    118	spin_unlock(&gman->lock);
    119	ida_free(&gman->gmr_ida, id);
    120	ttm_resource_fini(man, *res);
    121	kfree(*res);
    122	return -ENOSPC;
    123}
    124
    125static void vmw_gmrid_man_put_node(struct ttm_resource_manager *man,
    126				   struct ttm_resource *res)
    127{
    128	struct vmwgfx_gmrid_man *gman = to_gmrid_manager(man);
    129
    130	ida_free(&gman->gmr_ida, res->start);
    131	spin_lock(&gman->lock);
    132	gman->used_gmr_pages -= res->num_pages;
    133	spin_unlock(&gman->lock);
    134	ttm_resource_fini(man, res);
    135	kfree(res);
    136}
    137
    138static void vmw_gmrid_man_debug(struct ttm_resource_manager *man,
    139				struct drm_printer *printer)
    140{
    141	struct vmwgfx_gmrid_man *gman = to_gmrid_manager(man);
    142
    143	BUG_ON(gman->type != VMW_PL_GMR && gman->type != VMW_PL_MOB);
    144
    145	drm_printf(printer, "%s's used: %u pages, max: %u pages, %u id's\n",
    146		   (gman->type == VMW_PL_MOB) ? "Mob" : "GMR",
    147		   gman->used_gmr_pages, gman->max_gmr_pages, gman->max_gmr_ids);
    148}
    149
    150static const struct ttm_resource_manager_func vmw_gmrid_manager_func;
    151
    152int vmw_gmrid_man_init(struct vmw_private *dev_priv, int type)
    153{
    154	struct ttm_resource_manager *man;
    155	struct vmwgfx_gmrid_man *gman =
    156		kzalloc(sizeof(*gman), GFP_KERNEL);
    157
    158	if (unlikely(!gman))
    159		return -ENOMEM;
    160
    161	man = &gman->manager;
    162
    163	man->func = &vmw_gmrid_manager_func;
    164	man->use_tt = true;
    165	ttm_resource_manager_init(man, &dev_priv->bdev, 0);
    166	spin_lock_init(&gman->lock);
    167	gman->used_gmr_pages = 0;
    168	ida_init(&gman->gmr_ida);
    169	gman->type = type;
    170
    171	switch (type) {
    172	case VMW_PL_GMR:
    173		gman->max_gmr_ids = dev_priv->max_gmr_ids;
    174		gman->max_gmr_pages = dev_priv->max_gmr_pages;
    175		break;
    176	case VMW_PL_MOB:
    177		gman->max_gmr_ids = VMWGFX_NUM_MOB;
    178		gman->max_gmr_pages = dev_priv->max_mob_pages;
    179		break;
    180	default:
    181		BUG();
    182	}
    183	ttm_set_driver_manager(&dev_priv->bdev, type, &gman->manager);
    184	ttm_resource_manager_set_used(man, true);
    185	return 0;
    186}
    187
    188void vmw_gmrid_man_fini(struct vmw_private *dev_priv, int type)
    189{
    190	struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, type);
    191	struct vmwgfx_gmrid_man *gman = to_gmrid_manager(man);
    192
    193	ttm_resource_manager_set_used(man, false);
    194
    195	ttm_resource_manager_evict_all(&dev_priv->bdev, man);
    196
    197	ttm_resource_manager_cleanup(man);
    198
    199	ttm_set_driver_manager(&dev_priv->bdev, type, NULL);
    200	ida_destroy(&gman->gmr_ida);
    201	kfree(gman);
    202
    203}
    204
    205static const struct ttm_resource_manager_func vmw_gmrid_manager_func = {
    206	.alloc = vmw_gmrid_man_get_node,
    207	.free = vmw_gmrid_man_put_node,
    208	.debug = vmw_gmrid_man_debug
    209};