cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

xen_drm_front_kms.c (10225B)


      1// SPDX-License-Identifier: GPL-2.0 OR MIT
      2
      3/*
      4 *  Xen para-virtual DRM device
      5 *
      6 * Copyright (C) 2016-2018 EPAM Systems Inc.
      7 *
      8 * Author: Oleksandr Andrushchenko <oleksandr_andrushchenko@epam.com>
      9 */
     10
     11#include <drm/drm_atomic.h>
     12#include <drm/drm_atomic_helper.h>
     13#include <drm/drm_drv.h>
     14#include <drm/drm_fourcc.h>
     15#include <drm/drm_gem.h>
     16#include <drm/drm_gem_atomic_helper.h>
     17#include <drm/drm_gem_framebuffer_helper.h>
     18#include <drm/drm_probe_helper.h>
     19#include <drm/drm_vblank.h>
     20
     21#include "xen_drm_front.h"
     22#include "xen_drm_front_conn.h"
     23#include "xen_drm_front_kms.h"
     24
     25/*
     26 * Timeout in ms to wait for frame done event from the backend:
     27 * must be a bit more than IO time-out
     28 */
     29#define FRAME_DONE_TO_MS	(XEN_DRM_FRONT_WAIT_BACK_MS + 100)
     30
     31static struct xen_drm_front_drm_pipeline *
     32to_xen_drm_pipeline(struct drm_simple_display_pipe *pipe)
     33{
     34	return container_of(pipe, struct xen_drm_front_drm_pipeline, pipe);
     35}
     36
     37static void fb_destroy(struct drm_framebuffer *fb)
     38{
     39	struct xen_drm_front_drm_info *drm_info = fb->dev->dev_private;
     40	int idx;
     41
     42	if (drm_dev_enter(fb->dev, &idx)) {
     43		xen_drm_front_fb_detach(drm_info->front_info,
     44					xen_drm_front_fb_to_cookie(fb));
     45		drm_dev_exit(idx);
     46	}
     47	drm_gem_fb_destroy(fb);
     48}
     49
     50static const struct drm_framebuffer_funcs fb_funcs = {
     51	.destroy = fb_destroy,
     52};
     53
     54static struct drm_framebuffer *
     55fb_create(struct drm_device *dev, struct drm_file *filp,
     56	  const struct drm_mode_fb_cmd2 *mode_cmd)
     57{
     58	struct xen_drm_front_drm_info *drm_info = dev->dev_private;
     59	struct drm_framebuffer *fb;
     60	struct drm_gem_object *gem_obj;
     61	int ret;
     62
     63	fb = drm_gem_fb_create_with_funcs(dev, filp, mode_cmd, &fb_funcs);
     64	if (IS_ERR(fb))
     65		return fb;
     66
     67	gem_obj = fb->obj[0];
     68
     69	ret = xen_drm_front_fb_attach(drm_info->front_info,
     70				      xen_drm_front_dbuf_to_cookie(gem_obj),
     71				      xen_drm_front_fb_to_cookie(fb),
     72				      fb->width, fb->height,
     73				      fb->format->format);
     74	if (ret < 0) {
     75		DRM_ERROR("Back failed to attach FB %p: %d\n", fb, ret);
     76		goto fail;
     77	}
     78
     79	return fb;
     80
     81fail:
     82	drm_gem_fb_destroy(fb);
     83	return ERR_PTR(ret);
     84}
     85
     86static const struct drm_mode_config_funcs mode_config_funcs = {
     87	.fb_create = fb_create,
     88	.atomic_check = drm_atomic_helper_check,
     89	.atomic_commit = drm_atomic_helper_commit,
     90};
     91
     92static void send_pending_event(struct xen_drm_front_drm_pipeline *pipeline)
     93{
     94	struct drm_crtc *crtc = &pipeline->pipe.crtc;
     95	struct drm_device *dev = crtc->dev;
     96	unsigned long flags;
     97
     98	spin_lock_irqsave(&dev->event_lock, flags);
     99	if (pipeline->pending_event)
    100		drm_crtc_send_vblank_event(crtc, pipeline->pending_event);
    101	pipeline->pending_event = NULL;
    102	spin_unlock_irqrestore(&dev->event_lock, flags);
    103}
    104
    105static void display_enable(struct drm_simple_display_pipe *pipe,
    106			   struct drm_crtc_state *crtc_state,
    107			   struct drm_plane_state *plane_state)
    108{
    109	struct xen_drm_front_drm_pipeline *pipeline =
    110			to_xen_drm_pipeline(pipe);
    111	struct drm_crtc *crtc = &pipe->crtc;
    112	struct drm_framebuffer *fb = plane_state->fb;
    113	int ret, idx;
    114
    115	if (!drm_dev_enter(pipe->crtc.dev, &idx))
    116		return;
    117
    118	ret = xen_drm_front_mode_set(pipeline, crtc->x, crtc->y,
    119				     fb->width, fb->height,
    120				     fb->format->cpp[0] * 8,
    121				     xen_drm_front_fb_to_cookie(fb));
    122
    123	if (ret) {
    124		DRM_ERROR("Failed to enable display: %d\n", ret);
    125		pipeline->conn_connected = false;
    126	}
    127
    128	drm_dev_exit(idx);
    129}
    130
    131static void display_disable(struct drm_simple_display_pipe *pipe)
    132{
    133	struct xen_drm_front_drm_pipeline *pipeline =
    134			to_xen_drm_pipeline(pipe);
    135	int ret = 0, idx;
    136
    137	if (drm_dev_enter(pipe->crtc.dev, &idx)) {
    138		ret = xen_drm_front_mode_set(pipeline, 0, 0, 0, 0, 0,
    139					     xen_drm_front_fb_to_cookie(NULL));
    140		drm_dev_exit(idx);
    141	}
    142	if (ret)
    143		DRM_ERROR("Failed to disable display: %d\n", ret);
    144
    145	/* Make sure we can restart with enabled connector next time */
    146	pipeline->conn_connected = true;
    147
    148	/* release stalled event if any */
    149	send_pending_event(pipeline);
    150}
    151
    152void xen_drm_front_kms_on_frame_done(struct xen_drm_front_drm_pipeline *pipeline,
    153				     u64 fb_cookie)
    154{
    155	/*
    156	 * This runs in interrupt context, e.g. under
    157	 * drm_info->front_info->io_lock, so we cannot call _sync version
    158	 * to cancel the work
    159	 */
    160	cancel_delayed_work(&pipeline->pflip_to_worker);
    161
    162	send_pending_event(pipeline);
    163}
    164
    165static void pflip_to_worker(struct work_struct *work)
    166{
    167	struct delayed_work *delayed_work = to_delayed_work(work);
    168	struct xen_drm_front_drm_pipeline *pipeline =
    169			container_of(delayed_work,
    170				     struct xen_drm_front_drm_pipeline,
    171				     pflip_to_worker);
    172
    173	DRM_ERROR("Frame done timed-out, releasing");
    174	send_pending_event(pipeline);
    175}
    176
    177static bool display_send_page_flip(struct drm_simple_display_pipe *pipe,
    178				   struct drm_plane_state *old_plane_state)
    179{
    180	struct drm_plane_state *plane_state =
    181			drm_atomic_get_new_plane_state(old_plane_state->state,
    182						       &pipe->plane);
    183
    184	/*
    185	 * If old_plane_state->fb is NULL and plane_state->fb is not,
    186	 * then this is an atomic commit which will enable display.
    187	 * If old_plane_state->fb is not NULL and plane_state->fb is,
    188	 * then this is an atomic commit which will disable display.
    189	 * Ignore these and do not send page flip as this framebuffer will be
    190	 * sent to the backend as a part of display_set_config call.
    191	 */
    192	if (old_plane_state->fb && plane_state->fb) {
    193		struct xen_drm_front_drm_pipeline *pipeline =
    194				to_xen_drm_pipeline(pipe);
    195		struct xen_drm_front_drm_info *drm_info = pipeline->drm_info;
    196		int ret;
    197
    198		schedule_delayed_work(&pipeline->pflip_to_worker,
    199				      msecs_to_jiffies(FRAME_DONE_TO_MS));
    200
    201		ret = xen_drm_front_page_flip(drm_info->front_info,
    202					      pipeline->index,
    203					      xen_drm_front_fb_to_cookie(plane_state->fb));
    204		if (ret) {
    205			DRM_ERROR("Failed to send page flip request to backend: %d\n", ret);
    206
    207			pipeline->conn_connected = false;
    208			/*
    209			 * Report the flip not handled, so pending event is
    210			 * sent, unblocking user-space.
    211			 */
    212			return false;
    213		}
    214		/*
    215		 * Signal that page flip was handled, pending event will be sent
    216		 * on frame done event from the backend.
    217		 */
    218		return true;
    219	}
    220
    221	return false;
    222}
    223
    224static int display_check(struct drm_simple_display_pipe *pipe,
    225			 struct drm_plane_state *plane_state,
    226			 struct drm_crtc_state *crtc_state)
    227{
    228	/*
    229	 * Xen doesn't initialize vblanking via drm_vblank_init(), so
    230	 * DRM helpers assume that it doesn't handle vblanking and start
    231	 * sending out fake VBLANK events automatically.
    232	 *
    233	 * As xen contains it's own logic for sending out VBLANK events
    234	 * in send_pending_event(), disable no_vblank (i.e., the xen
    235	 * driver has vblanking support).
    236	 */
    237	crtc_state->no_vblank = false;
    238
    239	return 0;
    240}
    241
    242static void display_update(struct drm_simple_display_pipe *pipe,
    243			   struct drm_plane_state *old_plane_state)
    244{
    245	struct xen_drm_front_drm_pipeline *pipeline =
    246			to_xen_drm_pipeline(pipe);
    247	struct drm_crtc *crtc = &pipe->crtc;
    248	struct drm_pending_vblank_event *event;
    249	int idx;
    250
    251	event = crtc->state->event;
    252	if (event) {
    253		struct drm_device *dev = crtc->dev;
    254		unsigned long flags;
    255
    256		WARN_ON(pipeline->pending_event);
    257
    258		spin_lock_irqsave(&dev->event_lock, flags);
    259		crtc->state->event = NULL;
    260
    261		pipeline->pending_event = event;
    262		spin_unlock_irqrestore(&dev->event_lock, flags);
    263	}
    264
    265	if (!drm_dev_enter(pipe->crtc.dev, &idx)) {
    266		send_pending_event(pipeline);
    267		return;
    268	}
    269
    270	/*
    271	 * Send page flip request to the backend *after* we have event cached
    272	 * above, so on page flip done event from the backend we can
    273	 * deliver it and there is no race condition between this code and
    274	 * event from the backend.
    275	 * If this is not a page flip, e.g. no flip done event from the backend
    276	 * is expected, then send now.
    277	 */
    278	if (!display_send_page_flip(pipe, old_plane_state))
    279		send_pending_event(pipeline);
    280
    281	drm_dev_exit(idx);
    282}
    283
    284static enum drm_mode_status
    285display_mode_valid(struct drm_simple_display_pipe *pipe,
    286		   const struct drm_display_mode *mode)
    287{
    288	struct xen_drm_front_drm_pipeline *pipeline =
    289			container_of(pipe, struct xen_drm_front_drm_pipeline,
    290				     pipe);
    291
    292	if (mode->hdisplay != pipeline->width)
    293		return MODE_ERROR;
    294
    295	if (mode->vdisplay != pipeline->height)
    296		return MODE_ERROR;
    297
    298	return MODE_OK;
    299}
    300
    301static const struct drm_simple_display_pipe_funcs display_funcs = {
    302	.mode_valid = display_mode_valid,
    303	.enable = display_enable,
    304	.disable = display_disable,
    305	.check = display_check,
    306	.update = display_update,
    307};
    308
    309static int display_pipe_init(struct xen_drm_front_drm_info *drm_info,
    310			     int index, struct xen_drm_front_cfg_connector *cfg,
    311			     struct xen_drm_front_drm_pipeline *pipeline)
    312{
    313	struct drm_device *dev = drm_info->drm_dev;
    314	const u32 *formats;
    315	int format_count;
    316	int ret;
    317
    318	pipeline->drm_info = drm_info;
    319	pipeline->index = index;
    320	pipeline->height = cfg->height;
    321	pipeline->width = cfg->width;
    322
    323	INIT_DELAYED_WORK(&pipeline->pflip_to_worker, pflip_to_worker);
    324
    325	ret = xen_drm_front_conn_init(drm_info, &pipeline->conn);
    326	if (ret)
    327		return ret;
    328
    329	formats = xen_drm_front_conn_get_formats(&format_count);
    330
    331	return drm_simple_display_pipe_init(dev, &pipeline->pipe,
    332					    &display_funcs, formats,
    333					    format_count, NULL,
    334					    &pipeline->conn);
    335}
    336
    337int xen_drm_front_kms_init(struct xen_drm_front_drm_info *drm_info)
    338{
    339	struct drm_device *dev = drm_info->drm_dev;
    340	int i, ret;
    341
    342	drm_mode_config_init(dev);
    343
    344	dev->mode_config.min_width = 0;
    345	dev->mode_config.min_height = 0;
    346	dev->mode_config.max_width = 4095;
    347	dev->mode_config.max_height = 2047;
    348	dev->mode_config.funcs = &mode_config_funcs;
    349
    350	for (i = 0; i < drm_info->front_info->cfg.num_connectors; i++) {
    351		struct xen_drm_front_cfg_connector *cfg =
    352				&drm_info->front_info->cfg.connectors[i];
    353		struct xen_drm_front_drm_pipeline *pipeline =
    354				&drm_info->pipeline[i];
    355
    356		ret = display_pipe_init(drm_info, i, cfg, pipeline);
    357		if (ret) {
    358			drm_mode_config_cleanup(dev);
    359			return ret;
    360		}
    361	}
    362
    363	drm_mode_config_reset(dev);
    364	drm_kms_helper_poll_init(dev);
    365	return 0;
    366}
    367
    368void xen_drm_front_kms_fini(struct xen_drm_front_drm_info *drm_info)
    369{
    370	int i;
    371
    372	for (i = 0; i < drm_info->front_info->cfg.num_connectors; i++) {
    373		struct xen_drm_front_drm_pipeline *pipeline =
    374				&drm_info->pipeline[i];
    375
    376		cancel_delayed_work_sync(&pipeline->pflip_to_worker);
    377
    378		send_pending_event(pipeline);
    379	}
    380}