cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

xen-tpmfront.c (9654B)


      1// SPDX-License-Identifier: GPL-2.0-only
      2/*
      3 * Implementation of the Xen vTPM device frontend
      4 *
      5 * Author:  Daniel De Graaf <dgdegra@tycho.nsa.gov>
      6 */
      7#include <linux/errno.h>
      8#include <linux/err.h>
      9#include <linux/interrupt.h>
     10#include <linux/freezer.h>
     11#include <xen/xen.h>
     12#include <xen/events.h>
     13#include <xen/interface/io/tpmif.h>
     14#include <xen/grant_table.h>
     15#include <xen/xenbus.h>
     16#include <xen/page.h>
     17#include "tpm.h"
     18#include <xen/platform_pci.h>
     19
     20struct tpm_private {
     21	struct tpm_chip *chip;
     22	struct xenbus_device *dev;
     23
     24	struct vtpm_shared_page *shr;
     25
     26	unsigned int evtchn;
     27	int ring_ref;
     28	domid_t backend_id;
     29	int irq;
     30	wait_queue_head_t read_queue;
     31};
     32
     33enum status_bits {
     34	VTPM_STATUS_RUNNING  = 0x1,
     35	VTPM_STATUS_IDLE     = 0x2,
     36	VTPM_STATUS_RESULT   = 0x4,
     37	VTPM_STATUS_CANCELED = 0x8,
     38};
     39
     40static bool wait_for_tpm_stat_cond(struct tpm_chip *chip, u8 mask,
     41					bool check_cancel, bool *canceled)
     42{
     43	u8 status = chip->ops->status(chip);
     44
     45	*canceled = false;
     46	if ((status & mask) == mask)
     47		return true;
     48	if (check_cancel && chip->ops->req_canceled(chip, status)) {
     49		*canceled = true;
     50		return true;
     51	}
     52	return false;
     53}
     54
     55static int wait_for_tpm_stat(struct tpm_chip *chip, u8 mask,
     56		unsigned long timeout, wait_queue_head_t *queue,
     57		bool check_cancel)
     58{
     59	unsigned long stop;
     60	long rc;
     61	u8 status;
     62	bool canceled = false;
     63
     64	/* check current status */
     65	status = chip->ops->status(chip);
     66	if ((status & mask) == mask)
     67		return 0;
     68
     69	stop = jiffies + timeout;
     70
     71	if (chip->flags & TPM_CHIP_FLAG_IRQ) {
     72again:
     73		timeout = stop - jiffies;
     74		if ((long)timeout <= 0)
     75			return -ETIME;
     76		rc = wait_event_interruptible_timeout(*queue,
     77			wait_for_tpm_stat_cond(chip, mask, check_cancel,
     78					       &canceled),
     79			timeout);
     80		if (rc > 0) {
     81			if (canceled)
     82				return -ECANCELED;
     83			return 0;
     84		}
     85		if (rc == -ERESTARTSYS && freezing(current)) {
     86			clear_thread_flag(TIF_SIGPENDING);
     87			goto again;
     88		}
     89	} else {
     90		do {
     91			tpm_msleep(TPM_TIMEOUT);
     92			status = chip->ops->status(chip);
     93			if ((status & mask) == mask)
     94				return 0;
     95		} while (time_before(jiffies, stop));
     96	}
     97	return -ETIME;
     98}
     99
    100static u8 vtpm_status(struct tpm_chip *chip)
    101{
    102	struct tpm_private *priv = dev_get_drvdata(&chip->dev);
    103	switch (priv->shr->state) {
    104	case VTPM_STATE_IDLE:
    105		return VTPM_STATUS_IDLE | VTPM_STATUS_CANCELED;
    106	case VTPM_STATE_FINISH:
    107		return VTPM_STATUS_IDLE | VTPM_STATUS_RESULT;
    108	case VTPM_STATE_SUBMIT:
    109	case VTPM_STATE_CANCEL: /* cancel requested, not yet canceled */
    110		return VTPM_STATUS_RUNNING;
    111	default:
    112		return 0;
    113	}
    114}
    115
    116static bool vtpm_req_canceled(struct tpm_chip *chip, u8 status)
    117{
    118	return status & VTPM_STATUS_CANCELED;
    119}
    120
    121static void vtpm_cancel(struct tpm_chip *chip)
    122{
    123	struct tpm_private *priv = dev_get_drvdata(&chip->dev);
    124	priv->shr->state = VTPM_STATE_CANCEL;
    125	wmb();
    126	notify_remote_via_evtchn(priv->evtchn);
    127}
    128
    129static size_t shr_data_offset(struct vtpm_shared_page *shr)
    130{
    131	return struct_size(shr, extra_pages, shr->nr_extra_pages);
    132}
    133
    134static int vtpm_send(struct tpm_chip *chip, u8 *buf, size_t count)
    135{
    136	struct tpm_private *priv = dev_get_drvdata(&chip->dev);
    137	struct vtpm_shared_page *shr = priv->shr;
    138	size_t offset = shr_data_offset(shr);
    139
    140	u32 ordinal;
    141	unsigned long duration;
    142
    143	if (offset > PAGE_SIZE)
    144		return -EINVAL;
    145
    146	if (offset + count > PAGE_SIZE)
    147		return -EINVAL;
    148
    149	/* Wait for completion of any existing command or cancellation */
    150	if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, chip->timeout_c,
    151			&priv->read_queue, true) < 0) {
    152		vtpm_cancel(chip);
    153		return -ETIME;
    154	}
    155
    156	memcpy(offset + (u8 *)shr, buf, count);
    157	shr->length = count;
    158	barrier();
    159	shr->state = VTPM_STATE_SUBMIT;
    160	wmb();
    161	notify_remote_via_evtchn(priv->evtchn);
    162
    163	ordinal = be32_to_cpu(((struct tpm_header *)buf)->ordinal);
    164	duration = tpm_calc_ordinal_duration(chip, ordinal);
    165
    166	if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, duration,
    167			&priv->read_queue, true) < 0) {
    168		/* got a signal or timeout, try to cancel */
    169		vtpm_cancel(chip);
    170		return -ETIME;
    171	}
    172
    173	return 0;
    174}
    175
    176static int vtpm_recv(struct tpm_chip *chip, u8 *buf, size_t count)
    177{
    178	struct tpm_private *priv = dev_get_drvdata(&chip->dev);
    179	struct vtpm_shared_page *shr = priv->shr;
    180	size_t offset = shr_data_offset(shr);
    181	size_t length = shr->length;
    182
    183	if (shr->state == VTPM_STATE_IDLE)
    184		return -ECANCELED;
    185
    186	/* In theory the wait at the end of _send makes this one unnecessary */
    187	if (wait_for_tpm_stat(chip, VTPM_STATUS_RESULT, chip->timeout_c,
    188			&priv->read_queue, true) < 0) {
    189		vtpm_cancel(chip);
    190		return -ETIME;
    191	}
    192
    193	if (offset > PAGE_SIZE)
    194		return -EIO;
    195
    196	if (offset + length > PAGE_SIZE)
    197		length = PAGE_SIZE - offset;
    198
    199	if (length > count)
    200		length = count;
    201
    202	memcpy(buf, offset + (u8 *)shr, length);
    203
    204	return length;
    205}
    206
    207static const struct tpm_class_ops tpm_vtpm = {
    208	.status = vtpm_status,
    209	.recv = vtpm_recv,
    210	.send = vtpm_send,
    211	.cancel = vtpm_cancel,
    212	.req_complete_mask = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
    213	.req_complete_val  = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
    214	.req_canceled      = vtpm_req_canceled,
    215};
    216
    217static irqreturn_t tpmif_interrupt(int dummy, void *dev_id)
    218{
    219	struct tpm_private *priv = dev_id;
    220
    221	switch (priv->shr->state) {
    222	case VTPM_STATE_IDLE:
    223	case VTPM_STATE_FINISH:
    224		wake_up_interruptible(&priv->read_queue);
    225		break;
    226	case VTPM_STATE_SUBMIT:
    227	case VTPM_STATE_CANCEL:
    228	default:
    229		break;
    230	}
    231	return IRQ_HANDLED;
    232}
    233
    234static int setup_chip(struct device *dev, struct tpm_private *priv)
    235{
    236	struct tpm_chip *chip;
    237
    238	chip = tpmm_chip_alloc(dev, &tpm_vtpm);
    239	if (IS_ERR(chip))
    240		return PTR_ERR(chip);
    241
    242	init_waitqueue_head(&priv->read_queue);
    243
    244	priv->chip = chip;
    245	dev_set_drvdata(&chip->dev, priv);
    246
    247	return 0;
    248}
    249
    250/* caller must clean up in case of errors */
    251static int setup_ring(struct xenbus_device *dev, struct tpm_private *priv)
    252{
    253	struct xenbus_transaction xbt;
    254	const char *message = NULL;
    255	int rv;
    256
    257	rv = xenbus_setup_ring(dev, GFP_KERNEL, (void **)&priv->shr, 1,
    258			       &priv->ring_ref);
    259	if (rv < 0)
    260		return rv;
    261
    262	rv = xenbus_alloc_evtchn(dev, &priv->evtchn);
    263	if (rv)
    264		return rv;
    265
    266	rv = bind_evtchn_to_irqhandler(priv->evtchn, tpmif_interrupt, 0,
    267				       "tpmif", priv);
    268	if (rv <= 0) {
    269		xenbus_dev_fatal(dev, rv, "allocating TPM irq");
    270		return rv;
    271	}
    272	priv->irq = rv;
    273
    274 again:
    275	rv = xenbus_transaction_start(&xbt);
    276	if (rv) {
    277		xenbus_dev_fatal(dev, rv, "starting transaction");
    278		return rv;
    279	}
    280
    281	rv = xenbus_printf(xbt, dev->nodename,
    282			"ring-ref", "%u", priv->ring_ref);
    283	if (rv) {
    284		message = "writing ring-ref";
    285		goto abort_transaction;
    286	}
    287
    288	rv = xenbus_printf(xbt, dev->nodename, "event-channel", "%u",
    289			priv->evtchn);
    290	if (rv) {
    291		message = "writing event-channel";
    292		goto abort_transaction;
    293	}
    294
    295	rv = xenbus_printf(xbt, dev->nodename, "feature-protocol-v2", "1");
    296	if (rv) {
    297		message = "writing feature-protocol-v2";
    298		goto abort_transaction;
    299	}
    300
    301	rv = xenbus_transaction_end(xbt, 0);
    302	if (rv == -EAGAIN)
    303		goto again;
    304	if (rv) {
    305		xenbus_dev_fatal(dev, rv, "completing transaction");
    306		return rv;
    307	}
    308
    309	xenbus_switch_state(dev, XenbusStateInitialised);
    310
    311	return 0;
    312
    313 abort_transaction:
    314	xenbus_transaction_end(xbt, 1);
    315	if (message)
    316		xenbus_dev_error(dev, rv, "%s", message);
    317
    318	return rv;
    319}
    320
    321static void ring_free(struct tpm_private *priv)
    322{
    323	if (!priv)
    324		return;
    325
    326	xenbus_teardown_ring((void **)&priv->shr, 1, &priv->ring_ref);
    327
    328	if (priv->irq)
    329		unbind_from_irqhandler(priv->irq, priv);
    330
    331	kfree(priv);
    332}
    333
    334static int tpmfront_probe(struct xenbus_device *dev,
    335		const struct xenbus_device_id *id)
    336{
    337	struct tpm_private *priv;
    338	int rv;
    339
    340	priv = kzalloc(sizeof(*priv), GFP_KERNEL);
    341	if (!priv) {
    342		xenbus_dev_fatal(dev, -ENOMEM, "allocating priv structure");
    343		return -ENOMEM;
    344	}
    345
    346	rv = setup_chip(&dev->dev, priv);
    347	if (rv) {
    348		kfree(priv);
    349		return rv;
    350	}
    351
    352	rv = setup_ring(dev, priv);
    353	if (rv) {
    354		ring_free(priv);
    355		return rv;
    356	}
    357
    358	tpm_get_timeouts(priv->chip);
    359
    360	return tpm_chip_register(priv->chip);
    361}
    362
    363static int tpmfront_remove(struct xenbus_device *dev)
    364{
    365	struct tpm_chip *chip = dev_get_drvdata(&dev->dev);
    366	struct tpm_private *priv = dev_get_drvdata(&chip->dev);
    367	tpm_chip_unregister(chip);
    368	ring_free(priv);
    369	dev_set_drvdata(&chip->dev, NULL);
    370	return 0;
    371}
    372
    373static int tpmfront_resume(struct xenbus_device *dev)
    374{
    375	/* A suspend/resume/migrate will interrupt a vTPM anyway */
    376	tpmfront_remove(dev);
    377	return tpmfront_probe(dev, NULL);
    378}
    379
    380static void backend_changed(struct xenbus_device *dev,
    381		enum xenbus_state backend_state)
    382{
    383	switch (backend_state) {
    384	case XenbusStateInitialised:
    385	case XenbusStateConnected:
    386		if (dev->state == XenbusStateConnected)
    387			break;
    388
    389		if (!xenbus_read_unsigned(dev->otherend, "feature-protocol-v2",
    390					  0)) {
    391			xenbus_dev_fatal(dev, -EINVAL,
    392					"vTPM protocol 2 required");
    393			return;
    394		}
    395		xenbus_switch_state(dev, XenbusStateConnected);
    396		break;
    397
    398	case XenbusStateClosing:
    399	case XenbusStateClosed:
    400		device_unregister(&dev->dev);
    401		xenbus_frontend_closed(dev);
    402		break;
    403	default:
    404		break;
    405	}
    406}
    407
    408static const struct xenbus_device_id tpmfront_ids[] = {
    409	{ "vtpm" },
    410	{ "" }
    411};
    412MODULE_ALIAS("xen:vtpm");
    413
    414static struct xenbus_driver tpmfront_driver = {
    415	.ids = tpmfront_ids,
    416	.probe = tpmfront_probe,
    417	.remove = tpmfront_remove,
    418	.resume = tpmfront_resume,
    419	.otherend_changed = backend_changed,
    420};
    421
    422static int __init xen_tpmfront_init(void)
    423{
    424	if (!xen_domain())
    425		return -ENODEV;
    426
    427	if (!xen_has_pv_devices())
    428		return -ENODEV;
    429
    430	return xenbus_register_frontend(&tpmfront_driver);
    431}
    432module_init(xen_tpmfront_init);
    433
    434static void __exit xen_tpmfront_exit(void)
    435{
    436	xenbus_unregister_driver(&tpmfront_driver);
    437}
    438module_exit(xen_tpmfront_exit);
    439
    440MODULE_AUTHOR("Daniel De Graaf <dgdegra@tycho.nsa.gov>");
    441MODULE_DESCRIPTION("Xen vTPM Driver");
    442MODULE_LICENSE("GPL");