cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

sev-guest.c (20034B)


      1// SPDX-License-Identifier: GPL-2.0-only
      2/*
      3 * AMD Secure Encrypted Virtualization (SEV) guest driver interface
      4 *
      5 * Copyright (C) 2021 Advanced Micro Devices, Inc.
      6 *
      7 * Author: Brijesh Singh <brijesh.singh@amd.com>
      8 */
      9
     10#include <linux/module.h>
     11#include <linux/kernel.h>
     12#include <linux/types.h>
     13#include <linux/mutex.h>
     14#include <linux/io.h>
     15#include <linux/platform_device.h>
     16#include <linux/miscdevice.h>
     17#include <linux/set_memory.h>
     18#include <linux/fs.h>
     19#include <crypto/aead.h>
     20#include <linux/scatterlist.h>
     21#include <linux/psp-sev.h>
     22#include <uapi/linux/sev-guest.h>
     23#include <uapi/linux/psp-sev.h>
     24
     25#include <asm/svm.h>
     26#include <asm/sev.h>
     27
     28#include "sev-guest.h"
     29
     30#define DEVICE_NAME	"sev-guest"
     31#define AAD_LEN		48
     32#define MSG_HDR_VER	1
     33
     34struct snp_guest_crypto {
     35	struct crypto_aead *tfm;
     36	u8 *iv, *authtag;
     37	int iv_len, a_len;
     38};
     39
     40struct snp_guest_dev {
     41	struct device *dev;
     42	struct miscdevice misc;
     43
     44	void *certs_data;
     45	struct snp_guest_crypto *crypto;
     46	struct snp_guest_msg *request, *response;
     47	struct snp_secrets_page_layout *layout;
     48	struct snp_req_data input;
     49	u32 *os_area_msg_seqno;
     50	u8 *vmpck;
     51};
     52
     53static u32 vmpck_id;
     54module_param(vmpck_id, uint, 0444);
     55MODULE_PARM_DESC(vmpck_id, "The VMPCK ID to use when communicating with the PSP.");
     56
     57/* Mutex to serialize the shared buffer access and command handling. */
     58static DEFINE_MUTEX(snp_cmd_mutex);
     59
     60static bool is_vmpck_empty(struct snp_guest_dev *snp_dev)
     61{
     62	char zero_key[VMPCK_KEY_LEN] = {0};
     63
     64	if (snp_dev->vmpck)
     65		return !memcmp(snp_dev->vmpck, zero_key, VMPCK_KEY_LEN);
     66
     67	return true;
     68}
     69
     70static void snp_disable_vmpck(struct snp_guest_dev *snp_dev)
     71{
     72	memzero_explicit(snp_dev->vmpck, VMPCK_KEY_LEN);
     73	snp_dev->vmpck = NULL;
     74}
     75
     76static inline u64 __snp_get_msg_seqno(struct snp_guest_dev *snp_dev)
     77{
     78	u64 count;
     79
     80	lockdep_assert_held(&snp_cmd_mutex);
     81
     82	/* Read the current message sequence counter from secrets pages */
     83	count = *snp_dev->os_area_msg_seqno;
     84
     85	return count + 1;
     86}
     87
     88/* Return a non-zero on success */
     89static u64 snp_get_msg_seqno(struct snp_guest_dev *snp_dev)
     90{
     91	u64 count = __snp_get_msg_seqno(snp_dev);
     92
     93	/*
     94	 * The message sequence counter for the SNP guest request is a  64-bit
     95	 * value but the version 2 of GHCB specification defines a 32-bit storage
     96	 * for it. If the counter exceeds the 32-bit value then return zero.
     97	 * The caller should check the return value, but if the caller happens to
     98	 * not check the value and use it, then the firmware treats zero as an
     99	 * invalid number and will fail the  message request.
    100	 */
    101	if (count >= UINT_MAX) {
    102		dev_err(snp_dev->dev, "request message sequence counter overflow\n");
    103		return 0;
    104	}
    105
    106	return count;
    107}
    108
    109static void snp_inc_msg_seqno(struct snp_guest_dev *snp_dev)
    110{
    111	/*
    112	 * The counter is also incremented by the PSP, so increment it by 2
    113	 * and save in secrets page.
    114	 */
    115	*snp_dev->os_area_msg_seqno += 2;
    116}
    117
    118static inline struct snp_guest_dev *to_snp_dev(struct file *file)
    119{
    120	struct miscdevice *dev = file->private_data;
    121
    122	return container_of(dev, struct snp_guest_dev, misc);
    123}
    124
    125static struct snp_guest_crypto *init_crypto(struct snp_guest_dev *snp_dev, u8 *key, size_t keylen)
    126{
    127	struct snp_guest_crypto *crypto;
    128
    129	crypto = kzalloc(sizeof(*crypto), GFP_KERNEL_ACCOUNT);
    130	if (!crypto)
    131		return NULL;
    132
    133	crypto->tfm = crypto_alloc_aead("gcm(aes)", 0, 0);
    134	if (IS_ERR(crypto->tfm))
    135		goto e_free;
    136
    137	if (crypto_aead_setkey(crypto->tfm, key, keylen))
    138		goto e_free_crypto;
    139
    140	crypto->iv_len = crypto_aead_ivsize(crypto->tfm);
    141	crypto->iv = kmalloc(crypto->iv_len, GFP_KERNEL_ACCOUNT);
    142	if (!crypto->iv)
    143		goto e_free_crypto;
    144
    145	if (crypto_aead_authsize(crypto->tfm) > MAX_AUTHTAG_LEN) {
    146		if (crypto_aead_setauthsize(crypto->tfm, MAX_AUTHTAG_LEN)) {
    147			dev_err(snp_dev->dev, "failed to set authsize to %d\n", MAX_AUTHTAG_LEN);
    148			goto e_free_iv;
    149		}
    150	}
    151
    152	crypto->a_len = crypto_aead_authsize(crypto->tfm);
    153	crypto->authtag = kmalloc(crypto->a_len, GFP_KERNEL_ACCOUNT);
    154	if (!crypto->authtag)
    155		goto e_free_auth;
    156
    157	return crypto;
    158
    159e_free_auth:
    160	kfree(crypto->authtag);
    161e_free_iv:
    162	kfree(crypto->iv);
    163e_free_crypto:
    164	crypto_free_aead(crypto->tfm);
    165e_free:
    166	kfree(crypto);
    167
    168	return NULL;
    169}
    170
    171static void deinit_crypto(struct snp_guest_crypto *crypto)
    172{
    173	crypto_free_aead(crypto->tfm);
    174	kfree(crypto->iv);
    175	kfree(crypto->authtag);
    176	kfree(crypto);
    177}
    178
    179static int enc_dec_message(struct snp_guest_crypto *crypto, struct snp_guest_msg *msg,
    180			   u8 *src_buf, u8 *dst_buf, size_t len, bool enc)
    181{
    182	struct snp_guest_msg_hdr *hdr = &msg->hdr;
    183	struct scatterlist src[3], dst[3];
    184	DECLARE_CRYPTO_WAIT(wait);
    185	struct aead_request *req;
    186	int ret;
    187
    188	req = aead_request_alloc(crypto->tfm, GFP_KERNEL);
    189	if (!req)
    190		return -ENOMEM;
    191
    192	/*
    193	 * AEAD memory operations:
    194	 * +------ AAD -------+------- DATA -----+---- AUTHTAG----+
    195	 * |  msg header      |  plaintext       |  hdr->authtag  |
    196	 * | bytes 30h - 5Fh  |    or            |                |
    197	 * |                  |   cipher         |                |
    198	 * +------------------+------------------+----------------+
    199	 */
    200	sg_init_table(src, 3);
    201	sg_set_buf(&src[0], &hdr->algo, AAD_LEN);
    202	sg_set_buf(&src[1], src_buf, hdr->msg_sz);
    203	sg_set_buf(&src[2], hdr->authtag, crypto->a_len);
    204
    205	sg_init_table(dst, 3);
    206	sg_set_buf(&dst[0], &hdr->algo, AAD_LEN);
    207	sg_set_buf(&dst[1], dst_buf, hdr->msg_sz);
    208	sg_set_buf(&dst[2], hdr->authtag, crypto->a_len);
    209
    210	aead_request_set_ad(req, AAD_LEN);
    211	aead_request_set_tfm(req, crypto->tfm);
    212	aead_request_set_callback(req, 0, crypto_req_done, &wait);
    213
    214	aead_request_set_crypt(req, src, dst, len, crypto->iv);
    215	ret = crypto_wait_req(enc ? crypto_aead_encrypt(req) : crypto_aead_decrypt(req), &wait);
    216
    217	aead_request_free(req);
    218	return ret;
    219}
    220
    221static int __enc_payload(struct snp_guest_dev *snp_dev, struct snp_guest_msg *msg,
    222			 void *plaintext, size_t len)
    223{
    224	struct snp_guest_crypto *crypto = snp_dev->crypto;
    225	struct snp_guest_msg_hdr *hdr = &msg->hdr;
    226
    227	memset(crypto->iv, 0, crypto->iv_len);
    228	memcpy(crypto->iv, &hdr->msg_seqno, sizeof(hdr->msg_seqno));
    229
    230	return enc_dec_message(crypto, msg, plaintext, msg->payload, len, true);
    231}
    232
    233static int dec_payload(struct snp_guest_dev *snp_dev, struct snp_guest_msg *msg,
    234		       void *plaintext, size_t len)
    235{
    236	struct snp_guest_crypto *crypto = snp_dev->crypto;
    237	struct snp_guest_msg_hdr *hdr = &msg->hdr;
    238
    239	/* Build IV with response buffer sequence number */
    240	memset(crypto->iv, 0, crypto->iv_len);
    241	memcpy(crypto->iv, &hdr->msg_seqno, sizeof(hdr->msg_seqno));
    242
    243	return enc_dec_message(crypto, msg, msg->payload, plaintext, len, false);
    244}
    245
    246static int verify_and_dec_payload(struct snp_guest_dev *snp_dev, void *payload, u32 sz)
    247{
    248	struct snp_guest_crypto *crypto = snp_dev->crypto;
    249	struct snp_guest_msg *resp = snp_dev->response;
    250	struct snp_guest_msg *req = snp_dev->request;
    251	struct snp_guest_msg_hdr *req_hdr = &req->hdr;
    252	struct snp_guest_msg_hdr *resp_hdr = &resp->hdr;
    253
    254	dev_dbg(snp_dev->dev, "response [seqno %lld type %d version %d sz %d]\n",
    255		resp_hdr->msg_seqno, resp_hdr->msg_type, resp_hdr->msg_version, resp_hdr->msg_sz);
    256
    257	/* Verify that the sequence counter is incremented by 1 */
    258	if (unlikely(resp_hdr->msg_seqno != (req_hdr->msg_seqno + 1)))
    259		return -EBADMSG;
    260
    261	/* Verify response message type and version number. */
    262	if (resp_hdr->msg_type != (req_hdr->msg_type + 1) ||
    263	    resp_hdr->msg_version != req_hdr->msg_version)
    264		return -EBADMSG;
    265
    266	/*
    267	 * If the message size is greater than our buffer length then return
    268	 * an error.
    269	 */
    270	if (unlikely((resp_hdr->msg_sz + crypto->a_len) > sz))
    271		return -EBADMSG;
    272
    273	/* Decrypt the payload */
    274	return dec_payload(snp_dev, resp, payload, resp_hdr->msg_sz + crypto->a_len);
    275}
    276
    277static int enc_payload(struct snp_guest_dev *snp_dev, u64 seqno, int version, u8 type,
    278			void *payload, size_t sz)
    279{
    280	struct snp_guest_msg *req = snp_dev->request;
    281	struct snp_guest_msg_hdr *hdr = &req->hdr;
    282
    283	memset(req, 0, sizeof(*req));
    284
    285	hdr->algo = SNP_AEAD_AES_256_GCM;
    286	hdr->hdr_version = MSG_HDR_VER;
    287	hdr->hdr_sz = sizeof(*hdr);
    288	hdr->msg_type = type;
    289	hdr->msg_version = version;
    290	hdr->msg_seqno = seqno;
    291	hdr->msg_vmpck = vmpck_id;
    292	hdr->msg_sz = sz;
    293
    294	/* Verify the sequence number is non-zero */
    295	if (!hdr->msg_seqno)
    296		return -ENOSR;
    297
    298	dev_dbg(snp_dev->dev, "request [seqno %lld type %d version %d sz %d]\n",
    299		hdr->msg_seqno, hdr->msg_type, hdr->msg_version, hdr->msg_sz);
    300
    301	return __enc_payload(snp_dev, req, payload, sz);
    302}
    303
    304static int handle_guest_request(struct snp_guest_dev *snp_dev, u64 exit_code, int msg_ver,
    305				u8 type, void *req_buf, size_t req_sz, void *resp_buf,
    306				u32 resp_sz, __u64 *fw_err)
    307{
    308	unsigned long err;
    309	u64 seqno;
    310	int rc;
    311
    312	/* Get message sequence and verify that its a non-zero */
    313	seqno = snp_get_msg_seqno(snp_dev);
    314	if (!seqno)
    315		return -EIO;
    316
    317	memset(snp_dev->response, 0, sizeof(struct snp_guest_msg));
    318
    319	/* Encrypt the userspace provided payload */
    320	rc = enc_payload(snp_dev, seqno, msg_ver, type, req_buf, req_sz);
    321	if (rc)
    322		return rc;
    323
    324	/* Call firmware to process the request */
    325	rc = snp_issue_guest_request(exit_code, &snp_dev->input, &err);
    326	if (fw_err)
    327		*fw_err = err;
    328
    329	if (rc)
    330		return rc;
    331
    332	/*
    333	 * The verify_and_dec_payload() will fail only if the hypervisor is
    334	 * actively modifying the message header or corrupting the encrypted payload.
    335	 * This hints that hypervisor is acting in a bad faith. Disable the VMPCK so that
    336	 * the key cannot be used for any communication. The key is disabled to ensure
    337	 * that AES-GCM does not use the same IV while encrypting the request payload.
    338	 */
    339	rc = verify_and_dec_payload(snp_dev, resp_buf, resp_sz);
    340	if (rc) {
    341		dev_alert(snp_dev->dev,
    342			  "Detected unexpected decode failure, disabling the vmpck_id %d\n",
    343			  vmpck_id);
    344		snp_disable_vmpck(snp_dev);
    345		return rc;
    346	}
    347
    348	/* Increment to new message sequence after payload decryption was successful. */
    349	snp_inc_msg_seqno(snp_dev);
    350
    351	return 0;
    352}
    353
    354static int get_report(struct snp_guest_dev *snp_dev, struct snp_guest_request_ioctl *arg)
    355{
    356	struct snp_guest_crypto *crypto = snp_dev->crypto;
    357	struct snp_report_resp *resp;
    358	struct snp_report_req req;
    359	int rc, resp_len;
    360
    361	lockdep_assert_held(&snp_cmd_mutex);
    362
    363	if (!arg->req_data || !arg->resp_data)
    364		return -EINVAL;
    365
    366	if (copy_from_user(&req, (void __user *)arg->req_data, sizeof(req)))
    367		return -EFAULT;
    368
    369	/*
    370	 * The intermediate response buffer is used while decrypting the
    371	 * response payload. Make sure that it has enough space to cover the
    372	 * authtag.
    373	 */
    374	resp_len = sizeof(resp->data) + crypto->a_len;
    375	resp = kzalloc(resp_len, GFP_KERNEL_ACCOUNT);
    376	if (!resp)
    377		return -ENOMEM;
    378
    379	rc = handle_guest_request(snp_dev, SVM_VMGEXIT_GUEST_REQUEST, arg->msg_version,
    380				  SNP_MSG_REPORT_REQ, &req, sizeof(req), resp->data,
    381				  resp_len, &arg->fw_err);
    382	if (rc)
    383		goto e_free;
    384
    385	if (copy_to_user((void __user *)arg->resp_data, resp, sizeof(*resp)))
    386		rc = -EFAULT;
    387
    388e_free:
    389	kfree(resp);
    390	return rc;
    391}
    392
    393static int get_derived_key(struct snp_guest_dev *snp_dev, struct snp_guest_request_ioctl *arg)
    394{
    395	struct snp_guest_crypto *crypto = snp_dev->crypto;
    396	struct snp_derived_key_resp resp = {0};
    397	struct snp_derived_key_req req;
    398	int rc, resp_len;
    399	/* Response data is 64 bytes and max authsize for GCM is 16 bytes. */
    400	u8 buf[64 + 16];
    401
    402	lockdep_assert_held(&snp_cmd_mutex);
    403
    404	if (!arg->req_data || !arg->resp_data)
    405		return -EINVAL;
    406
    407	/*
    408	 * The intermediate response buffer is used while decrypting the
    409	 * response payload. Make sure that it has enough space to cover the
    410	 * authtag.
    411	 */
    412	resp_len = sizeof(resp.data) + crypto->a_len;
    413	if (sizeof(buf) < resp_len)
    414		return -ENOMEM;
    415
    416	if (copy_from_user(&req, (void __user *)arg->req_data, sizeof(req)))
    417		return -EFAULT;
    418
    419	rc = handle_guest_request(snp_dev, SVM_VMGEXIT_GUEST_REQUEST, arg->msg_version,
    420				  SNP_MSG_KEY_REQ, &req, sizeof(req), buf, resp_len,
    421				  &arg->fw_err);
    422	if (rc)
    423		return rc;
    424
    425	memcpy(resp.data, buf, sizeof(resp.data));
    426	if (copy_to_user((void __user *)arg->resp_data, &resp, sizeof(resp)))
    427		rc = -EFAULT;
    428
    429	/* The response buffer contains the sensitive data, explicitly clear it. */
    430	memzero_explicit(buf, sizeof(buf));
    431	memzero_explicit(&resp, sizeof(resp));
    432	return rc;
    433}
    434
    435static int get_ext_report(struct snp_guest_dev *snp_dev, struct snp_guest_request_ioctl *arg)
    436{
    437	struct snp_guest_crypto *crypto = snp_dev->crypto;
    438	struct snp_ext_report_req req;
    439	struct snp_report_resp *resp;
    440	int ret, npages = 0, resp_len;
    441
    442	lockdep_assert_held(&snp_cmd_mutex);
    443
    444	if (!arg->req_data || !arg->resp_data)
    445		return -EINVAL;
    446
    447	if (copy_from_user(&req, (void __user *)arg->req_data, sizeof(req)))
    448		return -EFAULT;
    449
    450	/* userspace does not want certificate data */
    451	if (!req.certs_len || !req.certs_address)
    452		goto cmd;
    453
    454	if (req.certs_len > SEV_FW_BLOB_MAX_SIZE ||
    455	    !IS_ALIGNED(req.certs_len, PAGE_SIZE))
    456		return -EINVAL;
    457
    458	if (!access_ok((const void __user *)req.certs_address, req.certs_len))
    459		return -EFAULT;
    460
    461	/*
    462	 * Initialize the intermediate buffer with all zeros. This buffer
    463	 * is used in the guest request message to get the certs blob from
    464	 * the host. If host does not supply any certs in it, then copy
    465	 * zeros to indicate that certificate data was not provided.
    466	 */
    467	memset(snp_dev->certs_data, 0, req.certs_len);
    468	npages = req.certs_len >> PAGE_SHIFT;
    469cmd:
    470	/*
    471	 * The intermediate response buffer is used while decrypting the
    472	 * response payload. Make sure that it has enough space to cover the
    473	 * authtag.
    474	 */
    475	resp_len = sizeof(resp->data) + crypto->a_len;
    476	resp = kzalloc(resp_len, GFP_KERNEL_ACCOUNT);
    477	if (!resp)
    478		return -ENOMEM;
    479
    480	snp_dev->input.data_npages = npages;
    481	ret = handle_guest_request(snp_dev, SVM_VMGEXIT_EXT_GUEST_REQUEST, arg->msg_version,
    482				   SNP_MSG_REPORT_REQ, &req.data,
    483				   sizeof(req.data), resp->data, resp_len, &arg->fw_err);
    484
    485	/* If certs length is invalid then copy the returned length */
    486	if (arg->fw_err == SNP_GUEST_REQ_INVALID_LEN) {
    487		req.certs_len = snp_dev->input.data_npages << PAGE_SHIFT;
    488
    489		if (copy_to_user((void __user *)arg->req_data, &req, sizeof(req)))
    490			ret = -EFAULT;
    491	}
    492
    493	if (ret)
    494		goto e_free;
    495
    496	if (npages &&
    497	    copy_to_user((void __user *)req.certs_address, snp_dev->certs_data,
    498			 req.certs_len)) {
    499		ret = -EFAULT;
    500		goto e_free;
    501	}
    502
    503	if (copy_to_user((void __user *)arg->resp_data, resp, sizeof(*resp)))
    504		ret = -EFAULT;
    505
    506e_free:
    507	kfree(resp);
    508	return ret;
    509}
    510
    511static long snp_guest_ioctl(struct file *file, unsigned int ioctl, unsigned long arg)
    512{
    513	struct snp_guest_dev *snp_dev = to_snp_dev(file);
    514	void __user *argp = (void __user *)arg;
    515	struct snp_guest_request_ioctl input;
    516	int ret = -ENOTTY;
    517
    518	if (copy_from_user(&input, argp, sizeof(input)))
    519		return -EFAULT;
    520
    521	input.fw_err = 0xff;
    522
    523	/* Message version must be non-zero */
    524	if (!input.msg_version)
    525		return -EINVAL;
    526
    527	mutex_lock(&snp_cmd_mutex);
    528
    529	/* Check if the VMPCK is not empty */
    530	if (is_vmpck_empty(snp_dev)) {
    531		dev_err_ratelimited(snp_dev->dev, "VMPCK is disabled\n");
    532		mutex_unlock(&snp_cmd_mutex);
    533		return -ENOTTY;
    534	}
    535
    536	switch (ioctl) {
    537	case SNP_GET_REPORT:
    538		ret = get_report(snp_dev, &input);
    539		break;
    540	case SNP_GET_DERIVED_KEY:
    541		ret = get_derived_key(snp_dev, &input);
    542		break;
    543	case SNP_GET_EXT_REPORT:
    544		ret = get_ext_report(snp_dev, &input);
    545		break;
    546	default:
    547		break;
    548	}
    549
    550	mutex_unlock(&snp_cmd_mutex);
    551
    552	if (input.fw_err && copy_to_user(argp, &input, sizeof(input)))
    553		return -EFAULT;
    554
    555	return ret;
    556}
    557
    558static void free_shared_pages(void *buf, size_t sz)
    559{
    560	unsigned int npages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
    561	int ret;
    562
    563	if (!buf)
    564		return;
    565
    566	ret = set_memory_encrypted((unsigned long)buf, npages);
    567	if (ret) {
    568		WARN_ONCE(ret, "failed to restore encryption mask (leak it)\n");
    569		return;
    570	}
    571
    572	__free_pages(virt_to_page(buf), get_order(sz));
    573}
    574
    575static void *alloc_shared_pages(struct device *dev, size_t sz)
    576{
    577	unsigned int npages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
    578	struct page *page;
    579	int ret;
    580
    581	page = alloc_pages(GFP_KERNEL_ACCOUNT, get_order(sz));
    582	if (!page)
    583		return NULL;
    584
    585	ret = set_memory_decrypted((unsigned long)page_address(page), npages);
    586	if (ret) {
    587		dev_err(dev, "failed to mark page shared, ret=%d\n", ret);
    588		__free_pages(page, get_order(sz));
    589		return NULL;
    590	}
    591
    592	return page_address(page);
    593}
    594
    595static const struct file_operations snp_guest_fops = {
    596	.owner	= THIS_MODULE,
    597	.unlocked_ioctl = snp_guest_ioctl,
    598};
    599
    600static u8 *get_vmpck(int id, struct snp_secrets_page_layout *layout, u32 **seqno)
    601{
    602	u8 *key = NULL;
    603
    604	switch (id) {
    605	case 0:
    606		*seqno = &layout->os_area.msg_seqno_0;
    607		key = layout->vmpck0;
    608		break;
    609	case 1:
    610		*seqno = &layout->os_area.msg_seqno_1;
    611		key = layout->vmpck1;
    612		break;
    613	case 2:
    614		*seqno = &layout->os_area.msg_seqno_2;
    615		key = layout->vmpck2;
    616		break;
    617	case 3:
    618		*seqno = &layout->os_area.msg_seqno_3;
    619		key = layout->vmpck3;
    620		break;
    621	default:
    622		break;
    623	}
    624
    625	return key;
    626}
    627
    628static int __init sev_guest_probe(struct platform_device *pdev)
    629{
    630	struct snp_secrets_page_layout *layout;
    631	struct sev_guest_platform_data *data;
    632	struct device *dev = &pdev->dev;
    633	struct snp_guest_dev *snp_dev;
    634	struct miscdevice *misc;
    635	int ret;
    636
    637	if (!dev->platform_data)
    638		return -ENODEV;
    639
    640	data = (struct sev_guest_platform_data *)dev->platform_data;
    641	layout = (__force void *)ioremap_encrypted(data->secrets_gpa, PAGE_SIZE);
    642	if (!layout)
    643		return -ENODEV;
    644
    645	ret = -ENOMEM;
    646	snp_dev = devm_kzalloc(&pdev->dev, sizeof(struct snp_guest_dev), GFP_KERNEL);
    647	if (!snp_dev)
    648		goto e_unmap;
    649
    650	ret = -EINVAL;
    651	snp_dev->vmpck = get_vmpck(vmpck_id, layout, &snp_dev->os_area_msg_seqno);
    652	if (!snp_dev->vmpck) {
    653		dev_err(dev, "invalid vmpck id %d\n", vmpck_id);
    654		goto e_unmap;
    655	}
    656
    657	/* Verify that VMPCK is not zero. */
    658	if (is_vmpck_empty(snp_dev)) {
    659		dev_err(dev, "vmpck id %d is null\n", vmpck_id);
    660		goto e_unmap;
    661	}
    662
    663	platform_set_drvdata(pdev, snp_dev);
    664	snp_dev->dev = dev;
    665	snp_dev->layout = layout;
    666
    667	/* Allocate the shared page used for the request and response message. */
    668	snp_dev->request = alloc_shared_pages(dev, sizeof(struct snp_guest_msg));
    669	if (!snp_dev->request)
    670		goto e_unmap;
    671
    672	snp_dev->response = alloc_shared_pages(dev, sizeof(struct snp_guest_msg));
    673	if (!snp_dev->response)
    674		goto e_free_request;
    675
    676	snp_dev->certs_data = alloc_shared_pages(dev, SEV_FW_BLOB_MAX_SIZE);
    677	if (!snp_dev->certs_data)
    678		goto e_free_response;
    679
    680	ret = -EIO;
    681	snp_dev->crypto = init_crypto(snp_dev, snp_dev->vmpck, VMPCK_KEY_LEN);
    682	if (!snp_dev->crypto)
    683		goto e_free_cert_data;
    684
    685	misc = &snp_dev->misc;
    686	misc->minor = MISC_DYNAMIC_MINOR;
    687	misc->name = DEVICE_NAME;
    688	misc->fops = &snp_guest_fops;
    689
    690	/* initial the input address for guest request */
    691	snp_dev->input.req_gpa = __pa(snp_dev->request);
    692	snp_dev->input.resp_gpa = __pa(snp_dev->response);
    693	snp_dev->input.data_gpa = __pa(snp_dev->certs_data);
    694
    695	ret =  misc_register(misc);
    696	if (ret)
    697		goto e_free_cert_data;
    698
    699	dev_info(dev, "Initialized SEV guest driver (using vmpck_id %d)\n", vmpck_id);
    700	return 0;
    701
    702e_free_cert_data:
    703	free_shared_pages(snp_dev->certs_data, SEV_FW_BLOB_MAX_SIZE);
    704e_free_response:
    705	free_shared_pages(snp_dev->response, sizeof(struct snp_guest_msg));
    706e_free_request:
    707	free_shared_pages(snp_dev->request, sizeof(struct snp_guest_msg));
    708e_unmap:
    709	iounmap(layout);
    710	return ret;
    711}
    712
    713static int __exit sev_guest_remove(struct platform_device *pdev)
    714{
    715	struct snp_guest_dev *snp_dev = platform_get_drvdata(pdev);
    716
    717	free_shared_pages(snp_dev->certs_data, SEV_FW_BLOB_MAX_SIZE);
    718	free_shared_pages(snp_dev->response, sizeof(struct snp_guest_msg));
    719	free_shared_pages(snp_dev->request, sizeof(struct snp_guest_msg));
    720	deinit_crypto(snp_dev->crypto);
    721	misc_deregister(&snp_dev->misc);
    722
    723	return 0;
    724}
    725
    726/*
    727 * This driver is meant to be a common SEV guest interface driver and to
    728 * support any SEV guest API. As such, even though it has been introduced
    729 * with the SEV-SNP support, it is named "sev-guest".
    730 */
    731static struct platform_driver sev_guest_driver = {
    732	.remove		= __exit_p(sev_guest_remove),
    733	.driver		= {
    734		.name = "sev-guest",
    735	},
    736};
    737
    738module_platform_driver_probe(sev_guest_driver, sev_guest_probe);
    739
    740MODULE_AUTHOR("Brijesh Singh <brijesh.singh@amd.com>");
    741MODULE_LICENSE("GPL");
    742MODULE_VERSION("1.0.0");
    743MODULE_DESCRIPTION("AMD SEV Guest Driver");