cipher.c (21711B)
1// SPDX-License-Identifier: GPL-2.0-only 2/* 3 * Cipher algorithms supported by the CESA: DES, 3DES and AES. 4 * 5 * Author: Boris Brezillon <boris.brezillon@free-electrons.com> 6 * Author: Arnaud Ebalard <arno@natisbad.org> 7 * 8 * This work is based on an initial version written by 9 * Sebastian Andrzej Siewior < sebastian at breakpoint dot cc > 10 */ 11 12#include <crypto/aes.h> 13#include <crypto/internal/des.h> 14#include <linux/device.h> 15#include <linux/dma-mapping.h> 16 17#include "cesa.h" 18 19struct mv_cesa_des_ctx { 20 struct mv_cesa_ctx base; 21 u8 key[DES_KEY_SIZE]; 22}; 23 24struct mv_cesa_des3_ctx { 25 struct mv_cesa_ctx base; 26 u8 key[DES3_EDE_KEY_SIZE]; 27}; 28 29struct mv_cesa_aes_ctx { 30 struct mv_cesa_ctx base; 31 struct crypto_aes_ctx aes; 32}; 33 34struct mv_cesa_skcipher_dma_iter { 35 struct mv_cesa_dma_iter base; 36 struct mv_cesa_sg_dma_iter src; 37 struct mv_cesa_sg_dma_iter dst; 38}; 39 40static inline void 41mv_cesa_skcipher_req_iter_init(struct mv_cesa_skcipher_dma_iter *iter, 42 struct skcipher_request *req) 43{ 44 mv_cesa_req_dma_iter_init(&iter->base, req->cryptlen); 45 mv_cesa_sg_dma_iter_init(&iter->src, req->src, DMA_TO_DEVICE); 46 mv_cesa_sg_dma_iter_init(&iter->dst, req->dst, DMA_FROM_DEVICE); 47} 48 49static inline bool 50mv_cesa_skcipher_req_iter_next_op(struct mv_cesa_skcipher_dma_iter *iter) 51{ 52 iter->src.op_offset = 0; 53 iter->dst.op_offset = 0; 54 55 return mv_cesa_req_dma_iter_next_op(&iter->base); 56} 57 58static inline void 59mv_cesa_skcipher_dma_cleanup(struct skcipher_request *req) 60{ 61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 62 63 if (req->dst != req->src) { 64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, 65 DMA_FROM_DEVICE); 66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, 67 DMA_TO_DEVICE); 68 } else { 69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, 70 DMA_BIDIRECTIONAL); 71 } 72 mv_cesa_dma_cleanup(&creq->base); 73} 74 75static inline void mv_cesa_skcipher_cleanup(struct skcipher_request *req) 76{ 77 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 78 79 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) 80 mv_cesa_skcipher_dma_cleanup(req); 81} 82 83static void mv_cesa_skcipher_std_step(struct skcipher_request *req) 84{ 85 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 86 struct mv_cesa_skcipher_std_req *sreq = &creq->std; 87 struct mv_cesa_engine *engine = creq->base.engine; 88 size_t len = min_t(size_t, req->cryptlen - sreq->offset, 89 CESA_SA_SRAM_PAYLOAD_SIZE); 90 91 mv_cesa_adjust_op(engine, &sreq->op); 92 if (engine->pool) 93 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op)); 94 else 95 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); 96 97 len = mv_cesa_sg_copy_to_sram(engine, req->src, creq->src_nents, 98 CESA_SA_DATA_SRAM_OFFSET, len, 99 sreq->offset); 100 101 sreq->size = len; 102 mv_cesa_set_crypt_op_len(&sreq->op, len); 103 104 /* FIXME: only update enc_len field */ 105 if (!sreq->skip_ctx) { 106 if (engine->pool) 107 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op)); 108 else 109 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); 110 sreq->skip_ctx = true; 111 } else if (engine->pool) 112 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op.desc)); 113 else 114 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op.desc)); 115 116 mv_cesa_set_int_mask(engine, CESA_SA_INT_ACCEL0_DONE); 117 writel_relaxed(CESA_SA_CFG_PARA_DIS, engine->regs + CESA_SA_CFG); 118 WARN_ON(readl(engine->regs + CESA_SA_CMD) & 119 CESA_SA_CMD_EN_CESA_SA_ACCL0); 120 writel(CESA_SA_CMD_EN_CESA_SA_ACCL0, engine->regs + CESA_SA_CMD); 121} 122 123static int mv_cesa_skcipher_std_process(struct skcipher_request *req, 124 u32 status) 125{ 126 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 127 struct mv_cesa_skcipher_std_req *sreq = &creq->std; 128 struct mv_cesa_engine *engine = creq->base.engine; 129 size_t len; 130 131 len = mv_cesa_sg_copy_from_sram(engine, req->dst, creq->dst_nents, 132 CESA_SA_DATA_SRAM_OFFSET, sreq->size, 133 sreq->offset); 134 135 sreq->offset += len; 136 if (sreq->offset < req->cryptlen) 137 return -EINPROGRESS; 138 139 return 0; 140} 141 142static int mv_cesa_skcipher_process(struct crypto_async_request *req, 143 u32 status) 144{ 145 struct skcipher_request *skreq = skcipher_request_cast(req); 146 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); 147 struct mv_cesa_req *basereq = &creq->base; 148 149 if (mv_cesa_req_get_type(basereq) == CESA_STD_REQ) 150 return mv_cesa_skcipher_std_process(skreq, status); 151 152 return mv_cesa_dma_process(basereq, status); 153} 154 155static void mv_cesa_skcipher_step(struct crypto_async_request *req) 156{ 157 struct skcipher_request *skreq = skcipher_request_cast(req); 158 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); 159 160 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) 161 mv_cesa_dma_step(&creq->base); 162 else 163 mv_cesa_skcipher_std_step(skreq); 164} 165 166static inline void 167mv_cesa_skcipher_dma_prepare(struct skcipher_request *req) 168{ 169 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 170 struct mv_cesa_req *basereq = &creq->base; 171 172 mv_cesa_dma_prepare(basereq, basereq->engine); 173} 174 175static inline void 176mv_cesa_skcipher_std_prepare(struct skcipher_request *req) 177{ 178 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 179 struct mv_cesa_skcipher_std_req *sreq = &creq->std; 180 181 sreq->size = 0; 182 sreq->offset = 0; 183} 184 185static inline void mv_cesa_skcipher_prepare(struct crypto_async_request *req, 186 struct mv_cesa_engine *engine) 187{ 188 struct skcipher_request *skreq = skcipher_request_cast(req); 189 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); 190 191 creq->base.engine = engine; 192 193 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) 194 mv_cesa_skcipher_dma_prepare(skreq); 195 else 196 mv_cesa_skcipher_std_prepare(skreq); 197} 198 199static inline void 200mv_cesa_skcipher_req_cleanup(struct crypto_async_request *req) 201{ 202 struct skcipher_request *skreq = skcipher_request_cast(req); 203 204 mv_cesa_skcipher_cleanup(skreq); 205} 206 207static void 208mv_cesa_skcipher_complete(struct crypto_async_request *req) 209{ 210 struct skcipher_request *skreq = skcipher_request_cast(req); 211 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); 212 struct mv_cesa_engine *engine = creq->base.engine; 213 unsigned int ivsize; 214 215 atomic_sub(skreq->cryptlen, &engine->load); 216 ivsize = crypto_skcipher_ivsize(crypto_skcipher_reqtfm(skreq)); 217 218 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) { 219 struct mv_cesa_req *basereq; 220 221 basereq = &creq->base; 222 memcpy(skreq->iv, basereq->chain.last->op->ctx.skcipher.iv, 223 ivsize); 224 } else if (engine->pool) 225 memcpy(skreq->iv, 226 engine->sram_pool + CESA_SA_CRYPT_IV_SRAM_OFFSET, 227 ivsize); 228 else 229 memcpy_fromio(skreq->iv, 230 engine->sram + CESA_SA_CRYPT_IV_SRAM_OFFSET, 231 ivsize); 232} 233 234static const struct mv_cesa_req_ops mv_cesa_skcipher_req_ops = { 235 .step = mv_cesa_skcipher_step, 236 .process = mv_cesa_skcipher_process, 237 .cleanup = mv_cesa_skcipher_req_cleanup, 238 .complete = mv_cesa_skcipher_complete, 239}; 240 241static void mv_cesa_skcipher_cra_exit(struct crypto_tfm *tfm) 242{ 243 void *ctx = crypto_tfm_ctx(tfm); 244 245 memzero_explicit(ctx, tfm->__crt_alg->cra_ctxsize); 246} 247 248static int mv_cesa_skcipher_cra_init(struct crypto_tfm *tfm) 249{ 250 struct mv_cesa_ctx *ctx = crypto_tfm_ctx(tfm); 251 252 ctx->ops = &mv_cesa_skcipher_req_ops; 253 254 crypto_skcipher_set_reqsize(__crypto_skcipher_cast(tfm), 255 sizeof(struct mv_cesa_skcipher_req)); 256 257 return 0; 258} 259 260static int mv_cesa_aes_setkey(struct crypto_skcipher *cipher, const u8 *key, 261 unsigned int len) 262{ 263 struct crypto_tfm *tfm = crypto_skcipher_tfm(cipher); 264 struct mv_cesa_aes_ctx *ctx = crypto_tfm_ctx(tfm); 265 int remaining; 266 int offset; 267 int ret; 268 int i; 269 270 ret = aes_expandkey(&ctx->aes, key, len); 271 if (ret) 272 return ret; 273 274 remaining = (ctx->aes.key_length - 16) / 4; 275 offset = ctx->aes.key_length + 24 - remaining; 276 for (i = 0; i < remaining; i++) 277 ctx->aes.key_dec[4 + i] = ctx->aes.key_enc[offset + i]; 278 279 return 0; 280} 281 282static int mv_cesa_des_setkey(struct crypto_skcipher *cipher, const u8 *key, 283 unsigned int len) 284{ 285 struct mv_cesa_des_ctx *ctx = crypto_skcipher_ctx(cipher); 286 int err; 287 288 err = verify_skcipher_des_key(cipher, key); 289 if (err) 290 return err; 291 292 memcpy(ctx->key, key, DES_KEY_SIZE); 293 294 return 0; 295} 296 297static int mv_cesa_des3_ede_setkey(struct crypto_skcipher *cipher, 298 const u8 *key, unsigned int len) 299{ 300 struct mv_cesa_des_ctx *ctx = crypto_skcipher_ctx(cipher); 301 int err; 302 303 err = verify_skcipher_des3_key(cipher, key); 304 if (err) 305 return err; 306 307 memcpy(ctx->key, key, DES3_EDE_KEY_SIZE); 308 309 return 0; 310} 311 312static int mv_cesa_skcipher_dma_req_init(struct skcipher_request *req, 313 const struct mv_cesa_op_ctx *op_templ) 314{ 315 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 316 gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? 317 GFP_KERNEL : GFP_ATOMIC; 318 struct mv_cesa_req *basereq = &creq->base; 319 struct mv_cesa_skcipher_dma_iter iter; 320 bool skip_ctx = false; 321 int ret; 322 323 basereq->chain.first = NULL; 324 basereq->chain.last = NULL; 325 326 if (req->src != req->dst) { 327 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents, 328 DMA_TO_DEVICE); 329 if (!ret) 330 return -ENOMEM; 331 332 ret = dma_map_sg(cesa_dev->dev, req->dst, creq->dst_nents, 333 DMA_FROM_DEVICE); 334 if (!ret) { 335 ret = -ENOMEM; 336 goto err_unmap_src; 337 } 338 } else { 339 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents, 340 DMA_BIDIRECTIONAL); 341 if (!ret) 342 return -ENOMEM; 343 } 344 345 mv_cesa_tdma_desc_iter_init(&basereq->chain); 346 mv_cesa_skcipher_req_iter_init(&iter, req); 347 348 do { 349 struct mv_cesa_op_ctx *op; 350 351 op = mv_cesa_dma_add_op(&basereq->chain, op_templ, skip_ctx, 352 flags); 353 if (IS_ERR(op)) { 354 ret = PTR_ERR(op); 355 goto err_free_tdma; 356 } 357 skip_ctx = true; 358 359 mv_cesa_set_crypt_op_len(op, iter.base.op_len); 360 361 /* Add input transfers */ 362 ret = mv_cesa_dma_add_op_transfers(&basereq->chain, &iter.base, 363 &iter.src, flags); 364 if (ret) 365 goto err_free_tdma; 366 367 /* Add dummy desc to launch the crypto operation */ 368 ret = mv_cesa_dma_add_dummy_launch(&basereq->chain, flags); 369 if (ret) 370 goto err_free_tdma; 371 372 /* Add output transfers */ 373 ret = mv_cesa_dma_add_op_transfers(&basereq->chain, &iter.base, 374 &iter.dst, flags); 375 if (ret) 376 goto err_free_tdma; 377 378 } while (mv_cesa_skcipher_req_iter_next_op(&iter)); 379 380 /* Add output data for IV */ 381 ret = mv_cesa_dma_add_result_op(&basereq->chain, 382 CESA_SA_CFG_SRAM_OFFSET, 383 CESA_SA_DATA_SRAM_OFFSET, 384 CESA_TDMA_SRC_IN_SRAM, flags); 385 386 if (ret) 387 goto err_free_tdma; 388 389 basereq->chain.last->flags |= CESA_TDMA_END_OF_REQ; 390 391 return 0; 392 393err_free_tdma: 394 mv_cesa_dma_cleanup(basereq); 395 if (req->dst != req->src) 396 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, 397 DMA_FROM_DEVICE); 398 399err_unmap_src: 400 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, 401 req->dst != req->src ? DMA_TO_DEVICE : DMA_BIDIRECTIONAL); 402 403 return ret; 404} 405 406static inline int 407mv_cesa_skcipher_std_req_init(struct skcipher_request *req, 408 const struct mv_cesa_op_ctx *op_templ) 409{ 410 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 411 struct mv_cesa_skcipher_std_req *sreq = &creq->std; 412 struct mv_cesa_req *basereq = &creq->base; 413 414 sreq->op = *op_templ; 415 sreq->skip_ctx = false; 416 basereq->chain.first = NULL; 417 basereq->chain.last = NULL; 418 419 return 0; 420} 421 422static int mv_cesa_skcipher_req_init(struct skcipher_request *req, 423 struct mv_cesa_op_ctx *tmpl) 424{ 425 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 426 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); 427 unsigned int blksize = crypto_skcipher_blocksize(tfm); 428 int ret; 429 430 if (!IS_ALIGNED(req->cryptlen, blksize)) 431 return -EINVAL; 432 433 creq->src_nents = sg_nents_for_len(req->src, req->cryptlen); 434 if (creq->src_nents < 0) { 435 dev_err(cesa_dev->dev, "Invalid number of src SG"); 436 return creq->src_nents; 437 } 438 creq->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); 439 if (creq->dst_nents < 0) { 440 dev_err(cesa_dev->dev, "Invalid number of dst SG"); 441 return creq->dst_nents; 442 } 443 444 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_OP_CRYPT_ONLY, 445 CESA_SA_DESC_CFG_OP_MSK); 446 447 if (cesa_dev->caps->has_tdma) 448 ret = mv_cesa_skcipher_dma_req_init(req, tmpl); 449 else 450 ret = mv_cesa_skcipher_std_req_init(req, tmpl); 451 452 return ret; 453} 454 455static int mv_cesa_skcipher_queue_req(struct skcipher_request *req, 456 struct mv_cesa_op_ctx *tmpl) 457{ 458 int ret; 459 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); 460 struct mv_cesa_engine *engine; 461 462 ret = mv_cesa_skcipher_req_init(req, tmpl); 463 if (ret) 464 return ret; 465 466 engine = mv_cesa_select_engine(req->cryptlen); 467 mv_cesa_skcipher_prepare(&req->base, engine); 468 469 ret = mv_cesa_queue_req(&req->base, &creq->base); 470 471 if (mv_cesa_req_needs_cleanup(&req->base, ret)) 472 mv_cesa_skcipher_cleanup(req); 473 474 return ret; 475} 476 477static int mv_cesa_des_op(struct skcipher_request *req, 478 struct mv_cesa_op_ctx *tmpl) 479{ 480 struct mv_cesa_des_ctx *ctx = crypto_tfm_ctx(req->base.tfm); 481 482 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_DES, 483 CESA_SA_DESC_CFG_CRYPTM_MSK); 484 485 memcpy(tmpl->ctx.skcipher.key, ctx->key, DES_KEY_SIZE); 486 487 return mv_cesa_skcipher_queue_req(req, tmpl); 488} 489 490static int mv_cesa_ecb_des_encrypt(struct skcipher_request *req) 491{ 492 struct mv_cesa_op_ctx tmpl; 493 494 mv_cesa_set_op_cfg(&tmpl, 495 CESA_SA_DESC_CFG_CRYPTCM_ECB | 496 CESA_SA_DESC_CFG_DIR_ENC); 497 498 return mv_cesa_des_op(req, &tmpl); 499} 500 501static int mv_cesa_ecb_des_decrypt(struct skcipher_request *req) 502{ 503 struct mv_cesa_op_ctx tmpl; 504 505 mv_cesa_set_op_cfg(&tmpl, 506 CESA_SA_DESC_CFG_CRYPTCM_ECB | 507 CESA_SA_DESC_CFG_DIR_DEC); 508 509 return mv_cesa_des_op(req, &tmpl); 510} 511 512struct skcipher_alg mv_cesa_ecb_des_alg = { 513 .setkey = mv_cesa_des_setkey, 514 .encrypt = mv_cesa_ecb_des_encrypt, 515 .decrypt = mv_cesa_ecb_des_decrypt, 516 .min_keysize = DES_KEY_SIZE, 517 .max_keysize = DES_KEY_SIZE, 518 .base = { 519 .cra_name = "ecb(des)", 520 .cra_driver_name = "mv-ecb-des", 521 .cra_priority = 300, 522 .cra_flags = CRYPTO_ALG_KERN_DRIVER_ONLY | CRYPTO_ALG_ASYNC | 523 CRYPTO_ALG_ALLOCATES_MEMORY, 524 .cra_blocksize = DES_BLOCK_SIZE, 525 .cra_ctxsize = sizeof(struct mv_cesa_des_ctx), 526 .cra_alignmask = 0, 527 .cra_module = THIS_MODULE, 528 .cra_init = mv_cesa_skcipher_cra_init, 529 .cra_exit = mv_cesa_skcipher_cra_exit, 530 }, 531}; 532 533static int mv_cesa_cbc_des_op(struct skcipher_request *req, 534 struct mv_cesa_op_ctx *tmpl) 535{ 536 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTCM_CBC, 537 CESA_SA_DESC_CFG_CRYPTCM_MSK); 538 539 memcpy(tmpl->ctx.skcipher.iv, req->iv, DES_BLOCK_SIZE); 540 541 return mv_cesa_des_op(req, tmpl); 542} 543 544static int mv_cesa_cbc_des_encrypt(struct skcipher_request *req) 545{ 546 struct mv_cesa_op_ctx tmpl; 547 548 mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_ENC); 549 550 return mv_cesa_cbc_des_op(req, &tmpl); 551} 552 553static int mv_cesa_cbc_des_decrypt(struct skcipher_request *req) 554{ 555 struct mv_cesa_op_ctx tmpl; 556 557 mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_DEC); 558 559 return mv_cesa_cbc_des_op(req, &tmpl); 560} 561 562struct skcipher_alg mv_cesa_cbc_des_alg = { 563 .setkey = mv_cesa_des_setkey, 564 .encrypt = mv_cesa_cbc_des_encrypt, 565 .decrypt = mv_cesa_cbc_des_decrypt, 566 .min_keysize = DES_KEY_SIZE, 567 .max_keysize = DES_KEY_SIZE, 568 .ivsize = DES_BLOCK_SIZE, 569 .base = { 570 .cra_name = "cbc(des)", 571 .cra_driver_name = "mv-cbc-des", 572 .cra_priority = 300, 573 .cra_flags = CRYPTO_ALG_KERN_DRIVER_ONLY | CRYPTO_ALG_ASYNC | 574 CRYPTO_ALG_ALLOCATES_MEMORY, 575 .cra_blocksize = DES_BLOCK_SIZE, 576 .cra_ctxsize = sizeof(struct mv_cesa_des_ctx), 577 .cra_alignmask = 0, 578 .cra_module = THIS_MODULE, 579 .cra_init = mv_cesa_skcipher_cra_init, 580 .cra_exit = mv_cesa_skcipher_cra_exit, 581 }, 582}; 583 584static int mv_cesa_des3_op(struct skcipher_request *req, 585 struct mv_cesa_op_ctx *tmpl) 586{ 587 struct mv_cesa_des3_ctx *ctx = crypto_tfm_ctx(req->base.tfm); 588 589 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_3DES, 590 CESA_SA_DESC_CFG_CRYPTM_MSK); 591 592 memcpy(tmpl->ctx.skcipher.key, ctx->key, DES3_EDE_KEY_SIZE); 593 594 return mv_cesa_skcipher_queue_req(req, tmpl); 595} 596 597static int mv_cesa_ecb_des3_ede_encrypt(struct skcipher_request *req) 598{ 599 struct mv_cesa_op_ctx tmpl; 600 601 mv_cesa_set_op_cfg(&tmpl, 602 CESA_SA_DESC_CFG_CRYPTCM_ECB | 603 CESA_SA_DESC_CFG_3DES_EDE | 604 CESA_SA_DESC_CFG_DIR_ENC); 605 606 return mv_cesa_des3_op(req, &tmpl); 607} 608 609static int mv_cesa_ecb_des3_ede_decrypt(struct skcipher_request *req) 610{ 611 struct mv_cesa_op_ctx tmpl; 612 613 mv_cesa_set_op_cfg(&tmpl, 614 CESA_SA_DESC_CFG_CRYPTCM_ECB | 615 CESA_SA_DESC_CFG_3DES_EDE | 616 CESA_SA_DESC_CFG_DIR_DEC); 617 618 return mv_cesa_des3_op(req, &tmpl); 619} 620 621struct skcipher_alg mv_cesa_ecb_des3_ede_alg = { 622 .setkey = mv_cesa_des3_ede_setkey, 623 .encrypt = mv_cesa_ecb_des3_ede_encrypt, 624 .decrypt = mv_cesa_ecb_des3_ede_decrypt, 625 .min_keysize = DES3_EDE_KEY_SIZE, 626 .max_keysize = DES3_EDE_KEY_SIZE, 627 .base = { 628 .cra_name = "ecb(des3_ede)", 629 .cra_driver_name = "mv-ecb-des3-ede", 630 .cra_priority = 300, 631 .cra_flags = CRYPTO_ALG_KERN_DRIVER_ONLY | CRYPTO_ALG_ASYNC | 632 CRYPTO_ALG_ALLOCATES_MEMORY, 633 .cra_blocksize = DES3_EDE_BLOCK_SIZE, 634 .cra_ctxsize = sizeof(struct mv_cesa_des3_ctx), 635 .cra_alignmask = 0, 636 .cra_module = THIS_MODULE, 637 .cra_init = mv_cesa_skcipher_cra_init, 638 .cra_exit = mv_cesa_skcipher_cra_exit, 639 }, 640}; 641 642static int mv_cesa_cbc_des3_op(struct skcipher_request *req, 643 struct mv_cesa_op_ctx *tmpl) 644{ 645 memcpy(tmpl->ctx.skcipher.iv, req->iv, DES3_EDE_BLOCK_SIZE); 646 647 return mv_cesa_des3_op(req, tmpl); 648} 649 650static int mv_cesa_cbc_des3_ede_encrypt(struct skcipher_request *req) 651{ 652 struct mv_cesa_op_ctx tmpl; 653 654 mv_cesa_set_op_cfg(&tmpl, 655 CESA_SA_DESC_CFG_CRYPTCM_CBC | 656 CESA_SA_DESC_CFG_3DES_EDE | 657 CESA_SA_DESC_CFG_DIR_ENC); 658 659 return mv_cesa_cbc_des3_op(req, &tmpl); 660} 661 662static int mv_cesa_cbc_des3_ede_decrypt(struct skcipher_request *req) 663{ 664 struct mv_cesa_op_ctx tmpl; 665 666 mv_cesa_set_op_cfg(&tmpl, 667 CESA_SA_DESC_CFG_CRYPTCM_CBC | 668 CESA_SA_DESC_CFG_3DES_EDE | 669 CESA_SA_DESC_CFG_DIR_DEC); 670 671 return mv_cesa_cbc_des3_op(req, &tmpl); 672} 673 674struct skcipher_alg mv_cesa_cbc_des3_ede_alg = { 675 .setkey = mv_cesa_des3_ede_setkey, 676 .encrypt = mv_cesa_cbc_des3_ede_encrypt, 677 .decrypt = mv_cesa_cbc_des3_ede_decrypt, 678 .min_keysize = DES3_EDE_KEY_SIZE, 679 .max_keysize = DES3_EDE_KEY_SIZE, 680 .ivsize = DES3_EDE_BLOCK_SIZE, 681 .base = { 682 .cra_name = "cbc(des3_ede)", 683 .cra_driver_name = "mv-cbc-des3-ede", 684 .cra_priority = 300, 685 .cra_flags = CRYPTO_ALG_KERN_DRIVER_ONLY | CRYPTO_ALG_ASYNC | 686 CRYPTO_ALG_ALLOCATES_MEMORY, 687 .cra_blocksize = DES3_EDE_BLOCK_SIZE, 688 .cra_ctxsize = sizeof(struct mv_cesa_des3_ctx), 689 .cra_alignmask = 0, 690 .cra_module = THIS_MODULE, 691 .cra_init = mv_cesa_skcipher_cra_init, 692 .cra_exit = mv_cesa_skcipher_cra_exit, 693 }, 694}; 695 696static int mv_cesa_aes_op(struct skcipher_request *req, 697 struct mv_cesa_op_ctx *tmpl) 698{ 699 struct mv_cesa_aes_ctx *ctx = crypto_tfm_ctx(req->base.tfm); 700 int i; 701 u32 *key; 702 u32 cfg; 703 704 cfg = CESA_SA_DESC_CFG_CRYPTM_AES; 705 706 if (mv_cesa_get_op_cfg(tmpl) & CESA_SA_DESC_CFG_DIR_DEC) 707 key = ctx->aes.key_dec; 708 else 709 key = ctx->aes.key_enc; 710 711 for (i = 0; i < ctx->aes.key_length / sizeof(u32); i++) 712 tmpl->ctx.skcipher.key[i] = cpu_to_le32(key[i]); 713 714 if (ctx->aes.key_length == 24) 715 cfg |= CESA_SA_DESC_CFG_AES_LEN_192; 716 else if (ctx->aes.key_length == 32) 717 cfg |= CESA_SA_DESC_CFG_AES_LEN_256; 718 719 mv_cesa_update_op_cfg(tmpl, cfg, 720 CESA_SA_DESC_CFG_CRYPTM_MSK | 721 CESA_SA_DESC_CFG_AES_LEN_MSK); 722 723 return mv_cesa_skcipher_queue_req(req, tmpl); 724} 725 726static int mv_cesa_ecb_aes_encrypt(struct skcipher_request *req) 727{ 728 struct mv_cesa_op_ctx tmpl; 729 730 mv_cesa_set_op_cfg(&tmpl, 731 CESA_SA_DESC_CFG_CRYPTCM_ECB | 732 CESA_SA_DESC_CFG_DIR_ENC); 733 734 return mv_cesa_aes_op(req, &tmpl); 735} 736 737static int mv_cesa_ecb_aes_decrypt(struct skcipher_request *req) 738{ 739 struct mv_cesa_op_ctx tmpl; 740 741 mv_cesa_set_op_cfg(&tmpl, 742 CESA_SA_DESC_CFG_CRYPTCM_ECB | 743 CESA_SA_DESC_CFG_DIR_DEC); 744 745 return mv_cesa_aes_op(req, &tmpl); 746} 747 748struct skcipher_alg mv_cesa_ecb_aes_alg = { 749 .setkey = mv_cesa_aes_setkey, 750 .encrypt = mv_cesa_ecb_aes_encrypt, 751 .decrypt = mv_cesa_ecb_aes_decrypt, 752 .min_keysize = AES_MIN_KEY_SIZE, 753 .max_keysize = AES_MAX_KEY_SIZE, 754 .base = { 755 .cra_name = "ecb(aes)", 756 .cra_driver_name = "mv-ecb-aes", 757 .cra_priority = 300, 758 .cra_flags = CRYPTO_ALG_KERN_DRIVER_ONLY | CRYPTO_ALG_ASYNC | 759 CRYPTO_ALG_ALLOCATES_MEMORY, 760 .cra_blocksize = AES_BLOCK_SIZE, 761 .cra_ctxsize = sizeof(struct mv_cesa_aes_ctx), 762 .cra_alignmask = 0, 763 .cra_module = THIS_MODULE, 764 .cra_init = mv_cesa_skcipher_cra_init, 765 .cra_exit = mv_cesa_skcipher_cra_exit, 766 }, 767}; 768 769static int mv_cesa_cbc_aes_op(struct skcipher_request *req, 770 struct mv_cesa_op_ctx *tmpl) 771{ 772 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTCM_CBC, 773 CESA_SA_DESC_CFG_CRYPTCM_MSK); 774 memcpy(tmpl->ctx.skcipher.iv, req->iv, AES_BLOCK_SIZE); 775 776 return mv_cesa_aes_op(req, tmpl); 777} 778 779static int mv_cesa_cbc_aes_encrypt(struct skcipher_request *req) 780{ 781 struct mv_cesa_op_ctx tmpl; 782 783 mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_ENC); 784 785 return mv_cesa_cbc_aes_op(req, &tmpl); 786} 787 788static int mv_cesa_cbc_aes_decrypt(struct skcipher_request *req) 789{ 790 struct mv_cesa_op_ctx tmpl; 791 792 mv_cesa_set_op_cfg(&tmpl, CESA_SA_DESC_CFG_DIR_DEC); 793 794 return mv_cesa_cbc_aes_op(req, &tmpl); 795} 796 797struct skcipher_alg mv_cesa_cbc_aes_alg = { 798 .setkey = mv_cesa_aes_setkey, 799 .encrypt = mv_cesa_cbc_aes_encrypt, 800 .decrypt = mv_cesa_cbc_aes_decrypt, 801 .min_keysize = AES_MIN_KEY_SIZE, 802 .max_keysize = AES_MAX_KEY_SIZE, 803 .ivsize = AES_BLOCK_SIZE, 804 .base = { 805 .cra_name = "cbc(aes)", 806 .cra_driver_name = "mv-cbc-aes", 807 .cra_priority = 300, 808 .cra_flags = CRYPTO_ALG_KERN_DRIVER_ONLY | CRYPTO_ALG_ASYNC | 809 CRYPTO_ALG_ALLOCATES_MEMORY, 810 .cra_blocksize = AES_BLOCK_SIZE, 811 .cra_ctxsize = sizeof(struct mv_cesa_aes_ctx), 812 .cra_alignmask = 0, 813 .cra_module = THIS_MODULE, 814 .cra_init = mv_cesa_skcipher_cra_init, 815 .cra_exit = mv_cesa_skcipher_cra_exit, 816 }, 817};