Loading arch/s390/crypto/aes_s390.c +60 −53 Original line number Original line Diff line number Diff line Loading @@ -22,6 +22,7 @@ #include <crypto/aes.h> #include <crypto/aes.h> #include <crypto/algapi.h> #include <crypto/algapi.h> #include <crypto/internal/skcipher.h> #include <linux/err.h> #include <linux/err.h> #include <linux/module.h> #include <linux/module.h> #include <linux/cpufeature.h> #include <linux/cpufeature.h> Loading @@ -44,7 +45,7 @@ struct s390_aes_ctx { long dec; long dec; int key_len; int key_len; union { union { struct crypto_blkcipher *blk; struct crypto_skcipher *blk; struct crypto_cipher *cip; struct crypto_cipher *cip; } fallback; } fallback; }; }; Loading @@ -63,7 +64,7 @@ struct s390_xts_ctx { long enc; long enc; long dec; long dec; int key_len; int key_len; struct crypto_blkcipher *fallback; struct crypto_skcipher *fallback; }; }; /* /* Loading Loading @@ -237,16 +238,16 @@ static int setkey_fallback_blk(struct crypto_tfm *tfm, const u8 *key, struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); unsigned int ret; unsigned int ret; sctx->fallback.blk->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; crypto_skcipher_clear_flags(sctx->fallback.blk, CRYPTO_TFM_REQ_MASK); sctx->fallback.blk->base.crt_flags |= (tfm->crt_flags & crypto_skcipher_set_flags(sctx->fallback.blk, tfm->crt_flags & CRYPTO_TFM_REQ_MASK); CRYPTO_TFM_REQ_MASK); ret = crypto_blkcipher_setkey(sctx->fallback.blk, key, len); ret = crypto_skcipher_setkey(sctx->fallback.blk, key, len); if (ret) { tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags |= (sctx->fallback.blk->base.crt_flags & tfm->crt_flags |= crypto_skcipher_get_flags(sctx->fallback.blk) & CRYPTO_TFM_RES_MASK); CRYPTO_TFM_RES_MASK; } return ret; return ret; } } Loading @@ -255,15 +256,17 @@ static int fallback_blk_dec(struct blkcipher_desc *desc, unsigned int nbytes) unsigned int nbytes) { { unsigned int ret; unsigned int ret; struct crypto_blkcipher *tfm; struct crypto_blkcipher *tfm = desc->tfm; struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, sctx->fallback.blk); tfm = desc->tfm; skcipher_request_set_tfm(req, sctx->fallback.blk); desc->tfm = sctx->fallback.blk; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_decrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_decrypt(req); desc->tfm = tfm; skcipher_request_zero(req); return ret; return ret; } } Loading @@ -272,15 +275,15 @@ static int fallback_blk_enc(struct blkcipher_desc *desc, unsigned int nbytes) unsigned int nbytes) { { unsigned int ret; unsigned int ret; struct crypto_blkcipher *tfm; struct crypto_blkcipher *tfm = desc->tfm; struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, sctx->fallback.blk); tfm = desc->tfm; skcipher_request_set_tfm(req, sctx->fallback.blk); desc->tfm = sctx->fallback.blk; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_encrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_encrypt(req); desc->tfm = tfm; return ret; return ret; } } Loading Loading @@ -370,8 +373,9 @@ static int fallback_init_blk(struct crypto_tfm *tfm) const char *name = tfm->__crt_alg->cra_name; const char *name = tfm->__crt_alg->cra_name; struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); sctx->fallback.blk = crypto_alloc_blkcipher(name, 0, sctx->fallback.blk = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); if (IS_ERR(sctx->fallback.blk)) { if (IS_ERR(sctx->fallback.blk)) { pr_err("Allocating AES fallback algorithm %s failed\n", pr_err("Allocating AES fallback algorithm %s failed\n", Loading @@ -386,8 +390,7 @@ static void fallback_exit_blk(struct crypto_tfm *tfm) { { struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); crypto_free_blkcipher(sctx->fallback.blk); crypto_free_skcipher(sctx->fallback.blk); sctx->fallback.blk = NULL; } } static struct crypto_alg ecb_aes_alg = { static struct crypto_alg ecb_aes_alg = { Loading Loading @@ -536,16 +539,16 @@ static int xts_fallback_setkey(struct crypto_tfm *tfm, const u8 *key, struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); unsigned int ret; unsigned int ret; xts_ctx->fallback->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; crypto_skcipher_clear_flags(xts_ctx->fallback, CRYPTO_TFM_REQ_MASK); xts_ctx->fallback->base.crt_flags |= (tfm->crt_flags & crypto_skcipher_set_flags(xts_ctx->fallback, tfm->crt_flags & CRYPTO_TFM_REQ_MASK); CRYPTO_TFM_REQ_MASK); ret = crypto_blkcipher_setkey(xts_ctx->fallback, key, len); ret = crypto_skcipher_setkey(xts_ctx->fallback, key, len); if (ret) { tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags |= (xts_ctx->fallback->base.crt_flags & tfm->crt_flags |= crypto_skcipher_get_flags(xts_ctx->fallback) & CRYPTO_TFM_RES_MASK); CRYPTO_TFM_RES_MASK; } return ret; return ret; } } Loading @@ -553,16 +556,18 @@ static int xts_fallback_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) unsigned int nbytes) { { struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(desc->tfm); struct crypto_blkcipher *tfm = desc->tfm; struct crypto_blkcipher *tfm; struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, xts_ctx->fallback); unsigned int ret; unsigned int ret; tfm = desc->tfm; skcipher_request_set_tfm(req, xts_ctx->fallback); desc->tfm = xts_ctx->fallback; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_decrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_decrypt(req); desc->tfm = tfm; skcipher_request_zero(req); return ret; return ret; } } Loading @@ -570,16 +575,18 @@ static int xts_fallback_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) unsigned int nbytes) { { struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(desc->tfm); struct crypto_blkcipher *tfm = desc->tfm; struct crypto_blkcipher *tfm; struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, xts_ctx->fallback); unsigned int ret; unsigned int ret; tfm = desc->tfm; skcipher_request_set_tfm(req, xts_ctx->fallback); desc->tfm = xts_ctx->fallback; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_encrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_encrypt(req); desc->tfm = tfm; skcipher_request_zero(req); return ret; return ret; } } Loading Loading @@ -700,8 +707,9 @@ static int xts_fallback_init(struct crypto_tfm *tfm) const char *name = tfm->__crt_alg->cra_name; const char *name = tfm->__crt_alg->cra_name; struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); xts_ctx->fallback = crypto_alloc_blkcipher(name, 0, xts_ctx->fallback = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); if (IS_ERR(xts_ctx->fallback)) { if (IS_ERR(xts_ctx->fallback)) { pr_err("Allocating XTS fallback algorithm %s failed\n", pr_err("Allocating XTS fallback algorithm %s failed\n", Loading @@ -715,8 +723,7 @@ static void xts_fallback_exit(struct crypto_tfm *tfm) { { struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); crypto_free_blkcipher(xts_ctx->fallback); crypto_free_skcipher(xts_ctx->fallback); xts_ctx->fallback = NULL; } } static struct crypto_alg xts_aes_alg = { static struct crypto_alg xts_aes_alg = { Loading Loading
arch/s390/crypto/aes_s390.c +60 −53 Original line number Original line Diff line number Diff line Loading @@ -22,6 +22,7 @@ #include <crypto/aes.h> #include <crypto/aes.h> #include <crypto/algapi.h> #include <crypto/algapi.h> #include <crypto/internal/skcipher.h> #include <linux/err.h> #include <linux/err.h> #include <linux/module.h> #include <linux/module.h> #include <linux/cpufeature.h> #include <linux/cpufeature.h> Loading @@ -44,7 +45,7 @@ struct s390_aes_ctx { long dec; long dec; int key_len; int key_len; union { union { struct crypto_blkcipher *blk; struct crypto_skcipher *blk; struct crypto_cipher *cip; struct crypto_cipher *cip; } fallback; } fallback; }; }; Loading @@ -63,7 +64,7 @@ struct s390_xts_ctx { long enc; long enc; long dec; long dec; int key_len; int key_len; struct crypto_blkcipher *fallback; struct crypto_skcipher *fallback; }; }; /* /* Loading Loading @@ -237,16 +238,16 @@ static int setkey_fallback_blk(struct crypto_tfm *tfm, const u8 *key, struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); unsigned int ret; unsigned int ret; sctx->fallback.blk->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; crypto_skcipher_clear_flags(sctx->fallback.blk, CRYPTO_TFM_REQ_MASK); sctx->fallback.blk->base.crt_flags |= (tfm->crt_flags & crypto_skcipher_set_flags(sctx->fallback.blk, tfm->crt_flags & CRYPTO_TFM_REQ_MASK); CRYPTO_TFM_REQ_MASK); ret = crypto_blkcipher_setkey(sctx->fallback.blk, key, len); ret = crypto_skcipher_setkey(sctx->fallback.blk, key, len); if (ret) { tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags |= (sctx->fallback.blk->base.crt_flags & tfm->crt_flags |= crypto_skcipher_get_flags(sctx->fallback.blk) & CRYPTO_TFM_RES_MASK); CRYPTO_TFM_RES_MASK; } return ret; return ret; } } Loading @@ -255,15 +256,17 @@ static int fallback_blk_dec(struct blkcipher_desc *desc, unsigned int nbytes) unsigned int nbytes) { { unsigned int ret; unsigned int ret; struct crypto_blkcipher *tfm; struct crypto_blkcipher *tfm = desc->tfm; struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, sctx->fallback.blk); tfm = desc->tfm; skcipher_request_set_tfm(req, sctx->fallback.blk); desc->tfm = sctx->fallback.blk; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_decrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_decrypt(req); desc->tfm = tfm; skcipher_request_zero(req); return ret; return ret; } } Loading @@ -272,15 +275,15 @@ static int fallback_blk_enc(struct blkcipher_desc *desc, unsigned int nbytes) unsigned int nbytes) { { unsigned int ret; unsigned int ret; struct crypto_blkcipher *tfm; struct crypto_blkcipher *tfm = desc->tfm; struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, sctx->fallback.blk); tfm = desc->tfm; skcipher_request_set_tfm(req, sctx->fallback.blk); desc->tfm = sctx->fallback.blk; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_encrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_encrypt(req); desc->tfm = tfm; return ret; return ret; } } Loading Loading @@ -370,8 +373,9 @@ static int fallback_init_blk(struct crypto_tfm *tfm) const char *name = tfm->__crt_alg->cra_name; const char *name = tfm->__crt_alg->cra_name; struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); sctx->fallback.blk = crypto_alloc_blkcipher(name, 0, sctx->fallback.blk = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); if (IS_ERR(sctx->fallback.blk)) { if (IS_ERR(sctx->fallback.blk)) { pr_err("Allocating AES fallback algorithm %s failed\n", pr_err("Allocating AES fallback algorithm %s failed\n", Loading @@ -386,8 +390,7 @@ static void fallback_exit_blk(struct crypto_tfm *tfm) { { struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); crypto_free_blkcipher(sctx->fallback.blk); crypto_free_skcipher(sctx->fallback.blk); sctx->fallback.blk = NULL; } } static struct crypto_alg ecb_aes_alg = { static struct crypto_alg ecb_aes_alg = { Loading Loading @@ -536,16 +539,16 @@ static int xts_fallback_setkey(struct crypto_tfm *tfm, const u8 *key, struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); unsigned int ret; unsigned int ret; xts_ctx->fallback->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; crypto_skcipher_clear_flags(xts_ctx->fallback, CRYPTO_TFM_REQ_MASK); xts_ctx->fallback->base.crt_flags |= (tfm->crt_flags & crypto_skcipher_set_flags(xts_ctx->fallback, tfm->crt_flags & CRYPTO_TFM_REQ_MASK); CRYPTO_TFM_REQ_MASK); ret = crypto_blkcipher_setkey(xts_ctx->fallback, key, len); ret = crypto_skcipher_setkey(xts_ctx->fallback, key, len); if (ret) { tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; tfm->crt_flags |= (xts_ctx->fallback->base.crt_flags & tfm->crt_flags |= crypto_skcipher_get_flags(xts_ctx->fallback) & CRYPTO_TFM_RES_MASK); CRYPTO_TFM_RES_MASK; } return ret; return ret; } } Loading @@ -553,16 +556,18 @@ static int xts_fallback_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) unsigned int nbytes) { { struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(desc->tfm); struct crypto_blkcipher *tfm = desc->tfm; struct crypto_blkcipher *tfm; struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, xts_ctx->fallback); unsigned int ret; unsigned int ret; tfm = desc->tfm; skcipher_request_set_tfm(req, xts_ctx->fallback); desc->tfm = xts_ctx->fallback; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_decrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_decrypt(req); desc->tfm = tfm; skcipher_request_zero(req); return ret; return ret; } } Loading @@ -570,16 +575,18 @@ static int xts_fallback_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) unsigned int nbytes) { { struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(desc->tfm); struct crypto_blkcipher *tfm = desc->tfm; struct crypto_blkcipher *tfm; struct s390_xts_ctx *xts_ctx = crypto_blkcipher_ctx(tfm); SKCIPHER_REQUEST_ON_STACK(req, xts_ctx->fallback); unsigned int ret; unsigned int ret; tfm = desc->tfm; skcipher_request_set_tfm(req, xts_ctx->fallback); desc->tfm = xts_ctx->fallback; skcipher_request_set_callback(req, desc->flags, NULL, NULL); skcipher_request_set_crypt(req, src, dst, nbytes, desc->info); ret = crypto_blkcipher_encrypt_iv(desc, dst, src, nbytes); ret = crypto_skcipher_encrypt(req); desc->tfm = tfm; skcipher_request_zero(req); return ret; return ret; } } Loading Loading @@ -700,8 +707,9 @@ static int xts_fallback_init(struct crypto_tfm *tfm) const char *name = tfm->__crt_alg->cra_name; const char *name = tfm->__crt_alg->cra_name; struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); xts_ctx->fallback = crypto_alloc_blkcipher(name, 0, xts_ctx->fallback = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); if (IS_ERR(xts_ctx->fallback)) { if (IS_ERR(xts_ctx->fallback)) { pr_err("Allocating XTS fallback algorithm %s failed\n", pr_err("Allocating XTS fallback algorithm %s failed\n", Loading @@ -715,8 +723,7 @@ static void xts_fallback_exit(struct crypto_tfm *tfm) { { struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); struct s390_xts_ctx *xts_ctx = crypto_tfm_ctx(tfm); crypto_free_blkcipher(xts_ctx->fallback); crypto_free_skcipher(xts_ctx->fallback); xts_ctx->fallback = NULL; } } static struct crypto_alg xts_aes_alg = { static struct crypto_alg xts_aes_alg = { Loading