Loading arch/arm/crypto/aes-ce-glue.c +1 −1 Original line number Diff line number Diff line Loading @@ -284,7 +284,7 @@ static int ctr_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, err = blkcipher_walk_done(desc, &walk, walk.nbytes % AES_BLOCK_SIZE); } if (nbytes) { if (walk.nbytes % AES_BLOCK_SIZE) { u8 *tdst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; u8 *tsrc = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; u8 __aligned(8) tail[AES_BLOCK_SIZE]; Loading arch/arm64/crypto/aes-glue.c +1 −1 Original line number Diff line number Diff line Loading @@ -216,7 +216,7 @@ static int ctr_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, err = blkcipher_walk_done(desc, &walk, walk.nbytes % AES_BLOCK_SIZE); } if (nbytes) { if (walk.nbytes % AES_BLOCK_SIZE) { u8 *tdst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; u8 *tsrc = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; u8 __aligned(8) tail[AES_BLOCK_SIZE]; Loading arch/x86/crypto/sha256-mb/sha256_mb.c +2 −2 Original line number Diff line number Diff line Loading @@ -485,10 +485,10 @@ static int sha_complete_job(struct mcryptd_hash_request_ctx *rctx, req = cast_mcryptd_ctx_to_req(req_ctx); if (irqs_disabled()) rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); else { local_bh_disable(); rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); local_bh_enable(); } } Loading arch/x86/crypto/sha256-mb/sha256_mb_mgr_flush_avx2.S +4 −3 Original line number Diff line number Diff line Loading @@ -265,13 +265,14 @@ ENTRY(sha256_mb_mgr_get_comp_job_avx2) vpinsrd $1, _args_digest+1*32(state, idx, 4), %xmm0, %xmm0 vpinsrd $2, _args_digest+2*32(state, idx, 4), %xmm0, %xmm0 vpinsrd $3, _args_digest+3*32(state, idx, 4), %xmm0, %xmm0 movl _args_digest+4*32(state, idx, 4), tmp2_w vmovd _args_digest(state , idx, 4) , %xmm0 vpinsrd $1, _args_digest+5*32(state, idx, 4), %xmm1, %xmm1 vpinsrd $2, _args_digest+6*32(state, idx, 4), %xmm1, %xmm1 vpinsrd $3, _args_digest+7*32(state, idx, 4), %xmm1, %xmm1 vmovdqu %xmm0, _result_digest(job_rax) movl tmp2_w, _result_digest+1*16(job_rax) offset = (_result_digest + 1*16) vmovdqu %xmm1, offset(job_rax) pop %rbx Loading arch/x86/crypto/sha512-mb/sha512_mb.c +2 −2 Original line number Diff line number Diff line Loading @@ -497,10 +497,10 @@ static int sha_complete_job(struct mcryptd_hash_request_ctx *rctx, req = cast_mcryptd_ctx_to_req(req_ctx); if (irqs_disabled()) rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); else { local_bh_disable(); rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); local_bh_enable(); } } Loading Loading
arch/arm/crypto/aes-ce-glue.c +1 −1 Original line number Diff line number Diff line Loading @@ -284,7 +284,7 @@ static int ctr_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, err = blkcipher_walk_done(desc, &walk, walk.nbytes % AES_BLOCK_SIZE); } if (nbytes) { if (walk.nbytes % AES_BLOCK_SIZE) { u8 *tdst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; u8 *tsrc = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; u8 __aligned(8) tail[AES_BLOCK_SIZE]; Loading
arch/arm64/crypto/aes-glue.c +1 −1 Original line number Diff line number Diff line Loading @@ -216,7 +216,7 @@ static int ctr_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, err = blkcipher_walk_done(desc, &walk, walk.nbytes % AES_BLOCK_SIZE); } if (nbytes) { if (walk.nbytes % AES_BLOCK_SIZE) { u8 *tdst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; u8 *tsrc = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; u8 __aligned(8) tail[AES_BLOCK_SIZE]; Loading
arch/x86/crypto/sha256-mb/sha256_mb.c +2 −2 Original line number Diff line number Diff line Loading @@ -485,10 +485,10 @@ static int sha_complete_job(struct mcryptd_hash_request_ctx *rctx, req = cast_mcryptd_ctx_to_req(req_ctx); if (irqs_disabled()) rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); else { local_bh_disable(); rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); local_bh_enable(); } } Loading
arch/x86/crypto/sha256-mb/sha256_mb_mgr_flush_avx2.S +4 −3 Original line number Diff line number Diff line Loading @@ -265,13 +265,14 @@ ENTRY(sha256_mb_mgr_get_comp_job_avx2) vpinsrd $1, _args_digest+1*32(state, idx, 4), %xmm0, %xmm0 vpinsrd $2, _args_digest+2*32(state, idx, 4), %xmm0, %xmm0 vpinsrd $3, _args_digest+3*32(state, idx, 4), %xmm0, %xmm0 movl _args_digest+4*32(state, idx, 4), tmp2_w vmovd _args_digest(state , idx, 4) , %xmm0 vpinsrd $1, _args_digest+5*32(state, idx, 4), %xmm1, %xmm1 vpinsrd $2, _args_digest+6*32(state, idx, 4), %xmm1, %xmm1 vpinsrd $3, _args_digest+7*32(state, idx, 4), %xmm1, %xmm1 vmovdqu %xmm0, _result_digest(job_rax) movl tmp2_w, _result_digest+1*16(job_rax) offset = (_result_digest + 1*16) vmovdqu %xmm1, offset(job_rax) pop %rbx Loading
arch/x86/crypto/sha512-mb/sha512_mb.c +2 −2 Original line number Diff line number Diff line Loading @@ -497,10 +497,10 @@ static int sha_complete_job(struct mcryptd_hash_request_ctx *rctx, req = cast_mcryptd_ctx_to_req(req_ctx); if (irqs_disabled()) rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); else { local_bh_disable(); rctx->complete(&req->base, ret); req_ctx->complete(&req->base, ret); local_bh_enable(); } } Loading