crypto: x86/glue_helper - add skcipher_walk functions
authorEric Biggers <ebiggers@google.com>
Tue, 20 Feb 2018 07:48:00 +0000 (23:48 -0800)
committerHerbert Xu <herbert@gondor.apana.org.au>
Fri, 2 Mar 2018 16:03:18 +0000 (00:03 +0800)
Add ECB, CBC, and CTR functions to glue_helper which use skcipher_walk
rather than blkcipher_walk.  This will allow converting the remaining
x86 algorithms from the blkcipher interface over to the skcipher
interface, after which we'll be able to remove the blkcipher_walk
versions of these functions.

Signed-off-by: Eric Biggers <ebiggers@google.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
arch/x86/crypto/glue_helper.c
arch/x86/include/asm/crypto/glue_helper.h

index d61e579..5b90979 100644 (file)
@@ -90,6 +90,52 @@ int glue_ecb_crypt_128bit(const struct common_glue_ctx *gctx,
 }
 EXPORT_SYMBOL_GPL(glue_ecb_crypt_128bit);
 
+int glue_ecb_req_128bit(const struct common_glue_ctx *gctx,
+                       struct skcipher_request *req)
+{
+       void *ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req));
+       const unsigned int bsize = 128 / 8;
+       struct skcipher_walk walk;
+       bool fpu_enabled = false;
+       unsigned int nbytes;
+       int err;
+
+       err = skcipher_walk_virt(&walk, req, false);
+
+       while ((nbytes = walk.nbytes)) {
+               const u8 *src = walk.src.virt.addr;
+               u8 *dst = walk.dst.virt.addr;
+               unsigned int func_bytes;
+               unsigned int i;
+
+               fpu_enabled = glue_skwalk_fpu_begin(bsize,
+                                                   gctx->fpu_blocks_limit,
+                                                   &walk, fpu_enabled, nbytes);
+               for (i = 0; i < gctx->num_funcs; i++) {
+                       func_bytes = bsize * gctx->funcs[i].num_blocks;
+
+                       if (nbytes < func_bytes)
+                               continue;
+
+                       /* Process multi-block batch */
+                       do {
+                               gctx->funcs[i].fn_u.ecb(ctx, dst, src);
+                               src += func_bytes;
+                               dst += func_bytes;
+                               nbytes -= func_bytes;
+                       } while (nbytes >= func_bytes);
+
+                       if (nbytes < bsize)
+                               break;
+               }
+               err = skcipher_walk_done(&walk, nbytes);
+       }
+
+       glue_fpu_end(fpu_enabled);
+       return err;
+}
+EXPORT_SYMBOL_GPL(glue_ecb_req_128bit);
+
 static unsigned int __glue_cbc_encrypt_128bit(const common_glue_func_t fn,
                                              struct blkcipher_desc *desc,
                                              struct blkcipher_walk *walk)
@@ -135,6 +181,38 @@ int glue_cbc_encrypt_128bit(const common_glue_func_t fn,
 }
 EXPORT_SYMBOL_GPL(glue_cbc_encrypt_128bit);
 
+int glue_cbc_encrypt_req_128bit(const common_glue_func_t fn,
+                               struct skcipher_request *req)
+{
+       void *ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req));
+       const unsigned int bsize = 128 / 8;
+       struct skcipher_walk walk;
+       unsigned int nbytes;
+       int err;
+
+       err = skcipher_walk_virt(&walk, req, false);
+
+       while ((nbytes = walk.nbytes)) {
+               const u128 *src = (u128 *)walk.src.virt.addr;
+               u128 *dst = (u128 *)walk.dst.virt.addr;
+               u128 *iv = (u128 *)walk.iv;
+
+               do {
+                       u128_xor(dst, src, iv);
+                       fn(ctx, (u8 *)dst, (u8 *)dst);
+                       iv = dst;
+                       src++;
+                       dst++;
+                       nbytes -= bsize;
+               } while (nbytes >= bsize);
+
+               *(u128 *)walk.iv = *iv;
+               err = skcipher_walk_done(&walk, nbytes);
+       }
+       return err;
+}
+EXPORT_SYMBOL_GPL(glue_cbc_encrypt_req_128bit);
+
 static unsigned int
 __glue_cbc_decrypt_128bit(const struct common_glue_ctx *gctx,
                          struct blkcipher_desc *desc,
@@ -211,6 +289,67 @@ int glue_cbc_decrypt_128bit(const struct common_glue_ctx *gctx,
 }
 EXPORT_SYMBOL_GPL(glue_cbc_decrypt_128bit);
 
+int glue_cbc_decrypt_req_128bit(const struct common_glue_ctx *gctx,
+                               struct skcipher_request *req)
+{
+       void *ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req));
+       const unsigned int bsize = 128 / 8;
+       struct skcipher_walk walk;
+       bool fpu_enabled = false;
+       unsigned int nbytes;
+       int err;
+
+       err = skcipher_walk_virt(&walk, req, false);
+
+       while ((nbytes = walk.nbytes)) {
+               const u128 *src = walk.src.virt.addr;
+               u128 *dst = walk.dst.virt.addr;
+               unsigned int func_bytes, num_blocks;
+               unsigned int i;
+               u128 last_iv;
+
+               fpu_enabled = glue_skwalk_fpu_begin(bsize,
+                                                   gctx->fpu_blocks_limit,
+                                                   &walk, fpu_enabled, nbytes);
+               /* Start of the last block. */
+               src += nbytes / bsize - 1;
+               dst += nbytes / bsize - 1;
+
+               last_iv = *src;
+
+               for (i = 0; i < gctx->num_funcs; i++) {
+                       num_blocks = gctx->funcs[i].num_blocks;
+                       func_bytes = bsize * num_blocks;
+
+                       if (nbytes < func_bytes)
+                               continue;
+
+                       /* Process multi-block batch */
+                       do {
+                               src -= num_blocks - 1;
+                               dst -= num_blocks - 1;
+
+                               gctx->funcs[i].fn_u.cbc(ctx, dst, src);
+
+                               nbytes -= func_bytes;
+                               if (nbytes < bsize)
+                                       goto done;
+
+                               u128_xor(dst, dst, --src);
+                               dst--;
+                       } while (nbytes >= func_bytes);
+               }
+done:
+               u128_xor(dst, dst, (u128 *)walk.iv);
+               *(u128 *)walk.iv = last_iv;
+               err = skcipher_walk_done(&walk, nbytes);
+       }
+
+       glue_fpu_end(fpu_enabled);
+       return err;
+}
+EXPORT_SYMBOL_GPL(glue_cbc_decrypt_req_128bit);
+
 static void glue_ctr_crypt_final_128bit(const common_glue_ctr_func_t fn_ctr,
                                        struct blkcipher_desc *desc,
                                        struct blkcipher_walk *walk)
@@ -301,6 +440,74 @@ int glue_ctr_crypt_128bit(const struct common_glue_ctx *gctx,
 }
 EXPORT_SYMBOL_GPL(glue_ctr_crypt_128bit);
 
+int glue_ctr_req_128bit(const struct common_glue_ctx *gctx,
+                       struct skcipher_request *req)
+{
+       void *ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req));
+       const unsigned int bsize = 128 / 8;
+       struct skcipher_walk walk;
+       bool fpu_enabled = false;
+       unsigned int nbytes;
+       int err;
+
+       err = skcipher_walk_virt(&walk, req, false);
+
+       while ((nbytes = walk.nbytes) >= bsize) {
+               const u128 *src = walk.src.virt.addr;
+               u128 *dst = walk.dst.virt.addr;
+               unsigned int func_bytes, num_blocks;
+               unsigned int i;
+               le128 ctrblk;
+
+               fpu_enabled = glue_skwalk_fpu_begin(bsize,
+                                                   gctx->fpu_blocks_limit,
+                                                   &walk, fpu_enabled, nbytes);
+
+               be128_to_le128(&ctrblk, (be128 *)walk.iv);
+
+               for (i = 0; i < gctx->num_funcs; i++) {
+                       num_blocks = gctx->funcs[i].num_blocks;
+                       func_bytes = bsize * num_blocks;
+
+                       if (nbytes < func_bytes)
+                               continue;
+
+                       /* Process multi-block batch */
+                       do {
+                               gctx->funcs[i].fn_u.ctr(ctx, dst, src, &ctrblk);
+                               src += num_blocks;
+                               dst += num_blocks;
+                               nbytes -= func_bytes;
+                       } while (nbytes >= func_bytes);
+
+                       if (nbytes < bsize)
+                               break;
+               }
+
+               le128_to_be128((be128 *)walk.iv, &ctrblk);
+               err = skcipher_walk_done(&walk, nbytes);
+       }
+
+       glue_fpu_end(fpu_enabled);
+
+       if (nbytes) {
+               le128 ctrblk;
+               u128 tmp;
+
+               be128_to_le128(&ctrblk, (be128 *)walk.iv);
+               memcpy(&tmp, walk.src.virt.addr, nbytes);
+               gctx->funcs[gctx->num_funcs - 1].fn_u.ctr(ctx, &tmp, &tmp,
+                                                         &ctrblk);
+               memcpy(walk.dst.virt.addr, &tmp, nbytes);
+               le128_to_be128((be128 *)walk.iv, &ctrblk);
+
+               err = skcipher_walk_done(&walk, 0);
+       }
+
+       return err;
+}
+EXPORT_SYMBOL_GPL(glue_ctr_req_128bit);
+
 static unsigned int __glue_xts_crypt_128bit(const struct common_glue_ctx *gctx,
                                            void *ctx,
                                            struct blkcipher_desc *desc,
index 553a03d..91c623e 100644 (file)
@@ -131,23 +131,35 @@ extern int glue_ecb_crypt_128bit(const struct common_glue_ctx *gctx,
                                 struct scatterlist *dst,
                                 struct scatterlist *src, unsigned int nbytes);
 
+extern int glue_ecb_req_128bit(const struct common_glue_ctx *gctx,
+                              struct skcipher_request *req);
+
 extern int glue_cbc_encrypt_128bit(const common_glue_func_t fn,
                                   struct blkcipher_desc *desc,
                                   struct scatterlist *dst,
                                   struct scatterlist *src,
                                   unsigned int nbytes);
 
+extern int glue_cbc_encrypt_req_128bit(const common_glue_func_t fn,
+                                      struct skcipher_request *req);
+
 extern int glue_cbc_decrypt_128bit(const struct common_glue_ctx *gctx,
                                   struct blkcipher_desc *desc,
                                   struct scatterlist *dst,
                                   struct scatterlist *src,
                                   unsigned int nbytes);
 
+extern int glue_cbc_decrypt_req_128bit(const struct common_glue_ctx *gctx,
+                                      struct skcipher_request *req);
+
 extern int glue_ctr_crypt_128bit(const struct common_glue_ctx *gctx,
                                 struct blkcipher_desc *desc,
                                 struct scatterlist *dst,
                                 struct scatterlist *src, unsigned int nbytes);
 
+extern int glue_ctr_req_128bit(const struct common_glue_ctx *gctx,
+                              struct skcipher_request *req);
+
 extern int glue_xts_crypt_128bit(const struct common_glue_ctx *gctx,
                                 struct blkcipher_desc *desc,
                                 struct scatterlist *dst,