crypto: ixp4xx - Correct functions alignment
authorCorentin Labbe <clabbe@baylibre.com>
Wed, 5 May 2021 20:26:16 +0000 (20:26 +0000)
committerHerbert Xu <herbert@gondor.apana.org.au>
Fri, 14 May 2021 11:07:57 +0000 (19:07 +0800)
This patch fixes all alignment issues reported by checkpatch.

Signed-off-by: Corentin Labbe <clabbe@baylibre.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
drivers/crypto/ixp4xx_crypto.c

index 03ae9c3..b38650b 100644 (file)
@@ -354,8 +354,8 @@ static void finish_scattered_hmac(struct crypt_ctl *crypt)
        int decryptlen = req->assoclen + req->cryptlen - authsize;
 
        if (req_ctx->encrypt) {
-               scatterwalk_map_and_copy(req_ctx->hmac_virt,
-                       req->dst, decryptlen, authsize, 1);
+               scatterwalk_map_and_copy(req_ctx->hmac_virt, req->dst,
+                                        decryptlen, authsize, 1);
        }
        dma_pool_free(buffer_pool, req_ctx->hmac_virt, crypt->icv_rev_aes);
 }
@@ -412,7 +412,7 @@ static void one_packet(dma_addr_t phys)
        case CTL_FLAG_GEN_ICV:
                ctx = crypto_tfm_ctx(crypt->data.tfm);
                dma_pool_free(ctx_pool, crypt->regist_ptr,
-                               crypt->regist_buf->phys_addr);
+                             crypt->regist_buf->phys_addr);
                dma_pool_free(buffer_pool, crypt->regist_buf, crypt->src_buf);
                if (atomic_dec_and_test(&ctx->configuring))
                        complete(&ctx->completion);
@@ -494,14 +494,13 @@ static int init_ixp_crypto(struct device *dev)
         * so assure it is large enough
         */
        BUILD_BUG_ON(SHA1_DIGEST_SIZE > sizeof(struct buffer_desc));
-       buffer_pool = dma_pool_create("buffer", dev,
-                       sizeof(struct buffer_desc), 32, 0);
+       buffer_pool = dma_pool_create("buffer", dev, sizeof(struct buffer_desc),
+                                     32, 0);
        ret = -ENOMEM;
        if (!buffer_pool)
                goto err;
 
-       ctx_pool = dma_pool_create("context", dev,
-                       NPE_CTX_LEN, 16, 0);
+       ctx_pool = dma_pool_create("context", dev, NPE_CTX_LEN, 16, 0);
        if (!ctx_pool)
                goto err;
 
@@ -546,9 +545,8 @@ static void release_ixp_crypto(struct device *dev)
        npe_release(npe_c);
 
        if (crypt_virt)
-               dma_free_coherent(dev,
-                       NPE_QLEN * sizeof(struct crypt_ctl),
-                       crypt_virt, crypt_phys);
+               dma_free_coherent(dev, NPE_QLEN * sizeof(struct crypt_ctl),
+                                 crypt_virt, crypt_phys);
 }
 
 static void reset_sa_dir(struct ix_sa_dir *dir)
@@ -641,7 +639,8 @@ static void exit_tfm_aead(struct crypto_aead *tfm)
 }
 
 static int register_chain_var(struct crypto_tfm *tfm, u8 xpad, u32 target,
-               int init_len, u32 ctx_addr, const u8 *key, int key_len)
+                             int init_len, u32 ctx_addr, const u8 *key,
+                             int key_len)
 {
        struct ixp_ctx *ctx = crypto_tfm_ctx(tfm);
        struct crypt_ctl *crypt;
@@ -735,11 +734,11 @@ static int setup_auth(struct crypto_tfm *tfm, int encrypt, unsigned int authsize
                dir->npe_mode |= NPE_OP_HASH_VERIFY;
 
        ret = register_chain_var(tfm, HMAC_OPAD_VALUE, otarget,
-                       init_len, npe_ctx_addr, key, key_len);
+                                init_len, npe_ctx_addr, key, key_len);
        if (ret)
                return ret;
        return register_chain_var(tfm, HMAC_IPAD_VALUE, itarget,
-                       init_len, npe_ctx_addr, key, key_len);
+                                 init_len, npe_ctx_addr, key, key_len);
 }
 
 static int gen_rev_aes_key(struct crypto_tfm *tfm)
@@ -770,8 +769,8 @@ static int gen_rev_aes_key(struct crypto_tfm *tfm)
        return 0;
 }
 
-static int setup_cipher(struct crypto_tfm *tfm, int encrypt,
-               const u8 *key, int key_len)
+static int setup_cipher(struct crypto_tfm *tfm, int encrypt, const u8 *key,
+                       int key_len)
 {
        u8 *cinfo;
        u32 cipher_cfg;
@@ -791,9 +790,15 @@ static int setup_cipher(struct crypto_tfm *tfm, int encrypt,
        }
        if (cipher_cfg & MOD_AES) {
                switch (key_len) {
-               case 16: keylen_cfg = MOD_AES128; break;
-               case 24: keylen_cfg = MOD_AES192; break;
-               case 32: keylen_cfg = MOD_AES256; break;
+               case 16:
+                       keylen_cfg = MOD_AES128;
+                       break;
+               case 24:
+                       keylen_cfg = MOD_AES192;
+                       break;
+               case 32:
+                       keylen_cfg = MOD_AES256;
+                       break;
                default:
                        return -EINVAL;
                }
@@ -855,7 +860,7 @@ static struct buffer_desc *chainup_buffers(struct device *dev,
 }
 
 static int ablk_setkey(struct crypto_skcipher *tfm, const u8 *key,
-                       unsigned int key_len)
+                      unsigned int key_len)
 {
        struct ixp_ctx *ctx = crypto_skcipher_ctx(tfm);
        int ret;
@@ -892,7 +897,7 @@ static int ablk_des3_setkey(struct crypto_skcipher *tfm, const u8 *key,
 }
 
 static int ablk_rfc3686_setkey(struct crypto_skcipher *tfm, const u8 *key,
-               unsigned int key_len)
+                              unsigned int key_len)
 {
        struct ixp_ctx *ctx = crypto_skcipher_ctx(tfm);
 
@@ -901,7 +906,7 @@ static int ablk_rfc3686_setkey(struct crypto_skcipher *tfm, const u8 *key,
                return -EINVAL;
 
        memcpy(ctx->nonce, key + (key_len - CTR_RFC3686_NONCE_SIZE),
-                       CTR_RFC3686_NONCE_SIZE);
+              CTR_RFC3686_NONCE_SIZE);
 
        key_len -= CTR_RFC3686_NONCE_SIZE;
        return ablk_setkey(tfm, key, key_len);
@@ -979,7 +984,7 @@ static int ablk_perform(struct skcipher_request *req, int encrypt)
                 * for more than one dst buffer, I think. */
                req_ctx->dst = NULL;
                if (!chainup_buffers(dev, req->dst, nbytes, &dst_hook,
-                                       flags, DMA_FROM_DEVICE))
+                                    flags, DMA_FROM_DEVICE))
                        goto free_buf_dest;
                src_direction = DMA_TO_DEVICE;
                req_ctx->dst = dst_hook.next;
@@ -988,8 +993,8 @@ static int ablk_perform(struct skcipher_request *req, int encrypt)
                req_ctx->dst = NULL;
        }
        req_ctx->src = NULL;
-       if (!chainup_buffers(dev, req->src, nbytes, &src_hook,
-                               flags, src_direction))
+       if (!chainup_buffers(dev, req->src, nbytes, &src_hook, flags,
+                            src_direction))
                goto free_buf_src;
 
        req_ctx->src = src_hook.next;
@@ -1042,7 +1047,7 @@ static int ablk_rfc3686_crypt(struct skcipher_request *req)
 }
 
 static int aead_perform(struct aead_request *req, int encrypt,
-               int cryptoffset, int eff_cryptlen, u8 *iv)
+                       int cryptoffset, int eff_cryptlen, u8 *iv)
 {
        struct crypto_aead *tfm = crypto_aead_reqtfm(req);
        struct ixp_ctx *ctx = crypto_aead_ctx(tfm);
@@ -1130,12 +1135,12 @@ static int aead_perform(struct aead_request *req, int encrypt,
                /* The 12 hmac bytes are scattered,
                 * we need to copy them into a safe buffer */
                req_ctx->hmac_virt = dma_pool_alloc(buffer_pool, flags,
-                               &crypt->icv_rev_aes);
+                                                   &crypt->icv_rev_aes);
                if (unlikely(!req_ctx->hmac_virt))
                        goto free_buf_dst;
                if (!encrypt) {
                        scatterwalk_map_and_copy(req_ctx->hmac_virt,
-                               req->src, cryptlen, authsize, 0);
+                                                req->src, cryptlen, authsize, 0);
                }
                req_ctx->encrypt = encrypt;
        } else {
@@ -1176,11 +1181,11 @@ static int aead_setup(struct crypto_aead *tfm, unsigned int authsize)
        if (ret)
                goto out;
        ret = setup_auth(&tfm->base, 0, authsize, ctx->authkey,
-                       ctx->authkey_len, digest_len);
+                        ctx->authkey_len, digest_len);
        if (ret)
                goto out;
        ret = setup_auth(&tfm->base, 1, authsize,  ctx->authkey,
-                       ctx->authkey_len, digest_len);
+                        ctx->authkey_len, digest_len);
 out:
        if (!atomic_dec_and_test(&ctx->configuring))
                wait_for_completion(&ctx->completion);
@@ -1197,7 +1202,7 @@ static int aead_setauthsize(struct crypto_aead *tfm, unsigned int authsize)
 }
 
 static int aead_setkey(struct crypto_aead *tfm, const u8 *key,
-                       unsigned int keylen)
+                      unsigned int keylen)
 {
        struct ixp_ctx *ctx = crypto_aead_ctx(tfm);
        struct crypto_authenc_keys keys;