| // SPDX-License-Identifier: GPL-2.0-or-later | 
 | /* | 
 |  * CBC: Cipher Block Chaining mode | 
 |  * | 
 |  * Copyright (c) 2006-2016 Herbert Xu <herbert@gondor.apana.org.au> | 
 |  */ | 
 |  | 
 | #include <crypto/algapi.h> | 
 | #include <crypto/internal/cipher.h> | 
 | #include <crypto/internal/skcipher.h> | 
 | #include <linux/err.h> | 
 | #include <linux/init.h> | 
 | #include <linux/kernel.h> | 
 | #include <linux/log2.h> | 
 | #include <linux/module.h> | 
 |  | 
 | static int crypto_cbc_encrypt_segment(struct skcipher_walk *walk, | 
 | 				      struct crypto_skcipher *skcipher) | 
 | { | 
 | 	unsigned int bsize = crypto_skcipher_blocksize(skcipher); | 
 | 	void (*fn)(struct crypto_tfm *, u8 *, const u8 *); | 
 | 	unsigned int nbytes = walk->nbytes; | 
 | 	u8 *src = walk->src.virt.addr; | 
 | 	u8 *dst = walk->dst.virt.addr; | 
 | 	struct crypto_cipher *cipher; | 
 | 	struct crypto_tfm *tfm; | 
 | 	u8 *iv = walk->iv; | 
 |  | 
 | 	cipher = skcipher_cipher_simple(skcipher); | 
 | 	tfm = crypto_cipher_tfm(cipher); | 
 | 	fn = crypto_cipher_alg(cipher)->cia_encrypt; | 
 |  | 
 | 	do { | 
 | 		crypto_xor(iv, src, bsize); | 
 | 		fn(tfm, dst, iv); | 
 | 		memcpy(iv, dst, bsize); | 
 |  | 
 | 		src += bsize; | 
 | 		dst += bsize; | 
 | 	} while ((nbytes -= bsize) >= bsize); | 
 |  | 
 | 	return nbytes; | 
 | } | 
 |  | 
 | static int crypto_cbc_encrypt_inplace(struct skcipher_walk *walk, | 
 | 				      struct crypto_skcipher *skcipher) | 
 | { | 
 | 	unsigned int bsize = crypto_skcipher_blocksize(skcipher); | 
 | 	void (*fn)(struct crypto_tfm *, u8 *, const u8 *); | 
 | 	unsigned int nbytes = walk->nbytes; | 
 | 	u8 *src = walk->src.virt.addr; | 
 | 	struct crypto_cipher *cipher; | 
 | 	struct crypto_tfm *tfm; | 
 | 	u8 *iv = walk->iv; | 
 |  | 
 | 	cipher = skcipher_cipher_simple(skcipher); | 
 | 	tfm = crypto_cipher_tfm(cipher); | 
 | 	fn = crypto_cipher_alg(cipher)->cia_encrypt; | 
 |  | 
 | 	do { | 
 | 		crypto_xor(src, iv, bsize); | 
 | 		fn(tfm, src, src); | 
 | 		iv = src; | 
 |  | 
 | 		src += bsize; | 
 | 	} while ((nbytes -= bsize) >= bsize); | 
 |  | 
 | 	memcpy(walk->iv, iv, bsize); | 
 |  | 
 | 	return nbytes; | 
 | } | 
 |  | 
 | static int crypto_cbc_encrypt(struct skcipher_request *req) | 
 | { | 
 | 	struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); | 
 | 	struct skcipher_walk walk; | 
 | 	int err; | 
 |  | 
 | 	err = skcipher_walk_virt(&walk, req, false); | 
 |  | 
 | 	while (walk.nbytes) { | 
 | 		if (walk.src.virt.addr == walk.dst.virt.addr) | 
 | 			err = crypto_cbc_encrypt_inplace(&walk, skcipher); | 
 | 		else | 
 | 			err = crypto_cbc_encrypt_segment(&walk, skcipher); | 
 | 		err = skcipher_walk_done(&walk, err); | 
 | 	} | 
 |  | 
 | 	return err; | 
 | } | 
 |  | 
 | static int crypto_cbc_decrypt_segment(struct skcipher_walk *walk, | 
 | 				      struct crypto_skcipher *skcipher) | 
 | { | 
 | 	unsigned int bsize = crypto_skcipher_blocksize(skcipher); | 
 | 	void (*fn)(struct crypto_tfm *, u8 *, const u8 *); | 
 | 	unsigned int nbytes = walk->nbytes; | 
 | 	u8 *src = walk->src.virt.addr; | 
 | 	u8 *dst = walk->dst.virt.addr; | 
 | 	struct crypto_cipher *cipher; | 
 | 	struct crypto_tfm *tfm; | 
 | 	u8 *iv = walk->iv; | 
 |  | 
 | 	cipher = skcipher_cipher_simple(skcipher); | 
 | 	tfm = crypto_cipher_tfm(cipher); | 
 | 	fn = crypto_cipher_alg(cipher)->cia_decrypt; | 
 |  | 
 | 	do { | 
 | 		fn(tfm, dst, src); | 
 | 		crypto_xor(dst, iv, bsize); | 
 | 		iv = src; | 
 |  | 
 | 		src += bsize; | 
 | 		dst += bsize; | 
 | 	} while ((nbytes -= bsize) >= bsize); | 
 |  | 
 | 	memcpy(walk->iv, iv, bsize); | 
 |  | 
 | 	return nbytes; | 
 | } | 
 |  | 
 | static int crypto_cbc_decrypt_inplace(struct skcipher_walk *walk, | 
 | 				      struct crypto_skcipher *skcipher) | 
 | { | 
 | 	unsigned int bsize = crypto_skcipher_blocksize(skcipher); | 
 | 	void (*fn)(struct crypto_tfm *, u8 *, const u8 *); | 
 | 	unsigned int nbytes = walk->nbytes; | 
 | 	u8 *src = walk->src.virt.addr; | 
 | 	u8 last_iv[MAX_CIPHER_BLOCKSIZE]; | 
 | 	struct crypto_cipher *cipher; | 
 | 	struct crypto_tfm *tfm; | 
 |  | 
 | 	cipher = skcipher_cipher_simple(skcipher); | 
 | 	tfm = crypto_cipher_tfm(cipher); | 
 | 	fn = crypto_cipher_alg(cipher)->cia_decrypt; | 
 |  | 
 | 	/* Start of the last block. */ | 
 | 	src += nbytes - (nbytes & (bsize - 1)) - bsize; | 
 | 	memcpy(last_iv, src, bsize); | 
 |  | 
 | 	for (;;) { | 
 | 		fn(tfm, src, src); | 
 | 		if ((nbytes -= bsize) < bsize) | 
 | 			break; | 
 | 		crypto_xor(src, src - bsize, bsize); | 
 | 		src -= bsize; | 
 | 	} | 
 |  | 
 | 	crypto_xor(src, walk->iv, bsize); | 
 | 	memcpy(walk->iv, last_iv, bsize); | 
 |  | 
 | 	return nbytes; | 
 | } | 
 |  | 
 | static int crypto_cbc_decrypt(struct skcipher_request *req) | 
 | { | 
 | 	struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); | 
 | 	struct skcipher_walk walk; | 
 | 	int err; | 
 |  | 
 | 	err = skcipher_walk_virt(&walk, req, false); | 
 |  | 
 | 	while (walk.nbytes) { | 
 | 		if (walk.src.virt.addr == walk.dst.virt.addr) | 
 | 			err = crypto_cbc_decrypt_inplace(&walk, skcipher); | 
 | 		else | 
 | 			err = crypto_cbc_decrypt_segment(&walk, skcipher); | 
 | 		err = skcipher_walk_done(&walk, err); | 
 | 	} | 
 |  | 
 | 	return err; | 
 | } | 
 |  | 
 | static int crypto_cbc_create(struct crypto_template *tmpl, struct rtattr **tb) | 
 | { | 
 | 	struct skcipher_instance *inst; | 
 | 	struct crypto_alg *alg; | 
 | 	int err; | 
 |  | 
 | 	inst = skcipher_alloc_instance_simple(tmpl, tb); | 
 | 	if (IS_ERR(inst)) | 
 | 		return PTR_ERR(inst); | 
 |  | 
 | 	alg = skcipher_ialg_simple(inst); | 
 |  | 
 | 	err = -EINVAL; | 
 | 	if (!is_power_of_2(alg->cra_blocksize)) | 
 | 		goto out_free_inst; | 
 |  | 
 | 	inst->alg.encrypt = crypto_cbc_encrypt; | 
 | 	inst->alg.decrypt = crypto_cbc_decrypt; | 
 |  | 
 | 	err = skcipher_register_instance(tmpl, inst); | 
 | 	if (err) { | 
 | out_free_inst: | 
 | 		inst->free(inst); | 
 | 	} | 
 |  | 
 | 	return err; | 
 | } | 
 |  | 
 | static struct crypto_template crypto_cbc_tmpl = { | 
 | 	.name = "cbc", | 
 | 	.create = crypto_cbc_create, | 
 | 	.module = THIS_MODULE, | 
 | }; | 
 |  | 
 | static int __init crypto_cbc_module_init(void) | 
 | { | 
 | 	return crypto_register_template(&crypto_cbc_tmpl); | 
 | } | 
 |  | 
 | static void __exit crypto_cbc_module_exit(void) | 
 | { | 
 | 	crypto_unregister_template(&crypto_cbc_tmpl); | 
 | } | 
 |  | 
 | subsys_initcall(crypto_cbc_module_init); | 
 | module_exit(crypto_cbc_module_exit); | 
 |  | 
 | MODULE_LICENSE("GPL"); | 
 | MODULE_DESCRIPTION("CBC block cipher mode of operation"); | 
 | MODULE_ALIAS_CRYPTO("cbc"); |