| /* | 
 |  * Cryptographic API. | 
 |  * | 
 |  * Cipher operations. | 
 |  * | 
 |  * Copyright (c) 2002 James Morris <jmorris@intercode.com.au> | 
 |  * | 
 |  * This program is free software; you can redistribute it and/or modify it | 
 |  * under the terms of the GNU General Public License as published by the Free | 
 |  * Software Foundation; either version 2 of the License, or (at your option)  | 
 |  * any later version. | 
 |  * | 
 |  */ | 
 | #include <linux/compiler.h> | 
 | #include <linux/kernel.h> | 
 | #include <linux/crypto.h> | 
 | #include <linux/errno.h> | 
 | #include <linux/mm.h> | 
 | #include <linux/slab.h> | 
 | #include <linux/string.h> | 
 | #include <asm/scatterlist.h> | 
 | #include "internal.h" | 
 | #include "scatterwalk.h" | 
 |  | 
 | typedef void (cryptfn_t)(void *, u8 *, const u8 *); | 
 | typedef void (procfn_t)(struct crypto_tfm *, u8 *, | 
 |                         u8*, cryptfn_t, void *); | 
 |  | 
 | static inline void xor_64(u8 *a, const u8 *b) | 
 | { | 
 | 	((u32 *)a)[0] ^= ((u32 *)b)[0]; | 
 | 	((u32 *)a)[1] ^= ((u32 *)b)[1]; | 
 | } | 
 |  | 
 | static inline void xor_128(u8 *a, const u8 *b) | 
 | { | 
 | 	((u32 *)a)[0] ^= ((u32 *)b)[0]; | 
 | 	((u32 *)a)[1] ^= ((u32 *)b)[1]; | 
 | 	((u32 *)a)[2] ^= ((u32 *)b)[2]; | 
 | 	((u32 *)a)[3] ^= ((u32 *)b)[3]; | 
 | } | 
 |   | 
 | static inline void *prepare_src(struct scatter_walk *walk, int bsize, | 
 | 				void *tmp, int in_place) | 
 | { | 
 | 	void *src = walk->data; | 
 | 	int n = bsize; | 
 |  | 
 | 	if (unlikely(scatterwalk_across_pages(walk, bsize))) { | 
 | 		src = tmp; | 
 | 		n = scatterwalk_copychunks(src, walk, bsize, 0); | 
 | 	} | 
 | 	scatterwalk_advance(walk, n); | 
 | 	return src; | 
 | } | 
 |  | 
 | static inline void *prepare_dst(struct scatter_walk *walk, int bsize, | 
 | 				void *tmp, int in_place) | 
 | { | 
 | 	void *dst = walk->data; | 
 |  | 
 | 	if (unlikely(scatterwalk_across_pages(walk, bsize)) || in_place) | 
 | 		dst = tmp; | 
 | 	return dst; | 
 | } | 
 |  | 
 | static inline void complete_src(struct scatter_walk *walk, int bsize, | 
 | 				void *src, int in_place) | 
 | { | 
 | } | 
 |  | 
 | static inline void complete_dst(struct scatter_walk *walk, int bsize, | 
 | 				void *dst, int in_place) | 
 | { | 
 | 	int n = bsize; | 
 |  | 
 | 	if (unlikely(scatterwalk_across_pages(walk, bsize))) | 
 | 		n = scatterwalk_copychunks(dst, walk, bsize, 1); | 
 | 	else if (in_place) | 
 | 		memcpy(walk->data, dst, bsize); | 
 | 	scatterwalk_advance(walk, n); | 
 | } | 
 |  | 
 | /*  | 
 |  * Generic encrypt/decrypt wrapper for ciphers, handles operations across | 
 |  * multiple page boundaries by using temporary blocks.  In user context, | 
 |  * the kernel is given a chance to schedule us once per block. | 
 |  */ | 
 | static int crypt(struct crypto_tfm *tfm, | 
 | 		 struct scatterlist *dst, | 
 | 		 struct scatterlist *src, | 
 |                  unsigned int nbytes, cryptfn_t crfn, | 
 |                  procfn_t prfn, void *info) | 
 | { | 
 | 	struct scatter_walk walk_in, walk_out; | 
 | 	const unsigned int bsize = crypto_tfm_alg_blocksize(tfm); | 
 | 	u8 tmp_src[bsize]; | 
 | 	u8 tmp_dst[bsize]; | 
 |  | 
 | 	if (!nbytes) | 
 | 		return 0; | 
 |  | 
 | 	if (nbytes % bsize) { | 
 | 		tfm->crt_flags |= CRYPTO_TFM_RES_BAD_BLOCK_LEN; | 
 | 		return -EINVAL; | 
 | 	} | 
 |  | 
 | 	scatterwalk_start(&walk_in, src); | 
 | 	scatterwalk_start(&walk_out, dst); | 
 |  | 
 | 	for(;;) { | 
 | 		u8 *src_p, *dst_p; | 
 | 		int in_place; | 
 |  | 
 | 		scatterwalk_map(&walk_in, 0); | 
 | 		scatterwalk_map(&walk_out, 1); | 
 |  | 
 | 		in_place = scatterwalk_samebuf(&walk_in, &walk_out); | 
 |  | 
 | 		do { | 
 | 			src_p = prepare_src(&walk_in, bsize, tmp_src, | 
 | 					    in_place); | 
 | 			dst_p = prepare_dst(&walk_out, bsize, tmp_dst, | 
 | 					    in_place); | 
 |  | 
 | 			prfn(tfm, dst_p, src_p, crfn, info); | 
 |  | 
 | 			complete_src(&walk_in, bsize, src_p, in_place); | 
 | 			complete_dst(&walk_out, bsize, dst_p, in_place); | 
 |  | 
 | 			nbytes -= bsize; | 
 | 		} while (nbytes && | 
 | 			 !scatterwalk_across_pages(&walk_in, bsize) && | 
 | 			 !scatterwalk_across_pages(&walk_out, bsize)); | 
 |  | 
 | 		scatterwalk_done(&walk_in, 0, nbytes); | 
 | 		scatterwalk_done(&walk_out, 1, nbytes); | 
 |  | 
 | 		if (!nbytes) | 
 | 			return 0; | 
 |  | 
 | 		crypto_yield(tfm); | 
 | 	} | 
 | } | 
 |  | 
 | static void cbc_process_encrypt(struct crypto_tfm *tfm, u8 *dst, u8 *src, | 
 | 				cryptfn_t fn, void *info) | 
 | { | 
 | 	u8 *iv = info; | 
 |  | 
 | 	tfm->crt_u.cipher.cit_xor_block(iv, src); | 
 | 	fn(crypto_tfm_ctx(tfm), dst, iv); | 
 | 	memcpy(iv, dst, crypto_tfm_alg_blocksize(tfm)); | 
 | } | 
 |  | 
 | static void cbc_process_decrypt(struct crypto_tfm *tfm, u8 *dst, u8 *src, | 
 | 				cryptfn_t fn, void *info) | 
 | { | 
 | 	u8 *iv = info; | 
 |  | 
 | 	fn(crypto_tfm_ctx(tfm), dst, src); | 
 | 	tfm->crt_u.cipher.cit_xor_block(dst, iv); | 
 | 	memcpy(iv, src, crypto_tfm_alg_blocksize(tfm)); | 
 | } | 
 |  | 
 | static void ecb_process(struct crypto_tfm *tfm, u8 *dst, u8 *src, | 
 | 			cryptfn_t fn, void *info) | 
 | { | 
 | 	fn(crypto_tfm_ctx(tfm), dst, src); | 
 | } | 
 |  | 
 | static int setkey(struct crypto_tfm *tfm, const u8 *key, unsigned int keylen) | 
 | { | 
 | 	struct cipher_alg *cia = &tfm->__crt_alg->cra_cipher; | 
 | 	 | 
 | 	if (keylen < cia->cia_min_keysize || keylen > cia->cia_max_keysize) { | 
 | 		tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN; | 
 | 		return -EINVAL; | 
 | 	} else | 
 | 		return cia->cia_setkey(crypto_tfm_ctx(tfm), key, keylen, | 
 | 		                       &tfm->crt_flags); | 
 | } | 
 |  | 
 | static int ecb_encrypt(struct crypto_tfm *tfm, | 
 | 		       struct scatterlist *dst, | 
 |                        struct scatterlist *src, unsigned int nbytes) | 
 | { | 
 | 	return crypt(tfm, dst, src, nbytes, | 
 | 	             tfm->__crt_alg->cra_cipher.cia_encrypt, | 
 | 	             ecb_process, NULL); | 
 | } | 
 |  | 
 | static int ecb_decrypt(struct crypto_tfm *tfm, | 
 |                        struct scatterlist *dst, | 
 |                        struct scatterlist *src, | 
 | 		       unsigned int nbytes) | 
 | { | 
 | 	return crypt(tfm, dst, src, nbytes, | 
 | 	             tfm->__crt_alg->cra_cipher.cia_decrypt, | 
 | 	             ecb_process, NULL); | 
 | } | 
 |  | 
 | static int cbc_encrypt(struct crypto_tfm *tfm, | 
 |                        struct scatterlist *dst, | 
 |                        struct scatterlist *src, | 
 | 		       unsigned int nbytes) | 
 | { | 
 | 	return crypt(tfm, dst, src, nbytes, | 
 | 	             tfm->__crt_alg->cra_cipher.cia_encrypt, | 
 | 	             cbc_process_encrypt, tfm->crt_cipher.cit_iv); | 
 | } | 
 |  | 
 | static int cbc_encrypt_iv(struct crypto_tfm *tfm, | 
 |                           struct scatterlist *dst, | 
 |                           struct scatterlist *src, | 
 |                           unsigned int nbytes, u8 *iv) | 
 | { | 
 | 	return crypt(tfm, dst, src, nbytes, | 
 | 	             tfm->__crt_alg->cra_cipher.cia_encrypt, | 
 | 	             cbc_process_encrypt, iv); | 
 | } | 
 |  | 
 | static int cbc_decrypt(struct crypto_tfm *tfm, | 
 |                        struct scatterlist *dst, | 
 |                        struct scatterlist *src, | 
 | 		       unsigned int nbytes) | 
 | { | 
 | 	return crypt(tfm, dst, src, nbytes, | 
 | 	             tfm->__crt_alg->cra_cipher.cia_decrypt, | 
 | 	             cbc_process_decrypt, tfm->crt_cipher.cit_iv); | 
 | } | 
 |  | 
 | static int cbc_decrypt_iv(struct crypto_tfm *tfm, | 
 |                           struct scatterlist *dst, | 
 |                           struct scatterlist *src, | 
 |                           unsigned int nbytes, u8 *iv) | 
 | { | 
 | 	return crypt(tfm, dst, src, nbytes, | 
 | 	             tfm->__crt_alg->cra_cipher.cia_decrypt, | 
 | 	             cbc_process_decrypt, iv); | 
 | } | 
 |  | 
 | static int nocrypt(struct crypto_tfm *tfm, | 
 |                    struct scatterlist *dst, | 
 |                    struct scatterlist *src, | 
 | 		   unsigned int nbytes) | 
 | { | 
 | 	return -ENOSYS; | 
 | } | 
 |  | 
 | static int nocrypt_iv(struct crypto_tfm *tfm, | 
 |                       struct scatterlist *dst, | 
 |                       struct scatterlist *src, | 
 |                       unsigned int nbytes, u8 *iv) | 
 | { | 
 | 	return -ENOSYS; | 
 | } | 
 |  | 
 | int crypto_init_cipher_flags(struct crypto_tfm *tfm, u32 flags) | 
 | { | 
 | 	u32 mode = flags & CRYPTO_TFM_MODE_MASK; | 
 | 	 | 
 | 	tfm->crt_cipher.cit_mode = mode ? mode : CRYPTO_TFM_MODE_ECB; | 
 | 	if (flags & CRYPTO_TFM_REQ_WEAK_KEY) | 
 | 		tfm->crt_flags = CRYPTO_TFM_REQ_WEAK_KEY; | 
 | 	 | 
 | 	return 0; | 
 | } | 
 |  | 
 | int crypto_init_cipher_ops(struct crypto_tfm *tfm) | 
 | { | 
 | 	int ret = 0; | 
 | 	struct cipher_tfm *ops = &tfm->crt_cipher; | 
 |  | 
 | 	ops->cit_setkey = setkey; | 
 |  | 
 | 	switch (tfm->crt_cipher.cit_mode) { | 
 | 	case CRYPTO_TFM_MODE_ECB: | 
 | 		ops->cit_encrypt = ecb_encrypt; | 
 | 		ops->cit_decrypt = ecb_decrypt; | 
 | 		break; | 
 | 		 | 
 | 	case CRYPTO_TFM_MODE_CBC: | 
 | 		ops->cit_encrypt = cbc_encrypt; | 
 | 		ops->cit_decrypt = cbc_decrypt; | 
 | 		ops->cit_encrypt_iv = cbc_encrypt_iv; | 
 | 		ops->cit_decrypt_iv = cbc_decrypt_iv; | 
 | 		break; | 
 | 		 | 
 | 	case CRYPTO_TFM_MODE_CFB: | 
 | 		ops->cit_encrypt = nocrypt; | 
 | 		ops->cit_decrypt = nocrypt; | 
 | 		ops->cit_encrypt_iv = nocrypt_iv; | 
 | 		ops->cit_decrypt_iv = nocrypt_iv; | 
 | 		break; | 
 | 	 | 
 | 	case CRYPTO_TFM_MODE_CTR: | 
 | 		ops->cit_encrypt = nocrypt; | 
 | 		ops->cit_decrypt = nocrypt; | 
 | 		ops->cit_encrypt_iv = nocrypt_iv; | 
 | 		ops->cit_decrypt_iv = nocrypt_iv; | 
 | 		break; | 
 |  | 
 | 	default: | 
 | 		BUG(); | 
 | 	} | 
 | 	 | 
 | 	if (ops->cit_mode == CRYPTO_TFM_MODE_CBC) { | 
 | 	    	 | 
 | 	    	switch (crypto_tfm_alg_blocksize(tfm)) { | 
 | 	    	case 8: | 
 | 	    		ops->cit_xor_block = xor_64; | 
 | 	    		break; | 
 | 	    		 | 
 | 	    	case 16: | 
 | 	    		ops->cit_xor_block = xor_128; | 
 | 	    		break; | 
 | 	    		 | 
 | 	    	default: | 
 | 	    		printk(KERN_WARNING "%s: block size %u not supported\n", | 
 | 	    		       crypto_tfm_alg_name(tfm), | 
 | 	    		       crypto_tfm_alg_blocksize(tfm)); | 
 | 	    		ret = -EINVAL; | 
 | 	    		goto out; | 
 | 	    	} | 
 | 	    	 | 
 | 		ops->cit_ivsize = crypto_tfm_alg_blocksize(tfm); | 
 | 	    	ops->cit_iv = kmalloc(ops->cit_ivsize, GFP_KERNEL); | 
 | 		if (ops->cit_iv == NULL) | 
 | 			ret = -ENOMEM; | 
 | 	} | 
 |  | 
 | out:	 | 
 | 	return ret; | 
 | } | 
 |  | 
 | void crypto_exit_cipher_ops(struct crypto_tfm *tfm) | 
 | { | 
 | 	if (tfm->crt_cipher.cit_iv) | 
 | 		kfree(tfm->crt_cipher.cit_iv); | 
 | } |