/arch/x86/crypto/ |
D | blowfish_glue.c | 267 u8 *ctrblk = walk->iv; in ctr_crypt_final() local 273 blowfish_enc_blk(ctx, keystream, ctrblk); in ctr_crypt_final() 277 crypto_inc(ctrblk, BF_BLOCK_SIZE); in ctr_crypt_final() 288 u64 ctrblk = be64_to_cpu(*(__be64 *)walk->iv); in __ctr_crypt() local 302 ctrblocks[0] = cpu_to_be64(ctrblk++); in __ctr_crypt() 303 ctrblocks[1] = cpu_to_be64(ctrblk++); in __ctr_crypt() 304 ctrblocks[2] = cpu_to_be64(ctrblk++); in __ctr_crypt() 305 ctrblocks[3] = cpu_to_be64(ctrblk++); in __ctr_crypt() 323 ctrblocks[0] = cpu_to_be64(ctrblk++); in __ctr_crypt() 332 *(__be64 *)walk->iv = cpu_to_be64(ctrblk); in __ctr_crypt()
|
D | des3_ede_glue.c | 273 u8 *ctrblk = walk->iv; in ctr_crypt_final() local 279 des3_ede_enc_blk(ctx, keystream, ctrblk); in ctr_crypt_final() 283 crypto_inc(ctrblk, DES3_EDE_BLOCK_SIZE); in ctr_crypt_final() 294 u64 ctrblk = be64_to_cpu(*(__be64 *)walk->iv); in __ctr_crypt() local 301 ctrblocks[0] = cpu_to_be64(ctrblk++); in __ctr_crypt() 302 ctrblocks[1] = cpu_to_be64(ctrblk++); in __ctr_crypt() 303 ctrblocks[2] = cpu_to_be64(ctrblk++); in __ctr_crypt() 322 ctrblocks[0] = cpu_to_be64(ctrblk++); in __ctr_crypt() 333 *(__be64 *)walk->iv = cpu_to_be64(ctrblk); in __ctr_crypt()
|
D | glue_helper.c | 224 le128 ctrblk; in glue_ctr_crypt_final_128bit() local 227 be128_to_le128(&ctrblk, (be128 *)walk->iv); in glue_ctr_crypt_final_128bit() 230 fn_ctr(ctx, &tmp, &tmp, &ctrblk); in glue_ctr_crypt_final_128bit() 233 le128_to_be128((be128 *)walk->iv, &ctrblk); in glue_ctr_crypt_final_128bit() 245 le128 ctrblk; in __glue_ctr_crypt_128bit() local 249 be128_to_le128(&ctrblk, (be128 *)walk->iv); in __glue_ctr_crypt_128bit() 258 gctx->funcs[i].fn_u.ctr(ctx, dst, src, &ctrblk); in __glue_ctr_crypt_128bit() 271 le128_to_be128((be128 *)walk->iv, &ctrblk); in __glue_ctr_crypt_128bit()
|
D | cast5_avx_glue.c | 251 u8 *ctrblk = walk->iv; in ctr_crypt_final() local 257 __cast5_encrypt(ctx, keystream, ctrblk); in ctr_crypt_final() 261 crypto_inc(ctrblk, CAST5_BLOCK_SIZE); in ctr_crypt_final() 290 u64 ctrblk; in __ctr_crypt() local 295 ctrblk = *(u64 *)walk->iv; in __ctr_crypt() 298 __cast5_encrypt(ctx, (u8 *)&ctrblk, (u8 *)&ctrblk); in __ctr_crypt() 299 *dst ^= ctrblk; in __ctr_crypt()
|
D | twofish_glue_3way.c | 67 be128 ctrblk; in twofish_enc_blk_ctr() local 72 le128_to_be128(&ctrblk, iv); in twofish_enc_blk_ctr() 75 twofish_enc_blk(ctx, (u8 *)&ctrblk, (u8 *)&ctrblk); in twofish_enc_blk_ctr() 76 u128_xor(dst, dst, (u128 *)&ctrblk); in twofish_enc_blk_ctr()
|
D | serpent_avx_glue.c | 70 be128 ctrblk; in __serpent_crypt_ctr() local 72 le128_to_be128(&ctrblk, iv); in __serpent_crypt_ctr() 75 __serpent_encrypt(ctx, (u8 *)&ctrblk, (u8 *)&ctrblk); in __serpent_crypt_ctr() 76 u128_xor(dst, src, (u128 *)&ctrblk); in __serpent_crypt_ctr()
|
D | serpent_sse2_glue.c | 64 be128 ctrblk; in serpent_crypt_ctr() local 66 le128_to_be128(&ctrblk, iv); in serpent_crypt_ctr() 69 __serpent_encrypt(ctx, (u8 *)&ctrblk, (u8 *)&ctrblk); in serpent_crypt_ctr() 70 u128_xor(dst, src, (u128 *)&ctrblk); in serpent_crypt_ctr()
|
D | cast6_avx_glue.c | 73 be128 ctrblk; in cast6_crypt_ctr() local 75 le128_to_be128(&ctrblk, iv); in cast6_crypt_ctr() 78 __cast6_encrypt(ctx, (u8 *)&ctrblk, (u8 *)&ctrblk); in cast6_crypt_ctr() 79 u128_xor(dst, src, (u128 *)&ctrblk); in cast6_crypt_ctr()
|
D | camellia_glue.c | 1294 be128 ctrblk; in camellia_crypt_ctr() local 1299 le128_to_be128(&ctrblk, iv); in camellia_crypt_ctr() 1302 camellia_enc_blk_xor(ctx, (u8 *)dst, (u8 *)&ctrblk); in camellia_crypt_ctr()
|
D | aesni-intel_glue.c | 463 u8 *ctrblk = walk->iv; in ctr_crypt_final() local 469 aesni_enc(ctx, keystream, ctrblk); in ctr_crypt_final() 472 crypto_inc(ctrblk, AES_BLOCK_SIZE); in ctr_crypt_final()
|
/arch/s390/crypto/ |
D | des_s390.c | 27 static u8 *ctrblk; variable 387 n = __ctrblk_init(ctrblk, walk->iv, nbytes); in ctr_desall_crypt() 388 ctrptr = (n > DES_BLOCK_SIZE) ? ctrblk : walk->iv; in ctr_desall_crypt() 391 if (ctrptr == ctrblk) in ctr_desall_crypt() 510 if (ctrblk) in des_s390_exit() 511 free_page((unsigned long) ctrblk); in des_s390_exit() 552 ctrblk = (u8 *) __get_free_page(GFP_KERNEL); in des_s390_init() 553 if (!ctrblk) { in des_s390_init()
|
D | aes_s390.c | 35 static u8 *ctrblk; variable 694 n = __ctrblk_init(ctrblk, walk->iv, nbytes); in ctr_aes_crypt() 695 ctrptr = (n > AES_BLOCK_SIZE) ? ctrblk : walk->iv; in ctr_aes_crypt() 699 if (ctrptr == ctrblk) in ctr_aes_crypt() 791 if (ctrblk) in aes_s390_fini() 792 free_page((unsigned long) ctrblk); in aes_s390_fini() 833 ctrblk = (u8 *) __get_free_page(GFP_KERNEL); in aes_s390_init() 834 if (!ctrblk) { in aes_s390_init()
|
/arch/sparc/crypto/ |
D | aes_glue.c | 339 u8 *ctrblk = walk->iv; in ctr_crypt_final() local 345 ctx->ops->ecb_encrypt(&ctx->key[0], (const u64 *)ctrblk, in ctr_crypt_final() 349 crypto_inc(ctrblk, AES_BLOCK_SIZE); in ctr_crypt_final()
|