/kernel/linux/linux-5.10/net/xfrm/ |
D | xfrm_ipcomp.c | 29 struct crypto_comp * __percpu *tfms; member 46 struct crypto_comp *tfm = *per_cpu_ptr(ipcd->tfms, cpu); in ipcomp_decompress() 146 tfm = *this_cpu_ptr(ipcd->tfms); in ipcomp_compress() 244 static void ipcomp_free_tfms(struct crypto_comp * __percpu *tfms) in ipcomp_free_tfms() argument 250 if (pos->tfms == tfms) in ipcomp_free_tfms() 262 if (!tfms) in ipcomp_free_tfms() 266 struct crypto_comp *tfm = *per_cpu_ptr(tfms, cpu); in ipcomp_free_tfms() 269 free_percpu(tfms); in ipcomp_free_tfms() 275 struct crypto_comp * __percpu *tfms; in ipcomp_alloc_tfms() local 283 tfm = this_cpu_read(*pos->tfms); in ipcomp_alloc_tfms() [all …]
|
/kernel/linux/linux-5.10/arch/x86/boot/ |
D | cpuflags.c | 90 u32 tfms; in get_cpuflags() local 106 cpuid(0x1, &tfms, &ignored, &cpu.flags[4], in get_cpuflags() 108 cpu.level = (tfms >> 8) & 15; in get_cpuflags() 110 cpu.model = (tfms >> 4) & 15; in get_cpuflags() 112 cpu.model += ((tfms >> 16) & 0xf) << 4; in get_cpuflags()
|
/kernel/linux/linux-5.10/block/ |
D | blk-crypto-fallback.c | 77 struct crypto_skcipher *tfms[BLK_ENCRYPTION_MODE_MAX]; member 99 err = crypto_skcipher_setkey(slotp->tfms[crypto_mode], blank_key, in blk_crypto_evict_keyslot() 119 err = crypto_skcipher_setkey(slotp->tfms[crypto_mode], key->raw, in blk_crypto_keyslot_program() 195 ciph_req = skcipher_request_alloc(slotp->tfms[slotp->crypto_mode], in blk_crypto_alloc_cipher_req() 625 slotp->tfms[mode_num] = crypto_alloc_skcipher(cipher_str, 0, 0); in blk_crypto_fallback_start_using_mode() 626 if (IS_ERR(slotp->tfms[mode_num])) { in blk_crypto_fallback_start_using_mode() 627 err = PTR_ERR(slotp->tfms[mode_num]); in blk_crypto_fallback_start_using_mode() 633 slotp->tfms[mode_num] = NULL; in blk_crypto_fallback_start_using_mode() 637 crypto_skcipher_set_flags(slotp->tfms[mode_num], in blk_crypto_fallback_start_using_mode() 651 crypto_free_skcipher(slotp->tfms[mode_num]); in blk_crypto_fallback_start_using_mode() [all …]
|
/kernel/linux/linux-5.10/net/ipv6/ |
D | seg6_hmac.c | 131 tfm = *this_cpu_ptr(algo->tfms); in __do_hmac() 366 algo->tfms = alloc_percpu(struct crypto_shash *); in seg6_hmac_init_algo() 367 if (!algo->tfms) in seg6_hmac_init_algo() 374 p_tfm = per_cpu_ptr(algo->tfms, cpu); in seg6_hmac_init_algo() 378 p_tfm = raw_cpu_ptr(algo->tfms); in seg6_hmac_init_algo() 427 tfm = *per_cpu_ptr(algo->tfms, cpu); in seg6_hmac_exit() 430 free_percpu(algo->tfms); in seg6_hmac_exit()
|
/kernel/linux/linux-5.10/include/net/ |
D | ipcomp.h | 13 struct crypto_comp * __percpu *tfms; member
|
D | seg6_hmac.h | 38 struct crypto_shash * __percpu *tfms; member
|
/kernel/linux/linux-5.10/drivers/md/ |
D | dm-crypt.c | 174 struct crypto_skcipher **tfms; member 245 return cc->cipher_tfm.tfms[0]; in any_tfm() 1468 skcipher_request_set_tfm(ctx->r.req, cc->cipher_tfm.tfms[key_index]); in crypt_alloc_req_skcipher() 2261 if (!cc->cipher_tfm.tfms) in crypt_free_tfms_skcipher() 2265 if (cc->cipher_tfm.tfms[i] && !IS_ERR(cc->cipher_tfm.tfms[i])) { in crypt_free_tfms_skcipher() 2266 crypto_free_skcipher(cc->cipher_tfm.tfms[i]); in crypt_free_tfms_skcipher() 2267 cc->cipher_tfm.tfms[i] = NULL; in crypt_free_tfms_skcipher() 2270 kfree(cc->cipher_tfm.tfms); in crypt_free_tfms_skcipher() 2271 cc->cipher_tfm.tfms = NULL; in crypt_free_tfms_skcipher() 2287 cc->cipher_tfm.tfms = kcalloc(cc->tfms_count, in crypt_alloc_tfms_skcipher() [all …]
|
/kernel/linux/linux-5.10/arch/x86/kernel/cpu/ |
D | common.c | 834 u32 junk, tfms, cap0, misc; in cpu_detect() local 836 cpuid(0x00000001, &tfms, &misc, &junk, &cap0); in cpu_detect() 837 c->x86 = x86_family(tfms); in cpu_detect() 838 c->x86_model = x86_model(tfms); in cpu_detect() 839 c->x86_stepping = x86_stepping(tfms); in cpu_detect()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/gma500/ |
D | mmu.c | 473 uint32_t tfms, misc, cap0, cap4, clflush_size; in psb_mmu_driver_init() local 480 cpuid(0x00000001, &tfms, &misc, &cap0, &cap4); in psb_mmu_driver_init()
|
/kernel/linux/patches/linux-5.10/imx8mm_patch/patches/drivers/ |
D | 0016_linux_drivers_crypto.patch | 1357 * distribute tfms across job rings to ensure in-order
|