/arch/x86/crypto/ |
D | ecb_cbc_helpers.h | 32 #define ECB_WALK_ADVANCE(blocks) do { \ argument 33 dst += (blocks) * __bsize; \ 34 src += (blocks) * __bsize; \ 35 nbytes -= (blocks) * __bsize; \ 38 #define ECB_BLOCK(blocks, func) do { \ argument 39 const int __blocks = (blocks); \ 46 ECB_WALK_ADVANCE(blocks); \ 61 #define CBC_DEC_BLOCK(blocks, func) do { \ argument 62 const int __blocks = (blocks); \ 68 const u8 *__iv = src + ((blocks) - 1) * __bsize; \ [all …]
|
D | blake2s-glue.c | 40 const size_t blocks = min_t(size_t, nblocks, in blake2s_compress() local 46 blake2s_compress_avx512(state, block, blocks, inc); in blake2s_compress() 48 blake2s_compress_ssse3(state, block, blocks, inc); in blake2s_compress() 51 nblocks -= blocks; in blake2s_compress() 52 block += blocks * BLAKE2S_BLOCK_SIZE; in blake2s_compress()
|
D | sha1_ssse3_glue.c | 76 const u8 *data, int blocks); 126 const u8 *data, int blocks); 188 const u8 *data, int blocks); 201 const u8 *data, int blocks) in sha1_apply_transform_avx2() argument 204 if (blocks >= SHA1_AVX2_BLOCK_OPTSIZE) in sha1_apply_transform_avx2() 205 sha1_transform_avx2(state, data, blocks); in sha1_apply_transform_avx2() 207 sha1_transform_avx(state, data, blocks); in sha1_apply_transform_avx2()
|
D | Kconfig | 95 Processes 16 blocks in parallel. 113 Processes eight blocks in parallel. 127 Processes one or three blocks in parallel. 143 Processes eight blocks in parallel. 159 Processes four blocks in parallel. 176 Processes eight blocks in parallel. 189 Processes 16 blocks in parallel. 268 Processes three blocks in parallel, better utilizing resources of 287 Processes eight blocks in parallel. 305 Processes 16 blocks in parallel. [all …]
|
D | sha512_ssse3_glue.c | 43 const u8 *data, int blocks); 146 const u8 *data, int blocks); 222 const u8 *data, int blocks);
|
D | chacha-avx512vl-x86_64.S | 29 # %rsi: up to 2 data blocks output, o 30 # %rdx: up to 2 data blocks input, i 34 # This function encrypts two ChaCha blocks by loading the state 194 # %rsi: up to 4 data blocks output, o 195 # %rdx: up to 4 data blocks input, i 199 # This function encrypts four ChaCha blocks by loading the state 460 # %rsi: up to 8 data blocks output, o 461 # %rdx: up to 8 data blocks input, i 465 # This function encrypts eight consecutive ChaCha blocks by loading 705 # xor/write first four blocks [all …]
|
/arch/arm64/crypto/ |
D | sha512-ce-glue.c | 30 int blocks); 32 asmlinkage void sha512_block_data_order(u64 *digest, u8 const *src, int blocks); 35 int blocks) in __sha512_ce_transform() argument 37 while (blocks) { in __sha512_ce_transform() 41 rem = sha512_ce_transform(sst, src, blocks); in __sha512_ce_transform() 43 src += (blocks - rem) * SHA512_BLOCK_SIZE; in __sha512_ce_transform() 44 blocks = rem; in __sha512_ce_transform() 49 int blocks) in __sha512_block_data_order() argument 51 sha512_block_data_order(sst->state, src, blocks); in __sha512_block_data_order()
|
D | aes-neonbs-glue.c | 29 int rounds, int blocks); 31 int rounds, int blocks); 34 int rounds, int blocks, u8 iv[]); 37 int rounds, int blocks, u8 iv[]); 40 int rounds, int blocks, u8 iv[]); 42 int rounds, int blocks, u8 iv[]); 46 int rounds, int blocks); 48 int rounds, int blocks, u8 iv[]); 96 int rounds, int blocks)) in __ecb_crypt() argument 106 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in __ecb_crypt() local [all …]
|
D | sha2-ce-glue.c | 34 int blocks); 42 int blocks) in sha256_ce_transform() argument 44 while (blocks) { in sha256_ce_transform() 50 sst), src, blocks); in sha256_ce_transform() 52 src += (blocks - rem) * SHA256_BLOCK_SIZE; in sha256_ce_transform() 53 blocks = rem; in sha256_ce_transform() 62 asmlinkage void sha256_block_data_order(u32 *digest, u8 const *src, int blocks); 65 int blocks) in sha256_arm64_transform() argument 67 sha256_block_data_order(sst->state, src, blocks); in sha256_arm64_transform()
|
D | sha1-ce-glue.c | 33 int blocks); 36 int blocks) in __sha1_ce_transform() argument 38 while (blocks) { in __sha1_ce_transform() 43 sst), src, blocks); in __sha1_ce_transform() 45 src += (blocks - rem) * SHA1_BLOCK_SIZE; in __sha1_ce_transform() 46 blocks = rem; in __sha1_ce_transform()
|
D | sha3-ce-glue.c | 31 asmlinkage int sha3_ce_transform(u64 *st, const u8 *data, int blocks, 44 int blocks; in sha3_update() local 59 blocks = len / sctx->rsiz; in sha3_update() 62 while (blocks) { in sha3_update() 66 rem = sha3_ce_transform(sctx->st, data, blocks, in sha3_update() 69 data += (blocks - rem) * sctx->rsiz; in sha3_update() 70 blocks = rem; in sha3_update()
|
D | aes-glue.c | 78 int rounds, int blocks); 80 int rounds, int blocks); 83 int rounds, int blocks, u8 iv[]); 85 int rounds, int blocks, u8 iv[]); 106 int rounds, int blocks, u8 iv[], 109 int rounds, int blocks, u8 iv[], 113 int blocks, u8 dg[], int enc_before, 185 unsigned int blocks; in ecb_encrypt() local 189 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt() 192 ctx->key_enc, rounds, blocks); in ecb_encrypt() [all …]
|
D | ghash-ce-glue.c | 52 asmlinkage void pmull_ghash_update_p64(int blocks, u64 dg[], const char *src, 55 asmlinkage void pmull_ghash_update_p8(int blocks, u64 dg[], const char *src, 74 static void ghash_do_update(int blocks, u64 dg[], const char *src, in ghash_do_update() argument 84 blocks++; in ghash_do_update() 92 } while (--blocks); in ghash_do_update() 99 void ghash_do_simd_update(int blocks, u64 dg[], const char *src, in ghash_do_simd_update() argument 101 void (*simd_update)(int blocks, u64 dg[], in ghash_do_simd_update() argument 108 simd_update(blocks, dg, src, key->h, head); in ghash_do_simd_update() 111 ghash_do_update(blocks, dg, src, key, head); in ghash_do_simd_update() 128 int blocks; in ghash_update() local [all …]
|
D | sha256-glue.c | 31 int blocks) in __sha256_block_data_order() argument 33 sha256_block_data_order(sst->state, src, blocks); in __sha256_block_data_order() 40 int blocks) in __sha256_block_neon() argument 42 sha256_block_neon(sst->state, src, blocks); in __sha256_block_neon()
|
D | sha512-glue.c | 27 int blocks) in __sha512_block_data_order() argument 29 sha512_block_data_order(sst->state, src, blocks); in __sha512_block_data_order()
|
/arch/m68k/emu/ |
D | nfblock.c | 40 static inline s32 nfhd_get_capacity(u32 major, u32 minor, u32 *blocks, in nfhd_get_capacity() argument 44 virt_to_phys(blocks), virt_to_phys(blocksize)); in nfhd_get_capacity() 55 u32 blocks, bsize; member 84 geo->cylinders = dev->blocks >> (6 - dev->bshift); in nfhd_getgeo() 97 static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) in nfhd_init_one() argument 104 blocks, bsize); in nfhd_init_one() 116 dev->blocks = blocks; in nfhd_init_one() 130 set_capacity(dev->disk, (sector_t)blocks * (bsize / 512)); in nfhd_init_one() 150 u32 blocks, bsize; in nfhd_init() local 168 if (nfhd_get_capacity(i, 0, &blocks, &bsize)) in nfhd_init() [all …]
|
/arch/arm/crypto/ |
D | aes-neonbs-glue.c | 32 int rounds, int blocks); 34 int rounds, int blocks); 37 int rounds, int blocks, u8 iv[]); 40 int rounds, int blocks, u8 ctr[]); 43 int rounds, int blocks, u8 iv[], int); 45 int rounds, int blocks, u8 iv[], int); 90 int rounds, int blocks)) in __ecb_crypt() argument 100 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in __ecb_crypt() local 103 blocks = round_down(blocks, in __ecb_crypt() 108 ctx->rounds, blocks); in __ecb_crypt() [all …]
|
D | blake2b-neon-glue.c | 30 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress_arch() local 34 blake2b_compress_neon(state, block, blocks, inc); in blake2b_compress_arch() 37 nblocks -= blocks; in blake2b_compress_arch() 38 block += blocks * BLAKE2B_BLOCK_SIZE; in blake2b_compress_arch()
|
D | aes-ce-glue.c | 30 int rounds, int blocks); 32 int rounds, int blocks); 35 int rounds, int blocks, u8 iv[]); 37 int rounds, int blocks, u8 iv[]); 44 int rounds, int blocks, u8 ctr[]); 172 unsigned int blocks; in ecb_encrypt() local 177 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt() 180 ctx->key_enc, num_rounds(ctx), blocks); in ecb_encrypt() 192 unsigned int blocks; in ecb_decrypt() local 197 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_decrypt() [all …]
|
D | ghash-ce-glue.c | 62 asmlinkage void pmull_ghash_update_p64(int blocks, u64 dg[], const char *src, 65 asmlinkage void pmull_ghash_update_p8(int blocks, u64 dg[], const char *src, 78 static void ghash_do_update(int blocks, u64 dg[], const char *src, in ghash_do_update() argument 84 pmull_ghash_update_p64(blocks, dg, src, key->h, head); in ghash_do_update() 86 pmull_ghash_update_p8(blocks, dg, src, key->h, head); in ghash_do_update() 96 blocks++; in ghash_do_update() 104 } while (--blocks); in ghash_do_update() 121 int blocks; in ghash_update() local 131 blocks = len / GHASH_BLOCK_SIZE; in ghash_update() 134 ghash_do_update(blocks, ctx->digest, src, key, in ghash_update() [all …]
|
/arch/powerpc/kernel/ |
D | rtas_flash.c | 110 struct flash_block blocks[FLASH_BLOCKS_PER_NODE]; member 173 if (f->blocks[i].data == NULL) { in flash_list_valid() 176 block_size = f->blocks[i].length; in flash_list_valid() 201 kmem_cache_free(flash_block_cache, f->blocks[i].data); in free_flash_list() 357 fl->blocks[next_free].data = p; in rtas_flash_write() 358 fl->blocks[next_free].length = count; in rtas_flash_write() 610 f->blocks[i].data = (char *)cpu_to_be64(__pa(f->blocks[i].data)); in rtas_flash_firmware() 611 image_size += f->blocks[i].length; in rtas_flash_firmware() 612 f->blocks[i].length = cpu_to_be64(f->blocks[i].length); in rtas_flash_firmware()
|
/arch/x86/kernel/cpu/mce/ |
D | amd.c | 886 first_block = bp[bank]->blocks; in amd_threshold_interrupt() 1117 if (tb->blocks) in allocate_threshold_blocks() 1118 list_add(&b->miscj, &tb->blocks->miscj); in allocate_threshold_blocks() 1120 tb->blocks = b; in allocate_threshold_blocks() 1149 struct list_head *head = &b->blocks->miscj; in __threshold_add_blocks() 1154 err = kobject_add(&b->blocks->kobj, b->kobj, b->blocks->kobj.name); in __threshold_add_blocks() 1251 list_for_each_entry_safe(pos, tmp, &bank->blocks->miscj, miscj) { in deallocate_threshold_blocks() 1256 kobject_put(&bank->blocks->kobj); in deallocate_threshold_blocks() 1266 list_for_each_entry_safe(pos, tmp, &b->blocks->miscj, miscj) in __threshold_remove_blocks() 1274 if (!bank->blocks) in threshold_remove_bank()
|
/arch/arm/boot/dts/microchip/ |
D | sama5d3_tcb1.dtsi | 4 * 2 TC blocks.
|
/arch/powerpc/platforms/pseries/ |
D | papr_scm.c | 94 uint64_t blocks; member 181 p->blocks, BIND_ANY_ADDR, token); in drc_pmem_bind() 248 p->drc_index, p->blocks - 1); in drc_pmem_query_n_bind() 253 if ((end_addr - start_addr) != ((p->blocks - 1) * p->block_size)) in drc_pmem_query_n_bind() 1291 mapping.size = p->blocks * p->block_size; // XXX: potential overflow? in papr_scm_nvdimm_init() 1402 u64 blocks, block_size; in papr_scm_probe() local 1421 if (of_property_read_u64(dn, "ibm,number-of-blocks", &blocks)) { in papr_scm_probe() 1452 p->blocks = blocks; in papr_scm_probe() 1498 p->res.end = p->bound_addr + p->blocks * p->block_size - 1; in papr_scm_probe()
|
/arch/powerpc/crypto/ |
D | aes-gcm-p10.S | 33 # v15, v16, v17, v18, v19, v20, v21, v22 for 8 blocks (encrypted) 36 # AES is implemented with 8x blocks and GHASH is using 2 4x blocks. 290 # v15 - v22 - input blocks 650 divdu 10, 12, 10 # n 128 bytes-blocks 825 # loop last few blocks 990 # Handle multiple partial blocks for encrypt and decrypt 1191 divdu 10, 12, 10 # n 128 bytes-blocks 1375 # loop last few blocks
|