/crypto/ |
D | scatterwalk.c | 39 walk->offset = sg->offset; in scatterwalk_start() 46 offset_in_page(walk->offset); in scatterwalk_map() 56 page = sg_page(walk->sg) + ((walk->offset - 1) >> PAGE_SHIFT); in scatterwalk_pagedone() 62 walk->offset += PAGE_SIZE - 1; in scatterwalk_pagedone() 63 walk->offset &= PAGE_MASK; in scatterwalk_pagedone() 64 if (walk->offset >= walk->sg->offset + walk->sg->length) in scatterwalk_pagedone() 71 if (!more || walk->offset >= walk->sg->offset + walk->sg->length || in scatterwalk_done() 72 !(walk->offset & (PAGE_SIZE - 1))) in scatterwalk_done() 108 unsigned int offset = 0; in scatterwalk_map_and_copy() local 116 if (start < offset + sg->length) in scatterwalk_map_and_copy() [all …]
|
D | ahash.c | 47 unsigned int offset = walk->offset; in hash_walk_next() local 49 ((unsigned int)(PAGE_SIZE)) - offset); in hash_walk_next() 55 walk->data += offset; in hash_walk_next() 57 if (offset & alignmask) { in hash_walk_next() 58 unsigned int unaligned = alignmask + 1 - (offset & alignmask); in hash_walk_next() 72 walk->offset = sg->offset; in hash_walk_new_entry() 73 walk->pg = sg_page(walk->sg) + (walk->offset >> PAGE_SHIFT); in hash_walk_new_entry() 74 walk->offset = offset_in_page(walk->offset); in hash_walk_new_entry() 89 walk->data -= walk->offset; in crypto_hash_walk_done() 91 if (nbytes && walk->offset & alignmask && !err) { in crypto_hash_walk_done() [all …]
|
D | cts.c | 77 unsigned int offset, in cts_cbc_encrypt() argument 96 scatterwalk_map_and_copy(s, src, offset, nbytes, 0); in cts_cbc_encrypt() 118 scatterwalk_map_and_copy(d, dst, offset, nbytes, 1); in cts_cbc_encrypt() 163 unsigned int offset, in cts_cbc_decrypt() argument 181 scatterwalk_map_and_copy(s, src, offset, nbytes, 0); in cts_cbc_decrypt() 214 scatterwalk_map_and_copy(d, dst, offset, nbytes, 1); in cts_cbc_decrypt()
|
D | shash.c | 280 unsigned int offset; in shash_ahash_digest() local 284 (sg = req->src, offset = sg->offset, in shash_ahash_digest() 285 nbytes < min(sg->length, ((unsigned int)(PAGE_SIZE)) - offset))) { in shash_ahash_digest() 289 err = crypto_shash_digest(desc, data + offset, nbytes, in shash_ahash_digest() 418 unsigned int offset = sg->offset; in shash_compat_digest() local 421 if (nbytes < min(sg->length, ((unsigned int)(PAGE_SIZE)) - offset)) { in shash_compat_digest() 429 err = crypto_shash_digest(desc, data + offset, nbytes, out); in shash_compat_digest()
|
D | md5.c | 98 const unsigned int offset = mctx->byte_count & 0x3f; in md5_final() local 99 char *p = (char *)mctx->block + offset; in md5_final() 100 int padding = 56 - (offset + 1); in md5_final()
|
D | algif_skcipher.c | 126 sg[i].offset += plen; in skcipher_pull_sgl() 306 PAGE_SIZE - sg->offset - sg->length); in skcipher_sendmsg() 309 sg->offset + sg->length, in skcipher_sendmsg() 315 ctx->merge = (sg->offset + sg->length) & in skcipher_sendmsg() 380 int offset, size_t size, int flags) in skcipher_sendpage() argument 412 sg_set_page(sgl->sg + sgl->cur, page, size, offset); in skcipher_sendpage() 605 int offset, size_t size, int flags) in skcipher_sendpage_nokey() argument 613 return skcipher_sendpage(sock, page, offset, size, flags); in skcipher_sendpage_nokey()
|
D | blkcipher.c | 202 walk->src.phys.offset = offset_in_page(walk->in.offset); in blkcipher_next_fast() 204 walk->dst.phys.offset = offset_in_page(walk->out.offset); in blkcipher_next_fast() 209 diff = walk->src.phys.offset - walk->dst.phys.offset; in blkcipher_next_fast() 270 walk->src.phys.offset &= PAGE_SIZE - 1; in blkcipher_walk_next() 271 walk->dst.phys.offset &= PAGE_SIZE - 1; in blkcipher_walk_next()
|
D | xcbc.c | 153 unsigned int offset = 0; in crypto_xcbc_digest_final() local 166 offset += bs; in crypto_xcbc_digest_final() 170 crypto_xor(prev, consts + offset, bs); in crypto_xcbc_digest_final()
|
D | ablkcipher.c | 217 walk->src.offset = offset_in_page(walk->in.offset); in ablkcipher_next_fast() 219 walk->dst.offset = offset_in_page(walk->out.offset); in ablkcipher_next_fast() 262 walk->src.offset = ((unsigned long)src & (PAGE_SIZE - 1)); in ablkcipher_walk_next() 263 walk->dst.offset = ((unsigned long)dst & (PAGE_SIZE - 1)); in ablkcipher_walk_next()
|
D | cmac.c | 179 unsigned int offset = 0; in crypto_cmac_digest_final() local 192 offset += bs; in crypto_cmac_digest_final() 196 crypto_xor(prev, consts + offset, bs); in crypto_cmac_digest_final()
|
D | authencesn.c | 405 vdst = PageHighMem(dstp) ? NULL : page_address(dstp) + dst->offset; in crypto_authenc_esn_genicv() 427 sg_set_page(hsg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_esn_genicv() 428 sg_set_page(hsg + 1, sg_page(assoc2), assoc2->length, assoc2->offset); in crypto_authenc_esn_genicv() 431 sg_set_page(tsg, sg_page(assoc1), assoc1->length, assoc1->offset); in crypto_authenc_esn_genicv() 573 vsrc = PageHighMem(srcp) ? NULL : page_address(srcp) + src->offset; in crypto_authenc_esn_iverify() 595 sg_set_page(hsg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_esn_iverify() 596 sg_set_page(hsg + 1, sg_page(assoc2), assoc2->length, assoc2->offset); in crypto_authenc_esn_iverify() 599 sg_set_page(tsg, sg_page(assoc1), assoc1->length, assoc1->offset); in crypto_authenc_esn_iverify()
|
D | algif_hash.c | 116 int offset, size_t size, int flags) in hash_sendpage() argument 128 sg_set_page(ctx->sgl.sg, page, size, offset); in hash_sendpage() 299 int offset, size_t size, int flags) in hash_sendpage_nokey() argument 307 return hash_sendpage(sock, page, offset, size, flags); in hash_sendpage_nokey()
|
D | md4.c | 204 const unsigned int offset = mctx->byte_count & 0x3f; in md4_final() local 205 char *p = (char *)mctx->block + offset; in md4_final() 206 int padding = 56 - (offset + 1); in md4_final()
|
D | eseqiv.c | 96 vsrc = PageHighMem(srcp) ? NULL : page_address(srcp) + osrc->offset; in eseqiv_givencrypt() 97 vdst = PageHighMem(dstp) ? NULL : page_address(dstp) + odst->offset; in eseqiv_givencrypt()
|
D | authenc.c | 341 vdst = PageHighMem(dstp) ? NULL : page_address(dstp) + dst->offset; in crypto_authenc_genicv() 354 sg_set_page(asg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_genicv() 496 vsrc = PageHighMem(srcp) ? NULL : page_address(srcp) + src->offset; in crypto_authenc_iverify() 509 sg_set_page(asg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_iverify()
|
D | algapi.c | 896 void *__crypto_dequeue_request(struct crypto_queue *queue, unsigned int offset) in __crypto_dequeue_request() argument 912 offset; in __crypto_dequeue_request()
|
D | gcm.c | 1151 vsrc = PageHighMem(srcp) ? NULL : page_address(srcp) + src->offset; in crypto_rfc4543_crypt() 1161 req->assoc->offset); in crypto_rfc4543_crypt()
|
D | testmgr.c | 674 if (WARN_ON(sg[k - 1].offset + in __test_aead()
|
/crypto/async_tx/ |
D | async_pq.c | 122 do_sync_gen_syndrome(struct page **blocks, unsigned int offset, int disks, in do_sync_gen_syndrome() argument 138 srcs[i] = page_address(blocks[i]) + offset; in do_sync_gen_syndrome() 166 async_gen_syndrome(struct page **blocks, unsigned int offset, int disks, in async_gen_syndrome() argument 184 is_dma_pq_aligned(device, offset, 0, len)) { in async_gen_syndrome() 201 unmap->addr[j] = dma_map_page(device->dev, blocks[i], offset, in async_gen_syndrome() 215 offset, len, DMA_BIDIRECTIONAL); in async_gen_syndrome() 224 offset, len, DMA_BIDIRECTIONAL); in async_gen_syndrome() 245 BUG_ON(len + offset > PAGE_SIZE); in async_gen_syndrome() 249 BUG_ON(len + offset > PAGE_SIZE); in async_gen_syndrome() 251 do_sync_gen_syndrome(blocks, offset, disks, len, submit); in async_gen_syndrome() [all …]
|
D | async_xor.c | 113 do_sync_xor(struct page *dest, struct page **src_list, unsigned int offset, in do_sync_xor() argument 130 srcs[xor_src_cnt++] = page_address(src_list[i]) + offset; in do_sync_xor() 133 dest_buf = page_address(dest) + offset; in do_sync_xor() 173 async_xor(struct page *dest, struct page **src_list, unsigned int offset, in async_xor() argument 187 if (unmap && is_dma_xor_aligned(device, offset, 0, len)) { in async_xor() 200 offset, len, DMA_TO_DEVICE); in async_xor() 204 unmap->addr[j] = dma_map_page(device->dev, dest, offset, len, in async_xor() 229 do_sync_xor(dest, src_list, offset, src_cnt, len, submit); in async_xor() 236 static int page_is_zero(struct page *p, unsigned int offset, size_t len) in page_is_zero() argument 238 return !memchr_inv(page_address(p) + offset, 0, len); in page_is_zero() [all …]
|