Home
last modified time | relevance | path

Searched refs:bv (Results 1 – 7 of 7) sorted by relevance

/block/
Dbio.c155 void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned int idx) in bvec_free() argument
164 mempool_free(bv, pool); in bvec_free()
168 kmem_cache_free(bvs->slab, bv); in bvec_free()
537 struct bio_vec bv; in zero_fill_bio_iter() local
540 __bio_for_each_segment(bv, bio, iter, start) { in zero_fill_bio_iter()
541 char *data = bvec_kmap_irq(&bv, &flags); in zero_fill_bio_iter()
542 memset(data, 0, bv.bv_len); in zero_fill_bio_iter()
543 flush_dcache_page(bv.bv_page); in zero_fill_bio_iter()
561 struct bio_vec bv; in bio_truncate() local
572 bio_for_each_segment(bv, bio, iter) { in bio_truncate()
[all …]
Dblk-merge.c201 const struct bio_vec *bv, unsigned *nsegs, in bvec_split_segs() argument
206 unsigned len = min(bv->bv_len, max_len); in bvec_split_segs()
211 seg_size = get_max_segment_size(q, bv->bv_page, in bvec_split_segs()
212 bv->bv_offset + total_len); in bvec_split_segs()
219 if ((bv->bv_offset + total_len) & queue_virt_boundary(q)) in bvec_split_segs()
226 return len > 0 || bv->bv_len > max_len; in bvec_split_segs()
253 struct bio_vec bv, bvprv, *bvprvp = NULL; in blk_bio_segment_split() local
259 bio_for_each_bvec(bv, bio, iter) { in blk_bio_segment_split()
264 if (bvprvp && bvec_gap_to_prev(q, bvprvp, bv.bv_offset)) in blk_bio_segment_split()
268 sectors + (bv.bv_len >> 9) <= max_sectors && in blk_bio_segment_split()
[all …]
Dblk-crypto-fallback.c164 struct bio_vec bv; in blk_crypto_clone_bio() local
177 bio_for_each_segment(bv, bio_src, iter) in blk_crypto_clone_bio()
178 bio->bi_io_vec[bio->bi_vcnt++] = bv; in blk_crypto_clone_bio()
216 struct bio_vec bv; in blk_crypto_split_bio_if_needed() local
219 bio_for_each_segment(bv, bio, iter) { in blk_crypto_split_bio_if_needed()
220 num_sectors += bv.bv_len >> SECTOR_SHIFT; in blk_crypto_split_bio_if_needed()
387 struct bio_vec bv; in blk_crypto_fallback_decrypt_bio() local
415 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { in blk_crypto_fallback_decrypt_bio()
416 struct page *page = bv.bv_page; in blk_crypto_fallback_decrypt_bio()
418 sg_set_page(&sg, page, data_unit_size, bv.bv_offset); in blk_crypto_fallback_decrypt_bio()
[all …]
Dbio-integrity.c168 struct bio_vec bv; in bio_integrity_process() local
179 __bio_for_each_segment(bv, bio, bviter, *proc_iter) { in bio_integrity_process()
180 void *kaddr = kmap_atomic(bv.bv_page); in bio_integrity_process()
182 iter.data_buf = kaddr + bv.bv_offset; in bio_integrity_process()
183 iter.data_size = bv.bv_len; in bio_integrity_process()
Dblk-crypto.c211 struct bio_vec bv; in bio_crypt_check_alignment() local
213 bio_for_each_segment(bv, bio, iter) { in bio_crypt_check_alignment()
214 if (!IS_ALIGNED(bv.bv_len | bv.bv_offset, data_unit_size)) in bio_crypt_check_alignment()
Dbounce.c221 struct bio_vec bv; in bounce_clone_bio() local
265 bio_for_each_segment(bv, bio_src, iter) in bounce_clone_bio()
266 bio->bi_io_vec[bio->bi_vcnt++] = bv; in bounce_clone_bio()
Dblk-map.c528 struct bio_vec bv; in blk_rq_append_bio() local
533 bio_for_each_bvec(bv, *bio, iter) in blk_rq_append_bio()