Home
last modified time | relevance | path

Searched refs:bv_page (Results 1 – 25 of 108) sorted by relevance

12345

/kernel/linux/linux-5.10/block/
Dbounce.c80 vto = kmap_atomic(to->bv_page); in bounce_copy_vec()
88 memcpy(page_address((to)->bv_page) + (to)->bv_offset, vfrom, (to)->bv_len)
146 if (tovec.bv_page != fromvec.bv_page) { in copy_to_high_bio_irq()
152 vfrom = page_address(fromvec.bv_page) + in copy_to_high_bio_irq()
156 flush_dcache_page(tovec.bv_page); in copy_to_high_bio_irq()
174 if (bvec->bv_page != orig_vec.bv_page) { in bounce_end_io()
175 dec_zone_page_state(bvec->bv_page, NR_BOUNCE); in bounce_end_io()
176 mempool_free(bvec->bv_page, pool); in bounce_end_io()
302 if (page_to_pfn(from.bv_page) > q->limits.bounce_pfn) in __blk_queue_bounce()
323 struct page *page = to->bv_page; in __blk_queue_bounce()
[all …]
Dbio.c541 flush_dcache_page(bv.bv_page); in zero_fill_bio_iter()
578 zero_user(bv.bv_page, bv.bv_offset + offset, in bio_truncate()
743 phys_addr_t vec_end_addr = page_to_phys(bv->bv_page) + bv_end - 1; in page_is_mergeable()
754 return (bv->bv_page + bv_end / PAGE_SIZE) == (page + off / PAGE_SIZE); in page_is_mergeable()
768 phys_addr_t addr1 = page_to_phys(bv->bv_page) + bv->bv_offset; in bio_try_merge_hw_seg()
823 bvec->bv_page = page; in bio_add_hw_page()
912 bv->bv_page = page; in __bio_add_page()
957 if (mark_dirty && !PageCompound(bvec->bv_page)) in bio_release_pages()
958 set_page_dirty_lock(bvec->bv_page); in bio_release_pages()
959 put_page(bvec->bv_page); in bio_release_pages()
[all …]
Dbio-integrity.c112 kfree(page_address(bip->bip_vec->bv_page) + in bio_integrity_free()
147 iv->bv_page = page; in bio_integrity_add_page()
171 void *prot_buf = page_address(bip->bip_vec->bv_page) + in bio_integrity_process()
180 void *kaddr = kmap_atomic(bv.bv_page); in bio_integrity_process()
Dblk-crypto-fallback.c152 mempool_free(enc_bio->bi_io_vec[i].bv_page, in blk_crypto_fallback_encrypt_endio()
315 struct page *plaintext_page = enc_bvec->bv_page; in blk_crypto_fallback_encrypt_bio()
319 enc_bvec->bv_page = ciphertext_page; in blk_crypto_fallback_encrypt_bio()
356 mempool_free(enc_bio->bi_io_vec[--i].bv_page, in blk_crypto_fallback_encrypt_bio()
414 struct page *page = bv.bv_page; in blk_crypto_fallback_decrypt_bio()
Dt10-pi.c153 pmap = kmap_atomic(iv.bv_page); in t10_pi_type1_prepare()
201 pmap = kmap_atomic(iv.bv_page); in t10_pi_type1_complete()
Dblk.h58 phys_addr_t addr1 = page_to_phys(vec1->bv_page) + vec1->bv_offset; in biovec_phys_mergeable()
59 phys_addr_t addr2 = page_to_phys(vec2->bv_page) + vec2->bv_offset; in biovec_phys_mergeable()
63 if (xen_domain() && !xen_biovec_phys_mergeable(vec1, vec2->bv_page)) in biovec_phys_mergeable()
Dblk-map.c54 ret = copy_page_from_iter(bvec->bv_page, in bio_copy_from_iter()
85 ret = copy_page_to_iter(bvec->bv_page, in bio_copy_to_iter()
445 memcpy(p, page_address(bvec->bv_page), bvec->bv_len); in bio_copy_kern_endio_read()
Dblk-merge.c208 seg_size = get_max_segment_size(q, bv->bv_page, in bvec_split_segs()
426 unsigned len = min(get_max_segment_size(q, bvec->bv_page, in blk_bvec_map_sg()
428 struct page *page = bvec->bv_page; in blk_bvec_map_sg()
456 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in __blk_bvec_map_sg()
/kernel/linux/linux-5.10/include/linux/
Dbvec.h32 struct page *bv_page; member
62 (__bvec_iter_bvec((bvec), (iter))->bv_page)
76 .bv_page = mp_bvec_iter_page((bvec), (iter)), \
95 .bv_page = bvec_iter_page((bvec), (iter)), \
160 bv->bv_page++; in bvec_advance()
163 bv->bv_page = bvec->bv_page + (bvec->bv_offset >> PAGE_SHIFT); in bvec_advance()
Dbio.h289 return bio_first_bvec_all(bio)->bv_page; in bio_first_page_all()
516 addr = (unsigned long) kmap_atomic(bvec->bv_page); in bvec_kmap_irq()
534 return page_address(bvec->bv_page) + bvec->bv_offset; in bvec_kmap_irq()
/kernel/linux/linux-5.10/drivers/md/bcache/
Dutil.c247 bv->bv_page = is_vmalloc_addr(base) in bch_bio_map()
278 bv->bv_page = alloc_page(gfp_mask); in bch_bio_alloc_pages()
279 if (!bv->bv_page) { in bch_bio_alloc_pages()
281 __free_page(bv->bv_page); in bch_bio_alloc_pages()
Ddebug.c130 void *p1 = kmap_atomic(bv.bv_page); in bch_data_verify()
134 p2 = page_address(cbv.bv_page); in bch_data_verify()
/kernel/linux/linux-5.10/fs/squashfs/
Dblock.c53 page_address(bvec->bv_page) + bvec->bv_offset + offset, in copy_bio_to_actor()
180 data = page_address(bvec->bv_page) + bvec->bv_offset; in squashfs_read_data()
189 data = page_address(bvec->bv_page) + bvec->bv_offset; in squashfs_read_data()
Dzlib_wrapper.c79 data = page_address(bvec->bv_page) + bvec->bv_offset; in zlib_uncompress()
Dlzo_wrapper.c79 data = page_address(bvec->bv_page) + bvec->bv_offset; in lzo_uncompress()
Dzstd_wrapper.c97 data = page_address(bvec->bv_page) + bvec->bv_offset; in zstd_uncompress()
Dlz4_wrapper.c104 data = page_address(bvec->bv_page) + bvec->bv_offset; in lz4_uncompress()
/kernel/linux/linux-5.10/drivers/xen/
Dbiomerge.c12 unsigned long bfn1 = pfn_to_bfn(page_to_pfn(vec1->bv_page)); in xen_biovec_phys_mergeable()
/kernel/linux/linux-5.10/lib/
Diov_iter.c641 memcpy_to_page(v.bv_page, v.bv_offset, in _copy_to_iter()
743 rem = copy_mc_to_page(v.bv_page, v.bv_offset, in _copy_mc_to_iter()
778 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter()
804 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter_full()
824 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter_nocache()
858 memcpy_page_flushcache((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter_flushcache()
883 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter_full_nocache()
991 memzero_page(v.bv_page, v.bv_offset, v.bv_len), in iov_iter_zero()
1014 memcpy_from_page((p += v.bv_len) - v.bv_len, v.bv_page, in iov_iter_copy_from_user_atomic()
1354 get_page(*pages = v.bv_page); in iov_iter_get_pages()
[all …]
/kernel/linux/linux-5.10/drivers/md/
Ddm-ebs-target.c74 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec()
77 pa = page_address(bv->bv_page) + bv->bv_offset; in __ebs_rw_bvec()
100 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec()
102 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec()
Ddm-log-writes.c203 if (block->vecs[i].bv_page) in free_pending_block()
204 __free_page(block->vecs[i].bv_page); in free_pending_block()
384 ret = bio_add_page(bio, block->vecs[i].bv_page, in log_one_block()
401 ret = bio_add_page(bio, block->vecs[i].bv_page, in log_one_block()
768 src = kmap_atomic(bv.bv_page); in log_writes_map()
773 block->vecs[i].bv_page = page; in log_writes_map()
/kernel/linux/linux-5.10/fs/9p/
Dvfs_addr.c42 struct bio_vec bvec = {.bv_page = page, .bv_len = PAGE_SIZE}; in v9fs_fid_readpage()
162 bvec.bv_page = page; in v9fs_vfs_writepage_locked()
/kernel/linux/linux-5.10/drivers/block/zram/
Dzram_drv.c552 if (!bio_add_page(bio, bvec->bv_page, bvec->bv_len, bvec->bv_offset)) { in read_from_bdev_async()
612 bvec.bv_page = page; in writeback_store()
668 bio_add_page(&bio, bvec.bv_page, bvec.bv_len, in writeback_store()
1254 bvec.bv_page = page; in __zram_bvec_read()
1321 page = bvec->bv_page; in zram_bvec_read()
1334 void *dst = kmap_atomic(bvec->bv_page); in zram_bvec_read()
1357 struct page *page = bvec->bv_page; in __zram_bvec_write()
1490 src = kmap_atomic(bvec->bv_page); in zram_bvec_write()
1496 vec.bv_page = page; in zram_bvec_write()
1559 flush_dcache_page(bvec->bv_page); in zram_bvec_rw()
[all …]
/kernel/linux/linux-5.10/drivers/nvdimm/
Dblk.c92 iobuf = kmap_atomic(bv.bv_page); in nd_blk_rw_integrity()
187 err = nsblk_do_bvec(nsblk, bip, bvec.bv_page, len, in nd_blk_submit_bio()
/kernel/linux/linux-5.10/fs/btrfs/
Dcompression.c262 SetPageChecked(bvec->bv_page); in end_compressed_bio_read()
498 return page_offset(last->bv_page) + last->bv_len + last->bv_offset; in bio_end_offset()
1254 start_byte = page_offset(bvec.bv_page) - disk_start; in btrfs_decompress_buf2page()
1278 kaddr = kmap_atomic(bvec.bv_page); in btrfs_decompress_buf2page()
1281 flush_dcache_page(bvec.bv_page); in btrfs_decompress_buf2page()
1293 start_byte = page_offset(bvec.bv_page) - disk_start; in btrfs_decompress_buf2page()

12345