Home
last modified time | relevance | path

Searched refs:bv_len (Results 1 – 25 of 44) sorted by relevance

12

/drivers/block/
Dn64cart.c69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec()
79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec()
83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec()
97 pos += bvec.bv_len; in n64cart_submit_bio()
Dloop.c314 iov_iter_bvec(&i, WRITE, bvec, 1, bvec->bv_len); in lo_write_bvec()
320 if (likely(bw == bvec->bv_len)) in lo_write_bvec()
325 (unsigned long long)*ppos, bvec->bv_len); in lo_write_bvec()
367 bvec.bv_offset, bvec.bv_len, pos >> 9); in lo_write_transfer()
373 b.bv_len = bvec.bv_len; in lo_write_transfer()
392 iov_iter_bvec(&i, READ, &bvec, 1, bvec.bv_len); in lo_read_simple()
399 if (len != bvec.bv_len) { in lo_read_simple()
431 b.bv_len = bvec.bv_len; in lo_read_transfer()
433 iov_iter_bvec(&i, READ, &b, 1, b.bv_len); in lo_read_transfer()
447 if (len != bvec.bv_len) { in lo_read_transfer()
Dbrd.c291 unsigned int len = bvec.bv_len; in brd_submit_bio()
Dnbd.c649 req, bvec.bv_len); in nbd_send_cmd()
650 iov_iter_bvec(&from, WRITE, &bvec, 1, bvec.bv_len); in nbd_send_cmd()
771 iov_iter_bvec(&to, READ, &bvec, 1, bvec.bv_len); in nbd_read_stat()
790 req, bvec.bv_len); in nbd_read_stat()
/drivers/md/
Ddm-ebs-target.c69 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() local
74 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec()
80 while (bv_len) { in __ebs_rw_bvec()
81 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); in __ebs_rw_bvec()
84 if (rw == READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) in __ebs_rw_bvec()
111 bv_len -= cur_len; in __ebs_rw_bvec()
Ddm-log-writes.c384 block->vecs[i].bv_len, 0); in log_one_block()
385 if (ret != block->vecs[i].bv_len) { in log_one_block()
402 block->vecs[i].bv_len, 0); in log_one_block()
403 if (ret != block->vecs[i].bv_len) { in log_one_block()
409 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block()
770 memcpy(dst, src + bv.bv_offset, bv.bv_len); in log_writes_map()
774 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
Ddm-user.c483 bvec.bv_len, iter); in bio_copy_from_iter()
497 if (ret < bvec.bv_len) in bio_copy_from_iter()
514 bvec.bv_len, iter); in bio_copy_to_iter()
523 if (ret < bvec.bv_len) in bio_copy_to_iter()
Ddm-io.c215 *len = bvec.bv_len; in bio_get_page()
219 dp->context_bi.bi_sector = (sector_t)bvec.bv_len; in bio_get_page()
Ddm-integrity.c1779 } while (pos < bv_copy.bv_len && sectors_to_process && checksums != checksums_onstack); in integrity_metadata()
1800 if (unlikely(pos < bv_copy.bv_len)) { in integrity_metadata()
1802 bv_copy.bv_len -= pos; in integrity_metadata()
1825 this_len = min(biv.bv_len, data_to_process); in integrity_metadata()
1900 if (unlikely(bv.bv_len & ((ic->sectors_per_block << SECTOR_SHIFT) - 1))) { in dm_integrity_map()
1902 bv.bv_offset, bv.bv_len, ic->sectors_per_block); in dm_integrity_map()
1953 if (unlikely(bv.bv_len >> SECTOR_SHIFT > n_sectors)) in __journal_read_write()
1954 bv.bv_len = n_sectors << SECTOR_SHIFT; in __journal_read_write()
1955 n_sectors -= bv.bv_len >> SECTOR_SHIFT; in __journal_read_write()
1956 bio_advance_iter(bio, &bio->bi_iter, bv.bv_len); in __journal_read_write()
[all …]
Ddm-verity-target.c389 len = bv.bv_len; in verity_for_io_block()
434 len = bv.bv_len; in verity_for_bv_block()
/drivers/md/bcache/
Dutil.c244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map()
251 base += bv->bv_len; in bch_bio_map()
254 size -= bv->bv_len; in bch_bio_map()
Ddebug.c138 bv.bv_len), in bch_data_verify()
145 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
/drivers/s390/block/
Ddasd_fba.c462 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular()
465 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular()
466 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular()
467 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular()
509 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular()
513 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular()
587 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp()
598 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
Ddasd_diag.c536 if (bv.bv_len & (blksize - 1)) in dasd_diag_build_cp()
539 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_diag_build_cp()
556 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_diag_build_cp()
Ddcssblk.c899 if (unlikely((page_addr & 4095) != 0) || (bvec.bv_len & 4095) != 0) in dcssblk_submit_bio()
904 bvec.bv_len); in dcssblk_submit_bio()
907 bvec.bv_len); in dcssblk_submit_bio()
909 bytes_done += bvec.bv_len; in dcssblk_submit_bio()
Ddasd_eckd.c3295 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_eckd_ese_read()
3950 if (bv.bv_len & (blksize - 1)) in dasd_eckd_build_cp_cmd_single()
3953 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_eckd_build_cp_cmd_single()
3954 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_eckd_build_cp_cmd_single()
3955 cidaw += bv.bv_len >> (block->s2b_shift + 9); in dasd_eckd_build_cp_cmd_single()
4030 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_eckd_build_cp_cmd_single()
4034 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_eckd_build_cp_cmd_single()
4193 seg_len = bv.bv_len; in dasd_eckd_build_cp_cmd_track()
4536 seg_len = bv.bv_len; in dasd_eckd_build_cp_tpm_track()
4570 dst, bv.bv_len); in dasd_eckd_build_cp_tpm_track()
[all …]
/drivers/block/zram/
Dzram_drv.c151 return bvec->bv_len != PAGE_SIZE; in is_partial_io()
186 *index += (*offset + bvec->bv_len) / PAGE_SIZE; in update_position()
187 *offset = (*offset + bvec->bv_len) % PAGE_SIZE; in update_position()
596 if (!bio_add_page(bio, bvec->bv_page, bvec->bv_len, bvec->bv_offset)) { in read_from_bdev_async()
670 bvec.bv_len = PAGE_SIZE; in writeback_store()
725 bio_add_page(&bio, bvec.bv_page, bvec.bv_len, in writeback_store()
1243 bvec.bv_len = PAGE_SIZE; in __zram_bvec_read()
1312 memcpy(dst + bvec->bv_offset, src + offset, bvec->bv_len); in zram_bvec_read()
1466 memcpy(dst + offset, src + bvec->bv_offset, bvec->bv_len); in zram_bvec_write()
1471 vec.bv_len = PAGE_SIZE; in zram_bvec_write()
[all …]
/drivers/block/rsxx/
Ddma.c674 unsigned int bv_len; in rsxx_dma_queue_bio() local
692 bv_len = bio->bi_iter.bi_size; in rsxx_dma_queue_bio()
694 while (bv_len > 0) { in rsxx_dma_queue_bio()
706 bv_len -= RSXX_HW_BLK_SIZE; in rsxx_dma_queue_bio()
710 bv_len = bvec.bv_len; in rsxx_dma_queue_bio()
713 while (bv_len > 0) { in rsxx_dma_queue_bio()
717 dma_len = min(bv_len, in rsxx_dma_queue_bio()
732 bv_len -= dma_len; in rsxx_dma_queue_bio()
/drivers/xen/
Dbiomerge.c15 return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2; in xen_biovec_phys_mergeable()
/drivers/nvme/target/
Dio-cmd-file.c80 bv->bv_len = sg->length; in nvmet_file_init_bvec()
150 len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io()
151 total_len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io()
/drivers/nvdimm/
Dblk.c91 cur_len = min(len, bv.bv_len); in nd_blk_rw_integrity()
184 unsigned int len = bvec.bv_len; in nd_blk_submit_bio()
Dpmem.c213 iter.bi_sector, bvec.bv_len); in pmem_submit_bio()
216 iter.bi_sector, bvec.bv_len); in pmem_submit_bio()
/drivers/scsi/
Daha1542.c272 bv.bv_len); in aha1542_free_cmd()
273 buf += bv.bv_len; in aha1542_free_cmd()
458 bv.bv_len); in aha1542_queuecommand()
459 buf += bv.bv_len; in aha1542_queuecommand()
/drivers/target/
Dtarget_core_file.c281 aio_cmd->bvecs[i].bv_len = sg->length; in fd_execute_rw_aio()
329 bvec[i].bv_len = sg->length; in fd_do_rw()
466 bvec[i].bv_len = cmd->t_data_sg[0].length; in fd_execute_write_same()
/drivers/block/aoe/
Daoecmd.c304 bv.bv_offset, bv.bv_len); in skb_fillup()
1023 skb_copy_bits(skb, soff, p, bv.bv_len); in bvcpy()
1025 soff += bv.bv_len; in bvcpy()

12