/drivers/block/ |
D | n64cart.c | 69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec() 79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec() 83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec() 97 pos += bvec.bv_len; in n64cart_submit_bio()
|
D | loop.c | 314 iov_iter_bvec(&i, WRITE, bvec, 1, bvec->bv_len); in lo_write_bvec() 320 if (likely(bw == bvec->bv_len)) in lo_write_bvec() 325 (unsigned long long)*ppos, bvec->bv_len); in lo_write_bvec() 367 bvec.bv_offset, bvec.bv_len, pos >> 9); in lo_write_transfer() 373 b.bv_len = bvec.bv_len; in lo_write_transfer() 392 iov_iter_bvec(&i, READ, &bvec, 1, bvec.bv_len); in lo_read_simple() 399 if (len != bvec.bv_len) { in lo_read_simple() 431 b.bv_len = bvec.bv_len; in lo_read_transfer() 433 iov_iter_bvec(&i, READ, &b, 1, b.bv_len); in lo_read_transfer() 447 if (len != bvec.bv_len) { in lo_read_transfer()
|
D | brd.c | 291 unsigned int len = bvec.bv_len; in brd_submit_bio()
|
D | nbd.c | 649 req, bvec.bv_len); in nbd_send_cmd() 650 iov_iter_bvec(&from, WRITE, &bvec, 1, bvec.bv_len); in nbd_send_cmd() 771 iov_iter_bvec(&to, READ, &bvec, 1, bvec.bv_len); in nbd_read_stat() 790 req, bvec.bv_len); in nbd_read_stat()
|
/drivers/md/ |
D | dm-ebs-target.c | 69 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() local 74 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec() 80 while (bv_len) { in __ebs_rw_bvec() 81 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); in __ebs_rw_bvec() 84 if (rw == READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) in __ebs_rw_bvec() 111 bv_len -= cur_len; in __ebs_rw_bvec()
|
D | dm-log-writes.c | 384 block->vecs[i].bv_len, 0); in log_one_block() 385 if (ret != block->vecs[i].bv_len) { in log_one_block() 402 block->vecs[i].bv_len, 0); in log_one_block() 403 if (ret != block->vecs[i].bv_len) { in log_one_block() 409 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block() 770 memcpy(dst, src + bv.bv_offset, bv.bv_len); in log_writes_map() 774 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
|
D | dm-user.c | 483 bvec.bv_len, iter); in bio_copy_from_iter() 497 if (ret < bvec.bv_len) in bio_copy_from_iter() 514 bvec.bv_len, iter); in bio_copy_to_iter() 523 if (ret < bvec.bv_len) in bio_copy_to_iter()
|
D | dm-io.c | 215 *len = bvec.bv_len; in bio_get_page() 219 dp->context_bi.bi_sector = (sector_t)bvec.bv_len; in bio_get_page()
|
D | dm-integrity.c | 1779 } while (pos < bv_copy.bv_len && sectors_to_process && checksums != checksums_onstack); in integrity_metadata() 1800 if (unlikely(pos < bv_copy.bv_len)) { in integrity_metadata() 1802 bv_copy.bv_len -= pos; in integrity_metadata() 1825 this_len = min(biv.bv_len, data_to_process); in integrity_metadata() 1900 if (unlikely(bv.bv_len & ((ic->sectors_per_block << SECTOR_SHIFT) - 1))) { in dm_integrity_map() 1902 bv.bv_offset, bv.bv_len, ic->sectors_per_block); in dm_integrity_map() 1953 if (unlikely(bv.bv_len >> SECTOR_SHIFT > n_sectors)) in __journal_read_write() 1954 bv.bv_len = n_sectors << SECTOR_SHIFT; in __journal_read_write() 1955 n_sectors -= bv.bv_len >> SECTOR_SHIFT; in __journal_read_write() 1956 bio_advance_iter(bio, &bio->bi_iter, bv.bv_len); in __journal_read_write() [all …]
|
D | dm-verity-target.c | 389 len = bv.bv_len; in verity_for_io_block() 434 len = bv.bv_len; in verity_for_bv_block()
|
/drivers/md/bcache/ |
D | util.c | 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map() 251 base += bv->bv_len; in bch_bio_map() 254 size -= bv->bv_len; in bch_bio_map()
|
D | debug.c | 138 bv.bv_len), in bch_data_verify() 145 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
|
/drivers/s390/block/ |
D | dasd_fba.c | 462 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular() 465 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular() 466 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular() 467 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular() 509 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular() 513 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular() 587 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp() 598 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
|
D | dasd_diag.c | 536 if (bv.bv_len & (blksize - 1)) in dasd_diag_build_cp() 539 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_diag_build_cp() 556 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_diag_build_cp()
|
D | dcssblk.c | 899 if (unlikely((page_addr & 4095) != 0) || (bvec.bv_len & 4095) != 0) in dcssblk_submit_bio() 904 bvec.bv_len); in dcssblk_submit_bio() 907 bvec.bv_len); in dcssblk_submit_bio() 909 bytes_done += bvec.bv_len; in dcssblk_submit_bio()
|
D | dasd_eckd.c | 3295 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_eckd_ese_read() 3950 if (bv.bv_len & (blksize - 1)) in dasd_eckd_build_cp_cmd_single() 3953 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_eckd_build_cp_cmd_single() 3954 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_eckd_build_cp_cmd_single() 3955 cidaw += bv.bv_len >> (block->s2b_shift + 9); in dasd_eckd_build_cp_cmd_single() 4030 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_eckd_build_cp_cmd_single() 4034 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_eckd_build_cp_cmd_single() 4193 seg_len = bv.bv_len; in dasd_eckd_build_cp_cmd_track() 4536 seg_len = bv.bv_len; in dasd_eckd_build_cp_tpm_track() 4570 dst, bv.bv_len); in dasd_eckd_build_cp_tpm_track() [all …]
|
/drivers/block/zram/ |
D | zram_drv.c | 151 return bvec->bv_len != PAGE_SIZE; in is_partial_io() 186 *index += (*offset + bvec->bv_len) / PAGE_SIZE; in update_position() 187 *offset = (*offset + bvec->bv_len) % PAGE_SIZE; in update_position() 596 if (!bio_add_page(bio, bvec->bv_page, bvec->bv_len, bvec->bv_offset)) { in read_from_bdev_async() 670 bvec.bv_len = PAGE_SIZE; in writeback_store() 725 bio_add_page(&bio, bvec.bv_page, bvec.bv_len, in writeback_store() 1243 bvec.bv_len = PAGE_SIZE; in __zram_bvec_read() 1312 memcpy(dst + bvec->bv_offset, src + offset, bvec->bv_len); in zram_bvec_read() 1466 memcpy(dst + offset, src + bvec->bv_offset, bvec->bv_len); in zram_bvec_write() 1471 vec.bv_len = PAGE_SIZE; in zram_bvec_write() [all …]
|
/drivers/block/rsxx/ |
D | dma.c | 674 unsigned int bv_len; in rsxx_dma_queue_bio() local 692 bv_len = bio->bi_iter.bi_size; in rsxx_dma_queue_bio() 694 while (bv_len > 0) { in rsxx_dma_queue_bio() 706 bv_len -= RSXX_HW_BLK_SIZE; in rsxx_dma_queue_bio() 710 bv_len = bvec.bv_len; in rsxx_dma_queue_bio() 713 while (bv_len > 0) { in rsxx_dma_queue_bio() 717 dma_len = min(bv_len, in rsxx_dma_queue_bio() 732 bv_len -= dma_len; in rsxx_dma_queue_bio()
|
/drivers/xen/ |
D | biomerge.c | 15 return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2; in xen_biovec_phys_mergeable()
|
/drivers/nvme/target/ |
D | io-cmd-file.c | 80 bv->bv_len = sg->length; in nvmet_file_init_bvec() 150 len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io() 151 total_len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io()
|
/drivers/nvdimm/ |
D | blk.c | 91 cur_len = min(len, bv.bv_len); in nd_blk_rw_integrity() 184 unsigned int len = bvec.bv_len; in nd_blk_submit_bio()
|
D | pmem.c | 213 iter.bi_sector, bvec.bv_len); in pmem_submit_bio() 216 iter.bi_sector, bvec.bv_len); in pmem_submit_bio()
|
/drivers/scsi/ |
D | aha1542.c | 272 bv.bv_len); in aha1542_free_cmd() 273 buf += bv.bv_len; in aha1542_free_cmd() 458 bv.bv_len); in aha1542_queuecommand() 459 buf += bv.bv_len; in aha1542_queuecommand()
|
/drivers/target/ |
D | target_core_file.c | 281 aio_cmd->bvecs[i].bv_len = sg->length; in fd_execute_rw_aio() 329 bvec[i].bv_len = sg->length; in fd_do_rw() 466 bvec[i].bv_len = cmd->t_data_sg[0].length; in fd_execute_write_same()
|
/drivers/block/aoe/ |
D | aoecmd.c | 304 bv.bv_offset, bv.bv_len); in skb_fillup() 1023 skb_copy_bits(skb, soff, p, bv.bv_len); in bvcpy() 1025 soff += bv.bv_len; in bvcpy()
|