/drivers/md/ |
D | dm-bow.c | 110 return bi_iter->bi_sector + bi_iter->bi_size / SECTOR_SIZE; in bvec_top() 141 < bi_iter->bi_size >> SECTOR_SHIFT) in find_first_overlapping_range() 142 bi_iter->bi_size = (range_top(br) - bi_iter->bi_sector) in find_first_overlapping_range() 197 bi_iter->bi_size = (range_top(*br) - (*br)->sector) in split_range() 352 bi_iter.bi_size = bc->block_size; in backup_log_sector() 356 if (bi_iter.bi_size != bc->block_size) { in backup_log_sector() 430 bi_iter.bi_size = bc->block_size; in prepare_log() 445 bi_iter.bi_size = bc->block_size; in prepare_log() 464 bi_iter.bi_size = bc->block_size; in prepare_log() 500 bi_iter.bi_size = bc->block_size; in find_sector0_current() [all …]
|
D | dm.c | 488 return ((bio->bi_opf & REQ_PREFLUSH) && bio->bi_iter.bi_size); in bio_is_flush_with_data() 495 unsigned int bi_size; in dm_io_acct() local 500 bi_size = bio->bi_iter.bi_size; in dm_io_acct() 501 bio->bi_iter.bi_size = 0; in dm_io_acct() 516 bio->bi_iter.bi_size = bi_size; in dm_io_acct() 1180 unsigned bi_size = bio->bi_iter.bi_size >> SECTOR_SHIFT; in dm_accept_partial_bio() local 1185 BUG_ON(bi_size > *tio->len_ptr); in dm_accept_partial_bio() 1186 BUG_ON(n_sectors > bi_size); in dm_accept_partial_bio() 1188 *tio->len_ptr -= bi_size - n_sectors; in dm_accept_partial_bio() 1189 bio->bi_iter.bi_size = n_sectors << SECTOR_SHIFT; in dm_accept_partial_bio() [all …]
|
D | md-faulty.c | 68 b->bi_iter.bi_size = bio->bi_iter.bi_size; in faulty_fail()
|
D | dm-writecache.c | 1337 if (next_boundary < bio->bi_iter.bi_size >> SECTOR_SHIFT) in writecache_map_remap_origin() 1354 if (bio->bi_iter.bi_size) in writecache_map_read() 1367 wc->stats.reads += (bio->bi_iter.bi_size - wc->block_size) >> wc->block_size_bits; in writecache_map_read() 1381 while (bio_size < bio->bi_iter.bi_size) { in writecache_bio_copy_ssd() 1415 wc->stats.writes += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_bio_copy_ssd() 1416 wc->stats.writes_allocate += (bio->bi_iter.bi_size - wc->block_size) >> wc->block_size_bits; in writecache_bio_copy_ssd() 1434 wc->stats.writes += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_map_write() 1462 wc->stats.writes_around += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_map_write() 1463 wc->stats.writes += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_map_write() 1482 } while (bio->bi_iter.bi_size); in writecache_map_write() [all …]
|
D | dm-log-writes.c | 225 bio->bi_iter.bi_size = 0; in write_metadata() 284 bio->bi_iter.bi_size = 0; in write_inline_data() 371 bio->bi_iter.bi_size = 0; in log_one_block() 394 bio->bi_iter.bi_size = 0; in log_one_block()
|
D | dm-zoned-target.c | 136 clone->bi_iter.bi_size = dmz_blk2sect(nr_blocks) << SECTOR_SHIFT; in dmz_submit_bio() 140 bio_advance(bio, clone->bi_iter.bi_size); in dmz_submit_bio() 160 swap(bio->bi_iter.bi_size, size); in dmz_handle_read_zero() 162 swap(bio->bi_iter.bi_size, size); in dmz_handle_read_zero()
|
D | dm-io.c | 358 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region() 367 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region()
|
D | dm-crypt.c | 1152 bip->bip_iter.bi_size = tag_len; in dm_crypt_integrity_io_alloc() 1559 while (ctx->iter_in.bi_size && ctx->iter_out.bi_size) { in crypt_convert() 1964 BUG_ON(io->ctx.iter_out.bi_size); in kcryptd_crypt_write_io_submit() 2060 clone = crypt_alloc_buffer(io, io->base_bio->bi_iter.bi_size); in kcryptd_crypt_write_convert() 3399 if (unlikely(bio->bi_iter.bi_size > (BIO_MAX_VECS << PAGE_SHIFT)) && in crypt_map() 3410 if (unlikely(bio->bi_iter.bi_size & (cc->sector_size - 1))) in crypt_map()
|
D | dm-flakey.c | 315 (unsigned long long)bio->bi_iter.bi_sector, bio->bi_iter.bi_size); in corrupt_bio_data()
|
D | dm-integrity.c | 1731 unsigned bi_size = dio->bio_details.bi_iter.bi_size; in integrity_metadata() local 1736 while (bi_size) { in integrity_metadata() 1737 unsigned this_step_blocks = bi_size >> (SECTOR_SHIFT + ic->sb->log2_sectors_per_block); in integrity_metadata() 1752 bi_size -= this_step_blocks << (SECTOR_SHIFT + ic->sb->log2_sectors_per_block); in integrity_metadata() 1916 if (unlikely(wanted_tag_size != bip->bip_iter.bi_size)) { in dm_integrity_map() 1918 bip->bip_iter.bi_size, wanted_tag_size); in dm_integrity_map() 2079 if (unlikely(bio->bi_iter.bi_size)) { in __journal_read_write() 2265 bio->bi_iter.bi_size = dio->range.n_sectors << SECTOR_SHIFT; in dm_integrity_map_continue()
|
/drivers/block/aoe/ |
D | aoecmd.c | 343 ah->scnt = f->iter.bi_size >> 9; in ata_rw_frameinit() 355 skb->len += f->iter.bi_size; in ata_rw_frameinit() 356 skb->data_len = f->iter.bi_size; in ata_rw_frameinit() 357 skb->truesize += f->iter.bi_size; in ata_rw_frameinit() 386 f->iter.bi_size = min_t(unsigned long, in aoecmd_ata_rw() 388 f->iter.bi_size); in aoecmd_ata_rw() 389 bio_advance_iter(buf->bio, &buf->iter, f->iter.bi_size); in aoecmd_ata_rw() 391 if (!buf->iter.bi_size) in aoecmd_ata_rw() 601 f->iter.bi_size = t->d->maxbcnt ? t->d->maxbcnt : DEFAULTBCNT; in probe() 604 for (frag = 0, n = f->iter.bi_size; n > 0; ++frag, n -= m) { in probe() [all …]
|
/drivers/md/bcache/ |
D | debug.c | 55 bio->bi_iter.bi_size = KEY_SIZE(&v->key) << 9; in bch_btree_verify() 120 check->bi_iter.bi_size = bio->bi_iter.bi_size; in bch_data_verify() 128 citer.bi_size = UINT_MAX; in bch_data_verify()
|
D | request.c | 124 bio->bi_iter.bi_size -= sectors << 9; in bch_data_invalidate() 426 if (i->sequential + bio->bi_iter.bi_size > i->sequential) in check_should_bypass() 427 i->sequential += bio->bi_iter.bi_size; in check_should_bypass() 838 s->iop.bio->bi_iter.bi_size = s->insert_bio_sectors << 9; in cached_dev_read_done() 924 cache_bio->bi_iter.bi_size = s->insert_bio_sectors << 9; in cached_dev_cache_miss() 1212 if (!bio->bi_iter.bi_size) { in cached_dev_submit_bio() 1260 swap(bio->bi_iter.bi_size, bytes); in flash_dev_cache_miss() 1262 swap(bio->bi_iter.bi_size, bytes); in flash_dev_cache_miss() 1266 if (!bio->bi_iter.bi_size) in flash_dev_cache_miss() 1300 if (!bio->bi_iter.bi_size) { in flash_dev_submit_bio()
|
D | util.c | 233 size_t size = bio->bi_iter.bi_size; in bch_bio_map() 236 BUG_ON(!bio->bi_iter.bi_size); in bch_bio_map()
|
D | journal.c | 59 bio->bi_iter.bi_size = len << 9; in journal_read_bucket() 624 bio->bi_iter.bi_size = bucket_bytes(ca); in do_journal_discard() 800 bio->bi_iter.bi_size = sectors << 9; in journal_write_unlocked()
|
D | movinggc.c | 87 bio->bi_iter.bi_size = KEY_SIZE(&io->w->key) << 9; in moving_init()
|
/drivers/net/ethernet/intel/iavf/ |
D | iavf_txrx.c | 59 unsigned long bi_size; in iavf_clean_tx_ring() local 70 bi_size = sizeof(struct iavf_tx_buffer) * tx_ring->count; in iavf_clean_tx_ring() 71 memset(tx_ring->tx_bi, 0, bi_size); in iavf_clean_tx_ring() 620 int bi_size; in iavf_setup_tx_descriptors() local 627 bi_size = sizeof(struct iavf_tx_buffer) * tx_ring->count; in iavf_setup_tx_descriptors() 628 tx_ring->tx_bi = kzalloc(bi_size, GFP_KERNEL); in iavf_setup_tx_descriptors() 660 unsigned long bi_size; in iavf_clean_rx_ring() local 700 bi_size = sizeof(struct iavf_rx_buffer) * rx_ring->count; in iavf_clean_rx_ring() 701 memset(rx_ring->rx_bi, 0, bi_size); in iavf_clean_rx_ring() 739 int bi_size; in iavf_setup_rx_descriptors() local [all …]
|
/drivers/block/rsxx/ |
D | dev.c | 147 if (bio->bi_iter.bi_size == 0) { in rsxx_submit_bio() 167 (u64)bio->bi_iter.bi_sector << 9, bio->bi_iter.bi_size); in rsxx_submit_bio()
|
/drivers/block/rnbd/ |
D | rnbd-proto.h | 166 __le32 bi_size; member
|
D | README | 77 sector to be accessed, read-write flags and bi_size.
|
D | rnbd-srv.c | 163 bio->bi_iter.bi_size = le32_to_cpu(msg->bi_size); in process_rdma()
|
/drivers/target/ |
D | target_core_iblock.c | 683 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); in iblock_alloc_bip() 688 pr_debug("IBLOCK BIP Size: %u Sector: %llu\n", bip->bip_iter.bi_size, in iblock_alloc_bip() 691 resid = bip->bip_iter.bi_size; in iblock_alloc_bip()
|
/drivers/nvme/target/ |
D | io-cmd-bdev.c | 200 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); in nvmet_bdev_alloc_bip() 205 resid = bip->bip_iter.bi_size; in nvmet_bdev_alloc_bip()
|
/drivers/block/ |
D | pktcdvd.c | 898 bio->bi_iter.bi_size >> 10; in pkt_iosched_process_queue() 1005 int num_frames = bio->bi_iter.bi_size / CD_FRAMESIZE; in pkt_gather_data() 1178 pkt->write_size += bio->bi_iter.bi_size / CD_FRAMESIZE; in pkt_handle_queue() 1216 if (!src_iter.bi_size) { in bio_list_copy_data() 1224 if (!dst_iter.bi_size) { in bio_list_copy_data() 2342 bio->bi_iter.bi_size / CD_FRAMESIZE; in pkt_make_request_write() 2437 if (!bio->bi_iter.bi_size || (bio->bi_iter.bi_size % CD_FRAMESIZE)) { in pkt_submit_bio()
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 790 unsigned long bi_size; in i40e_clean_tx_ring() local 806 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; in i40e_clean_tx_ring() 807 memset(tx_ring->tx_bi, 0, bi_size); in i40e_clean_tx_ring() 1424 int bi_size; in i40e_setup_tx_descriptors() local 1431 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; in i40e_setup_tx_descriptors() 1432 tx_ring->tx_bi = kzalloc(bi_size, GFP_KERNEL); in i40e_setup_tx_descriptors()
|