Home
last modified time | relevance | path

Searched refs:bi_size (Results 1 – 25 of 53) sorted by relevance

123

/drivers/md/
Ddm-bow.c110 return bi_iter->bi_sector + bi_iter->bi_size / SECTOR_SIZE; in bvec_top()
141 < bi_iter->bi_size >> SECTOR_SHIFT) in find_first_overlapping_range()
142 bi_iter->bi_size = (range_top(br) - bi_iter->bi_sector) in find_first_overlapping_range()
197 bi_iter->bi_size = (range_top(*br) - (*br)->sector) in split_range()
352 bi_iter.bi_size = bc->block_size; in backup_log_sector()
356 if (bi_iter.bi_size != bc->block_size) { in backup_log_sector()
430 bi_iter.bi_size = bc->block_size; in prepare_log()
445 bi_iter.bi_size = bc->block_size; in prepare_log()
464 bi_iter.bi_size = bc->block_size; in prepare_log()
500 bi_iter.bi_size = bc->block_size; in find_sector0_current()
[all …]
Ddm.c488 return ((bio->bi_opf & REQ_PREFLUSH) && bio->bi_iter.bi_size); in bio_is_flush_with_data()
495 unsigned int bi_size; in dm_io_acct() local
500 bi_size = bio->bi_iter.bi_size; in dm_io_acct()
501 bio->bi_iter.bi_size = 0; in dm_io_acct()
516 bio->bi_iter.bi_size = bi_size; in dm_io_acct()
1180 unsigned bi_size = bio->bi_iter.bi_size >> SECTOR_SHIFT; in dm_accept_partial_bio() local
1185 BUG_ON(bi_size > *tio->len_ptr); in dm_accept_partial_bio()
1186 BUG_ON(n_sectors > bi_size); in dm_accept_partial_bio()
1188 *tio->len_ptr -= bi_size - n_sectors; in dm_accept_partial_bio()
1189 bio->bi_iter.bi_size = n_sectors << SECTOR_SHIFT; in dm_accept_partial_bio()
[all …]
Dmd-faulty.c68 b->bi_iter.bi_size = bio->bi_iter.bi_size; in faulty_fail()
Ddm-writecache.c1337 if (next_boundary < bio->bi_iter.bi_size >> SECTOR_SHIFT) in writecache_map_remap_origin()
1354 if (bio->bi_iter.bi_size) in writecache_map_read()
1367 wc->stats.reads += (bio->bi_iter.bi_size - wc->block_size) >> wc->block_size_bits; in writecache_map_read()
1381 while (bio_size < bio->bi_iter.bi_size) { in writecache_bio_copy_ssd()
1415 wc->stats.writes += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_bio_copy_ssd()
1416 wc->stats.writes_allocate += (bio->bi_iter.bi_size - wc->block_size) >> wc->block_size_bits; in writecache_bio_copy_ssd()
1434 wc->stats.writes += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_map_write()
1462 wc->stats.writes_around += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_map_write()
1463 wc->stats.writes += bio->bi_iter.bi_size >> wc->block_size_bits; in writecache_map_write()
1482 } while (bio->bi_iter.bi_size); in writecache_map_write()
[all …]
Ddm-log-writes.c225 bio->bi_iter.bi_size = 0; in write_metadata()
284 bio->bi_iter.bi_size = 0; in write_inline_data()
371 bio->bi_iter.bi_size = 0; in log_one_block()
394 bio->bi_iter.bi_size = 0; in log_one_block()
Ddm-zoned-target.c136 clone->bi_iter.bi_size = dmz_blk2sect(nr_blocks) << SECTOR_SHIFT; in dmz_submit_bio()
140 bio_advance(bio, clone->bi_iter.bi_size); in dmz_submit_bio()
160 swap(bio->bi_iter.bi_size, size); in dmz_handle_read_zero()
162 swap(bio->bi_iter.bi_size, size); in dmz_handle_read_zero()
Ddm-io.c358 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region()
367 bio->bi_iter.bi_size = num_sectors << SECTOR_SHIFT; in do_region()
Ddm-crypt.c1152 bip->bip_iter.bi_size = tag_len; in dm_crypt_integrity_io_alloc()
1559 while (ctx->iter_in.bi_size && ctx->iter_out.bi_size) { in crypt_convert()
1964 BUG_ON(io->ctx.iter_out.bi_size); in kcryptd_crypt_write_io_submit()
2060 clone = crypt_alloc_buffer(io, io->base_bio->bi_iter.bi_size); in kcryptd_crypt_write_convert()
3399 if (unlikely(bio->bi_iter.bi_size > (BIO_MAX_VECS << PAGE_SHIFT)) && in crypt_map()
3410 if (unlikely(bio->bi_iter.bi_size & (cc->sector_size - 1))) in crypt_map()
Ddm-flakey.c315 (unsigned long long)bio->bi_iter.bi_sector, bio->bi_iter.bi_size); in corrupt_bio_data()
Ddm-integrity.c1731 unsigned bi_size = dio->bio_details.bi_iter.bi_size; in integrity_metadata() local
1736 while (bi_size) { in integrity_metadata()
1737 unsigned this_step_blocks = bi_size >> (SECTOR_SHIFT + ic->sb->log2_sectors_per_block); in integrity_metadata()
1752 bi_size -= this_step_blocks << (SECTOR_SHIFT + ic->sb->log2_sectors_per_block); in integrity_metadata()
1916 if (unlikely(wanted_tag_size != bip->bip_iter.bi_size)) { in dm_integrity_map()
1918 bip->bip_iter.bi_size, wanted_tag_size); in dm_integrity_map()
2079 if (unlikely(bio->bi_iter.bi_size)) { in __journal_read_write()
2265 bio->bi_iter.bi_size = dio->range.n_sectors << SECTOR_SHIFT; in dm_integrity_map_continue()
/drivers/block/aoe/
Daoecmd.c343 ah->scnt = f->iter.bi_size >> 9; in ata_rw_frameinit()
355 skb->len += f->iter.bi_size; in ata_rw_frameinit()
356 skb->data_len = f->iter.bi_size; in ata_rw_frameinit()
357 skb->truesize += f->iter.bi_size; in ata_rw_frameinit()
386 f->iter.bi_size = min_t(unsigned long, in aoecmd_ata_rw()
388 f->iter.bi_size); in aoecmd_ata_rw()
389 bio_advance_iter(buf->bio, &buf->iter, f->iter.bi_size); in aoecmd_ata_rw()
391 if (!buf->iter.bi_size) in aoecmd_ata_rw()
601 f->iter.bi_size = t->d->maxbcnt ? t->d->maxbcnt : DEFAULTBCNT; in probe()
604 for (frag = 0, n = f->iter.bi_size; n > 0; ++frag, n -= m) { in probe()
[all …]
/drivers/md/bcache/
Ddebug.c55 bio->bi_iter.bi_size = KEY_SIZE(&v->key) << 9; in bch_btree_verify()
120 check->bi_iter.bi_size = bio->bi_iter.bi_size; in bch_data_verify()
128 citer.bi_size = UINT_MAX; in bch_data_verify()
Drequest.c124 bio->bi_iter.bi_size -= sectors << 9; in bch_data_invalidate()
426 if (i->sequential + bio->bi_iter.bi_size > i->sequential) in check_should_bypass()
427 i->sequential += bio->bi_iter.bi_size; in check_should_bypass()
838 s->iop.bio->bi_iter.bi_size = s->insert_bio_sectors << 9; in cached_dev_read_done()
924 cache_bio->bi_iter.bi_size = s->insert_bio_sectors << 9; in cached_dev_cache_miss()
1212 if (!bio->bi_iter.bi_size) { in cached_dev_submit_bio()
1260 swap(bio->bi_iter.bi_size, bytes); in flash_dev_cache_miss()
1262 swap(bio->bi_iter.bi_size, bytes); in flash_dev_cache_miss()
1266 if (!bio->bi_iter.bi_size) in flash_dev_cache_miss()
1300 if (!bio->bi_iter.bi_size) { in flash_dev_submit_bio()
Dutil.c233 size_t size = bio->bi_iter.bi_size; in bch_bio_map()
236 BUG_ON(!bio->bi_iter.bi_size); in bch_bio_map()
Djournal.c59 bio->bi_iter.bi_size = len << 9; in journal_read_bucket()
624 bio->bi_iter.bi_size = bucket_bytes(ca); in do_journal_discard()
800 bio->bi_iter.bi_size = sectors << 9; in journal_write_unlocked()
Dmovinggc.c87 bio->bi_iter.bi_size = KEY_SIZE(&io->w->key) << 9; in moving_init()
/drivers/net/ethernet/intel/iavf/
Diavf_txrx.c59 unsigned long bi_size; in iavf_clean_tx_ring() local
70 bi_size = sizeof(struct iavf_tx_buffer) * tx_ring->count; in iavf_clean_tx_ring()
71 memset(tx_ring->tx_bi, 0, bi_size); in iavf_clean_tx_ring()
620 int bi_size; in iavf_setup_tx_descriptors() local
627 bi_size = sizeof(struct iavf_tx_buffer) * tx_ring->count; in iavf_setup_tx_descriptors()
628 tx_ring->tx_bi = kzalloc(bi_size, GFP_KERNEL); in iavf_setup_tx_descriptors()
660 unsigned long bi_size; in iavf_clean_rx_ring() local
700 bi_size = sizeof(struct iavf_rx_buffer) * rx_ring->count; in iavf_clean_rx_ring()
701 memset(rx_ring->rx_bi, 0, bi_size); in iavf_clean_rx_ring()
739 int bi_size; in iavf_setup_rx_descriptors() local
[all …]
/drivers/block/rsxx/
Ddev.c147 if (bio->bi_iter.bi_size == 0) { in rsxx_submit_bio()
167 (u64)bio->bi_iter.bi_sector << 9, bio->bi_iter.bi_size); in rsxx_submit_bio()
/drivers/block/rnbd/
Drnbd-proto.h166 __le32 bi_size; member
DREADME77 sector to be accessed, read-write flags and bi_size.
Drnbd-srv.c163 bio->bi_iter.bi_size = le32_to_cpu(msg->bi_size); in process_rdma()
/drivers/target/
Dtarget_core_iblock.c683 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); in iblock_alloc_bip()
688 pr_debug("IBLOCK BIP Size: %u Sector: %llu\n", bip->bip_iter.bi_size, in iblock_alloc_bip()
691 resid = bip->bip_iter.bi_size; in iblock_alloc_bip()
/drivers/nvme/target/
Dio-cmd-bdev.c200 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); in nvmet_bdev_alloc_bip()
205 resid = bip->bip_iter.bi_size; in nvmet_bdev_alloc_bip()
/drivers/block/
Dpktcdvd.c898 bio->bi_iter.bi_size >> 10; in pkt_iosched_process_queue()
1005 int num_frames = bio->bi_iter.bi_size / CD_FRAMESIZE; in pkt_gather_data()
1178 pkt->write_size += bio->bi_iter.bi_size / CD_FRAMESIZE; in pkt_handle_queue()
1216 if (!src_iter.bi_size) { in bio_list_copy_data()
1224 if (!dst_iter.bi_size) { in bio_list_copy_data()
2342 bio->bi_iter.bi_size / CD_FRAMESIZE; in pkt_make_request_write()
2437 if (!bio->bi_iter.bi_size || (bio->bi_iter.bi_size % CD_FRAMESIZE)) { in pkt_submit_bio()
/drivers/net/ethernet/intel/i40e/
Di40e_txrx.c790 unsigned long bi_size; in i40e_clean_tx_ring() local
806 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; in i40e_clean_tx_ring()
807 memset(tx_ring->tx_bi, 0, bi_size); in i40e_clean_tx_ring()
1424 int bi_size; in i40e_setup_tx_descriptors() local
1431 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; in i40e_setup_tx_descriptors()
1432 tx_ring->tx_bi = kzalloc(bi_size, GFP_KERNEL); in i40e_setup_tx_descriptors()

123