/drivers/md/ |
D | dm-bio-record.h | 25 struct bvec_iter bi_iter; member 36 bd->bi_iter = bio->bi_iter; in dm_bio_record() 48 bio->bi_iter = bd->bi_iter; in dm_bio_restore()
|
D | md-faulty.c | 68 b->bi_iter.bi_size = bio->bi_iter.bi_size; in faulty_fail() 69 b->bi_iter.bi_sector = bio->bi_iter.bi_sector; in faulty_fail() 179 if (check_sector(conf, bio->bi_iter.bi_sector, in faulty_make_request() 183 add_sector(conf, bio->bi_iter.bi_sector, in faulty_make_request() 191 if (check_sector(conf, bio->bi_iter.bi_sector, in faulty_make_request() 197 add_sector(conf, bio->bi_iter.bi_sector, in faulty_make_request() 202 add_sector(conf, bio->bi_iter.bi_sector, in faulty_make_request()
|
D | dm-ebs-target.c | 49 sector_t end_sector = __block_mod(bio->bi_iter.bi_sector, ec->u_bs) + bio_sectors(bio); in __nr_blocks() 144 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio() 168 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio() 194 block1 = __sector_to_block(ec, bio->bi_iter.bi_sector); in __ebs_process_bios() 199 if (__block_mod(bio->bi_iter.bi_sector, ec->u_bs)) in __ebs_process_bios() 365 bio->bi_iter.bi_sector = ec->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in ebs_map() 374 if (likely(__block_mod(bio->bi_iter.bi_sector, ec->u_bs) || in ebs_map()
|
D | md-multipath.c | 93 (unsigned long long)bio->bi_iter.bi_sector); in multipath_end_request() 125 mp_bh->bio.bi_iter.bi_sector += multipath->rdev->data_offset; in multipath_make_request() 304 bio->bi_iter.bi_sector = mp_bh->master_bio->bi_iter.bi_sector; in multipathd() 309 (unsigned long long)bio->bi_iter.bi_sector); in multipathd() 314 (unsigned long long)bio->bi_iter.bi_sector); in multipathd() 316 bio->bi_iter.bi_sector += in multipathd()
|
D | dm-writecache.c | 1246 struct bio_vec bv = bio_iter_iovec(bio, bio->bi_iter); in bio_copy_block() 1296 writecache_discard(wc, bio->bi_iter.bi_sector, in writecache_flush_thread() 1333 read_original_sector(wc, e) - bio->bi_iter.bi_sector; in writecache_map_remap_origin() 1334 if (next_boundary < bio->bi_iter.bi_size >> SECTOR_SHIFT) in writecache_map_remap_origin() 1346 e = writecache_find_entry(wc, bio->bi_iter.bi_sector, WFE_RETURN_FOLLOWING); in writecache_map_read() 1347 if (e && read_original_sector(wc, e) == bio->bi_iter.bi_sector) { in writecache_map_read() 1351 if (bio->bi_iter.bi_size) in writecache_map_read() 1357 bio->bi_iter.bi_sector = cache_sector(wc, e); in writecache_map_read() 1364 wc->stats.reads += (bio->bi_iter.bi_size - wc->block_size) >> wc->block_size_bits; in writecache_map_read() 1378 while (bio_size < bio->bi_iter.bi_size) { in writecache_bio_copy_ssd() [all …]
|
D | dm-stripe.c | 251 stripe_map_range_sector(sc, bio->bi_iter.bi_sector, in stripe_map_range() 257 bio->bi_iter.bi_sector = begin + in stripe_map_range() 259 bio->bi_iter.bi_size = to_bytes(end - begin); in stripe_map_range() 288 stripe_map_sector(sc, bio->bi_iter.bi_sector, in stripe_map() 289 &stripe, &bio->bi_iter.bi_sector); in stripe_map() 291 bio->bi_iter.bi_sector += sc->stripe[stripe].physical_start; in stripe_map()
|
D | md-linear.c | 211 sector_t bio_sector = bio->bi_iter.bi_sector; in linear_make_request() 242 bio->bi_iter.bi_sector = bio->bi_iter.bi_sector - in linear_make_request() 261 (unsigned long long)bio->bi_iter.bi_sector, in linear_make_request()
|
D | dm-log-writes.c | 221 bio->bi_iter.bi_size = 0; in write_metadata() 222 bio->bi_iter.bi_sector = sector; in write_metadata() 274 bio->bi_iter.bi_size = 0; in write_inline_data() 275 bio->bi_iter.bi_sector = sector; in write_inline_data() 355 bio->bi_iter.bi_size = 0; in log_one_block() 356 bio->bi_iter.bi_sector = sector; in log_one_block() 373 bio->bi_iter.bi_size = 0; in log_one_block() 374 bio->bi_iter.bi_sector = sector; in log_one_block() 702 block->sector = bio_to_dev_sectors(lc, bio->bi_iter.bi_sector); in log_writes_map()
|
D | dm-zone.c | 393 if ((clone->bi_iter.bi_sector & (zsectors - 1)) != zwp_offset) in dm_zone_map_bio_begin() 404 clone->bi_iter.bi_sector += zwp_offset; in dm_zone_map_bio_begin() 608 orig_bio->bi_iter.bi_sector += in dm_zone_endio() 609 clone->bi_iter.bi_sector & mask; in dm_zone_endio() 643 orig_bio->bi_iter.bi_sector += in dm_zone_endio()
|
D | dm-flakey.c | 287 bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in flakey_map_bio() 317 (unsigned long long)bio->bi_iter.bi_sector, bio->bi_iter.bi_size); in corrupt_bio_data()
|
D | dm-io-rewind.c | 142 dm_bio_rewind_iter(bio, &bio->bi_iter, bytes); in dm_bio_rewind() 156 orig->bi_iter.bi_size)); in dm_io_rewind()
|
D | raid10.c | 1152 bio->bi_iter.bi_sector < conf->reshape_progress && in regular_request_wait() 1153 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait() 1161 conf->reshape_progress <= bio->bi_iter.bi_sector || in regular_request_wait() 1162 conf->reshape_progress >= bio->bi_iter.bi_sector + in regular_request_wait() 1249 read_bio->bi_iter.bi_sector = r10_bio->devs[slot].addr + in raid10_read_request() 1294 mbio->bi_iter.bi_sector = (r10_bio->devs[n_copy].addr + in raid10_write_one_disk() 1413 bio->bi_iter.bi_sector, in raid10_write_request() 1425 bio->bi_iter.bi_sector, bio_end_sector(bio))) in raid10_write_request() 1437 ? (bio->bi_iter.bi_sector < conf->reshape_safe && in raid10_write_request() 1438 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request() [all …]
|
D | dm-dust.c | 230 bio->bi_iter.bi_sector = dd->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in dust_map() 233 r = dust_map_read(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map() 235 r = dust_map_write(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
|
D | raid1.c | 321 (unsigned long long) bio->bi_iter.bi_sector, in raid_end_bio_io() 546 (unsigned long long) mbio->bi_iter.bi_sector, in raid1_end_write_request() 1117 int size = bio->bi_iter.bi_size; in alloc_behind_master_bio() 1129 behind_bio->bi_iter.bi_size = size; in alloc_behind_master_bio() 1156 bio->bi_iter.bi_size); in alloc_behind_master_bio() 1191 r1_bio->sector = bio->bi_iter.bi_sector; in init_r1bio() 1244 if (!wait_read_barrier(conf, bio->bi_iter.bi_sector, in raid1_read_request() 1312 read_bio->bi_iter.bi_sector = r1_bio->sector + in raid1_read_request() 1343 bio->bi_iter.bi_sector, bio_end_sector(bio))) { in raid1_write_request() 1354 bio->bi_iter.bi_sector, in raid1_write_request() [all …]
|
D | dm-unstripe.c | 119 sector_t sector = bio->bi_iter.bi_sector; in map_to_core() 139 bio->bi_iter.bi_sector = map_to_core(ti, bio) + uc->physical_start; in unstripe_map()
|
D | raid0.c | 464 sector_t start = bio->bi_iter.bi_sector; in raid0_handle_discard() 480 zone->zone_end - bio->bi_iter.bi_sector, GFP_NOIO, in raid0_handle_discard() 565 sector_t bio_sector = bio->bi_iter.bi_sector; in raid0_map_submit_bio() 591 bio->bi_iter.bi_sector = sector + zone->dev_start + in raid0_map_submit_bio() 616 sector = bio->bi_iter.bi_sector; in raid0_make_request()
|
D | dm-linear.c | 92 bio->bi_iter.bi_sector = linear_map_sector(ti, bio->bi_iter.bi_sector); in linear_map()
|
D | dm-zoned-target.c | 133 clone->bi_iter.bi_sector = in dmz_submit_bio() 135 clone->bi_iter.bi_size = dmz_blk2sect(nr_blocks) << SECTOR_SHIFT; in dmz_submit_bio() 139 bio_advance(bio, clone->bi_iter.bi_size); in dmz_submit_bio() 159 swap(bio->bi_iter.bi_size, size); in dmz_handle_read_zero() 161 swap(bio->bi_iter.bi_size, size); in dmz_handle_read_zero() 630 sector_t sector = bio->bi_iter.bi_sector; in dmz_map()
|
D | dm-thin.c | 677 sector_t block_nr = bio->bi_iter.bi_sector; in get_bio_block() 694 sector_t b = bio->bi_iter.bi_sector; in get_bio_block_range() 695 sector_t e = b + (bio->bi_iter.bi_size >> SECTOR_SHIFT); in get_bio_block_range() 718 sector_t bi_sector = bio->bi_iter.bi_sector; in remap() 722 bio->bi_iter.bi_sector = in remap() 726 bio->bi_iter.bi_sector = (block * pool->sectors_per_block) + in remap() 1235 return bio->bi_iter.bi_size == in io_overlaps_block() 1863 if (bio_data_dir(bio) == WRITE && bio->bi_iter.bi_size) { in process_shared_bio() 1888 if (!bio->bi_iter.bi_size) { in provision_block() 1960 else if (bio->bi_iter.bi_sector < tc->origin_size) { in process_cell() [all …]
|
/drivers/md/bcache/ |
D | request.c | 114 bio_sectors(bio), (uint64_t) bio->bi_iter.bi_sector); in bch_data_invalidate() 123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate() 124 bio->bi_iter.bi_size -= sectors << 9; in bch_data_invalidate() 128 bio->bi_iter.bi_sector, in bch_data_invalidate() 220 SET_KEY_OFFSET(k, bio->bi_iter.bi_sector); in bch_data_insert_start() 397 if (bio->bi_iter.bi_sector & (c->cache->sb.block_size - 1) || in check_should_bypass() 416 hlist_for_each_entry(i, iohash(dc, bio->bi_iter.bi_sector), hash) in check_should_bypass() 417 if (i->last == bio->bi_iter.bi_sector && in check_should_bypass() 426 if (i->sequential + bio->bi_iter.bi_size > i->sequential) in check_should_bypass() 427 i->sequential += bio->bi_iter.bi_size; in check_should_bypass() [all …]
|
D | debug.c | 54 bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); in bch_btree_verify() 55 bio->bi_iter.bi_size = KEY_SIZE(&v->key) << 9; in bch_btree_verify() 120 check->bi_iter.bi_sector = bio->bi_iter.bi_sector; in bch_data_verify() 121 check->bi_iter.bi_size = bio->bi_iter.bi_size; in bch_data_verify() 141 (uint64_t) bio->bi_iter.bi_sector); in bch_data_verify()
|
D | util.c | 233 size_t size = bio->bi_iter.bi_size; in bch_bio_map() 236 BUG_ON(!bio->bi_iter.bi_size); in bch_bio_map()
|
D | journal.c | 57 bio->bi_iter.bi_sector = bucket + offset; in journal_read_bucket() 58 bio->bi_iter.bi_size = len << 9; in journal_read_bucket() 619 bio->bi_iter.bi_sector = bucket_to_sector(ca->set, in do_journal_discard() 621 bio->bi_iter.bi_size = bucket_bytes(ca); in do_journal_discard() 796 bio->bi_iter.bi_sector = PTR_OFFSET(k, i); in journal_write_unlocked() 797 bio->bi_iter.bi_size = sectors << 9; in journal_write_unlocked()
|
/drivers/block/ |
D | pktcdvd.c | 639 if (s <= tmp->bio->bi_iter.bi_sector) in pkt_rbtree_find() 648 if (s > tmp->bio->bi_iter.bi_sector) { in pkt_rbtree_find() 653 BUG_ON(s > tmp->bio->bi_iter.bi_sector); in pkt_rbtree_find() 664 sector_t s = node->bio->bi_iter.bi_sector; in pkt_rbtree_insert() 670 if (s < tmp->bio->bi_iter.bi_sector) in pkt_rbtree_insert() 851 if (bio && (bio->bi_iter.bi_sector == in pkt_iosched_process_queue() 884 bio->bi_iter.bi_size >> 10; in pkt_iosched_process_queue() 941 (unsigned long long)bio->bi_iter.bi_sector, bio->bi_status); in pkt_end_io_read() 991 int first_frame = (bio->bi_iter.bi_sector - pkt->sector) / in pkt_gather_data() 993 int num_frames = bio->bi_iter.bi_size / CD_FRAMESIZE; in pkt_gather_data() [all …]
|
/drivers/nvdimm/ |
D | nd_virtio.c | 107 if (bio && bio->bi_iter.bi_sector != -1) { in async_pmem_flush() 114 child->bi_iter.bi_sector = -1; in async_pmem_flush()
|