/kernel/linux/linux-5.10/drivers/md/ |
D | dm-crypt.c | 530 if (bio_data_dir(dmreq->ctx->bio_in) == WRITE) { in crypt_iv_lmk_gen() 548 if (bio_data_dir(dmreq->ctx->bio_in) == WRITE) in crypt_iv_lmk_post() 680 if (bio_data_dir(dmreq->ctx->bio_in) != WRITE) { in crypt_iv_tcw_gen() 703 if (bio_data_dir(dmreq->ctx->bio_in) != WRITE) in crypt_iv_tcw_post() 979 if (bio_data_dir(dmreq->ctx->bio_in) == WRITE) { in crypt_iv_elephant() 986 if (bio_data_dir(dmreq->ctx->bio_in) != WRITE) { in crypt_iv_elephant() 996 if (bio_data_dir(dmreq->ctx->bio_in) == WRITE) { in crypt_iv_elephant() 1016 if (bio_data_dir(dmreq->ctx->bio_in) == WRITE) { in crypt_iv_elephant_gen() 1028 if (bio_data_dir(dmreq->ctx->bio_in) != WRITE) in crypt_iv_elephant_post() 1336 if (cc->integrity_iv_size && bio_data_dir(ctx->bio_in) != WRITE) { in crypt_convert_block_aead() [all …]
|
D | dm-flakey.c | 313 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf, in corrupt_bio_data() 343 if (bio_data_dir(bio) == READ) { in flakey_map() 394 if (!*error && pb->bio_submitted && (bio_data_dir(bio) == READ)) { in flakey_end_io()
|
D | dm-snap.c | 1990 if (bio_data_dir(bio) == WRITE) { in snapshot_map() 1999 bio_data_dir(bio) == WRITE)) { in snapshot_map() 2049 if (bio_data_dir(bio) == WRITE) { in snapshot_map() 2169 if (bio_data_dir(bio) == WRITE && in snapshot_merge_map() 2181 if (bio_data_dir(bio) == WRITE) in snapshot_merge_map() 2189 if (bio_data_dir(bio) == WRITE) { in snapshot_merge_map() 2686 if (bio_data_dir(bio) != WRITE) in origin_map()
|
D | dm-cache-target.c | 592 return bio_data_dir(bio) == WRITE ? in lock_level() 823 if (bio_data_dir(bio) == WRITE) in __remap_to_origin_clear_discard() 839 if (bio_data_dir(bio) == WRITE) { in remap_to_cache_dirty() 1127 return (bio_data_dir(bio) == WRITE) && in bio_writes_complete_block() 1657 atomic_inc(bio_data_dir(bio) == READ ? in inc_hit_counter() 1663 atomic_inc(bio_data_dir(bio) == READ ? in inc_miss_counter() 1690 data_dir = bio_data_dir(bio); in map_bio() 1751 if (bio_data_dir(bio) == WRITE) { in map_bio() 1758 if (bio_data_dir(bio) == WRITE && writethrough_mode(cache) && in map_bio()
|
D | md-faulty.c | 169 if (bio_data_dir(bio) == WRITE) { in faulty_make_request()
|
D | dm-raid1.c | 524 queue_bio(m->ms, bio, bio_data_dir(bio)); in read_callback() 1186 int r, rw = bio_data_dir(bio); in mirror_map() 1236 int rw = bio_data_dir(bio); in mirror_end_io()
|
D | dm-delay.c | 288 if (bio_data_dir(bio) == WRITE) { in delay_map()
|
D | dm-log-writes.c | 688 if (bio_data_dir(bio) == READ) in log_writes_map() 796 if (bio_data_dir(bio) == WRITE && pb->block) { in normal_end_io()
|
D | dm-io.c | 143 if (bio->bi_status && bio_data_dir(bio) == READ) in endio()
|
D | dm-dust.c | 232 if (bio_data_dir(bio) == READ) in dust_map()
|
D | dm-thin.c | 1252 return (bio_data_dir(bio) == WRITE) && in io_overwrites_block() 1821 if (bio_data_dir(bio) == WRITE || op_is_flush(bio->bi_opf) || in __remap_and_issue_shared_cell() 1874 if (bio_data_dir(bio) == WRITE && bio->bi_iter.bi_size) { in process_shared_bio() 1910 if (bio_data_dir(bio) == READ) { in provision_block() 1964 if (bio_data_dir(bio) == READ && tc->origin_dev) { in process_cell() 2015 int rw = bio_data_dir(bio); in __process_bio_read_only()
|
D | dm-writecache.c | 1210 int rw = bio_data_dir(bio); in bio_copy_block() 1337 if (bio_data_dir(bio) == READ) { in writecache_map() 1474 atomic_inc(&wc->bio_in_progress[bio_data_dir(bio)]); in writecache_map() 1498 int dir = bio_data_dir(bio); in writecache_end_io()
|
/kernel/linux/linux-5.10/block/ |
D | blk-throttle.c | 899 bool rw = bio_data_dir(bio); in tg_with_in_iops_limit() 947 bool rw = bio_data_dir(bio); in tg_with_in_bps_limit() 998 bool rw = bio_data_dir(bio); in tg_may_dispatch() 1058 bool rw = bio_data_dir(bio); in throtl_charge_bio() 1090 bool rw = bio_data_dir(bio); in throtl_add_bio_tg() 1201 tg_dispatch_one_bio(tg, bio_data_dir(bio)); in throtl_dispatch_tg() 1211 tg_dispatch_one_bio(tg, bio_data_dir(bio)); in throtl_dispatch_tg() 2224 bool rw = bio_data_dir(bio); in blk_throtl_charge_bio_split() 2245 bool rw = bio_data_dir(bio); in blk_throtl_bio() 2395 int rw = bio_data_dir(bio); in blk_throtl_bio_endio()
|
D | bio-integrity.c | 233 if (bio_data_dir(bio) == READ) { in bio_integrity_prep() 303 if (bio_data_dir(bio) == WRITE) { in bio_integrity_prep()
|
D | blk-map.c | 120 else if (bio_data_dir(bio) == READ) in bio_uncopy_user() 347 bio_release_pages(bio, bio_data_dir(bio) == READ); in bio_unmap_user()
|
D | bounce.c | 291 int rw = bio_data_dir(*bio_orig); in __blk_queue_bounce()
|
D | blk-crypto-fallback.c | 504 if (bio_data_dir(bio) == WRITE) in blk_crypto_fallback_bio_prep()
|
/kernel/linux/linux-5.10/drivers/md/bcache/ |
D | io.c | 141 int is_read = (bio_data_dir(bio) == READ ? 1 : 0); in bch_bbio_count_io_errors()
|
/kernel/linux/linux-5.10/arch/m68k/emu/ |
D | nfblock.c | 70 dir = bio_data_dir(bio); in nfhd_submit_bio()
|
/kernel/linux/linux-5.10/drivers/block/drbd/ |
D | drbd_req.c | 34 req->rq_state = (bio_data_dir(bio_src) == WRITE ? RQ_WRITE : 0) in drbd_req_new() 806 if (bio_data_dir(req->master_bio) == WRITE) in __req_mod() 1194 const int rw = bio_data_dir(bio); in drbd_request_prepare() 1307 const int rw = bio_data_dir(req->master_bio); in drbd_send_and_submit() 1435 const int rw = bio_data_dir(req->master_bio); in submit_fast_path()
|
/kernel/linux/linux-5.10/drivers/block/rsxx/ |
D | dev.c | 164 bio_data_dir(bio) ? 'W' : 'R', bio_meta, in rsxx_submit_bio()
|
/kernel/linux/linux-5.10/drivers/s390/block/ |
D | xpram.c | 214 if (bio_data_dir(bio) == READ) { in xpram_submit_bio()
|
D | dcssblk.c | 900 if (bio_data_dir(bio) == WRITE) { in dcssblk_submit_bio() 916 if (bio_data_dir(bio) == READ) { in dcssblk_submit_bio()
|
/kernel/linux/linux-5.10/drivers/nvdimm/ |
D | blk.c | 179 rw = bio_data_dir(bio); in nd_blk_submit_bio()
|
/kernel/linux/linux-5.10/arch/xtensa/platforms/iss/ |
D | simdisk.c | 116 bio_data_dir(bio) == WRITE); in simdisk_submit_bio()
|