/drivers/md/ |
D | dm-user.c | 376 typeof(bio->bi_opf) opf = bio->bi_opf & ~REQ_OP_MASK; in bio_flags_to_user_flags() local 378 if (opf & REQ_FAILFAST_DEV) { in bio_flags_to_user_flags() 379 opf &= ~REQ_FAILFAST_DEV; in bio_flags_to_user_flags() 383 if (opf & REQ_FAILFAST_TRANSPORT) { in bio_flags_to_user_flags() 384 opf &= ~REQ_FAILFAST_TRANSPORT; in bio_flags_to_user_flags() 388 if (opf & REQ_FAILFAST_DRIVER) { in bio_flags_to_user_flags() 389 opf &= ~REQ_FAILFAST_DRIVER; in bio_flags_to_user_flags() 393 if (opf & REQ_SYNC) { in bio_flags_to_user_flags() 394 opf &= ~REQ_SYNC; in bio_flags_to_user_flags() 398 if (opf & REQ_META) { in bio_flags_to_user_flags() [all …]
|
D | dm-io.c | 306 static void do_region(const blk_opf_t opf, unsigned int region, in do_region() argument 319 const enum req_op op = opf & REQ_OP_MASK; in do_region() 353 bio = bio_alloc_bioset(where->bdev, num_bvecs, opf, GFP_NOIO, in do_region() 385 static void dispatch_io(blk_opf_t opf, unsigned int num_regions, in dispatch_io() argument 395 opf |= REQ_SYNC; in dispatch_io() 403 if (where[i].count || (opf & REQ_PREFLUSH)) in dispatch_io() 404 do_region(opf, i, where + i, dp, io, ioprio); in dispatch_io() 428 struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, in sync_io() argument 434 if (num_regions > 1 && !op_is_write(opf)) { in sync_io() 451 dispatch_io(opf, num_regions, where, dp, io, 1, ioprio); in sync_io() [all …]
|
D | dm-snap-persistent.c | 232 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, blk_opf_t opf, in chunk_io() argument 241 .bi_opf = opf, in chunk_io() 288 static int area_io(struct pstore *ps, blk_opf_t opf) in area_io() argument 292 return chunk_io(ps, ps->area, chunk, opf, 0); in area_io()
|
D | dm-integrity.c | 543 static int sync_rw_sb(struct dm_integrity_c *ic, blk_opf_t opf) in sync_rw_sb() argument 547 const enum req_op op = opf & REQ_OP_MASK; in sync_rw_sb() 550 io_req.bi_opf = opf; in sync_rw_sb() 1050 static void rw_journal_sectors(struct dm_integrity_c *ic, blk_opf_t opf, in rw_journal_sectors() argument 1068 io_req.bi_opf = opf; in rw_journal_sectors() 1088 dm_integrity_io_error(ic, (opf & REQ_OP_MASK) == REQ_OP_READ ? in rw_journal_sectors() 1097 static void rw_journal(struct dm_integrity_c *ic, blk_opf_t opf, in rw_journal() argument 1106 rw_journal_sectors(ic, opf, sector, n_sectors, comp); in rw_journal()
|
D | md.h | 787 struct page *page, blk_opf_t opf, bool metadata_op);
|
D | md.c | 981 struct page *page, blk_opf_t opf, bool metadata_op) in sync_page_io() argument 987 bio_init(&bio, rdev->meta_bdev, &bvec, 1, opf); in sync_page_io() 989 bio_init(&bio, rdev->bdev, &bvec, 1, opf); in sync_page_io()
|
/drivers/nvme/target/ |
D | io-cmd-bdev.c | 247 blk_opf_t opf; in nvmet_bdev_execute_rw() local 262 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_rw() 264 opf |= REQ_FUA; in nvmet_bdev_execute_rw() 267 opf = REQ_OP_READ; in nvmet_bdev_execute_rw() 272 opf |= REQ_NOMERGE; in nvmet_bdev_execute_rw() 279 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_rw() 281 bio = bio_alloc(req->ns->bdev, bio_max_segs(sg_cnt), opf, in nvmet_bdev_execute_rw() 308 opf, GFP_KERNEL); in nvmet_bdev_execute_rw()
|
D | zns.c | 541 const blk_opf_t opf = REQ_OP_ZONE_APPEND | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_zone_append() local 572 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_zone_append() 574 bio = bio_alloc(req->ns->bdev, req->sg_cnt, opf, GFP_KERNEL); in nvmet_bdev_execute_zone_append()
|
/drivers/scsi/device_handler/ |
D | scsi_dh_hp_sw.c | 86 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_tur() local 93 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_tur() 127 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_start_stop() local 134 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_start_stop()
|
D | scsi_dh_emc.c | 242 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in send_trespass_cmd() local 269 err = scsi_execute_cmd(sdev, cdb, opf, csdev->buffer, len, in send_trespass_cmd()
|
D | scsi_dh_alua.c | 130 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in submit_rtpg() local 145 return scsi_execute_cmd(sdev, cdb, opf, buff, bufflen, in submit_rtpg() 163 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in submit_stpg() local 180 return scsi_execute_cmd(sdev, cdb, opf, stpg_data, in submit_stpg()
|
D | scsi_dh_rdac.c | 539 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in send_mode_select() local 561 if (scsi_execute_cmd(sdev, cdb, opf, &h->ctlr->mode_select, data_size, in send_mode_select()
|
/drivers/target/ |
D | target_core_iblock.c | 355 blk_opf_t opf) in iblock_get_bio() argument 364 bio = bio_alloc_bioset(ib_dev->ibd_bd, bio_max_segs(sg_num), opf, in iblock_get_bio() 734 blk_opf_t opf; in iblock_execute_rw() local 747 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in iblock_execute_rw() 755 opf |= REQ_FUA; in iblock_execute_rw() 757 opf |= REQ_FUA; in iblock_execute_rw() 760 opf = REQ_OP_READ; in iblock_execute_rw() 775 bio = iblock_get_bio(cmd, block_lba, sgl_nents, opf); in iblock_execute_rw() 808 bio = iblock_get_bio(cmd, block_lba, sg_num, opf); in iblock_execute_rw()
|
/drivers/block/ |
D | brd.c | 211 unsigned int len, unsigned int off, blk_opf_t opf, in brd_do_bvec() argument 217 if (op_is_write(opf)) { in brd_do_bvec() 222 gfp_t gfp = opf & REQ_NOWAIT ? GFP_NOWAIT : GFP_NOIO; in brd_do_bvec() 230 if (!op_is_write(opf)) { in brd_do_bvec()
|
/drivers/md/bcache/ |
D | super.c | 417 static void uuid_io(struct cache_set *c, blk_opf_t opf, struct bkey *k, in uuid_io() argument 432 bio->bi_opf = opf | REQ_SYNC | REQ_META; in uuid_io() 441 if ((opf & REQ_OP_MASK) != REQ_OP_WRITE) in uuid_io() 446 pr_debug("%s UUIDs at %s\n", (opf & REQ_OP_MASK) == REQ_OP_WRITE ? in uuid_io() 590 static void prio_io(struct cache *ca, uint64_t bucket, blk_opf_t opf) in prio_io() argument 603 bio->bi_opf = opf | REQ_SYNC | REQ_META; in prio_io()
|
/drivers/scsi/ |
D | scsi_lib.c | 202 blk_opf_t opf, void *buffer, unsigned int bufflen, in scsi_execute_cmd() argument 217 req = scsi_alloc_request(sdev->request_queue, opf, args->req_flags); in scsi_execute_cmd() 1131 struct request *scsi_alloc_request(struct request_queue *q, blk_opf_t opf, in scsi_alloc_request() argument 1136 rq = blk_mq_alloc_request(q, opf, flags); in scsi_alloc_request()
|
D | scsi_transport_spi.c | 113 blk_opf_t opf = op | REQ_FAILFAST_DEV | REQ_FAILFAST_TRANSPORT | in spi_execute() local 127 result = scsi_execute_cmd(sdev, cmd, opf, buffer, bufflen, in spi_execute()
|
/drivers/block/drbd/ |
D | drbd_int.h | 326 blk_opf_t opf; member 340 ((peer_req)->opf & REQ_OP_MASK)
|
D | drbd_receiver.c | 1688 drbd_err(device, "Invalid bio op received: 0x%x\n", peer_req->opf); in drbd_submit_peer_request() 1692 bio = bio_alloc(device->ldev->backing_bdev, nr_pages, peer_req->opf, GFP_NOIO); in drbd_submit_peer_request() 2076 peer_req->opf = REQ_OP_WRITE; in recv_resync_read() 2601 peer_req->opf = wire_flags_to_bio(connection, dp_flags); in receive_Data() 2870 peer_req->opf = REQ_OP_READ; in receive_DataRequest() 4974 peer_req->opf = REQ_OP_DISCARD; in receive_rs_deallocated()
|
D | drbd_worker.c | 403 peer_req->opf = REQ_OP_READ; in read_for_csum()
|