Home
last modified time | relevance | path

Searched refs:opf (Results 1 – 20 of 20) sorted by relevance

/drivers/md/
Ddm-user.c376 typeof(bio->bi_opf) opf = bio->bi_opf & ~REQ_OP_MASK; in bio_flags_to_user_flags() local
378 if (opf & REQ_FAILFAST_DEV) { in bio_flags_to_user_flags()
379 opf &= ~REQ_FAILFAST_DEV; in bio_flags_to_user_flags()
383 if (opf & REQ_FAILFAST_TRANSPORT) { in bio_flags_to_user_flags()
384 opf &= ~REQ_FAILFAST_TRANSPORT; in bio_flags_to_user_flags()
388 if (opf & REQ_FAILFAST_DRIVER) { in bio_flags_to_user_flags()
389 opf &= ~REQ_FAILFAST_DRIVER; in bio_flags_to_user_flags()
393 if (opf & REQ_SYNC) { in bio_flags_to_user_flags()
394 opf &= ~REQ_SYNC; in bio_flags_to_user_flags()
398 if (opf & REQ_META) { in bio_flags_to_user_flags()
[all …]
Ddm-io.c306 static void do_region(const blk_opf_t opf, unsigned int region, in do_region() argument
319 const enum req_op op = opf & REQ_OP_MASK; in do_region()
353 bio = bio_alloc_bioset(where->bdev, num_bvecs, opf, GFP_NOIO, in do_region()
385 static void dispatch_io(blk_opf_t opf, unsigned int num_regions, in dispatch_io() argument
395 opf |= REQ_SYNC; in dispatch_io()
403 if (where[i].count || (opf & REQ_PREFLUSH)) in dispatch_io()
404 do_region(opf, i, where + i, dp, io, ioprio); in dispatch_io()
428 struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, in sync_io() argument
434 if (num_regions > 1 && !op_is_write(opf)) { in sync_io()
451 dispatch_io(opf, num_regions, where, dp, io, 1, ioprio); in sync_io()
[all …]
Ddm-snap-persistent.c232 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, blk_opf_t opf, in chunk_io() argument
241 .bi_opf = opf, in chunk_io()
288 static int area_io(struct pstore *ps, blk_opf_t opf) in area_io() argument
292 return chunk_io(ps, ps->area, chunk, opf, 0); in area_io()
Ddm-integrity.c543 static int sync_rw_sb(struct dm_integrity_c *ic, blk_opf_t opf) in sync_rw_sb() argument
547 const enum req_op op = opf & REQ_OP_MASK; in sync_rw_sb()
550 io_req.bi_opf = opf; in sync_rw_sb()
1050 static void rw_journal_sectors(struct dm_integrity_c *ic, blk_opf_t opf, in rw_journal_sectors() argument
1068 io_req.bi_opf = opf; in rw_journal_sectors()
1088 dm_integrity_io_error(ic, (opf & REQ_OP_MASK) == REQ_OP_READ ? in rw_journal_sectors()
1097 static void rw_journal(struct dm_integrity_c *ic, blk_opf_t opf, in rw_journal() argument
1106 rw_journal_sectors(ic, opf, sector, n_sectors, comp); in rw_journal()
Dmd.h787 struct page *page, blk_opf_t opf, bool metadata_op);
Dmd.c981 struct page *page, blk_opf_t opf, bool metadata_op) in sync_page_io() argument
987 bio_init(&bio, rdev->meta_bdev, &bvec, 1, opf); in sync_page_io()
989 bio_init(&bio, rdev->bdev, &bvec, 1, opf); in sync_page_io()
/drivers/nvme/target/
Dio-cmd-bdev.c247 blk_opf_t opf; in nvmet_bdev_execute_rw() local
262 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_rw()
264 opf |= REQ_FUA; in nvmet_bdev_execute_rw()
267 opf = REQ_OP_READ; in nvmet_bdev_execute_rw()
272 opf |= REQ_NOMERGE; in nvmet_bdev_execute_rw()
279 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_rw()
281 bio = bio_alloc(req->ns->bdev, bio_max_segs(sg_cnt), opf, in nvmet_bdev_execute_rw()
308 opf, GFP_KERNEL); in nvmet_bdev_execute_rw()
Dzns.c541 const blk_opf_t opf = REQ_OP_ZONE_APPEND | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_zone_append() local
572 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_zone_append()
574 bio = bio_alloc(req->ns->bdev, req->sg_cnt, opf, GFP_KERNEL); in nvmet_bdev_execute_zone_append()
/drivers/scsi/device_handler/
Dscsi_dh_hp_sw.c86 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_tur() local
93 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_tur()
127 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_start_stop() local
134 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_start_stop()
Dscsi_dh_emc.c242 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in send_trespass_cmd() local
269 err = scsi_execute_cmd(sdev, cdb, opf, csdev->buffer, len, in send_trespass_cmd()
Dscsi_dh_alua.c130 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in submit_rtpg() local
145 return scsi_execute_cmd(sdev, cdb, opf, buff, bufflen, in submit_rtpg()
163 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in submit_stpg() local
180 return scsi_execute_cmd(sdev, cdb, opf, stpg_data, in submit_stpg()
Dscsi_dh_rdac.c539 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in send_mode_select() local
561 if (scsi_execute_cmd(sdev, cdb, opf, &h->ctlr->mode_select, data_size, in send_mode_select()
/drivers/target/
Dtarget_core_iblock.c355 blk_opf_t opf) in iblock_get_bio() argument
364 bio = bio_alloc_bioset(ib_dev->ibd_bd, bio_max_segs(sg_num), opf, in iblock_get_bio()
734 blk_opf_t opf; in iblock_execute_rw() local
747 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in iblock_execute_rw()
755 opf |= REQ_FUA; in iblock_execute_rw()
757 opf |= REQ_FUA; in iblock_execute_rw()
760 opf = REQ_OP_READ; in iblock_execute_rw()
775 bio = iblock_get_bio(cmd, block_lba, sgl_nents, opf); in iblock_execute_rw()
808 bio = iblock_get_bio(cmd, block_lba, sg_num, opf); in iblock_execute_rw()
/drivers/block/
Dbrd.c211 unsigned int len, unsigned int off, blk_opf_t opf, in brd_do_bvec() argument
217 if (op_is_write(opf)) { in brd_do_bvec()
222 gfp_t gfp = opf & REQ_NOWAIT ? GFP_NOWAIT : GFP_NOIO; in brd_do_bvec()
230 if (!op_is_write(opf)) { in brd_do_bvec()
/drivers/md/bcache/
Dsuper.c417 static void uuid_io(struct cache_set *c, blk_opf_t opf, struct bkey *k, in uuid_io() argument
432 bio->bi_opf = opf | REQ_SYNC | REQ_META; in uuid_io()
441 if ((opf & REQ_OP_MASK) != REQ_OP_WRITE) in uuid_io()
446 pr_debug("%s UUIDs at %s\n", (opf & REQ_OP_MASK) == REQ_OP_WRITE ? in uuid_io()
590 static void prio_io(struct cache *ca, uint64_t bucket, blk_opf_t opf) in prio_io() argument
603 bio->bi_opf = opf | REQ_SYNC | REQ_META; in prio_io()
/drivers/scsi/
Dscsi_lib.c202 blk_opf_t opf, void *buffer, unsigned int bufflen, in scsi_execute_cmd() argument
217 req = scsi_alloc_request(sdev->request_queue, opf, args->req_flags); in scsi_execute_cmd()
1131 struct request *scsi_alloc_request(struct request_queue *q, blk_opf_t opf, in scsi_alloc_request() argument
1136 rq = blk_mq_alloc_request(q, opf, flags); in scsi_alloc_request()
Dscsi_transport_spi.c113 blk_opf_t opf = op | REQ_FAILFAST_DEV | REQ_FAILFAST_TRANSPORT | in spi_execute() local
127 result = scsi_execute_cmd(sdev, cmd, opf, buffer, bufflen, in spi_execute()
/drivers/block/drbd/
Ddrbd_int.h326 blk_opf_t opf; member
340 ((peer_req)->opf & REQ_OP_MASK)
Ddrbd_receiver.c1688 drbd_err(device, "Invalid bio op received: 0x%x\n", peer_req->opf); in drbd_submit_peer_request()
1692 bio = bio_alloc(device->ldev->backing_bdev, nr_pages, peer_req->opf, GFP_NOIO); in drbd_submit_peer_request()
2076 peer_req->opf = REQ_OP_WRITE; in recv_resync_read()
2601 peer_req->opf = wire_flags_to_bio(connection, dp_flags); in receive_Data()
2870 peer_req->opf = REQ_OP_READ; in receive_DataRequest()
4974 peer_req->opf = REQ_OP_DISCARD; in receive_rs_deallocated()
Ddrbd_worker.c403 peer_req->opf = REQ_OP_READ; in read_for_csum()