Home
last modified time | relevance | path

Searched refs:op (Results 1 – 15 of 15) sorted by relevance

/block/
Dioctl.c17 struct blkpg_partition __user *upart, int op) in blkpg_do_ioctl() argument
33 if (op == BLKPG_DEL_PARTITION) in blkpg_do_ioctl()
45 switch (op) { in blkpg_do_ioctl()
59 int op; in blkpg_ioctl() local
61 if (get_user(op, &arg->op) || get_user(udata, &arg->data)) in blkpg_ioctl()
64 return blkpg_do_ioctl(bdev, udata, op); in blkpg_ioctl()
69 compat_int_t op; member
79 int op; in compat_blkpg_ioctl() local
81 if (get_user(op, &arg->op) || get_user(udata, &arg->data)) in compat_blkpg_ioctl()
84 return blkpg_do_ioctl(bdev, compat_ptr(udata), op); in compat_blkpg_ioctl()
Dblk-zoned.c258 int blkdev_zone_mgmt(struct block_device *bdev, enum req_opf op, in blkdev_zone_mgmt() argument
275 if (!op_is_zone_mgmt(op)) in blkdev_zone_mgmt()
295 if (op == REQ_OP_ZONE_RESET && sector == 0 && nr_sectors == capacity) { in blkdev_zone_mgmt()
304 bio->bi_opf = op | REQ_SYNC; in blkdev_zone_mgmt()
401 enum req_opf op; in blkdev_zone_mgmt_ioctl() local
422 op = REQ_OP_ZONE_RESET; in blkdev_zone_mgmt_ioctl()
431 op = REQ_OP_ZONE_OPEN; in blkdev_zone_mgmt_ioctl()
434 op = REQ_OP_ZONE_CLOSE; in blkdev_zone_mgmt_ioctl()
437 op = REQ_OP_ZONE_FINISH; in blkdev_zone_mgmt_ioctl()
443 ret = blkdev_zone_mgmt(bdev, op, zrange.sector, zrange.nr_sectors, in blkdev_zone_mgmt_ioctl()
Dblk-cgroup-rwstat.h62 unsigned int op, uint64_t val) in blkg_rwstat_add() argument
66 if (op_is_discard(op)) in blkg_rwstat_add()
68 else if (op_is_write(op)) in blkg_rwstat_add()
75 if (op_is_sync(op)) in blkg_rwstat_add()
Dblk-core.c162 inline const char *blk_op_str(unsigned int op) in blk_op_str() argument
166 if (op < ARRAY_SIZE(blk_op_name) && blk_op_name[op]) in blk_op_str()
167 op_str = blk_op_name[op]; in blk_op_str()
635 struct request *blk_get_request(struct request_queue *q, unsigned int op, in blk_get_request() argument
640 WARN_ON_ONCE(op & REQ_NOWAIT); in blk_get_request()
643 req = blk_mq_alloc_request(q, op, flags); in blk_get_request()
1290 unsigned int sectors, unsigned int op, in __part_start_io_acct() argument
1293 const int sgrp = op_stat_group(op); in __part_start_io_acct()
1299 part_stat_local_inc(part, in_flight[op_is_write(op)]); in __part_start_io_acct()
1331 unsigned int op) in disk_start_io_acct() argument
[all …]
Dbfq-cgroup.c223 unsigned int op) in bfqg_stats_update_io_add() argument
225 blkg_rwstat_add(&bfqg->stats.queued, op, 1); in bfqg_stats_update_io_add()
231 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, unsigned int op) in bfqg_stats_update_io_remove() argument
233 blkg_rwstat_add(&bfqg->stats.queued, op, -1); in bfqg_stats_update_io_remove()
236 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, unsigned int op) in bfqg_stats_update_io_merged() argument
238 blkg_rwstat_add(&bfqg->stats.merged, op, 1); in bfqg_stats_update_io_merged()
242 u64 io_start_time_ns, unsigned int op) in bfqg_stats_update_completion() argument
248 blkg_rwstat_add(&stats->service_time, op, in bfqg_stats_update_completion()
251 blkg_rwstat_add(&stats->wait_time, op, in bfqg_stats_update_completion()
258 unsigned int op) { } in bfqg_stats_update_io_add() argument
[all …]
Dblk-lib.c32 unsigned int op; in __blkdev_issue_discard() local
44 op = REQ_OP_SECURE_ERASE; in __blkdev_issue_discard()
48 op = REQ_OP_DISCARD; in __blkdev_issue_discard()
101 bio_set_op_attrs(bio, op, 0); in __blkdev_issue_discard()
Dblk-mq-debugfs.c151 char opbuf[16] = { }, *op; in queue_state_write() local
168 op = strstrip(opbuf); in queue_state_write()
169 if (strcmp(op, "run") == 0) { in queue_state_write()
171 } else if (strcmp(op, "start") == 0) { in queue_state_write()
173 } else if (strcmp(op, "kick") == 0) { in queue_state_write()
176 pr_err("%s: unsupported operation '%s'\n", __func__, op); in queue_state_write()
332 const unsigned int op = req_op(rq); in __blk_mq_debugfs_rq_show() local
333 const char *op_str = blk_op_str(op); in __blk_mq_debugfs_rq_show()
337 seq_printf(m, "%u", op); in __blk_mq_debugfs_rq_show()
Dkyber-iosched.c198 static unsigned int kyber_sched_domain(unsigned int op) in kyber_sched_domain() argument
200 switch (op & REQ_OP_MASK) { in kyber_sched_domain()
553 static void kyber_limit_depth(unsigned int op, struct blk_mq_alloc_data *data) in kyber_limit_depth() argument
559 if (!op_is_sync(op)) { in kyber_limit_depth()
889 #define KYBER_LAT_ATTR(op) __ATTR(op##_lat_nsec, 0644, kyber_##op##_lat_show, kyber_##op##_lat_stor… argument
Dbfq-iosched.h994 unsigned int op);
995 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, unsigned int op);
996 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, unsigned int op);
998 u64 io_start_time_ns, unsigned int op);
Dblk-wbt.c673 const int op = req_op(rq); in wbt_data_dir() local
675 if (op == REQ_OP_READ) in wbt_data_dir()
677 else if (op_is_write(op)) in wbt_data_dir()
Dfops.c36 unsigned int op = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in dio_bio_write_op() local
40 op |= REQ_FUA; in dio_bio_write_op()
41 return op; in dio_bio_write_op()
Dmq-deadline.c642 static void dd_limit_depth(unsigned int op, struct blk_mq_alloc_data *data) in dd_limit_depth() argument
647 if (op_is_sync(op) && !op_is_write(op)) in dd_limit_depth()
Dblk-throttle.c2333 int op, unsigned long time) in throtl_track_latency() argument
2339 !(op == REQ_OP_READ || op == REQ_OP_WRITE) || in throtl_track_latency()
2345 latency = get_cpu_ptr(td->latency_buckets[op]); in throtl_track_latency()
2348 put_cpu_ptr(td->latency_buckets[op]); in throtl_track_latency()
Dbfq-iosched.c584 static void bfq_limit_depth(unsigned int op, struct blk_mq_alloc_data *data) in bfq_limit_depth() argument
588 if (op_is_sync(op) && !op_is_write(op)) in bfq_limit_depth()
592 bfqd->word_depths[!!bfqd->wr_busy_queues][op_is_sync(op)]; in bfq_limit_depth()
595 __func__, bfqd->wr_busy_queues, op_is_sync(op), in bfq_limit_depth()
Dblk-mq.c410 struct request *blk_mq_alloc_request(struct request_queue *q, unsigned int op, in blk_mq_alloc_request() argument
416 .cmd_flags = op, in blk_mq_alloc_request()
439 unsigned int op, blk_mq_req_flags_t flags, unsigned int hctx_idx) in blk_mq_alloc_request_hctx() argument
444 .cmd_flags = op, in blk_mq_alloc_request_hctx()