Searched refs:ops (Results 1 – 15 of 15) sorted by relevance
/block/ |
D | blk-rq-qos.c | 29 if (rqos->ops->cleanup) in __rq_qos_cleanup() 30 rqos->ops->cleanup(rqos, bio); in __rq_qos_cleanup() 38 if (rqos->ops->done) in __rq_qos_done() 39 rqos->ops->done(rqos, rq); in __rq_qos_done() 47 if (rqos->ops->issue) in __rq_qos_issue() 48 rqos->ops->issue(rqos, rq); in __rq_qos_issue() 56 if (rqos->ops->requeue) in __rq_qos_requeue() 57 rqos->ops->requeue(rqos, rq); in __rq_qos_requeue() 65 if (rqos->ops->throttle) in __rq_qos_throttle() 66 rqos->ops->throttle(rqos, bio); in __rq_qos_throttle() [all …]
|
D | blk-mq-sched.h | 43 if (e->type->ops.allow_merge) in blk_mq_sched_allow_merge() 44 return e->type->ops.allow_merge(q, rq, bio); in blk_mq_sched_allow_merge() 54 if (e->type->ops.completed_request) in blk_mq_sched_completed_request() 55 e->type->ops.completed_request(rq, now); in blk_mq_sched_completed_request() 65 if (e->type->ops.requeue_request) in blk_mq_sched_requeue_request() 66 e->type->ops.requeue_request(rq); in blk_mq_sched_requeue_request() 74 if (e && e->type->ops.has_work) in blk_mq_sched_has_work() 75 return e->type->ops.has_work(hctx); in blk_mq_sched_has_work()
|
D | ioctl.c | 279 const struct pr_ops *ops = bdev->bd_disk->fops->pr_ops; in blkdev_pr_register() local 284 if (!ops || !ops->pr_register) in blkdev_pr_register() 291 return ops->pr_register(bdev, reg.old_key, reg.new_key, reg.flags); in blkdev_pr_register() 297 const struct pr_ops *ops = bdev->bd_disk->fops->pr_ops; in blkdev_pr_reserve() local 302 if (!ops || !ops->pr_reserve) in blkdev_pr_reserve() 309 return ops->pr_reserve(bdev, rsv.key, rsv.type, rsv.flags); in blkdev_pr_reserve() 315 const struct pr_ops *ops = bdev->bd_disk->fops->pr_ops; in blkdev_pr_release() local 320 if (!ops || !ops->pr_release) in blkdev_pr_release() 327 return ops->pr_release(bdev, rsv.key, rsv.type); in blkdev_pr_release() 333 const struct pr_ops *ops = bdev->bd_disk->fops->pr_ops; in blkdev_pr_preempt() local [all …]
|
D | blk-mq-sched.c | 106 if (e->type->ops.has_work && !e->type->ops.has_work(hctx)) in __blk_mq_do_dispatch_sched() 118 rq = e->type->ops.dispatch_request(hctx); in __blk_mq_do_dispatch_sched() 348 if (e && e->type->ops.bio_merge) { in blk_mq_sched_bio_merge() 349 ret = e->type->ops.bio_merge(q, bio, nr_segs); in blk_mq_sched_bio_merge() 471 ret = e->ops.init_sched(q, e); in blk_mq_init_sched() 480 if (e->ops.init_hctx) { in blk_mq_init_sched() 481 ret = e->ops.init_hctx(hctx, i); in blk_mq_init_sched() 537 if (e->type->ops.exit_hctx && hctx->sched_data) { in blk_mq_exit_sched() 538 e->type->ops.exit_hctx(hctx, i); in blk_mq_exit_sched() 548 if (e->type->ops.exit_sched) in blk_mq_exit_sched() [all …]
|
D | elevator.c | 65 if (e->type->ops.allow_merge) in elv_iosched_allow_bio_merge() 66 return e->type->ops.allow_merge(q, rq, bio); in elv_iosched_allow_bio_merge() 318 if (e->type->ops.request_merge) in elv_merge() 319 return e->type->ops.request_merge(q, req, bio); in elv_merge() 375 if (e->type->ops.request_merged) in elv_merged_request() 376 e->type->ops.request_merged(q, rq, type); in elv_merged_request() 389 if (e->type->ops.requests_merged) in elv_merge_requests() 390 e->type->ops.requests_merged(q, rq, next); in elv_merge_requests() 400 if (e->type->ops.next_request) in elv_latter_request() 401 return e->type->ops.next_request(q, rq); in elv_latter_request() [all …]
|
D | blk-mq.c | 397 if (e->type->ops.prepare_request) in blk_mq_rq_ctx_init() 398 e->type->ops.prepare_request(rq); in blk_mq_rq_ctx_init() 466 struct elevator_mq_ops *ops = &q->elevator->type->ops; in __blk_mq_alloc_requests() local 471 if (ops->limit_depth) in __blk_mq_alloc_requests() 472 ops->limit_depth(data->cmd_flags, data); in __blk_mq_alloc_requests() 689 q->elevator->type->ops.finish_request(rq); in blk_mq_finish_request() 2571 q->elevator->type->ops.insert_requests(hctx, &list, flags); in blk_mq_insert_request() 2792 this_hctx->queue->elevator->type->ops.insert_requests(this_hctx, in blk_mq_dispatch_plug_list() 3290 if (tags->static_rqs && set->ops->exit_request) { in blk_mq_free_rqs() 3298 set->ops->exit_request(set, rq, hctx_idx); in blk_mq_free_rqs() [all …]
|
D | blk-ioc.c | 53 if (et->ops.exit_icq) in ioc_exit_icq() 54 et->ops.exit_icq(icq); in ioc_exit_icq() 392 if (et->ops.init_icq) in ioc_create_icq() 393 et->ops.init_icq(icq); in ioc_create_icq()
|
D | blk-rq-qos.h | 28 const struct rq_qos_ops *ops; member 89 const struct rq_qos_ops *ops);
|
D | elevator.h | 76 struct elevator_mq_ops ops; member
|
D | blk-mq-debugfs.c | 805 if (rqos->debugfs_dir || !rqos->ops->debugfs_attrs) in blk_mq_debugfs_register_rqos() 813 debugfs_create_files(rqos->debugfs_dir, rqos, rqos->ops->debugfs_attrs); in blk_mq_debugfs_register_rqos()
|
D | bsg-lib.c | 377 set->ops = &bsg_mq_ops; in bsg_setup_queue()
|
D | bdev.c | 359 ctx->ops = &bdev_sops; in bd_init_fs_context()
|
D | kyber-iosched.c | 1015 .ops = {
|
D | mq-deadline.c | 1258 .ops = {
|
D | bfq-iosched.c | 7597 .ops = {
|