Home
last modified time | relevance | path

Searched refs:mq_hctx (Results 1 – 11 of 11) sorted by relevance

/kernel/linux/linux-5.10/block/
Dblk-mq-sched.c84 return rqa->mq_hctx > rqb->mq_hctx; in sched_rq_cmp()
90 list_first_entry(rq_list, struct request, queuelist)->mq_hctx; in blk_mq_dispatch_hctx_list()
96 if (rq->mq_hctx != hctx) { in blk_mq_dispatch_hctx_list()
167 if (rq->mq_hctx != hctx) in __blk_mq_do_dispatch_sched()
269 } while (blk_mq_dispatch_rq_list(rq->mq_hctx, &rq_list, 1)); in blk_mq_do_dispatch_ctx()
427 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in blk_mq_sched_insert_request()
Dblk-mq.c309 rq->mq_hctx = data->hctx; in blk_mq_rq_ctx_init()
507 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in __blk_mq_free_request()
512 rq->mq_hctx = NULL; in __blk_mq_free_request()
526 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in blk_mq_free_request()
1099 struct sbitmap_queue *bt = rq->mq_hctx->tags->bitmap_tags; in __blk_mq_get_driver_tag()
1100 unsigned int tag_offset = rq->mq_hctx->tags->nr_reserved_tags; in __blk_mq_get_driver_tag()
1103 blk_mq_tag_busy(rq->mq_hctx); in __blk_mq_get_driver_tag()
1105 if (blk_mq_tag_is_reserved(rq->mq_hctx->sched_tags, rq->internal_tag)) { in __blk_mq_get_driver_tag()
1106 bt = rq->mq_hctx->tags->breserved_tags; in __blk_mq_get_driver_tag()
1109 if (!hctx_may_queue(rq->mq_hctx, bt)) in __blk_mq_get_driver_tag()
[all …]
Dblk-flush.c320 flush_rq->mq_hctx = first_rq->mq_hctx; in blk_kick_flush()
354 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in mq_flush_data_end_io()
Dblk-mq-tag.c235 if (rq->q == hctx->queue && rq->mq_hctx == hctx) in bt_iter()
640 return (rq->mq_hctx->queue_num << BLK_MQ_UNIQUE_TAG_BITS) | in blk_mq_unique_tag()
Dblk-mq.h243 __blk_mq_put_driver_tag(rq->mq_hctx, rq); in blk_mq_put_driver_tag()
Dmq-deadline.c390 atomic_dec(&rq->mq_hctx->elevator_queued); in dd_dispatch_request()
580 blk_mq_sched_mark_restart_hctx(rq->mq_hctx); in dd_finish_request()
Dblk-mq-debugfs.c409 if (rq->mq_hctx == params->hctx) in hctx_show_busy_rq()
Dbfq-iosched.c5938 atomic_dec(&rq->mq_hctx->elevator_queued); in bfq_finish_requeue_request()
/kernel/linux/linux-5.10/include/linux/
Dblkdev.h138 struct blk_mq_hw_ctx *mq_hctx; member
/kernel/linux/linux-5.10/drivers/nvme/host/
Dpci.c1287 nvme_poll(req->mq_hctx); in nvme_timeout()
Dcore.c889 blk_poll(q, request_to_qc_t(rq->mq_hctx, rq), true); in nvme_execute_rq_polled()