Home
last modified time | relevance | path

Searched refs:queue_depth (Results 1 – 9 of 9) sorted by relevance

/block/
Dblk-rq-qos.c128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth()
144 rqd->queue_depth); in rq_depth_calc_max_depth()
148 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
Dblk-mq.c2397 for (i = 0; i < set->queue_depth; i++) { in blk_mq_clear_rq_mapping()
2724 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument
2735 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
2759 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
2935 set->queue_depth, set->reserved_tags, flags); in __blk_mq_alloc_map_and_request()
2940 set->queue_depth); in __blk_mq_alloc_map_and_request()
3222 unsigned int queue_depth, in blk_mq_init_sq_queue() argument
3232 set->queue_depth = queue_depth; in blk_mq_init_sq_queue()
3403 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
3475 depth = set->queue_depth; in blk_mq_alloc_map_and_requests()
[all …]
Dblk-mq-sched.c578 q->nr_requests = q->tag_set->queue_depth; in blk_mq_init_sched()
587 q->nr_requests = 2 * min_t(unsigned int, q->tag_set->queue_depth, in blk_mq_init_sched()
Dblk-rq-qos.h58 unsigned int queue_depth; member
Dblk-iolatency.c995 iolat->rq_depth.queue_depth = blkg->q->nr_requests; in iolatency_pd_init()
997 iolat->rq_depth.default_depth = iolat->rq_depth.queue_depth; in iolatency_pd_init()
Dbsg-lib.c383 set->queue_depth = 128; in bsg_setup_queue()
Dblk-mq-tag.c510 unsigned int depth = set->queue_depth - set->reserved_tags; in blk_mq_init_shared_sbitmap()
Dblk-settings.c769 q->queue_depth = depth; in blk_set_queue_depth()
Dblk-wbt.c684 RQWB(rqos)->rq_depth.queue_depth = blk_queue_depth(rqos->q); in wbt_queue_depth_changed()