Searched refs:queue_depth (Results 1 – 9 of 9) sorted by relevance
128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth()144 rqd->queue_depth); in rq_depth_calc_max_depth()148 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
2397 for (i = 0; i < set->queue_depth; i++) { in blk_mq_clear_rq_mapping()2724 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument2735 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()2759 set->queue_depth, flush_rq); in blk_mq_exit_hctx()2935 set->queue_depth, set->reserved_tags, flags); in __blk_mq_alloc_map_and_request()2940 set->queue_depth); in __blk_mq_alloc_map_and_request()3222 unsigned int queue_depth, in blk_mq_init_sq_queue() argument3232 set->queue_depth = queue_depth; in blk_mq_init_sq_queue()3403 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()3475 depth = set->queue_depth; in blk_mq_alloc_map_and_requests()[all …]
578 q->nr_requests = q->tag_set->queue_depth; in blk_mq_init_sched()587 q->nr_requests = 2 * min_t(unsigned int, q->tag_set->queue_depth, in blk_mq_init_sched()
58 unsigned int queue_depth; member
995 iolat->rq_depth.queue_depth = blkg->q->nr_requests; in iolatency_pd_init()997 iolat->rq_depth.default_depth = iolat->rq_depth.queue_depth; in iolatency_pd_init()
383 set->queue_depth = 128; in bsg_setup_queue()
510 unsigned int depth = set->queue_depth - set->reserved_tags; in blk_mq_init_shared_sbitmap()
769 q->queue_depth = depth; in blk_set_queue_depth()
684 RQWB(rqos)->rq_depth.queue_depth = blk_queue_depth(rqos->q); in wbt_queue_depth_changed()