Lines Matching refs:i
235 unsigned int i; in blk_mq_quiesce_queue() local
240 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_quiesce_queue()
274 unsigned int i; in blk_mq_wake_waiters() local
276 queue_for_each_hw_ctx(q, hctx, i) in blk_mq_wake_waiters()
838 int i; in blk_mq_timeout_work() local
864 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_timeout_work()
1271 int i; in blk_mq_run_hw_queues() local
1273 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_run_hw_queues()
1293 int i; in blk_mq_queue_stopped() local
1295 queue_for_each_hw_ctx(q, hctx, i) in blk_mq_queue_stopped()
1332 int i; in blk_mq_stop_hw_queues() local
1334 queue_for_each_hw_ctx(q, hctx, i) in blk_mq_stop_hw_queues()
1350 int i; in blk_mq_start_hw_queues() local
1352 queue_for_each_hw_ctx(q, hctx, i) in blk_mq_start_hw_queues()
1370 int i; in blk_mq_start_stopped_hw_queues() local
1372 queue_for_each_hw_ctx(q, hctx, i) in blk_mq_start_stopped_hw_queues()
1767 int i; in blk_mq_free_rqs() local
1769 for (i = 0; i < tags->nr_tags; i++) { in blk_mq_free_rqs()
1770 struct request *rq = tags->static_rqs[i]; in blk_mq_free_rqs()
1775 tags->static_rqs[i] = NULL; in blk_mq_free_rqs()
1846 unsigned int i, j, entries_per_page, max_order = 4; in blk_mq_alloc_rqs() local
1864 for (i = 0; i < depth; ) { in blk_mq_alloc_rqs()
1898 to_do = min(entries_per_page, depth - i); in blk_mq_alloc_rqs()
1903 tags->static_rqs[i] = rq; in blk_mq_alloc_rqs()
1907 tags->static_rqs[i] = NULL; in blk_mq_alloc_rqs()
1913 i++; in blk_mq_alloc_rqs()
1991 unsigned int i; in blk_mq_exit_hw_queues() local
1993 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_exit_hw_queues()
1994 if (i == nr_queue) in blk_mq_exit_hw_queues()
1996 blk_mq_exit_hctx(q, set, hctx, i); in blk_mq_exit_hw_queues()
2077 unsigned int i; in blk_mq_init_cpu_queues() local
2079 for_each_possible_cpu(i) { in blk_mq_init_cpu_queues()
2080 struct blk_mq_ctx *__ctx = per_cpu_ptr(q->queue_ctx, i); in blk_mq_init_cpu_queues()
2083 __ctx->cpu = i; in blk_mq_init_cpu_queues()
2089 if (!cpu_present(i)) in blk_mq_init_cpu_queues()
2092 hctx = blk_mq_map_queue(q, i); in blk_mq_init_cpu_queues()
2099 hctx->numa_node = local_memory_node(cpu_to_node(i)); in blk_mq_init_cpu_queues()
2134 unsigned int i, hctx_idx; in blk_mq_map_swqueue() local
2144 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_map_swqueue()
2154 for_each_present_cpu(i) { in blk_mq_map_swqueue()
2155 hctx_idx = q->mq_map[i]; in blk_mq_map_swqueue()
2165 q->mq_map[i] = 0; in blk_mq_map_swqueue()
2168 ctx = per_cpu_ptr(q->queue_ctx, i); in blk_mq_map_swqueue()
2169 hctx = blk_mq_map_queue(q, i); in blk_mq_map_swqueue()
2171 cpumask_set_cpu(i, hctx->cpumask); in blk_mq_map_swqueue()
2178 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_map_swqueue()
2188 if (i && set->tags[i]) in blk_mq_map_swqueue()
2189 blk_mq_free_map_and_requests(set, i); in blk_mq_map_swqueue()
2195 hctx->tags = set->tags[i]; in blk_mq_map_swqueue()
2220 int i; in queue_set_hctx_shared() local
2222 queue_for_each_hw_ctx(q, hctx, i) { in queue_set_hctx_shared()
2295 unsigned int i; in blk_mq_release() local
2298 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_release()
2350 int i, j; in blk_mq_realloc_hw_ctxs() local
2357 for (i = 0; i < set->nr_hw_queues; i++) { in blk_mq_realloc_hw_ctxs()
2360 if (hctxs[i]) in blk_mq_realloc_hw_ctxs()
2363 node = blk_mq_hw_queue_to_node(q->mq_map, i); in blk_mq_realloc_hw_ctxs()
2364 hctxs[i] = kzalloc_node(blk_mq_hw_ctx_size(set), in blk_mq_realloc_hw_ctxs()
2366 if (!hctxs[i]) in blk_mq_realloc_hw_ctxs()
2369 if (!zalloc_cpumask_var_node(&hctxs[i]->cpumask, GFP_KERNEL, in blk_mq_realloc_hw_ctxs()
2371 kfree(hctxs[i]); in blk_mq_realloc_hw_ctxs()
2372 hctxs[i] = NULL; in blk_mq_realloc_hw_ctxs()
2376 atomic_set(&hctxs[i]->nr_active, 0); in blk_mq_realloc_hw_ctxs()
2377 hctxs[i]->numa_node = node; in blk_mq_realloc_hw_ctxs()
2378 hctxs[i]->queue_num = i; in blk_mq_realloc_hw_ctxs()
2380 if (blk_mq_init_hctx(q, set, hctxs[i], i)) { in blk_mq_realloc_hw_ctxs()
2381 free_cpumask_var(hctxs[i]->cpumask); in blk_mq_realloc_hw_ctxs()
2382 kfree(hctxs[i]); in blk_mq_realloc_hw_ctxs()
2383 hctxs[i] = NULL; in blk_mq_realloc_hw_ctxs()
2386 blk_mq_hctx_kobj_init(hctxs[i]); in blk_mq_realloc_hw_ctxs()
2388 for (j = i; j < q->nr_hw_queues; j++) { in blk_mq_realloc_hw_ctxs()
2400 q->nr_hw_queues = i; in blk_mq_realloc_hw_ctxs()
2520 int i; in __blk_mq_alloc_rq_maps() local
2522 for (i = 0; i < set->nr_hw_queues; i++) in __blk_mq_alloc_rq_maps()
2523 if (!__blk_mq_alloc_rq_map(set, i)) in __blk_mq_alloc_rq_maps()
2529 while (--i >= 0) in __blk_mq_alloc_rq_maps()
2530 blk_mq_free_rq_map(set->tags[i]); in __blk_mq_alloc_rq_maps()
2675 int i; in blk_mq_free_tag_set() local
2677 for (i = 0; i < nr_cpu_ids; i++) in blk_mq_free_tag_set()
2678 blk_mq_free_map_and_requests(set, i); in blk_mq_free_tag_set()
2692 int i, ret; in blk_mq_update_nr_requests() local
2700 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_update_nr_requests()