Home
last modified time | relevance | path

Searched refs:list (Results 1 – 15 of 15) sorted by relevance

/block/
Dblk-softirq.c46 struct list_head *list; in trigger_softirq() local
49 list = this_cpu_ptr(&blk_cpu_done); in trigger_softirq()
50 list_add_tail(&rq->ipi_list, list); in trigger_softirq()
52 if (list->next == &rq->ipi_list) in trigger_softirq()
128 struct list_head *list; in __blk_complete_request() local
130 list = this_cpu_ptr(&blk_cpu_done); in __blk_complete_request()
131 list_add_tail(&req->ipi_list, list); in __blk_complete_request()
139 if (list->next == &req->ipi_list) in __blk_complete_request()
Dblk-mq-sched.c262 bool blk_mq_bio_list_merge(struct request_queue *q, struct list_head *list, in blk_mq_bio_list_merge() argument
268 list_for_each_entry_reverse(rq, list, queuelist) { in blk_mq_bio_list_merge()
398 LIST_HEAD(list); in blk_mq_sched_insert_request()
400 list_add(&rq->queuelist, &list); in blk_mq_sched_insert_request()
401 e->type->ops.insert_requests(hctx, &list, at_head); in blk_mq_sched_insert_request()
415 struct list_head *list, bool run_queue_async) in blk_mq_sched_insert_requests() argument
429 e->type->ops.insert_requests(hctx, list, false); in blk_mq_sched_insert_requests()
437 blk_mq_try_issue_list_directly(hctx, list); in blk_mq_sched_insert_requests()
438 if (list_empty(list)) in blk_mq_sched_insert_requests()
441 blk_mq_insert_requests(hctx, ctx, list); in blk_mq_sched_insert_requests()
Dblk-flush.c179 list_move_tail(&rq->flush.list, pending); in blk_flush_complete_seq()
183 list_move_tail(&rq->flush.list, &fq->flush_data_in_flight); in blk_flush_complete_seq()
195 list_del_init(&rq->flush.list); in blk_flush_complete_seq()
244 list_for_each_entry_safe(rq, n, running, flush.list) { in flush_end_io()
273 list_first_entry(pending, struct request, flush.list); in blk_kick_flush()
411 INIT_LIST_HEAD(&rq->flush.list); in blk_insert_flush()
Dblk-mq.c976 struct list_head *list; member
987 list_splice_tail_init(&ctx->rq_lists[type], flush_data->list); in flush_busy_ctx()
997 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list) in blk_mq_flush_busy_ctxs() argument
1001 .list = list, in blk_mq_flush_busy_ctxs()
1211 bool blk_mq_dispatch_rq_list(struct request_queue *q, struct list_head *list, in blk_mq_dispatch_rq_list() argument
1220 if (list_empty(list)) in blk_mq_dispatch_rq_list()
1223 WARN_ON(!list_is_singular(list) && got_budget); in blk_mq_dispatch_rq_list()
1232 rq = list_first_entry(list, struct request, queuelist); in blk_mq_dispatch_rq_list()
1266 if (list_empty(list)) in blk_mq_dispatch_rq_list()
1269 nxt = list_first_entry(list, struct request, queuelist); in blk_mq_dispatch_rq_list()
[all …]
Dblk-mq.h46 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list);
71 struct list_head *list);
76 struct list_head *list);
Dblk-stat.c64 list_for_each_entry_rcu(cb, &q->stats->callbacks, list) { in blk_stat_add()
150 list_add_tail_rcu(&cb->list, &q->stats->callbacks); in blk_stat_add_callback()
159 list_del_rcu(&cb->list); in blk_stat_remove_callback()
Dblk-core.c1396 void blk_steal_bios(struct bio_list *list, struct request *rq) in blk_steal_bios() argument
1399 if (list->tail) in blk_steal_bios()
1400 list->tail->bi_next = rq->bio; in blk_steal_bios()
1402 list->head = rq->bio; in blk_steal_bios()
1403 list->tail = rq->biotail; in blk_steal_bios()
1738 list); in flush_plug_callbacks()
1739 list_del(&cb->list); in flush_plug_callbacks()
1754 list_for_each_entry(cb, &plug->cb_list, list) in blk_check_plugged()
1764 list_add(&cb->list, &plug->cb_list); in blk_check_plugged()
Delevator.c127 list_for_each_entry(e, &elv_list, list) { in elevator_find()
553 list_add_tail(&e->list, &elv_list); in elv_register()
566 list_del_init(&e->list); in elv_unregister()
647 list_for_each_entry(e, &elv_list, list) { in elevator_get_by_features()
793 list_for_each_entry(__e, &elv_list, list) { in elv_iosched_show()
Dblk-stat.h23 struct list_head list; member
Dkyber-iosched.c660 struct list_head *list; member
670 flush_data->list); in flush_busy_kcq()
679 struct list_head *list) in kyber_flush_busy_kcqs() argument
684 .list = list, in kyber_flush_busy_kcqs()
Dblk-mq-sched.h26 struct list_head *list, bool run_queue_async);
Dmq-deadline.c524 struct list_head *list, bool at_head) in dd_insert_requests() argument
530 while (!list_empty(list)) { in dd_insert_requests()
533 rq = list_first_entry(list, struct request, queuelist); in dd_insert_requests()
Dbfq-iosched.c5534 struct list_head *list, bool at_head) in bfq_insert_requests() argument
5536 while (!list_empty(list)) { in bfq_insert_requests()
5539 rq = list_first_entry(list, struct request, queuelist); in bfq_insert_requests()
/block/partitions/
Dldm.c545 struct vblk *v = list_entry (item, struct vblk, list); in ldm_get_disk_objid()
591 vb = list_entry (item, struct vblk, list); in ldm_create_data_partitions()
1187 list_add (&vb->list, &ldb->v_dgrp); in ldm_ldmdb_add()
1191 list_add (&vb->list, &ldb->v_disk); in ldm_ldmdb_add()
1194 list_add (&vb->list, &ldb->v_volu); in ldm_ldmdb_add()
1197 list_add (&vb->list, &ldb->v_comp); in ldm_ldmdb_add()
1202 struct vblk *v = list_entry (item, struct vblk, list); in ldm_ldmdb_add()
1205 list_add_tail (&vb->list, &v->list); in ldm_ldmdb_add()
1209 list_add_tail (&vb->list, &ldb->v_part); in ldm_ldmdb_add()
1253 f = list_entry (item, struct frag, list); in ldm_frag_add()
[all …]
Dldm.h91 struct list_head list; member
182 struct list_head list; member