/block/ |
D | blk-softirq.c | 46 struct list_head *list; in trigger_softirq() local 49 list = this_cpu_ptr(&blk_cpu_done); in trigger_softirq() 50 list_add_tail(&rq->ipi_list, list); in trigger_softirq() 52 if (list->next == &rq->ipi_list) in trigger_softirq() 128 struct list_head *list; in __blk_complete_request() local 130 list = this_cpu_ptr(&blk_cpu_done); in __blk_complete_request() 131 list_add_tail(&req->ipi_list, list); in __blk_complete_request() 139 if (list->next == &req->ipi_list) in __blk_complete_request()
|
D | blk-mq-sched.c | 262 bool blk_mq_bio_list_merge(struct request_queue *q, struct list_head *list, in blk_mq_bio_list_merge() argument 268 list_for_each_entry_reverse(rq, list, queuelist) { in blk_mq_bio_list_merge() 398 LIST_HEAD(list); in blk_mq_sched_insert_request() 400 list_add(&rq->queuelist, &list); in blk_mq_sched_insert_request() 401 e->type->ops.insert_requests(hctx, &list, at_head); in blk_mq_sched_insert_request() 415 struct list_head *list, bool run_queue_async) in blk_mq_sched_insert_requests() argument 429 e->type->ops.insert_requests(hctx, list, false); in blk_mq_sched_insert_requests() 437 blk_mq_try_issue_list_directly(hctx, list); in blk_mq_sched_insert_requests() 438 if (list_empty(list)) in blk_mq_sched_insert_requests() 441 blk_mq_insert_requests(hctx, ctx, list); in blk_mq_sched_insert_requests()
|
D | blk-flush.c | 179 list_move_tail(&rq->flush.list, pending); in blk_flush_complete_seq() 183 list_move_tail(&rq->flush.list, &fq->flush_data_in_flight); in blk_flush_complete_seq() 195 list_del_init(&rq->flush.list); in blk_flush_complete_seq() 244 list_for_each_entry_safe(rq, n, running, flush.list) { in flush_end_io() 273 list_first_entry(pending, struct request, flush.list); in blk_kick_flush() 411 INIT_LIST_HEAD(&rq->flush.list); in blk_insert_flush()
|
D | blk-mq.c | 976 struct list_head *list; member 987 list_splice_tail_init(&ctx->rq_lists[type], flush_data->list); in flush_busy_ctx() 997 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list) in blk_mq_flush_busy_ctxs() argument 1001 .list = list, in blk_mq_flush_busy_ctxs() 1211 bool blk_mq_dispatch_rq_list(struct request_queue *q, struct list_head *list, in blk_mq_dispatch_rq_list() argument 1220 if (list_empty(list)) in blk_mq_dispatch_rq_list() 1223 WARN_ON(!list_is_singular(list) && got_budget); in blk_mq_dispatch_rq_list() 1232 rq = list_first_entry(list, struct request, queuelist); in blk_mq_dispatch_rq_list() 1266 if (list_empty(list)) in blk_mq_dispatch_rq_list() 1269 nxt = list_first_entry(list, struct request, queuelist); in blk_mq_dispatch_rq_list() [all …]
|
D | blk-mq.h | 46 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list); 71 struct list_head *list); 76 struct list_head *list);
|
D | blk-stat.c | 64 list_for_each_entry_rcu(cb, &q->stats->callbacks, list) { in blk_stat_add() 150 list_add_tail_rcu(&cb->list, &q->stats->callbacks); in blk_stat_add_callback() 159 list_del_rcu(&cb->list); in blk_stat_remove_callback()
|
D | blk-core.c | 1396 void blk_steal_bios(struct bio_list *list, struct request *rq) in blk_steal_bios() argument 1399 if (list->tail) in blk_steal_bios() 1400 list->tail->bi_next = rq->bio; in blk_steal_bios() 1402 list->head = rq->bio; in blk_steal_bios() 1403 list->tail = rq->biotail; in blk_steal_bios() 1738 list); in flush_plug_callbacks() 1739 list_del(&cb->list); in flush_plug_callbacks() 1754 list_for_each_entry(cb, &plug->cb_list, list) in blk_check_plugged() 1764 list_add(&cb->list, &plug->cb_list); in blk_check_plugged()
|
D | elevator.c | 127 list_for_each_entry(e, &elv_list, list) { in elevator_find() 553 list_add_tail(&e->list, &elv_list); in elv_register() 566 list_del_init(&e->list); in elv_unregister() 647 list_for_each_entry(e, &elv_list, list) { in elevator_get_by_features() 793 list_for_each_entry(__e, &elv_list, list) { in elv_iosched_show()
|
D | blk-stat.h | 23 struct list_head list; member
|
D | kyber-iosched.c | 660 struct list_head *list; member 670 flush_data->list); in flush_busy_kcq() 679 struct list_head *list) in kyber_flush_busy_kcqs() argument 684 .list = list, in kyber_flush_busy_kcqs()
|
D | blk-mq-sched.h | 26 struct list_head *list, bool run_queue_async);
|
D | mq-deadline.c | 524 struct list_head *list, bool at_head) in dd_insert_requests() argument 530 while (!list_empty(list)) { in dd_insert_requests() 533 rq = list_first_entry(list, struct request, queuelist); in dd_insert_requests()
|
D | bfq-iosched.c | 5534 struct list_head *list, bool at_head) in bfq_insert_requests() argument 5536 while (!list_empty(list)) { in bfq_insert_requests() 5539 rq = list_first_entry(list, struct request, queuelist); in bfq_insert_requests()
|
/block/partitions/ |
D | ldm.c | 545 struct vblk *v = list_entry (item, struct vblk, list); in ldm_get_disk_objid() 591 vb = list_entry (item, struct vblk, list); in ldm_create_data_partitions() 1187 list_add (&vb->list, &ldb->v_dgrp); in ldm_ldmdb_add() 1191 list_add (&vb->list, &ldb->v_disk); in ldm_ldmdb_add() 1194 list_add (&vb->list, &ldb->v_volu); in ldm_ldmdb_add() 1197 list_add (&vb->list, &ldb->v_comp); in ldm_ldmdb_add() 1202 struct vblk *v = list_entry (item, struct vblk, list); in ldm_ldmdb_add() 1205 list_add_tail (&vb->list, &v->list); in ldm_ldmdb_add() 1209 list_add_tail (&vb->list, &ldb->v_part); in ldm_ldmdb_add() 1253 f = list_entry (item, struct frag, list); in ldm_frag_add() [all …]
|
D | ldm.h | 91 struct list_head list; member 182 struct list_head list; member
|