/block/partitions/ |
D | ldm.c | 545 struct vblk *v = list_entry (item, struct vblk, list); in ldm_get_disk_objid() 591 vb = list_entry (item, struct vblk, list); in ldm_create_data_partitions() 1187 list_add (&vb->list, &ldb->v_dgrp); in ldm_ldmdb_add() 1191 list_add (&vb->list, &ldb->v_disk); in ldm_ldmdb_add() 1194 list_add (&vb->list, &ldb->v_volu); in ldm_ldmdb_add() 1197 list_add (&vb->list, &ldb->v_comp); in ldm_ldmdb_add() 1202 struct vblk *v = list_entry (item, struct vblk, list); in ldm_ldmdb_add() 1205 list_add_tail (&vb->list, &v->list); in ldm_ldmdb_add() 1209 list_add_tail (&vb->list, &ldb->v_part); in ldm_ldmdb_add() 1253 f = list_entry (item, struct frag, list); in ldm_frag_add() [all …]
|
D | ldm.h | 88 struct list_head list; member 179 struct list_head list; member
|
/block/ |
D | blk-flush.c | 187 list_move_tail(&rq->flush.list, pending); in blk_flush_complete_seq() 191 list_move_tail(&rq->flush.list, &fq->flush_data_in_flight); in blk_flush_complete_seq() 203 list_del_init(&rq->flush.list); in blk_flush_complete_seq() 258 list_for_each_entry_safe(rq, n, running, flush.list) { in flush_end_io() 291 list_first_entry(pending, struct request, flush.list); in blk_kick_flush() 436 INIT_LIST_HEAD(&rq->flush.list); in blk_insert_flush()
|
D | blk-mq-sched.c | 467 LIST_HEAD(list); in blk_mq_sched_insert_request() 469 list_add(&rq->queuelist, &list); in blk_mq_sched_insert_request() 470 e->type->ops.insert_requests(hctx, &list, at_head); in blk_mq_sched_insert_request() 484 struct list_head *list, bool run_queue_async) in blk_mq_sched_insert_requests() argument 498 e->type->ops.insert_requests(hctx, list, false); in blk_mq_sched_insert_requests() 506 blk_mq_try_issue_list_directly(hctx, list); in blk_mq_sched_insert_requests() 507 if (list_empty(list)) in blk_mq_sched_insert_requests() 510 blk_mq_insert_requests(hctx, ctx, list); in blk_mq_sched_insert_requests()
|
D | blk-mq.c | 607 struct list_head *list; in blk_mq_trigger_softirq() local 611 list = this_cpu_ptr(&blk_cpu_done); in blk_mq_trigger_softirq() 612 list_add_tail(&rq->ipi_list, list); in blk_mq_trigger_softirq() 619 if (list->next == &rq->ipi_list) in blk_mq_trigger_softirq() 1025 struct list_head *list; member 1036 list_splice_tail_init(&ctx->rq_lists[type], flush_data->list); in flush_busy_ctx() 1046 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list) in blk_mq_flush_busy_ctxs() argument 1050 .list = list, in blk_mq_flush_busy_ctxs() 1281 struct list_head *list) in blk_mq_handle_dev_resource() argument 1284 list_first_entry_or_null(list, struct request, queuelist); in blk_mq_handle_dev_resource() [all …]
|
D | blk-stat.c | 65 list_for_each_entry_rcu(cb, &q->stats->callbacks, list) { in blk_stat_add() 152 list_add_tail_rcu(&cb->list, &q->stats->callbacks); in blk_stat_add_callback() 163 list_del_rcu(&cb->list); in blk_stat_remove_callback()
|
D | blk-mq.h | 49 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list); 76 struct list_head *list); 81 struct list_head *list);
|
D | blk-core.c | 1400 void blk_steal_bios(struct bio_list *list, struct request *rq) in blk_steal_bios() argument 1403 if (list->tail) in blk_steal_bios() 1404 list->tail->bi_next = rq->bio; in blk_steal_bios() 1406 list->head = rq->bio; in blk_steal_bios() 1407 list->tail = rq->biotail; in blk_steal_bios() 1738 list); in flush_plug_callbacks() 1739 list_del(&cb->list); in flush_plug_callbacks() 1754 list_for_each_entry(cb, &plug->cb_list, list) in blk_check_plugged() 1764 list_add(&cb->list, &plug->cb_list); in blk_check_plugged()
|
D | elevator.c | 127 list_for_each_entry(e, &elv_list, list) { in elevator_find() 547 list_add_tail(&e->list, &elv_list); in elv_register() 560 list_del_init(&e->list); in elv_unregister() 644 list_for_each_entry(e, &elv_list, list) { in elevator_get_by_features() 790 list_for_each_entry(__e, &elv_list, list) { in elv_iosched_show()
|
D | kyber-iosched.c | 661 struct list_head *list; member 671 flush_data->list); in flush_busy_kcq() 680 struct list_head *list) in kyber_flush_busy_kcqs() argument 685 .list = list, in kyber_flush_busy_kcqs()
|
D | blk-stat.h | 23 struct list_head list; member
|
D | blk-mq-sched.h | 23 struct list_head *list, bool run_queue_async);
|
D | mq-deadline-main.c | 775 struct list_head *list, bool at_head) in dd_insert_requests() argument 781 while (!list_empty(list)) { in dd_insert_requests() 784 rq = list_first_entry(list, struct request, queuelist); in dd_insert_requests()
|
D | blk-merge.c | 1094 bool blk_bio_list_merge(struct request_queue *q, struct list_head *list, in blk_bio_list_merge() argument 1100 list_for_each_entry_reverse(rq, list, queuelist) { in blk_bio_list_merge()
|
D | blk.h | 185 bool blk_bio_list_merge(struct request_queue *q, struct list_head *list,
|
D | bfq-iosched.c | 5576 struct list_head *list, bool at_head) in bfq_insert_requests() argument 5578 while (!list_empty(list)) { in bfq_insert_requests() 5581 rq = list_first_entry(list, struct request, queuelist); in bfq_insert_requests()
|