/block/ |
D | blk-throttle.c | 53 struct list_head node; /* service_queue->queued[] */ member 84 #define rb_entry_tg(node) rb_entry((node), struct throtl_grp, rb_node) argument 308 INIT_LIST_HEAD(&qn->node); in throtl_qnode_init() 327 if (list_empty(&qn->node)) { in throtl_qnode_add_bio() 328 list_add_tail(&qn->node, queued); in throtl_qnode_add_bio() 339 struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); in throtl_peek_queued() 367 struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); in throtl_pop_queued() 377 list_del_init(&qn->node); in throtl_pop_queued() 383 list_move_tail(&qn->node, queued); in throtl_pop_queued() 597 struct rb_node **node = &parent_sq->pending_tree.rb_node; in tg_service_queue_add() local [all …]
|
D | blk-mq-tag.c | 463 int node, bool reserved) in bt_alloc() argument 493 GFP_KERNEL, node); in bt_alloc() 524 int node) in blk_mq_init_bitmap_tags() argument 528 if (bt_alloc(&tags->bitmap_tags, depth, node, false)) in blk_mq_init_bitmap_tags() 530 if (bt_alloc(&tags->breserved_tags, tags->nr_reserved_tags, node, true)) in blk_mq_init_bitmap_tags() 541 unsigned int reserved_tags, int node) in blk_mq_init_tags() argument 550 tags = kzalloc_node(sizeof(*tags), GFP_KERNEL, node); in blk_mq_init_tags() 557 return blk_mq_init_bitmap_tags(tags, node); in blk_mq_init_tags()
|
D | blk-ioc.c | 234 int create_task_io_context(struct task_struct *task, gfp_t gfp_flags, int node) in create_task_io_context() argument 240 node); in create_task_io_context() 288 gfp_t gfp_flags, int node) in get_task_io_context() argument 303 } while (!create_task_io_context(task, gfp_flags, node)); in get_task_io_context() 365 q->node); in ioc_create_icq()
|
D | blk.h | 58 int node, int cmd_size); 250 int create_task_io_context(struct task_struct *task, gfp_t gfp_mask, int node); 264 static inline struct io_context *create_io_context(gfp_t gfp_mask, int node) in create_io_context() argument 268 create_task_io_context(current, gfp_mask, node); in create_io_context()
|
D | deadline-iosched.c | 68 struct rb_node *node = rb_next(&rq->rb_node); in deadline_latter_request() local 70 if (node) in deadline_latter_request() 71 return rb_entry_rq(node); in deadline_latter_request() 349 dd = kzalloc_node(sizeof(*dd), GFP_KERNEL, q->node); in deadline_init_queue()
|
D | blk-mq.c | 1397 static int blk_mq_alloc_bitmap(struct blk_mq_ctxmap *bitmap, int node) in blk_mq_alloc_bitmap() argument 1405 GFP_KERNEL, node); in blk_mq_alloc_bitmap() 1529 int node; in blk_mq_init_hctx() local 1532 node = hctx->numa_node; in blk_mq_init_hctx() 1533 if (node == NUMA_NO_NODE) in blk_mq_init_hctx() 1534 node = hctx->numa_node = set->numa_node; in blk_mq_init_hctx() 1556 GFP_KERNEL, node); in blk_mq_init_hctx() 1560 if (blk_mq_alloc_bitmap(&hctx->ctx_map, node)) in blk_mq_init_hctx() 1576 flush_start_tag + hctx_idx, node)) in blk_mq_init_hctx() 1786 int node = blk_mq_hw_queue_to_node(map, i); in blk_mq_init_queue() local [all …]
|
D | blk-flush.c | 504 int node, int cmd_size) in blk_alloc_flush_queue() argument 509 fq = kzalloc_node(sizeof(*fq), GFP_KERNEL, node); in blk_alloc_flush_queue() 518 fq->flush_rq = kzalloc_node(rq_sz, GFP_KERNEL, node); in blk_alloc_flush_queue()
|
D | blk-mq-tag.h | 48 …n struct blk_mq_tags *blk_mq_init_tags(unsigned int nr_tags, unsigned int reserved_tags, int node);
|
D | noop-iosched.c | 71 nd = kmalloc_node(sizeof(*nd), GFP_KERNEL, q->node); in noop_init_queue()
|
D | cfq-iosched.c | 68 #define rb_entry_cfqg(node) rb_entry((node), struct cfq_group, rb_node) argument 1250 struct rb_node **node = &st->rb.rb_node; in __cfq_group_service_tree_add() local 1256 while (*node != NULL) { in __cfq_group_service_tree_add() 1257 parent = *node; in __cfq_group_service_tree_add() 1261 node = &parent->rb_left; in __cfq_group_service_tree_add() 1263 node = &parent->rb_right; in __cfq_group_service_tree_add() 1271 rb_link_node(&cfqg->rb_node, parent, node); in __cfq_group_service_tree_add() 2590 struct rb_node *parent, *node; in cfqq_close() local 2614 node = rb_next(&__cfqq->p_node); in cfqq_close() 2616 node = rb_prev(&__cfqq->p_node); in cfqq_close() [all …]
|
D | blk-cgroup.c | 76 blkg = kzalloc_node(sizeof(*blkg), gfp_mask, q->node); in blkg_alloc() 100 pd = kzalloc_node(pol->pd_size, gfp_mask, q->node); in blkg_alloc() 1012 pd = kzalloc_node(pol->pd_size, GFP_KERNEL, q->node); in blkcg_activate_policy()
|
D | genhd.c | 1409 struct list_head node; /* all disk_event's */ member 1778 list_for_each_entry(ev, &disk_events, node) in disk_events_set_dfl_poll_msecs() 1813 INIT_LIST_HEAD(&ev->node); in disk_alloc_events() 1835 list_add_tail(&disk->ev->node, &disk_events); in disk_add_events() 1853 list_del_init(&disk->ev->node); in disk_del_events()
|
D | blk-core.c | 557 gfp_mask, q->node); in blk_init_rl() 595 q->node = node_id; in blk_alloc_queue_node() 1166 create_io_context(gfp_mask, q->node); in blk_old_get_request() 1835 create_io_context(GFP_ATOMIC, q->node); in generic_make_request_checks()
|
D | elevator.c | 158 eq = kzalloc_node(sizeof(*eq), GFP_KERNEL, q->node); in elevator_alloc()
|