Home
last modified time | relevance | path

Searched refs:ioc (Results 1 – 8 of 8) sorted by relevance

/block/
Dblk-ioc.c24 void get_io_context(struct io_context *ioc) in get_io_context() argument
26 BUG_ON(atomic_long_read(&ioc->refcount) <= 0); in get_io_context()
27 atomic_long_inc(&ioc->refcount); in get_io_context()
55 struct io_context *ioc = icq->ioc; in ioc_destroy_icq() local
59 lockdep_assert_held(&ioc->lock); in ioc_destroy_icq()
62 radix_tree_delete(&ioc->icq_tree, icq->q->id); in ioc_destroy_icq()
71 if (rcu_access_pointer(ioc->icq_hint) == icq) in ioc_destroy_icq()
72 rcu_assign_pointer(ioc->icq_hint, NULL); in ioc_destroy_icq()
90 struct io_context *ioc = container_of(work, struct io_context, in ioc_release_fn() local
100 spin_lock_irqsave_nested(&ioc->lock, flags, 1); in ioc_release_fn()
[all …]
Dioprio.c35 struct io_context *ioc; in set_task_ioprio() local
51 ioc = get_task_io_context(task, GFP_ATOMIC, NUMA_NO_NODE); in set_task_ioprio()
52 if (ioc) { in set_task_ioprio()
53 ioc->ioprio = ioprio; in set_task_ioprio()
54 put_io_context(ioc); in set_task_ioprio()
Dblk-core.c899 put_io_context(rq->elv.icq->ioc); in blk_free_request()
909 static inline int ioc_batching(struct request_queue *q, struct io_context *ioc) in ioc_batching() argument
911 if (!ioc) in ioc_batching()
919 return ioc->nr_batch_requests == q->nr_batching || in ioc_batching()
920 (ioc->nr_batch_requests > 0 in ioc_batching()
921 && time_before(jiffies, ioc->last_waited + BLK_BATCH_TIME)); in ioc_batching()
930 static void ioc_set_batching(struct request_queue *q, struct io_context *ioc) in ioc_set_batching() argument
932 if (!ioc || ioc_batching(q, ioc)) in ioc_set_batching()
935 ioc->nr_batch_requests = q->nr_batching; in ioc_set_batching()
936 ioc->last_waited = jiffies; in ioc_set_batching()
[all …]
Dblk.h264 void get_io_context(struct io_context *ioc);
265 struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q);
266 struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
DMakefile6 blk-flush.o blk-settings.o blk-ioc.o blk-map.o \
Dblk-cgroup.c1142 struct io_context *ioc; in blkcg_can_attach() local
1148 ioc = task->io_context; in blkcg_can_attach()
1149 if (ioc && atomic_read(&ioc->nr_tasks) > 1) in blkcg_can_attach()
Dbio.c2010 struct io_context *ioc; in bio_associate_current() local
2015 ioc = current->io_context; in bio_associate_current()
2016 if (!ioc) in bio_associate_current()
2019 get_io_context_active(ioc); in bio_associate_current()
2020 bio->bi_ioc = ioc; in bio_associate_current()
Dcfq-iosched.c872 struct io_context *ioc) in cfq_cic_lookup() argument
874 if (ioc) in cfq_cic_lookup()
875 return icq_to_cic(ioc_lookup_icq(ioc, cfqd->queue)); in cfq_cic_lookup()
2682 put_io_context(cfqd->active_cic->icq.ioc); in __cfq_slice_expired()
2936 if (!cic || !atomic_read(&cic->icq.ioc->active_ref)) in cfq_arm_slice_timer()
3497 atomic_long_inc(&cic->icq.ioc->refcount); in cfq_dispatch_request()
3680 int ioprio = cic->icq.ioc->ioprio; in check_ioprio_changed()
3906 else if (!atomic_read(&cic->icq.ioc->active_ref) || in cfq_update_idle_window()