Home
last modified time | relevance | path

Searched refs:e (Results 1 – 11 of 11) sorted by relevance

/block/
Delevator.c63 struct elevator_queue *e = q->elevator; in elv_iosched_allow_bio_merge() local
65 if (e->type->ops.allow_merge) in elv_iosched_allow_bio_merge()
66 return e->type->ops.allow_merge(q, rq, bio); in elv_iosched_allow_bio_merge()
87 const struct elevator_type *e) in elv_support_features() argument
89 return (q->required_elevator_features & e->elevator_features) == in elv_support_features()
100 static bool elevator_match(const struct elevator_type *e, const char *name) in elevator_match() argument
102 return !strcmp(e->elevator_name, name) || in elevator_match()
103 (e->elevator_alias && !strcmp(e->elevator_alias, name)); in elevator_match()
108 struct elevator_type *e; in __elevator_find() local
110 list_for_each_entry(e, &elv_list, list) in __elevator_find()
[all …]
Dblk-mq-sched.h21 int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e);
22 void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e);
41 struct elevator_queue *e = q->elevator; in blk_mq_sched_allow_merge() local
43 if (e->type->ops.allow_merge) in blk_mq_sched_allow_merge()
44 return e->type->ops.allow_merge(q, rq, bio); in blk_mq_sched_allow_merge()
52 struct elevator_queue *e = rq->q->elevator; in blk_mq_sched_completed_request() local
54 if (e->type->ops.completed_request) in blk_mq_sched_completed_request()
55 e->type->ops.completed_request(rq, now); in blk_mq_sched_completed_request()
63 struct elevator_queue *e = q->elevator; in blk_mq_sched_requeue_request() local
65 if (e->type->ops.requeue_request) in blk_mq_sched_requeue_request()
[all …]
Dblk-mq-sched.c90 struct elevator_queue *e = q->elevator; in __blk_mq_do_dispatch_sched() local
106 if (e->type->ops.has_work && !e->type->ops.has_work(hctx)) in __blk_mq_do_dispatch_sched()
118 rq = e->type->ops.dispatch_request(hctx); in __blk_mq_do_dispatch_sched()
342 struct elevator_queue *e = q->elevator; in blk_mq_sched_bio_merge() local
348 if (e && e->type->ops.bio_merge) { in blk_mq_sched_bio_merge()
349 ret = e->type->ops.bio_merge(q, bio, nr_segs); in blk_mq_sched_bio_merge()
443 int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e) in blk_mq_init_sched() argument
471 ret = e->ops.init_sched(q, e); in blk_mq_init_sched()
480 if (e->ops.init_hctx) { in blk_mq_init_sched()
481 ret = e->ops.init_hctx(hctx, i); in blk_mq_init_sched()
[all …]
Dbadblocks.c207 sector_t e = a + BB_LEN(p[lo]); in badblocks_set() local
210 if (e >= s) { in badblocks_set()
212 if (s == a && s + sectors >= e) in badblocks_set()
218 if (e < s + sectors) in badblocks_set()
219 e = s + sectors; in badblocks_set()
220 if (e - a <= BB_MAX_LEN) { in badblocks_set()
221 p[lo] = BB_MAKE(a, e-a, ack); in badblocks_set()
222 s = e; in badblocks_set()
231 sectors = e - s; in badblocks_set()
239 sector_t e = a + BB_LEN(p[hi]); in badblocks_set() local
[all …]
Delevator.h98 static inline bool elevator_tryget(struct elevator_type *e) in elevator_tryget() argument
100 return try_module_get(e->elevator_owner); in elevator_tryget()
103 static inline void __elevator_get(struct elevator_type *e) in __elevator_get() argument
105 __module_get(e->elevator_owner); in __elevator_get()
108 static inline void elevator_put(struct elevator_type *e) in elevator_put() argument
110 module_put(e->elevator_owner); in elevator_put()
Dkyber-iosched.c405 static int kyber_init_sched(struct request_queue *q, struct elevator_type *e) in kyber_init_sched() argument
410 eq = elevator_alloc(q, e); in kyber_init_sched()
430 static void kyber_exit_sched(struct elevator_queue *e) in kyber_exit_sched() argument
432 struct kyber_queue_data *kqd = e->elevator_data; in kyber_exit_sched()
864 static ssize_t kyber_##name##_lat_show(struct elevator_queue *e, \
867 struct kyber_queue_data *kqd = e->elevator_data; \
872 static ssize_t kyber_##name##_lat_store(struct elevator_queue *e, \
875 struct kyber_queue_data *kqd = e->elevator_data; \
Dblk-mq-debugfs.c748 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched() local
759 if (!e->queue_debugfs_attrs) in blk_mq_debugfs_register_sched()
764 debugfs_create_files(q->sched_debugfs_dir, q, e->queue_debugfs_attrs); in blk_mq_debugfs_register_sched()
819 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched_hctx() local
831 if (!e->hctx_debugfs_attrs) in blk_mq_debugfs_register_sched_hctx()
837 e->hctx_debugfs_attrs); in blk_mq_debugfs_register_sched_hctx()
Dmq-deadline.c676 static void dd_exit_sched(struct elevator_queue *e) in dd_exit_sched() argument
678 struct deadline_data *dd = e->elevator_data; in dd_exit_sched()
705 static int dd_init_sched(struct request_queue *q, struct elevator_type *e) in dd_init_sched() argument
712 eq = elevator_alloc(q, e); in dd_init_sched()
985 static ssize_t __FUNC(struct elevator_queue *e, char *page) \
987 struct deadline_data *dd = e->elevator_data; \
1003 static ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count) \
1005 struct deadline_data *dd = e->elevator_data; \
Dbfq-iosched.c7138 static void bfq_exit_queue(struct elevator_queue *e) in bfq_exit_queue() argument
7140 struct bfq_data *bfqd = e->elevator_data; in bfq_exit_queue()
7170 clear_bit(ELEVATOR_FLAG_DISABLE_WBT, &e->flags); in bfq_exit_queue()
7192 static int bfq_init_queue(struct request_queue *q, struct elevator_type *e) in bfq_init_queue() argument
7199 eq = elevator_alloc(q, e); in bfq_init_queue()
7400 static ssize_t __FUNC(struct elevator_queue *e, char *page) \
7402 struct bfq_data *bfqd = e->elevator_data; \
7422 static ssize_t __FUNC(struct elevator_queue *e, char *page) \
7424 struct bfq_data *bfqd = e->elevator_data; \
7434 __FUNC(struct elevator_queue *e, const char *page, size_t count) \
[all …]
Dblk-mq.c392 struct elevator_queue *e = data->q->elevator; in blk_mq_rq_ctx_init() local
397 if (e->type->ops.prepare_request) in blk_mq_rq_ctx_init()
398 e->type->ops.prepare_request(rq); in blk_mq_rq_ctx_init()
/block/partitions/
DKconfig187 label, i.e. DOS partition table. It does not support GPT labelled