/block/ |
D | elevator.c | 63 struct elevator_queue *e = q->elevator; in elv_iosched_allow_bio_merge() local 65 if (e->type->ops.allow_merge) in elv_iosched_allow_bio_merge() 66 return e->type->ops.allow_merge(q, rq, bio); in elv_iosched_allow_bio_merge() 87 const struct elevator_type *e) in elv_support_features() argument 89 return (q->required_elevator_features & e->elevator_features) == in elv_support_features() 100 static bool elevator_match(const struct elevator_type *e, const char *name) in elevator_match() argument 102 return !strcmp(e->elevator_name, name) || in elevator_match() 103 (e->elevator_alias && !strcmp(e->elevator_alias, name)); in elevator_match() 108 struct elevator_type *e; in __elevator_find() local 110 list_for_each_entry(e, &elv_list, list) in __elevator_find() [all …]
|
D | blk-mq-sched.h | 21 int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e); 22 void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e); 41 struct elevator_queue *e = q->elevator; in blk_mq_sched_allow_merge() local 43 if (e->type->ops.allow_merge) in blk_mq_sched_allow_merge() 44 return e->type->ops.allow_merge(q, rq, bio); in blk_mq_sched_allow_merge() 52 struct elevator_queue *e = rq->q->elevator; in blk_mq_sched_completed_request() local 54 if (e->type->ops.completed_request) in blk_mq_sched_completed_request() 55 e->type->ops.completed_request(rq, now); in blk_mq_sched_completed_request() 63 struct elevator_queue *e = q->elevator; in blk_mq_sched_requeue_request() local 65 if (e->type->ops.requeue_request) in blk_mq_sched_requeue_request() [all …]
|
D | blk-mq-sched.c | 90 struct elevator_queue *e = q->elevator; in __blk_mq_do_dispatch_sched() local 106 if (e->type->ops.has_work && !e->type->ops.has_work(hctx)) in __blk_mq_do_dispatch_sched() 118 rq = e->type->ops.dispatch_request(hctx); in __blk_mq_do_dispatch_sched() 342 struct elevator_queue *e = q->elevator; in blk_mq_sched_bio_merge() local 348 if (e && e->type->ops.bio_merge) { in blk_mq_sched_bio_merge() 349 ret = e->type->ops.bio_merge(q, bio, nr_segs); in blk_mq_sched_bio_merge() 443 int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e) in blk_mq_init_sched() argument 471 ret = e->ops.init_sched(q, e); in blk_mq_init_sched() 480 if (e->ops.init_hctx) { in blk_mq_init_sched() 481 ret = e->ops.init_hctx(hctx, i); in blk_mq_init_sched() [all …]
|
D | badblocks.c | 207 sector_t e = a + BB_LEN(p[lo]); in badblocks_set() local 210 if (e >= s) { in badblocks_set() 212 if (s == a && s + sectors >= e) in badblocks_set() 218 if (e < s + sectors) in badblocks_set() 219 e = s + sectors; in badblocks_set() 220 if (e - a <= BB_MAX_LEN) { in badblocks_set() 221 p[lo] = BB_MAKE(a, e-a, ack); in badblocks_set() 222 s = e; in badblocks_set() 231 sectors = e - s; in badblocks_set() 239 sector_t e = a + BB_LEN(p[hi]); in badblocks_set() local [all …]
|
D | elevator.h | 98 static inline bool elevator_tryget(struct elevator_type *e) in elevator_tryget() argument 100 return try_module_get(e->elevator_owner); in elevator_tryget() 103 static inline void __elevator_get(struct elevator_type *e) in __elevator_get() argument 105 __module_get(e->elevator_owner); in __elevator_get() 108 static inline void elevator_put(struct elevator_type *e) in elevator_put() argument 110 module_put(e->elevator_owner); in elevator_put()
|
D | kyber-iosched.c | 405 static int kyber_init_sched(struct request_queue *q, struct elevator_type *e) in kyber_init_sched() argument 410 eq = elevator_alloc(q, e); in kyber_init_sched() 430 static void kyber_exit_sched(struct elevator_queue *e) in kyber_exit_sched() argument 432 struct kyber_queue_data *kqd = e->elevator_data; in kyber_exit_sched() 864 static ssize_t kyber_##name##_lat_show(struct elevator_queue *e, \ 867 struct kyber_queue_data *kqd = e->elevator_data; \ 872 static ssize_t kyber_##name##_lat_store(struct elevator_queue *e, \ 875 struct kyber_queue_data *kqd = e->elevator_data; \
|
D | blk-mq-debugfs.c | 748 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched() local 759 if (!e->queue_debugfs_attrs) in blk_mq_debugfs_register_sched() 764 debugfs_create_files(q->sched_debugfs_dir, q, e->queue_debugfs_attrs); in blk_mq_debugfs_register_sched() 819 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched_hctx() local 831 if (!e->hctx_debugfs_attrs) in blk_mq_debugfs_register_sched_hctx() 837 e->hctx_debugfs_attrs); in blk_mq_debugfs_register_sched_hctx()
|
D | mq-deadline.c | 676 static void dd_exit_sched(struct elevator_queue *e) in dd_exit_sched() argument 678 struct deadline_data *dd = e->elevator_data; in dd_exit_sched() 705 static int dd_init_sched(struct request_queue *q, struct elevator_type *e) in dd_init_sched() argument 712 eq = elevator_alloc(q, e); in dd_init_sched() 985 static ssize_t __FUNC(struct elevator_queue *e, char *page) \ 987 struct deadline_data *dd = e->elevator_data; \ 1003 static ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count) \ 1005 struct deadline_data *dd = e->elevator_data; \
|
D | bfq-iosched.c | 7138 static void bfq_exit_queue(struct elevator_queue *e) in bfq_exit_queue() argument 7140 struct bfq_data *bfqd = e->elevator_data; in bfq_exit_queue() 7170 clear_bit(ELEVATOR_FLAG_DISABLE_WBT, &e->flags); in bfq_exit_queue() 7192 static int bfq_init_queue(struct request_queue *q, struct elevator_type *e) in bfq_init_queue() argument 7199 eq = elevator_alloc(q, e); in bfq_init_queue() 7400 static ssize_t __FUNC(struct elevator_queue *e, char *page) \ 7402 struct bfq_data *bfqd = e->elevator_data; \ 7422 static ssize_t __FUNC(struct elevator_queue *e, char *page) \ 7424 struct bfq_data *bfqd = e->elevator_data; \ 7434 __FUNC(struct elevator_queue *e, const char *page, size_t count) \ [all …]
|
D | blk-mq.c | 392 struct elevator_queue *e = data->q->elevator; in blk_mq_rq_ctx_init() local 397 if (e->type->ops.prepare_request) in blk_mq_rq_ctx_init() 398 e->type->ops.prepare_request(rq); in blk_mq_rq_ctx_init()
|
/block/partitions/ |
D | Kconfig | 187 label, i.e. DOS partition table. It does not support GPT labelled
|