/block/partitions/ |
D | cmdline.c | 74 char *next = strchr(++partdef, ')'); in parse_subpart() local 76 if (!next) { in parse_subpart() 82 length = min_t(int, next - partdef, in parse_subpart() 87 partdef = ++next; in parse_subpart() 124 char *next; in parse_parts() local 136 next = strchr(bdevdef, ':'); in parse_parts() 137 if (!next) { in parse_parts() 142 length = min_t(int, next - bdevdef, sizeof(newparts->name) - 1); in parse_parts() 149 while (next && *(++next)) { in parse_parts() 150 bdevdef = next; in parse_parts() [all …]
|
D | msdos.c | 150 if (state->next == state->limit) in parse_extended() 174 sector_t offs, size, next; in parse_extended() local 183 next = this_sector + offs; in parse_extended() 187 if (next < first_sector) in parse_extended() 189 if (next + size > first_sector + first_size) in parse_extended() 193 put_partition(state, state->next, next, size); in parse_extended() 194 set_info(state, state->next, disksig); in parse_extended() 196 state->parts[state->next].flags = ADDPART_FLAG_RAID; in parse_extended() 198 if (++state->next == state->limit) in parse_extended() 283 for (i = 0; i < max_nparts && state->next < state->limit; i++) { in parse_solaris_x86() [all …]
|
D | acorn.c | 529 sector_t next; in adfspart_check_EESOX() local 534 next = le32_to_cpu(p->start); in adfspart_check_EESOX() 536 put_partition(state, slot++, start, next - start); in adfspart_check_EESOX() 537 start = next; in adfspart_check_EESOX()
|
D | check.h | 21 int next; member
|
/block/ |
D | blk-merge.c | 18 struct request *prev_rq, struct bio *prev, struct bio *next) in bio_will_gap() argument 47 bio_get_first_bvec(next, &nb); in bio_will_gap() 640 struct request *next) in req_attempt_discard_merge() argument 646 if (blk_rq_sectors(req) + bio_sectors(next->bio) > in req_attempt_discard_merge() 650 req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next); in req_attempt_discard_merge() 658 struct request *next) in ll_merge_requests_fn() argument 662 if (req_gap_back_merge(req, next->bio)) in ll_merge_requests_fn() 668 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > in ll_merge_requests_fn() 672 total_phys_segments = req->nr_phys_segments + next->nr_phys_segments; in ll_merge_requests_fn() 676 if (!blk_cgroup_mergeable(req, next->bio)) in ll_merge_requests_fn() [all …]
|
D | blk-rq-qos.c | 37 rqos = rqos->next; in __rq_qos_cleanup() 46 rqos = rqos->next; in __rq_qos_done() 55 rqos = rqos->next; in __rq_qos_issue() 64 rqos = rqos->next; in __rq_qos_requeue() 73 rqos = rqos->next; in __rq_qos_throttle() 82 rqos = rqos->next; in __rq_qos_track() 91 rqos = rqos->next; in __rq_qos_merge() 100 rqos = rqos->next; in __rq_qos_done_bio() 109 rqos = rqos->next; in __rq_qos_queue_depth_changed() 301 q->rq_qos = rqos->next; in rq_qos_exit()
|
D | blk-integrity.c | 165 struct request *next) in blk_integrity_merge_rq() argument 167 if (blk_integrity_rq(req) == 0 && blk_integrity_rq(next) == 0) in blk_integrity_merge_rq() 170 if (blk_integrity_rq(req) == 0 || blk_integrity_rq(next) == 0) in blk_integrity_merge_rq() 174 bio_integrity(next->bio)->bip_flags) in blk_integrity_merge_rq() 177 if (req->nr_integrity_segments + next->nr_integrity_segments > in blk_integrity_merge_rq() 181 if (integrity_req_gap_back_merge(req, next->bio)) in blk_integrity_merge_rq() 191 struct bio *next = bio->bi_next; in blk_integrity_merge_bio() local 204 bio->bi_next = next; in blk_integrity_merge_bio()
|
D | blk-rq-qos.h | 32 struct rq_qos *next; member 66 for (rqos = q->rq_qos; rqos; rqos = rqos->next) { in rq_qos_id() 103 rqos->next = q->rq_qos; in rq_qos_add() 131 for (cur = &q->rq_qos; *cur; cur = &(*cur)->next) { in rq_qos_del() 133 *cur = rqos->next; in rq_qos_del()
|
D | blk-crypto-internal.h | 46 struct request *next) in bio_crypt_ctx_merge_rq() argument 49 next->crypt_ctx); in bio_crypt_ctx_merge_rq() 89 struct request *next) in bio_crypt_ctx_merge_rq() argument
|
D | mq-deadline.c | 221 struct request *next) in dd_merged_requests() argument 224 const u8 ioprio_class = dd_rq_ioclass(next); in dd_merged_requests() 235 if (!list_empty(&req->queuelist) && !list_empty(&next->queuelist)) { in dd_merged_requests() 236 if (time_before((unsigned long)next->fifo_time, in dd_merged_requests() 238 list_move(&req->queuelist, &next->queuelist); in dd_merged_requests() 239 req->fifo_time = next->fifo_time; in dd_merged_requests() 246 deadline_remove_request(q, &dd->per_prio[prio], next); in dd_merged_requests() 283 struct request *rq = rq_entry_fifo(per_prio->fifo_list[data_dir].next); in deadline_check_fifo() 341 rq = rq_entry_fifo(per_prio->fifo_list[data_dir].next); in deadline_fifo_request() 1026 .next = deadline_##name##_fifo_next, \ [all …]
|
D | blk.h | 114 struct bio *next) in integrity_req_gap_back_merge() argument 117 struct bio_integrity_payload *bip_next = bio_integrity(next); in integrity_req_gap_back_merge() 147 struct bio *next) in integrity_req_gap_back_merge() argument 233 struct request *next);
|
D | t10-pi.c | 68 goto next; in t10_pi_verify() 81 goto next; in t10_pi_verify() 94 next: in t10_pi_verify()
|
D | blk-mq.c | 587 struct request *rq, *next; in blk_complete_reqs() local 589 llist_for_each_entry_safe(rq, next, entry, ipi_list) in blk_complete_reqs() 781 struct request *rq, *next; in blk_mq_requeue_work() local 787 list_for_each_entry_safe(rq, next, &rq_list, queuelist) { in blk_mq_requeue_work() 805 rq = list_entry(rq_list.next, struct request, queuelist); in blk_mq_requeue_work() 904 static bool blk_mq_req_expired(struct request *rq, unsigned long *next) in blk_mq_req_expired() argument 917 if (*next == 0) in blk_mq_req_expired() 918 *next = deadline; in blk_mq_req_expired() 919 else if (time_after(*next, deadline)) in blk_mq_req_expired() 920 *next = deadline; in blk_mq_req_expired() [all …]
|
D | genhd.c | 182 struct blk_major_name *next; member 202 for (dp = major_names[major_to_index(offset)]; dp; dp = dp->next) in blkdev_show() 274 p->next = NULL; in __register_blkdev() 278 for (n = &major_names[index]; *n; n = &(*n)->next) { in __register_blkdev() 307 for (n = &major_names[index]; *n; n = &(*n)->next) in unregister_blkdev() 314 *n = p->next; in unregister_blkdev() 685 for (n = &major_names[major_to_index(major)]; *n; n = &(*n)->next) { in blk_request_module() 835 .next = disk_seqf_next, 1226 .next = disk_seqf_next,
|
D | badblocks.c | 177 sector_t next = s + sectors; in badblocks_set() local 180 next += (1<<bb->shift) - 1; in badblocks_set() 181 next >>= bb->shift; in badblocks_set() 182 sectors = next - s; in badblocks_set()
|
D | bfq-iosched.c | 930 rq = rq_entry_fifo(bfqq->fifo.next); in bfq_check_fifo() 945 struct request *next, *prev = NULL; in bfq_find_next_rq() local 948 next = bfq_check_fifo(bfqq, last); in bfq_find_next_rq() 949 if (next) in bfq_find_next_rq() 950 return next; in bfq_find_next_rq() 956 next = rb_entry_rq(rbnext); in bfq_find_next_rq() 960 next = rb_entry_rq(rbnext); in bfq_find_next_rq() 963 return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last)); in bfq_find_next_rq() 2445 struct request *next) in bfq_requests_merged() argument 2448 *next_bfqq = RQ_BFQQ(next); in bfq_requests_merged() [all …]
|
D | elevator.c | 232 struct hlist_node *next; in elv_rqhash_find() local 235 hash_for_each_possible_safe(e->hash, rq, next, hash, offset) { in elv_rqhash_find() 412 struct request *next) in elv_merge_requests() argument 417 e->type->ops.requests_merged(q, rq, next); in elv_merge_requests()
|
D | bfq-wf2q.c | 342 struct rb_node *next; in bfq_idle_extract() local 345 next = rb_next(&entity->rb_node); in bfq_idle_extract() 346 st->first_idle = bfq_entity_of(next); in bfq_idle_extract() 350 next = rb_prev(&entity->rb_node); in bfq_idle_extract() 351 st->last_idle = bfq_entity_of(next); in bfq_idle_extract()
|
D | blk-mq-debugfs.c | 70 .next = queue_requeue_list_next, 388 .next = hctx_dispatch_next, 657 .next = ctx_##name##_rq_list_next, \ 853 rqos = rqos->next; in blk_mq_debugfs_register()
|
D | blk-iolatency.c | 685 goto next; in blkiolatency_timer_fn() 691 goto next; in blkiolatency_timer_fn() 716 next: in blkiolatency_timer_fn()
|
D | bio.c | 1377 struct bio *bio, *next; in bio_dirty_fn() local 1380 next = bio_dirty_list; in bio_dirty_fn() 1384 while ((bio = next) != NULL) { in bio_dirty_fn() 1385 next = bio->bi_private; in bio_dirty_fn()
|
D | kyber-iosched.c | 593 struct request *rq, *next; in kyber_insert_requests() local 595 list_for_each_entry_safe(rq, next, rq_list, queuelist) { in kyber_insert_requests() 938 .next = kyber_##name##_rqs_next, \
|
D | blk-mq.h | 313 struct request *rq = list_entry_rq(list->next); in blk_mq_free_requests()
|
D | blk-ioc.c | 218 struct io_cq *icq = list_entry(icq_list->next, in __ioc_clear_queue()
|
D | sed-opal.c | 2121 struct opal_suspend_data *suspend, *next; in clean_opal_dev() local 2124 list_for_each_entry_safe(suspend, next, &dev->unlk_lst, node) { in clean_opal_dev()
|