Home
last modified time | relevance | path

Searched refs:next (Results 1 – 25 of 25) sorted by relevance

/block/partitions/
Dcmdline.c74 char *next = strchr(++partdef, ')'); in parse_subpart() local
76 if (!next) { in parse_subpart()
82 length = min_t(int, next - partdef, in parse_subpart()
87 partdef = ++next; in parse_subpart()
124 char *next; in parse_parts() local
136 next = strchr(bdevdef, ':'); in parse_parts()
137 if (!next) { in parse_parts()
142 length = min_t(int, next - bdevdef, sizeof(newparts->name) - 1); in parse_parts()
149 while (next && *(++next)) { in parse_parts()
150 bdevdef = next; in parse_parts()
[all …]
Dmsdos.c150 if (state->next == state->limit) in parse_extended()
174 sector_t offs, size, next; in parse_extended() local
183 next = this_sector + offs; in parse_extended()
187 if (next < first_sector) in parse_extended()
189 if (next + size > first_sector + first_size) in parse_extended()
193 put_partition(state, state->next, next, size); in parse_extended()
194 set_info(state, state->next, disksig); in parse_extended()
196 state->parts[state->next].flags = ADDPART_FLAG_RAID; in parse_extended()
198 if (++state->next == state->limit) in parse_extended()
283 for (i = 0; i < max_nparts && state->next < state->limit; i++) { in parse_solaris_x86()
[all …]
Dacorn.c529 sector_t next; in adfspart_check_EESOX() local
534 next = le32_to_cpu(p->start); in adfspart_check_EESOX()
536 put_partition(state, slot++, start, next - start); in adfspart_check_EESOX()
537 start = next; in adfspart_check_EESOX()
Dcheck.h21 int next; member
/block/
Dblk-merge.c18 struct request *prev_rq, struct bio *prev, struct bio *next) in bio_will_gap() argument
47 bio_get_first_bvec(next, &nb); in bio_will_gap()
640 struct request *next) in req_attempt_discard_merge() argument
646 if (blk_rq_sectors(req) + bio_sectors(next->bio) > in req_attempt_discard_merge()
650 req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next); in req_attempt_discard_merge()
658 struct request *next) in ll_merge_requests_fn() argument
662 if (req_gap_back_merge(req, next->bio)) in ll_merge_requests_fn()
668 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > in ll_merge_requests_fn()
672 total_phys_segments = req->nr_phys_segments + next->nr_phys_segments; in ll_merge_requests_fn()
676 if (!blk_cgroup_mergeable(req, next->bio)) in ll_merge_requests_fn()
[all …]
Dblk-rq-qos.c37 rqos = rqos->next; in __rq_qos_cleanup()
46 rqos = rqos->next; in __rq_qos_done()
55 rqos = rqos->next; in __rq_qos_issue()
64 rqos = rqos->next; in __rq_qos_requeue()
73 rqos = rqos->next; in __rq_qos_throttle()
82 rqos = rqos->next; in __rq_qos_track()
91 rqos = rqos->next; in __rq_qos_merge()
100 rqos = rqos->next; in __rq_qos_done_bio()
109 rqos = rqos->next; in __rq_qos_queue_depth_changed()
301 q->rq_qos = rqos->next; in rq_qos_exit()
Dblk-integrity.c165 struct request *next) in blk_integrity_merge_rq() argument
167 if (blk_integrity_rq(req) == 0 && blk_integrity_rq(next) == 0) in blk_integrity_merge_rq()
170 if (blk_integrity_rq(req) == 0 || blk_integrity_rq(next) == 0) in blk_integrity_merge_rq()
174 bio_integrity(next->bio)->bip_flags) in blk_integrity_merge_rq()
177 if (req->nr_integrity_segments + next->nr_integrity_segments > in blk_integrity_merge_rq()
181 if (integrity_req_gap_back_merge(req, next->bio)) in blk_integrity_merge_rq()
191 struct bio *next = bio->bi_next; in blk_integrity_merge_bio() local
204 bio->bi_next = next; in blk_integrity_merge_bio()
Dblk-rq-qos.h32 struct rq_qos *next; member
66 for (rqos = q->rq_qos; rqos; rqos = rqos->next) { in rq_qos_id()
103 rqos->next = q->rq_qos; in rq_qos_add()
131 for (cur = &q->rq_qos; *cur; cur = &(*cur)->next) { in rq_qos_del()
133 *cur = rqos->next; in rq_qos_del()
Dblk-crypto-internal.h46 struct request *next) in bio_crypt_ctx_merge_rq() argument
49 next->crypt_ctx); in bio_crypt_ctx_merge_rq()
89 struct request *next) in bio_crypt_ctx_merge_rq() argument
Dmq-deadline.c221 struct request *next) in dd_merged_requests() argument
224 const u8 ioprio_class = dd_rq_ioclass(next); in dd_merged_requests()
235 if (!list_empty(&req->queuelist) && !list_empty(&next->queuelist)) { in dd_merged_requests()
236 if (time_before((unsigned long)next->fifo_time, in dd_merged_requests()
238 list_move(&req->queuelist, &next->queuelist); in dd_merged_requests()
239 req->fifo_time = next->fifo_time; in dd_merged_requests()
246 deadline_remove_request(q, &dd->per_prio[prio], next); in dd_merged_requests()
283 struct request *rq = rq_entry_fifo(per_prio->fifo_list[data_dir].next); in deadline_check_fifo()
341 rq = rq_entry_fifo(per_prio->fifo_list[data_dir].next); in deadline_fifo_request()
1026 .next = deadline_##name##_fifo_next, \
[all …]
Dblk.h114 struct bio *next) in integrity_req_gap_back_merge() argument
117 struct bio_integrity_payload *bip_next = bio_integrity(next); in integrity_req_gap_back_merge()
147 struct bio *next) in integrity_req_gap_back_merge() argument
233 struct request *next);
Dt10-pi.c68 goto next; in t10_pi_verify()
81 goto next; in t10_pi_verify()
94 next: in t10_pi_verify()
Dblk-mq.c587 struct request *rq, *next; in blk_complete_reqs() local
589 llist_for_each_entry_safe(rq, next, entry, ipi_list) in blk_complete_reqs()
781 struct request *rq, *next; in blk_mq_requeue_work() local
787 list_for_each_entry_safe(rq, next, &rq_list, queuelist) { in blk_mq_requeue_work()
805 rq = list_entry(rq_list.next, struct request, queuelist); in blk_mq_requeue_work()
904 static bool blk_mq_req_expired(struct request *rq, unsigned long *next) in blk_mq_req_expired() argument
917 if (*next == 0) in blk_mq_req_expired()
918 *next = deadline; in blk_mq_req_expired()
919 else if (time_after(*next, deadline)) in blk_mq_req_expired()
920 *next = deadline; in blk_mq_req_expired()
[all …]
Dgenhd.c182 struct blk_major_name *next; member
202 for (dp = major_names[major_to_index(offset)]; dp; dp = dp->next) in blkdev_show()
274 p->next = NULL; in __register_blkdev()
278 for (n = &major_names[index]; *n; n = &(*n)->next) { in __register_blkdev()
307 for (n = &major_names[index]; *n; n = &(*n)->next) in unregister_blkdev()
314 *n = p->next; in unregister_blkdev()
685 for (n = &major_names[major_to_index(major)]; *n; n = &(*n)->next) { in blk_request_module()
835 .next = disk_seqf_next,
1226 .next = disk_seqf_next,
Dbadblocks.c177 sector_t next = s + sectors; in badblocks_set() local
180 next += (1<<bb->shift) - 1; in badblocks_set()
181 next >>= bb->shift; in badblocks_set()
182 sectors = next - s; in badblocks_set()
Dbfq-iosched.c930 rq = rq_entry_fifo(bfqq->fifo.next); in bfq_check_fifo()
945 struct request *next, *prev = NULL; in bfq_find_next_rq() local
948 next = bfq_check_fifo(bfqq, last); in bfq_find_next_rq()
949 if (next) in bfq_find_next_rq()
950 return next; in bfq_find_next_rq()
956 next = rb_entry_rq(rbnext); in bfq_find_next_rq()
960 next = rb_entry_rq(rbnext); in bfq_find_next_rq()
963 return bfq_choose_req(bfqd, next, prev, blk_rq_pos(last)); in bfq_find_next_rq()
2445 struct request *next) in bfq_requests_merged() argument
2448 *next_bfqq = RQ_BFQQ(next); in bfq_requests_merged()
[all …]
Delevator.c232 struct hlist_node *next; in elv_rqhash_find() local
235 hash_for_each_possible_safe(e->hash, rq, next, hash, offset) { in elv_rqhash_find()
412 struct request *next) in elv_merge_requests() argument
417 e->type->ops.requests_merged(q, rq, next); in elv_merge_requests()
Dbfq-wf2q.c342 struct rb_node *next; in bfq_idle_extract() local
345 next = rb_next(&entity->rb_node); in bfq_idle_extract()
346 st->first_idle = bfq_entity_of(next); in bfq_idle_extract()
350 next = rb_prev(&entity->rb_node); in bfq_idle_extract()
351 st->last_idle = bfq_entity_of(next); in bfq_idle_extract()
Dblk-mq-debugfs.c70 .next = queue_requeue_list_next,
388 .next = hctx_dispatch_next,
657 .next = ctx_##name##_rq_list_next, \
853 rqos = rqos->next; in blk_mq_debugfs_register()
Dblk-iolatency.c685 goto next; in blkiolatency_timer_fn()
691 goto next; in blkiolatency_timer_fn()
716 next: in blkiolatency_timer_fn()
Dbio.c1377 struct bio *bio, *next; in bio_dirty_fn() local
1380 next = bio_dirty_list; in bio_dirty_fn()
1384 while ((bio = next) != NULL) { in bio_dirty_fn()
1385 next = bio->bi_private; in bio_dirty_fn()
Dkyber-iosched.c593 struct request *rq, *next; in kyber_insert_requests() local
595 list_for_each_entry_safe(rq, next, rq_list, queuelist) { in kyber_insert_requests()
938 .next = kyber_##name##_rqs_next, \
Dblk-mq.h313 struct request *rq = list_entry_rq(list->next); in blk_mq_free_requests()
Dblk-ioc.c218 struct io_cq *icq = list_entry(icq_list->next, in __ioc_clear_queue()
Dsed-opal.c2121 struct opal_suspend_data *suspend, *next; in clean_opal_dev() local
2124 list_for_each_entry_safe(suspend, next, &dev->unlk_lst, node) { in clean_opal_dev()