Home
last modified time | relevance | path

Searched refs:iter (Results 1 – 12 of 12) sorted by relevance

/block/
Dt10-pi.c29 static blk_status_t t10_pi_generate(struct blk_integrity_iter *iter, in t10_pi_generate() argument
34 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_generate()
35 struct t10_pi_tuple *pi = iter->prot_buf; in t10_pi_generate()
37 pi->guard_tag = fn(iter->data_buf, iter->interval); in t10_pi_generate()
41 pi->ref_tag = cpu_to_be32(lower_32_bits(iter->seed)); in t10_pi_generate()
45 iter->data_buf += iter->interval; in t10_pi_generate()
46 iter->prot_buf += sizeof(struct t10_pi_tuple); in t10_pi_generate()
47 iter->seed++; in t10_pi_generate()
53 static blk_status_t t10_pi_verify(struct blk_integrity_iter *iter, in t10_pi_verify() argument
60 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_verify()
[all …]
Dblk-map.c21 struct bvec_iter iter; in blk_rq_append_bio() local
27 bio_for_each_bvec(bv, *bio, iter) in blk_rq_append_bio()
65 struct rq_map_data *map_data, struct iov_iter *iter, in __blk_rq_map_user_iov() argument
73 bio = bio_copy_user_iov(q, map_data, iter, gfp_mask); in __blk_rq_map_user_iov()
75 bio = bio_map_user_iov(q, iter, gfp_mask); in __blk_rq_map_user_iov()
122 const struct iov_iter *iter, gfp_t gfp_mask) in blk_rq_map_user_iov() argument
130 if (!iter_is_iovec(iter)) in blk_rq_map_user_iov()
135 else if (iov_iter_alignment(iter) & align) in blk_rq_map_user_iov()
138 copy = queue_virt_boundary(q) & iov_iter_gap_alignment(iter); in blk_rq_map_user_iov()
140 i = *iter; in blk_rq_map_user_iov()
Dbio.c533 struct bvec_iter iter; in zero_fill_bio_iter() local
535 __bio_for_each_segment(bv, bio, iter, start) { in zero_fill_bio_iter()
557 struct bvec_iter iter; in bio_truncate() local
567 bio_for_each_segment(bv, bio, iter) { in bio_truncate()
890 static int __bio_iov_bvec_add_pages(struct bio *bio, struct iov_iter *iter) in __bio_iov_bvec_add_pages() argument
892 const struct bio_vec *bv = iter->bvec; in __bio_iov_bvec_add_pages()
896 if (WARN_ON_ONCE(iter->iov_offset > bv->bv_len)) in __bio_iov_bvec_add_pages()
899 len = min_t(size_t, bv->bv_len - iter->iov_offset, iter->count); in __bio_iov_bvec_add_pages()
901 bv->bv_offset + iter->iov_offset); in __bio_iov_bvec_add_pages()
904 iov_iter_advance(iter, size); in __bio_iov_bvec_add_pages()
[all …]
Dbio-integrity.c157 struct blk_integrity_iter iter; in bio_integrity_process() local
165 iter.disk_name = bio->bi_disk->disk_name; in bio_integrity_process()
166 iter.interval = 1 << bi->interval_exp; in bio_integrity_process()
167 iter.seed = proc_iter->bi_sector; in bio_integrity_process()
168 iter.prot_buf = prot_buf; in bio_integrity_process()
173 iter.data_buf = kaddr + bv.bv_offset; in bio_integrity_process()
174 iter.data_size = bv.bv_len; in bio_integrity_process()
176 ret = proc_fn(&iter); in bio_integrity_process()
Dbounce.c136 struct bvec_iter iter; in copy_to_high_bio_irq() local
144 bio_for_each_segment(tovec, to, iter) { in copy_to_high_bio_irq()
220 struct bvec_iter iter; in bounce_clone_bio() local
265 bio_for_each_segment(bv, bio_src, iter) in bounce_clone_bio()
290 struct bvec_iter iter; in __blk_queue_bounce() local
296 bio_for_each_segment(from, *bio_orig, iter) { in __blk_queue_bounce()
Dgenhd.c947 struct class_dev_iter iter; in printk_all_partitions() local
950 class_dev_iter_init(&iter, &block_class, NULL, &disk_type); in printk_all_partitions()
951 while ((dev = class_dev_iter_next(&iter))) { in printk_all_partitions()
991 class_dev_iter_exit(&iter); in printk_all_partitions()
999 struct class_dev_iter *iter; in disk_seqf_start() local
1002 iter = kmalloc(sizeof(*iter), GFP_KERNEL); in disk_seqf_start()
1003 if (!iter) in disk_seqf_start()
1006 seqf->private = iter; in disk_seqf_start()
1007 class_dev_iter_init(iter, &block_class, NULL, &disk_type); in disk_seqf_start()
1009 dev = class_dev_iter_next(iter); in disk_seqf_start()
[all …]
Dblk-integrity.c32 struct bvec_iter iter; in blk_rq_count_integrity_sg() local
35 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_count_integrity_sg()
74 struct bvec_iter iter; in blk_rq_map_integrity_sg() local
77 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_map_integrity_sg()
366 static blk_status_t blk_integrity_nop_fn(struct blk_integrity_iter *iter) in blk_integrity_nop_fn() argument
Dblk-crypto-fallback.c170 struct bvec_iter iter; in blk_crypto_clone_bio() local
184 bio_for_each_segment(bv, bio_src, iter) in blk_crypto_clone_bio()
230 struct bvec_iter iter; in blk_crypto_split_bio_if_needed() local
232 bio_for_each_segment(bv, bio, iter) { in blk_crypto_split_bio_if_needed()
404 struct bvec_iter iter; in blk_crypto_decrypt_bio() local
435 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { in blk_crypto_decrypt_bio()
Dblk-crypto.c43 struct bvec_iter iter; in bio_crypt_check_alignment() local
46 bio_for_each_segment(bv, bio, iter) { in bio_crypt_check_alignment()
Dblk-merge.c246 struct bvec_iter iter; in blk_bio_segment_split() local
251 bio_for_each_bvec(bv, bio, iter) { in blk_bio_segment_split()
360 struct req_iterator iter; in blk_recalc_rq_segments() local
375 rq_for_each_bvec(bv, rq, iter) in blk_recalc_rq_segments()
467 struct bvec_iter iter; in __blk_bios_map_sg() local
472 bio_for_each_bvec(bvec, bio, iter) { in __blk_bios_map_sg()
Dsed-opal.c850 struct opal_resp_tok *iter; in response_parse() local
884 iter = resp->toks; in response_parse()
889 token_length = response_parse_tiny(iter, pos); in response_parse()
891 token_length = response_parse_short(iter, pos); in response_parse()
893 token_length = response_parse_medium(iter, pos); in response_parse()
895 token_length = response_parse_long(iter, pos); in response_parse()
897 token_length = response_parse_token(iter, pos); in response_parse()
904 iter++; in response_parse()
1071 struct opal_suspend_data *iter; in add_suspend_info() local
1073 list_for_each_entry(iter, &dev->unlk_lst, node) { in add_suspend_info()
[all …]
Dblk-core.c1533 struct req_iterator iter; in rq_flush_dcache_pages() local
1536 rq_for_each_segment(bvec, rq, iter) in rq_flush_dcache_pages()