/block/ |
D | t10-pi.c | 29 static blk_status_t t10_pi_generate(struct blk_integrity_iter *iter, in t10_pi_generate() argument 34 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_generate() 35 struct t10_pi_tuple *pi = iter->prot_buf; in t10_pi_generate() 37 pi->guard_tag = fn(iter->data_buf, iter->interval); in t10_pi_generate() 41 pi->ref_tag = cpu_to_be32(lower_32_bits(iter->seed)); in t10_pi_generate() 45 iter->data_buf += iter->interval; in t10_pi_generate() 46 iter->prot_buf += sizeof(struct t10_pi_tuple); in t10_pi_generate() 47 iter->seed++; in t10_pi_generate() 53 static blk_status_t t10_pi_verify(struct blk_integrity_iter *iter, in t10_pi_verify() argument 60 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_verify() [all …]
|
D | blk-map.c | 21 struct bvec_iter iter; in blk_rq_append_bio() local 27 bio_for_each_bvec(bv, *bio, iter) in blk_rq_append_bio() 65 struct rq_map_data *map_data, struct iov_iter *iter, in __blk_rq_map_user_iov() argument 73 bio = bio_copy_user_iov(q, map_data, iter, gfp_mask); in __blk_rq_map_user_iov() 75 bio = bio_map_user_iov(q, iter, gfp_mask); in __blk_rq_map_user_iov() 122 const struct iov_iter *iter, gfp_t gfp_mask) in blk_rq_map_user_iov() argument 130 if (!iter_is_iovec(iter)) in blk_rq_map_user_iov() 135 else if (iov_iter_alignment(iter) & align) in blk_rq_map_user_iov() 138 copy = queue_virt_boundary(q) & iov_iter_gap_alignment(iter); in blk_rq_map_user_iov() 140 i = *iter; in blk_rq_map_user_iov()
|
D | bio.c | 533 struct bvec_iter iter; in zero_fill_bio_iter() local 535 __bio_for_each_segment(bv, bio, iter, start) { in zero_fill_bio_iter() 557 struct bvec_iter iter; in bio_truncate() local 567 bio_for_each_segment(bv, bio, iter) { in bio_truncate() 890 static int __bio_iov_bvec_add_pages(struct bio *bio, struct iov_iter *iter) in __bio_iov_bvec_add_pages() argument 892 const struct bio_vec *bv = iter->bvec; in __bio_iov_bvec_add_pages() 896 if (WARN_ON_ONCE(iter->iov_offset > bv->bv_len)) in __bio_iov_bvec_add_pages() 899 len = min_t(size_t, bv->bv_len - iter->iov_offset, iter->count); in __bio_iov_bvec_add_pages() 901 bv->bv_offset + iter->iov_offset); in __bio_iov_bvec_add_pages() 904 iov_iter_advance(iter, size); in __bio_iov_bvec_add_pages() [all …]
|
D | bio-integrity.c | 157 struct blk_integrity_iter iter; in bio_integrity_process() local 165 iter.disk_name = bio->bi_disk->disk_name; in bio_integrity_process() 166 iter.interval = 1 << bi->interval_exp; in bio_integrity_process() 167 iter.seed = proc_iter->bi_sector; in bio_integrity_process() 168 iter.prot_buf = prot_buf; in bio_integrity_process() 173 iter.data_buf = kaddr + bv.bv_offset; in bio_integrity_process() 174 iter.data_size = bv.bv_len; in bio_integrity_process() 176 ret = proc_fn(&iter); in bio_integrity_process()
|
D | bounce.c | 136 struct bvec_iter iter; in copy_to_high_bio_irq() local 144 bio_for_each_segment(tovec, to, iter) { in copy_to_high_bio_irq() 220 struct bvec_iter iter; in bounce_clone_bio() local 265 bio_for_each_segment(bv, bio_src, iter) in bounce_clone_bio() 290 struct bvec_iter iter; in __blk_queue_bounce() local 296 bio_for_each_segment(from, *bio_orig, iter) { in __blk_queue_bounce()
|
D | genhd.c | 947 struct class_dev_iter iter; in printk_all_partitions() local 950 class_dev_iter_init(&iter, &block_class, NULL, &disk_type); in printk_all_partitions() 951 while ((dev = class_dev_iter_next(&iter))) { in printk_all_partitions() 991 class_dev_iter_exit(&iter); in printk_all_partitions() 999 struct class_dev_iter *iter; in disk_seqf_start() local 1002 iter = kmalloc(sizeof(*iter), GFP_KERNEL); in disk_seqf_start() 1003 if (!iter) in disk_seqf_start() 1006 seqf->private = iter; in disk_seqf_start() 1007 class_dev_iter_init(iter, &block_class, NULL, &disk_type); in disk_seqf_start() 1009 dev = class_dev_iter_next(iter); in disk_seqf_start() [all …]
|
D | blk-integrity.c | 32 struct bvec_iter iter; in blk_rq_count_integrity_sg() local 35 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_count_integrity_sg() 74 struct bvec_iter iter; in blk_rq_map_integrity_sg() local 77 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_map_integrity_sg() 366 static blk_status_t blk_integrity_nop_fn(struct blk_integrity_iter *iter) in blk_integrity_nop_fn() argument
|
D | blk-crypto-fallback.c | 170 struct bvec_iter iter; in blk_crypto_clone_bio() local 184 bio_for_each_segment(bv, bio_src, iter) in blk_crypto_clone_bio() 230 struct bvec_iter iter; in blk_crypto_split_bio_if_needed() local 232 bio_for_each_segment(bv, bio, iter) { in blk_crypto_split_bio_if_needed() 404 struct bvec_iter iter; in blk_crypto_decrypt_bio() local 435 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { in blk_crypto_decrypt_bio()
|
D | blk-crypto.c | 43 struct bvec_iter iter; in bio_crypt_check_alignment() local 46 bio_for_each_segment(bv, bio, iter) { in bio_crypt_check_alignment()
|
D | blk-merge.c | 246 struct bvec_iter iter; in blk_bio_segment_split() local 251 bio_for_each_bvec(bv, bio, iter) { in blk_bio_segment_split() 360 struct req_iterator iter; in blk_recalc_rq_segments() local 375 rq_for_each_bvec(bv, rq, iter) in blk_recalc_rq_segments() 467 struct bvec_iter iter; in __blk_bios_map_sg() local 472 bio_for_each_bvec(bvec, bio, iter) { in __blk_bios_map_sg()
|
D | sed-opal.c | 850 struct opal_resp_tok *iter; in response_parse() local 884 iter = resp->toks; in response_parse() 889 token_length = response_parse_tiny(iter, pos); in response_parse() 891 token_length = response_parse_short(iter, pos); in response_parse() 893 token_length = response_parse_medium(iter, pos); in response_parse() 895 token_length = response_parse_long(iter, pos); in response_parse() 897 token_length = response_parse_token(iter, pos); in response_parse() 904 iter++; in response_parse() 1071 struct opal_suspend_data *iter; in add_suspend_info() local 1073 list_for_each_entry(iter, &dev->unlk_lst, node) { in add_suspend_info() [all …]
|
D | blk-core.c | 1533 struct req_iterator iter; in rq_flush_dcache_pages() local 1536 rq_for_each_segment(bvec, rq, iter) in rq_flush_dcache_pages()
|