/block/ |
D | blk-lib.c | 39 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) in __blkdev_issue_discard() argument 67 bio = blk_next_bio(bio, bdev, 0, REQ_OP_DISCARD, gfp_mask); in __blkdev_issue_discard() 98 sector_t nr_sects, gfp_t gfp_mask) in blkdev_issue_discard() argument 105 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, &bio); in blkdev_issue_discard() 119 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_write_zeroes() argument 135 bio = blk_next_bio(bio, bdev, 0, REQ_OP_WRITE_ZEROES, gfp_mask); in __blkdev_issue_write_zeroes() 169 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_zero_pages() argument 181 REQ_OP_WRITE, gfp_mask); in __blkdev_issue_zero_pages() 219 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, in __blkdev_issue_zeroout() argument 229 ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp_mask, in __blkdev_issue_zeroout() [all …]
|
D | blk-map.c | 22 gfp_t gfp_mask) in bio_alloc_map_data() argument 29 bmd = kmalloc(struct_size(bmd, iov, data->nr_segs), gfp_mask); in bio_alloc_map_data() 130 struct iov_iter *iter, gfp_t gfp_mask) in bio_copy_user_iov() argument 140 bmd = bio_alloc_map_data(iter, gfp_mask); in bio_copy_user_iov() 155 bio = bio_kmalloc(nr_pages, gfp_mask); in bio_copy_user_iov() 183 page = alloc_page(GFP_NOIO | gfp_mask); in bio_copy_user_iov() 252 unsigned int nr_vecs, gfp_t gfp_mask) in blk_rq_map_bio_alloc() argument 259 bio = bio_alloc_bioset(NULL, nr_vecs, opf, gfp_mask, in blk_rq_map_bio_alloc() 264 bio = bio_kmalloc(nr_vecs, gfp_mask); in blk_rq_map_bio_alloc() 273 gfp_t gfp_mask) in bio_map_user_iov() argument [all …]
|
D | blk-crypto.c | 102 const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask) in bio_crypt_set_ctx() argument 110 WARN_ON_ONCE(!(gfp_mask & __GFP_DIRECT_RECLAIM)); in bio_crypt_set_ctx() 112 bc = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in bio_crypt_set_ctx() 127 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask) in __bio_crypt_clone() argument 129 dst->bi_crypt_context = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in __bio_crypt_clone() 315 gfp_t gfp_mask) in __blk_crypto_rq_bio_prep() argument 318 rq->crypt_ctx = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in __blk_crypto_rq_bio_prep()
|
D | bio.c | 179 gfp_t gfp_mask) in bvec_alloc() argument 200 bvl = kmem_cache_alloc(bvs->slab, bvec_alloc_gfp(gfp_mask)); in bvec_alloc() 201 if (likely(bvl) || !(gfp_mask & __GFP_DIRECT_RECLAIM)) in bvec_alloc() 206 return mempool_alloc(pool, gfp_mask); in bvec_alloc() 474 blk_opf_t opf, gfp_t gfp_mask, in bio_alloc_bioset() argument 477 gfp_t saved_gfp = gfp_mask; in bio_alloc_bioset() 488 gfp_mask, bs); in bio_alloc_bioset() 522 gfp_mask &= ~__GFP_DIRECT_RECLAIM; in bio_alloc_bioset() 524 p = mempool_alloc(&bs->bio_pool, gfp_mask); in bio_alloc_bioset() 525 if (!p && gfp_mask != saved_gfp) { in bio_alloc_bioset() [all …]
|
D | bio-integrity.c | 49 gfp_t gfp_mask, in bio_integrity_alloc() argument 60 bip = kmalloc(struct_size(bip, bip_inline_vecs, nr_vecs), gfp_mask); in bio_integrity_alloc() 63 bip = mempool_alloc(&bs->bio_integrity_pool, gfp_mask); in bio_integrity_alloc() 75 &bip->bip_max_vcnt, gfp_mask); in bio_integrity_alloc() 400 gfp_t gfp_mask) in bio_integrity_clone() argument 407 bip = bio_integrity_alloc(bio, gfp_mask, bip_src->bip_vcnt); in bio_integrity_clone()
|
D | blk-zoned.c | 187 gfp_t gfp_mask) in blkdev_zone_reset_all_emulated() argument 214 gfp_mask); in blkdev_zone_reset_all_emulated() 232 static int blkdev_zone_reset_all(struct block_device *bdev, gfp_t gfp_mask) in blkdev_zone_reset_all() argument 257 sector_t sector, sector_t nr_sectors, gfp_t gfp_mask) in blkdev_zone_mgmt() argument 294 return blkdev_zone_reset_all_emulated(bdev, gfp_mask); in blkdev_zone_mgmt() 295 return blkdev_zone_reset_all(bdev, gfp_mask); in blkdev_zone_mgmt() 299 bio = blk_next_bio(bio, bdev, 0, op | REQ_SYNC, gfp_mask); in blkdev_zone_mgmt()
|
D | blk-crypto-internal.h | 190 gfp_t gfp_mask); 202 gfp_t gfp_mask) in blk_crypto_rq_bio_prep() argument 205 return __blk_crypto_rq_bio_prep(rq, bio, gfp_mask); in blk_crypto_rq_bio_prep()
|
D | blk-cgroup.c | 212 gfp_t gfp_mask) in blkg_alloc() argument 218 blkg = kzalloc_node(sizeof(*blkg), gfp_mask, disk->queue->node); in blkg_alloc() 222 if (percpu_ref_init(&blkg->refcnt, blkg_release, 0, gfp_mask)) in blkg_alloc() 225 blkg->iostat_cpu = alloc_percpu_gfp(struct blkg_iostat_set, gfp_mask); in blkg_alloc() 251 pd = pol->pd_alloc_fn(gfp_mask, disk->queue, blkcg); in blkg_alloc()
|
D | blk.h | 91 gfp_t gfp_mask);
|
D | blk-mq.c | 3137 struct bio_set *bs, gfp_t gfp_mask, in blk_rq_prep_clone() argument 3147 bio = bio_alloc_clone(rq->q->disk->part0, bio_src, gfp_mask, in blk_rq_prep_clone() 3174 if (rq->bio && blk_crypto_rq_bio_prep(rq, rq->bio, gfp_mask) < 0) in blk_rq_prep_clone()
|