Home
last modified time | relevance | path

Searched refs:gfp_mask (Results 1 – 10 of 10) sorted by relevance

/block/
Dblk-lib.c39 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) in __blkdev_issue_discard() argument
67 bio = blk_next_bio(bio, bdev, 0, REQ_OP_DISCARD, gfp_mask); in __blkdev_issue_discard()
98 sector_t nr_sects, gfp_t gfp_mask) in blkdev_issue_discard() argument
105 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, &bio); in blkdev_issue_discard()
119 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_write_zeroes() argument
135 bio = blk_next_bio(bio, bdev, 0, REQ_OP_WRITE_ZEROES, gfp_mask); in __blkdev_issue_write_zeroes()
169 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_zero_pages() argument
181 REQ_OP_WRITE, gfp_mask); in __blkdev_issue_zero_pages()
219 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, in __blkdev_issue_zeroout() argument
229 ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp_mask, in __blkdev_issue_zeroout()
[all …]
Dblk-map.c22 gfp_t gfp_mask) in bio_alloc_map_data() argument
29 bmd = kmalloc(struct_size(bmd, iov, data->nr_segs), gfp_mask); in bio_alloc_map_data()
130 struct iov_iter *iter, gfp_t gfp_mask) in bio_copy_user_iov() argument
140 bmd = bio_alloc_map_data(iter, gfp_mask); in bio_copy_user_iov()
155 bio = bio_kmalloc(nr_pages, gfp_mask); in bio_copy_user_iov()
183 page = alloc_page(GFP_NOIO | gfp_mask); in bio_copy_user_iov()
252 unsigned int nr_vecs, gfp_t gfp_mask) in blk_rq_map_bio_alloc() argument
259 bio = bio_alloc_bioset(NULL, nr_vecs, opf, gfp_mask, in blk_rq_map_bio_alloc()
264 bio = bio_kmalloc(nr_vecs, gfp_mask); in blk_rq_map_bio_alloc()
273 gfp_t gfp_mask) in bio_map_user_iov() argument
[all …]
Dblk-crypto.c102 const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask) in bio_crypt_set_ctx() argument
110 WARN_ON_ONCE(!(gfp_mask & __GFP_DIRECT_RECLAIM)); in bio_crypt_set_ctx()
112 bc = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in bio_crypt_set_ctx()
127 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask) in __bio_crypt_clone() argument
129 dst->bi_crypt_context = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in __bio_crypt_clone()
315 gfp_t gfp_mask) in __blk_crypto_rq_bio_prep() argument
318 rq->crypt_ctx = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in __blk_crypto_rq_bio_prep()
Dbio.c179 gfp_t gfp_mask) in bvec_alloc() argument
200 bvl = kmem_cache_alloc(bvs->slab, bvec_alloc_gfp(gfp_mask)); in bvec_alloc()
201 if (likely(bvl) || !(gfp_mask & __GFP_DIRECT_RECLAIM)) in bvec_alloc()
206 return mempool_alloc(pool, gfp_mask); in bvec_alloc()
474 blk_opf_t opf, gfp_t gfp_mask, in bio_alloc_bioset() argument
477 gfp_t saved_gfp = gfp_mask; in bio_alloc_bioset()
488 gfp_mask, bs); in bio_alloc_bioset()
522 gfp_mask &= ~__GFP_DIRECT_RECLAIM; in bio_alloc_bioset()
524 p = mempool_alloc(&bs->bio_pool, gfp_mask); in bio_alloc_bioset()
525 if (!p && gfp_mask != saved_gfp) { in bio_alloc_bioset()
[all …]
Dbio-integrity.c49 gfp_t gfp_mask, in bio_integrity_alloc() argument
60 bip = kmalloc(struct_size(bip, bip_inline_vecs, nr_vecs), gfp_mask); in bio_integrity_alloc()
63 bip = mempool_alloc(&bs->bio_integrity_pool, gfp_mask); in bio_integrity_alloc()
75 &bip->bip_max_vcnt, gfp_mask); in bio_integrity_alloc()
400 gfp_t gfp_mask) in bio_integrity_clone() argument
407 bip = bio_integrity_alloc(bio, gfp_mask, bip_src->bip_vcnt); in bio_integrity_clone()
Dblk-zoned.c187 gfp_t gfp_mask) in blkdev_zone_reset_all_emulated() argument
214 gfp_mask); in blkdev_zone_reset_all_emulated()
232 static int blkdev_zone_reset_all(struct block_device *bdev, gfp_t gfp_mask) in blkdev_zone_reset_all() argument
257 sector_t sector, sector_t nr_sectors, gfp_t gfp_mask) in blkdev_zone_mgmt() argument
294 return blkdev_zone_reset_all_emulated(bdev, gfp_mask); in blkdev_zone_mgmt()
295 return blkdev_zone_reset_all(bdev, gfp_mask); in blkdev_zone_mgmt()
299 bio = blk_next_bio(bio, bdev, 0, op | REQ_SYNC, gfp_mask); in blkdev_zone_mgmt()
Dblk-crypto-internal.h190 gfp_t gfp_mask);
202 gfp_t gfp_mask) in blk_crypto_rq_bio_prep() argument
205 return __blk_crypto_rq_bio_prep(rq, bio, gfp_mask); in blk_crypto_rq_bio_prep()
Dblk-cgroup.c212 gfp_t gfp_mask) in blkg_alloc() argument
218 blkg = kzalloc_node(sizeof(*blkg), gfp_mask, disk->queue->node); in blkg_alloc()
222 if (percpu_ref_init(&blkg->refcnt, blkg_release, 0, gfp_mask)) in blkg_alloc()
225 blkg->iostat_cpu = alloc_percpu_gfp(struct blkg_iostat_set, gfp_mask); in blkg_alloc()
251 pd = pol->pd_alloc_fn(gfp_mask, disk->queue, blkcg); in blkg_alloc()
Dblk.h91 gfp_t gfp_mask);
Dblk-mq.c3137 struct bio_set *bs, gfp_t gfp_mask, in blk_rq_prep_clone() argument
3147 bio = bio_alloc_clone(rq->q->disk->part0, bio_src, gfp_mask, in blk_rq_prep_clone()
3174 if (rq->bio && blk_crypto_rq_bio_prep(rq, rq->bio, gfp_mask) < 0) in blk_rq_prep_clone()