Home
last modified time | relevance | path

Searched refs:gfp_mask (Results 1 – 25 of 36) sorted by relevance

12

/fs/nfs/blocklayout/
Ddev.c230 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask);
235 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_simple() argument
241 dev = bl_resolve_deviceid(server, v, gfp_mask); in bl_parse_simple()
352 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_scsi() argument
401 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_slice() argument
406 ret = bl_parse_deviceid(server, d, volumes, v->slice.volume, gfp_mask); in bl_parse_slice()
417 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_concat() argument
430 volumes, v->concat.volumes[i], gfp_mask); in bl_parse_concat()
446 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_stripe() argument
459 volumes, v->stripe.volumes[i], gfp_mask); in bl_parse_stripe()
[all …]
Dblocklayout.h177 struct pnfs_device *pdev, gfp_t gfp_mask);
194 struct pnfs_block_volume *b, gfp_t gfp_mask);
Dblocklayout.c558 gfp_t gfp_mask) in bl_alloc_extent() argument
578 lo->plh_lc_cred, gfp_mask); in bl_alloc_extent()
612 gfp_t gfp_mask) in bl_alloc_lseg() argument
632 lseg = kzalloc(sizeof(*lseg), gfp_mask); in bl_alloc_lseg()
637 scratch = alloc_page(gfp_mask); in bl_alloc_lseg()
658 status = bl_alloc_extent(&xdr, lo, &lv, &extents, gfp_mask); in bl_alloc_lseg()
Drpc_pipefs.c54 gfp_t gfp_mask) in bl_resolve_deviceid() argument
77 msg->data = kzalloc(msg->len, gfp_mask); in bl_resolve_deviceid()
/fs/btrfs/
Dulist.h55 struct ulist *ulist_alloc(gfp_t gfp_mask);
57 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask);
59 u64 *old_aux, gfp_t gfp_mask);
64 void **old_aux, gfp_t gfp_mask) in ulist_add_merge_ptr() argument
68 int ret = ulist_add_merge(ulist, val, (uintptr_t)aux, &old64, gfp_mask); in ulist_add_merge_ptr()
72 return ulist_add_merge(ulist, val, (u64)aux, (u64 *)old_aux, gfp_mask); in ulist_add_merge_ptr()
Dulist.c92 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() argument
94 struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); in ulist_alloc()
186 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() argument
188 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); in ulist_add()
192 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge() argument
203 node = kmalloc(sizeof(*node), gfp_mask); in ulist_add_merge()
Dextent_io.h444 struct bio *btrfs_io_bio_alloc(gfp_t gfp_mask, unsigned int nr_iovecs);
445 struct bio *btrfs_bio_clone(struct bio *bio, gfp_t gfp_mask);
/fs/ntfs/
Dmalloc.h42 static inline void *__ntfs_malloc(unsigned long size, gfp_t gfp_mask) in __ntfs_malloc() argument
47 return kmalloc(PAGE_SIZE, gfp_mask & ~__GFP_HIGHMEM); in __ntfs_malloc()
51 return __vmalloc(size, gfp_mask, PAGE_KERNEL); in __ntfs_malloc()
/fs/squashfs/
Dpage_actor.c137 struct page **alloc_page_array(int nr_pages, int gfp_mask) in alloc_page_array() argument
142 page = kcalloc(nr_pages, sizeof(struct page *), gfp_mask); in alloc_page_array()
146 page[i] = alloc_page(gfp_mask); in alloc_page_array()
/fs/nfs/
Dnfs4session.c104 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_new_slot() argument
108 slot = kzalloc(sizeof(*slot), gfp_mask); in nfs4_new_slot()
118 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_find_or_create_slot() argument
126 seq_init, gfp_mask); in nfs4_find_or_create_slot()
Dpnfs_dev.c189 gfp_t gfp_mask) in nfs4_find_get_deviceid() argument
198 new = nfs4_get_device_info(server, id, cred, gfp_mask); in nfs4_find_get_deviceid()
Dnfs4_fs.h259 extern int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait);
463 extern struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask);
Dnfs4state.c744 fmode_t fmode, gfp_t gfp_mask, int wait) in __nfs4_close() argument
785 nfs4_do_close(state, gfp_mask, wait); in __nfs4_close()
1021 struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask) in nfs_alloc_seqid() argument
1025 new = kmalloc(sizeof(*new), gfp_mask); in nfs_alloc_seqid()
Dpnfs.h335 gfp_t gfp_mask);
Dnfs4proc.c1189 gfp_t gfp_mask) in nfs4_opendata_alloc() argument
1197 p = kzalloc(sizeof(*p), gfp_mask); in nfs4_opendata_alloc()
1201 p->f_label = nfs4_label_alloc(server, gfp_mask); in nfs4_opendata_alloc()
1205 p->a_label = nfs4_label_alloc(server, gfp_mask); in nfs4_opendata_alloc()
1210 p->o_arg.seqid = alloc_seqid(&sp->so_seqid, gfp_mask); in nfs4_opendata_alloc()
3194 int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait) in nfs4_do_close() argument
3217 calldata = kzalloc(sizeof(*calldata), gfp_mask); in nfs4_do_close()
3226 calldata->arg.seqid = alloc_seqid(&state->owner->so_seqid, gfp_mask); in nfs4_do_close()
5986 gfp_t gfp_mask) in nfs4_alloc_lockdata() argument
5993 p = kzalloc(sizeof(*p), gfp_mask); in nfs4_alloc_lockdata()
[all …]
/fs/jbd2/
Dtransaction.c276 gfp_t gfp_mask) in start_this_handle() argument
307 if ((gfp_mask & __GFP_FS) == 0) in start_this_handle()
308 gfp_mask |= __GFP_NOFAIL; in start_this_handle()
310 gfp_mask); in start_this_handle()
426 gfp_t gfp_mask, unsigned int type, in jbd2__journal_start() argument
457 err = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_start()
636 int jbd2__journal_restart(handle_t *handle, int nblocks, gfp_t gfp_mask) in jbd2__journal_restart() argument
679 ret = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_restart()
1971 struct page *page, gfp_t gfp_mask) in jbd2_journal_try_to_free_buffers() argument
Drevoke.c144 gfp_t gfp_mask = GFP_NOFS; in insert_revoke_hash() local
147 gfp_mask |= __GFP_NOFAIL; in insert_revoke_hash()
148 record = kmem_cache_alloc(jbd2_revoke_record_cache, gfp_mask); in insert_revoke_hash()
/fs/nilfs2/
Dmdt.h89 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz);
Dmdt.c453 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz) in nilfs_mdt_init() argument
465 mapping_set_gfp_mask(inode->i_mapping, gfp_mask); in nilfs_mdt_init()
/fs/jfs/
Djfs_metapage.c186 static inline struct metapage *alloc_metapage(gfp_t gfp_mask) in alloc_metapage() argument
188 struct metapage *mp = mempool_alloc(metapage_mempool, gfp_mask); in alloc_metapage()
541 static int metapage_releasepage(struct page *page, gfp_t gfp_mask) in metapage_releasepage() argument
/fs/gfs2/
Dinode.h18 extern int gfs2_releasepage(struct page *page, gfp_t gfp_mask);
/fs/xfs/
Dxfs_iops.c1215 gfp_t gfp_mask; in xfs_setup_inode() local
1255 gfp_mask = mapping_gfp_mask(inode->i_mapping); in xfs_setup_inode()
1256 mapping_set_gfp_mask(inode->i_mapping, (gfp_mask & ~(__GFP_FS))); in xfs_setup_inode()
Dxfs_buf.c354 gfp_t gfp_mask = xb_to_gfp(flags); in xfs_buf_allocate_memory() local
403 page = alloc_page(gfp_mask); in xfs_buf_allocate_memory()
421 __func__, gfp_mask); in xfs_buf_allocate_memory()
/fs/
Dbuffer.c993 gfp_t gfp_mask; in grow_dev_page() local
995 gfp_mask = mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS) | gfp; in grow_dev_page()
1003 gfp_mask |= __GFP_NOFAIL; in grow_dev_page()
1005 page = find_or_create_page(inode->i_mapping, index, gfp_mask); in grow_dev_page()
Ddax.c576 vmf->gfp_mask | __GFP_ZERO); in dax_load_hole()
627 error = radix_tree_preload(vmf->gfp_mask & ~__GFP_HIGHMEM); in dax_insert_mapping_entry()

12