/fs/nfs/blocklayout/ |
D | dev.c | 230 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask); 235 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_simple() argument 241 dev = bl_resolve_deviceid(server, v, gfp_mask); in bl_parse_simple() 352 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_scsi() argument 401 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_slice() argument 406 ret = bl_parse_deviceid(server, d, volumes, v->slice.volume, gfp_mask); in bl_parse_slice() 417 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_concat() argument 430 volumes, v->concat.volumes[i], gfp_mask); in bl_parse_concat() 446 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_stripe() argument 459 volumes, v->stripe.volumes[i], gfp_mask); in bl_parse_stripe() [all …]
|
D | blocklayout.h | 177 struct pnfs_device *pdev, gfp_t gfp_mask); 194 struct pnfs_block_volume *b, gfp_t gfp_mask);
|
D | blocklayout.c | 558 gfp_t gfp_mask) in bl_alloc_extent() argument 578 lo->plh_lc_cred, gfp_mask); in bl_alloc_extent() 612 gfp_t gfp_mask) in bl_alloc_lseg() argument 632 lseg = kzalloc(sizeof(*lseg), gfp_mask); in bl_alloc_lseg() 637 scratch = alloc_page(gfp_mask); in bl_alloc_lseg() 658 status = bl_alloc_extent(&xdr, lo, &lv, &extents, gfp_mask); in bl_alloc_lseg()
|
D | rpc_pipefs.c | 54 gfp_t gfp_mask) in bl_resolve_deviceid() argument 77 msg->data = kzalloc(msg->len, gfp_mask); in bl_resolve_deviceid()
|
/fs/btrfs/ |
D | ulist.h | 55 struct ulist *ulist_alloc(gfp_t gfp_mask); 57 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask); 59 u64 *old_aux, gfp_t gfp_mask); 64 void **old_aux, gfp_t gfp_mask) in ulist_add_merge_ptr() argument 68 int ret = ulist_add_merge(ulist, val, (uintptr_t)aux, &old64, gfp_mask); in ulist_add_merge_ptr() 72 return ulist_add_merge(ulist, val, (u64)aux, (u64 *)old_aux, gfp_mask); in ulist_add_merge_ptr()
|
D | ulist.c | 92 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() argument 94 struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); in ulist_alloc() 186 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() argument 188 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); in ulist_add() 192 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge() argument 203 node = kmalloc(sizeof(*node), gfp_mask); in ulist_add_merge()
|
D | extent_io.h | 444 struct bio *btrfs_io_bio_alloc(gfp_t gfp_mask, unsigned int nr_iovecs); 445 struct bio *btrfs_bio_clone(struct bio *bio, gfp_t gfp_mask);
|
/fs/ntfs/ |
D | malloc.h | 42 static inline void *__ntfs_malloc(unsigned long size, gfp_t gfp_mask) in __ntfs_malloc() argument 47 return kmalloc(PAGE_SIZE, gfp_mask & ~__GFP_HIGHMEM); in __ntfs_malloc() 51 return __vmalloc(size, gfp_mask, PAGE_KERNEL); in __ntfs_malloc()
|
/fs/squashfs/ |
D | page_actor.c | 137 struct page **alloc_page_array(int nr_pages, int gfp_mask) in alloc_page_array() argument 142 page = kcalloc(nr_pages, sizeof(struct page *), gfp_mask); in alloc_page_array() 146 page[i] = alloc_page(gfp_mask); in alloc_page_array()
|
/fs/nfs/ |
D | nfs4session.c | 104 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_new_slot() argument 108 slot = kzalloc(sizeof(*slot), gfp_mask); in nfs4_new_slot() 118 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_find_or_create_slot() argument 126 seq_init, gfp_mask); in nfs4_find_or_create_slot()
|
D | pnfs_dev.c | 189 gfp_t gfp_mask) in nfs4_find_get_deviceid() argument 198 new = nfs4_get_device_info(server, id, cred, gfp_mask); in nfs4_find_get_deviceid()
|
D | nfs4_fs.h | 259 extern int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait); 463 extern struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask);
|
D | nfs4state.c | 744 fmode_t fmode, gfp_t gfp_mask, int wait) in __nfs4_close() argument 785 nfs4_do_close(state, gfp_mask, wait); in __nfs4_close() 1021 struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask) in nfs_alloc_seqid() argument 1025 new = kmalloc(sizeof(*new), gfp_mask); in nfs_alloc_seqid()
|
D | pnfs.h | 335 gfp_t gfp_mask);
|
D | nfs4proc.c | 1189 gfp_t gfp_mask) in nfs4_opendata_alloc() argument 1197 p = kzalloc(sizeof(*p), gfp_mask); in nfs4_opendata_alloc() 1201 p->f_label = nfs4_label_alloc(server, gfp_mask); in nfs4_opendata_alloc() 1205 p->a_label = nfs4_label_alloc(server, gfp_mask); in nfs4_opendata_alloc() 1210 p->o_arg.seqid = alloc_seqid(&sp->so_seqid, gfp_mask); in nfs4_opendata_alloc() 3194 int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait) in nfs4_do_close() argument 3217 calldata = kzalloc(sizeof(*calldata), gfp_mask); in nfs4_do_close() 3226 calldata->arg.seqid = alloc_seqid(&state->owner->so_seqid, gfp_mask); in nfs4_do_close() 5986 gfp_t gfp_mask) in nfs4_alloc_lockdata() argument 5993 p = kzalloc(sizeof(*p), gfp_mask); in nfs4_alloc_lockdata() [all …]
|
/fs/jbd2/ |
D | transaction.c | 276 gfp_t gfp_mask) in start_this_handle() argument 307 if ((gfp_mask & __GFP_FS) == 0) in start_this_handle() 308 gfp_mask |= __GFP_NOFAIL; in start_this_handle() 310 gfp_mask); in start_this_handle() 426 gfp_t gfp_mask, unsigned int type, in jbd2__journal_start() argument 457 err = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_start() 636 int jbd2__journal_restart(handle_t *handle, int nblocks, gfp_t gfp_mask) in jbd2__journal_restart() argument 679 ret = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_restart() 1971 struct page *page, gfp_t gfp_mask) in jbd2_journal_try_to_free_buffers() argument
|
D | revoke.c | 144 gfp_t gfp_mask = GFP_NOFS; in insert_revoke_hash() local 147 gfp_mask |= __GFP_NOFAIL; in insert_revoke_hash() 148 record = kmem_cache_alloc(jbd2_revoke_record_cache, gfp_mask); in insert_revoke_hash()
|
/fs/nilfs2/ |
D | mdt.h | 89 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz);
|
D | mdt.c | 453 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz) in nilfs_mdt_init() argument 465 mapping_set_gfp_mask(inode->i_mapping, gfp_mask); in nilfs_mdt_init()
|
/fs/jfs/ |
D | jfs_metapage.c | 186 static inline struct metapage *alloc_metapage(gfp_t gfp_mask) in alloc_metapage() argument 188 struct metapage *mp = mempool_alloc(metapage_mempool, gfp_mask); in alloc_metapage() 541 static int metapage_releasepage(struct page *page, gfp_t gfp_mask) in metapage_releasepage() argument
|
/fs/gfs2/ |
D | inode.h | 18 extern int gfs2_releasepage(struct page *page, gfp_t gfp_mask);
|
/fs/xfs/ |
D | xfs_iops.c | 1215 gfp_t gfp_mask; in xfs_setup_inode() local 1255 gfp_mask = mapping_gfp_mask(inode->i_mapping); in xfs_setup_inode() 1256 mapping_set_gfp_mask(inode->i_mapping, (gfp_mask & ~(__GFP_FS))); in xfs_setup_inode()
|
D | xfs_buf.c | 354 gfp_t gfp_mask = xb_to_gfp(flags); in xfs_buf_allocate_memory() local 403 page = alloc_page(gfp_mask); in xfs_buf_allocate_memory() 421 __func__, gfp_mask); in xfs_buf_allocate_memory()
|
/fs/ |
D | buffer.c | 993 gfp_t gfp_mask; in grow_dev_page() local 995 gfp_mask = mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS) | gfp; in grow_dev_page() 1003 gfp_mask |= __GFP_NOFAIL; in grow_dev_page() 1005 page = find_or_create_page(inode->i_mapping, index, gfp_mask); in grow_dev_page()
|
D | dax.c | 576 vmf->gfp_mask | __GFP_ZERO); in dax_load_hole() 627 error = radix_tree_preload(vmf->gfp_mask & ~__GFP_HIGHMEM); in dax_insert_mapping_entry()
|