/include/linux/ |
D | kmemleak.h | 19 gfp_t gfp) __ref; 21 gfp_t gfp) __ref; 23 gfp_t gfp) __ref; 30 extern void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) __ref; 33 gfp_t gfp) __ref; 40 gfp_t gfp) in kmemleak_alloc_recursive() argument 43 kmemleak_alloc(ptr, size, min_count, gfp); in kmemleak_alloc_recursive() 63 gfp_t gfp) in kmemleak_alloc() argument 68 gfp_t gfp) in kmemleak_alloc_recursive() argument 72 gfp_t gfp) in kmemleak_alloc_percpu() argument [all …]
|
D | devcoredump.h | 56 gfp_t gfp); 59 void *data, size_t datalen, gfp_t gfp, 65 size_t datalen, gfp_t gfp); 68 size_t datalen, gfp_t gfp) in dev_coredumpv() argument 75 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument 84 size_t datalen, gfp_t gfp) in dev_coredumpsg() argument
|
D | vmpressure.h | 33 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, 35 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio); 47 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, in vmpressure() argument 49 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure_prio() argument
|
D | memregion.h | 12 int memregion_alloc(gfp_t gfp); 15 static inline int memregion_alloc(gfp_t gfp) in memregion_alloc() argument
|
D | zpool.h | 43 gfp_t gfp, const struct zpool_ops *ops); 51 int zpool_malloc(struct zpool *pool, size_t size, gfp_t gfp, 91 gfp_t gfp, 97 int (*malloc)(void *pool, size_t size, gfp_t gfp,
|
D | xarray.h | 584 void *entry, gfp_t gfp) in xa_store_bh() argument 589 curr = __xa_store(xa, index, entry, gfp); in xa_store_bh() 610 void *entry, gfp_t gfp) in xa_store_irq() argument 615 curr = __xa_store(xa, index, entry, gfp); in xa_store_irq() 685 void *old, void *entry, gfp_t gfp) in xa_cmpxchg() argument 690 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg() 712 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_bh() argument 717 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_bh() 739 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_irq() argument 744 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_irq() [all …]
|
D | idr.h | 273 static inline int ida_alloc(struct ida *ida, gfp_t gfp) in ida_alloc() argument 275 return ida_alloc_range(ida, 0, ~0, gfp); in ida_alloc() 291 static inline int ida_alloc_min(struct ida *ida, unsigned int min, gfp_t gfp) in ida_alloc_min() argument 293 return ida_alloc_range(ida, min, ~0, gfp); in ida_alloc_min() 309 static inline int ida_alloc_max(struct ida *ida, unsigned int max, gfp_t gfp) in ida_alloc_max() argument 311 return ida_alloc_range(ida, 0, max, gfp); in ida_alloc_max() 323 #define ida_simple_get(ida, start, end, gfp) \ argument 324 ida_alloc_range(ida, start, (end) - 1, gfp)
|
D | skb_array.h | 180 static inline int skb_array_init(struct skb_array *a, int size, gfp_t gfp) in skb_array_init() argument 182 return ptr_ring_init(&a->ring, size, gfp); in skb_array_init() 196 static inline int skb_array_resize(struct skb_array *a, int size, gfp_t gfp) in skb_array_resize() argument 198 return ptr_ring_resize(&a->ring, size, gfp, __skb_array_destroy_skb); in skb_array_resize() 203 gfp_t gfp) in skb_array_resize_multiple() argument 207 nrings, size, gfp, in skb_array_resize_multiple()
|
D | fscache.h | 574 gfp_t gfp) in fscache_read_or_alloc_page() argument 578 context, gfp); in fscache_read_or_alloc_page() 625 gfp_t gfp) in fscache_read_or_alloc_pages() argument 630 context, gfp); in fscache_read_or_alloc_pages() 656 gfp_t gfp) in fscache_alloc_page() argument 659 return __fscache_alloc_page(cookie, page, gfp); in fscache_alloc_page() 707 gfp_t gfp) in fscache_write_page() argument 710 return __fscache_write_page(cookie, page, object_size, gfp); in fscache_write_page() 793 gfp_t gfp) in fscache_maybe_release_page() argument 796 return __fscache_maybe_release_page(cookie, page, gfp); in fscache_maybe_release_page()
|
D | btree-type.h | 32 gfp_t gfp) in BTREE_FN() 34 return btree_merge(&target->h, &victim->h, BTREE_TYPE_GEO, gfp); in BTREE_FN() 45 void *val, gfp_t gfp) in BTREE_FN() 48 return btree_insert(&head->h, BTREE_TYPE_GEO, &_key, val, gfp); in BTREE_FN() 88 void *val, gfp_t gfp) in BTREE_FN() 91 val, gfp); in BTREE_FN()
|
D | dma-mapping.h | 149 gfp_t gfp, unsigned long attrs); 167 enum dma_data_direction dir, gfp_t gfp, unsigned long attrs); 242 dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) in dmam_alloc_attrs() argument 295 size_t size, enum dma_data_direction dir, gfp_t gfp, in dma_alloc_noncontiguous() argument 320 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp); 327 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) in dma_alloc_noncoherent() argument 329 struct page *page = dma_alloc_pages(dev, size, dma_handle, dir, gfp); in dma_alloc_noncoherent() 433 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument 435 return dma_alloc_attrs(dev, size, dma_handle, gfp, in dma_alloc_coherent() 436 (gfp & __GFP_NOWARN) ? DMA_ATTR_NO_WARN : 0); in dma_alloc_coherent() [all …]
|
D | virtio.h | 40 gfp_t gfp); 45 gfp_t gfp); 51 gfp_t gfp); 58 gfp_t gfp);
|
D | btree-128.h | 42 void *val, gfp_t gfp) in btree_insert128() argument 46 (unsigned long *)&key, val, gfp); in btree_insert128() 79 gfp_t gfp) in btree_merge128() argument 81 return btree_merge(&target->h, &victim->h, &btree_geo128, gfp); in btree_merge128()
|
D | ptr_ring.h | 467 static inline void **__ptr_ring_init_queue_alloc(unsigned int size, gfp_t gfp) in __ptr_ring_init_queue_alloc() argument 471 return kvmalloc_array(size, sizeof(void *), gfp | __GFP_ZERO); in __ptr_ring_init_queue_alloc() 487 static inline int ptr_ring_init(struct ptr_ring *r, int size, gfp_t gfp) in ptr_ring_init() argument 489 r->queue = __ptr_ring_init_queue_alloc(size, gfp); in ptr_ring_init() 559 int size, gfp_t gfp, in __ptr_ring_swap_queue() argument 590 static inline int ptr_ring_resize(struct ptr_ring *r, int size, gfp_t gfp, in ptr_ring_resize() argument 594 void **queue = __ptr_ring_init_queue_alloc(size, gfp); in ptr_ring_resize() 603 old = __ptr_ring_swap_queue(r, queue, size, gfp, destroy); in ptr_ring_resize() 622 gfp_t gfp, void (*destroy)(void *)) in ptr_ring_resize_multiple() argument 628 queues = kmalloc_array(nrings, sizeof(*queues), gfp); in ptr_ring_resize_multiple() [all …]
|
D | percpu.h | 134 extern void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp); 139 #define alloc_percpu_gfp(type, gfp) \ argument 141 __alignof__(type), gfp)
|
/include/net/sctp/ |
D | ulpevent.h | 81 gfp_t gfp); 90 gfp_t gfp); 96 gfp_t gfp); 103 gfp_t gfp); 108 gfp_t gfp); 113 __u32 flags, gfp_t gfp); 116 const struct sctp_association *asoc, gfp_t gfp); 120 gfp_t gfp); 124 __u32 indication, gfp_t gfp); 127 const struct sctp_association *asoc, gfp_t gfp); [all …]
|
D | stream_interleave.h | 25 int len, __u8 flags, gfp_t gfp); 29 struct sctp_chunk *chunk, gfp_t gfp); 33 struct sctp_chunk *chunk, gfp_t gfp); 34 void (*start_pd)(struct sctp_ulpq *ulpq, gfp_t gfp); 35 void (*abort_pd)(struct sctp_ulpq *ulpq, gfp_t gfp);
|
D | auth.h | 71 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp); 73 int sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp); 79 gfp_t gfp); 80 int sctp_auth_init_hmacs(struct sctp_endpoint *ep, gfp_t gfp); 94 struct sctp_shared_key *ep_key, gfp_t gfp); 110 int sctp_auth_init(struct sctp_endpoint *ep, gfp_t gfp);
|
D | stream_sched.h | 22 gfp_t gfp); 28 int (*init_sid)(struct sctp_stream *stream, __u16 sid, gfp_t gfp); 50 __u16 value, gfp_t gfp); 56 int sctp_sched_init_sid(struct sctp_stream *stream, __u16 sid, gfp_t gfp);
|
/include/asm-generic/ |
D | pgalloc.h | 59 static inline pgtable_t __pte_alloc_one(struct mm_struct *mm, gfp_t gfp) in __pte_alloc_one() argument 63 pte = alloc_page(gfp); in __pte_alloc_one() 122 gfp_t gfp = GFP_PGTABLE_USER; in pmd_alloc_one() local 125 gfp = GFP_PGTABLE_KERNEL; in pmd_alloc_one() 126 page = alloc_pages(gfp, 0); in pmd_alloc_one() 162 gfp_t gfp = GFP_PGTABLE_USER; in pud_alloc_one() local 165 gfp = GFP_PGTABLE_KERNEL; in pud_alloc_one() 166 return (pud_t *)get_zeroed_page(gfp); in pud_alloc_one()
|
/include/drm/ |
D | drm_managed.h | 47 void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp) __malloc; 59 static inline void *drmm_kzalloc(struct drm_device *dev, size_t size, gfp_t gfp) in drmm_kzalloc() argument 61 return drmm_kmalloc(dev, size, gfp | __GFP_ZERO); in drmm_kzalloc() 103 char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp);
|
/include/linux/greybus/ |
D | operation.h | 149 gfp_t gfp); 154 size_t response_size, gfp_t gfp) in gb_operation_create() argument 157 response_size, 0, gfp); in gb_operation_create() 164 gfp_t gfp); 170 size_t response_size, gfp_t gfp); 175 gfp_t gfp);
|
/include/net/ |
D | page_pool.h | 140 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); 144 gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN); in page_pool_dev_alloc_pages() local 146 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages() 150 unsigned int size, gfp_t gfp); 156 gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN); in page_pool_dev_alloc_frag() local 158 return page_pool_alloc_frag(pool, offset, size, gfp); in page_pool_dev_alloc_frag()
|
D | hwbm.h | 21 int hwbm_pool_refill(struct hwbm_pool *bm_pool, gfp_t gfp); 26 static inline int hwbm_pool_refill(struct hwbm_pool *bm_pool, gfp_t gfp) in hwbm_pool_refill() argument
|
/include/net/netfilter/ |
D | nf_conntrack_timestamp.h | 27 struct nf_conn_tstamp *nf_ct_tstamp_ext_add(struct nf_conn *ct, gfp_t gfp) in nf_ct_tstamp_ext_add() argument 35 return nf_ct_ext_add(ct, NF_CT_EXT_TSTAMP, gfp); in nf_ct_tstamp_ext_add()
|