Home
last modified time | relevance | path

Searched refs:gfp (Results 1 – 25 of 103) sorted by relevance

12345

/include/linux/
Dkmemleak.h19 gfp_t gfp) __ref;
21 gfp_t gfp) __ref;
23 gfp_t gfp) __ref;
30 extern void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) __ref;
33 gfp_t gfp) __ref;
40 gfp_t gfp) in kmemleak_alloc_recursive() argument
43 kmemleak_alloc(ptr, size, min_count, gfp); in kmemleak_alloc_recursive()
63 gfp_t gfp) in kmemleak_alloc() argument
68 gfp_t gfp) in kmemleak_alloc_recursive() argument
72 gfp_t gfp) in kmemleak_alloc_percpu() argument
[all …]
Ddevcoredump.h56 gfp_t gfp);
59 void *data, size_t datalen, gfp_t gfp,
65 size_t datalen, gfp_t gfp);
68 size_t datalen, gfp_t gfp) in dev_coredumpv() argument
75 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument
84 size_t datalen, gfp_t gfp) in dev_coredumpsg() argument
Dvmpressure.h33 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
35 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio);
47 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, in vmpressure() argument
49 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure_prio() argument
Dmemregion.h12 int memregion_alloc(gfp_t gfp);
15 static inline int memregion_alloc(gfp_t gfp) in memregion_alloc() argument
Dzpool.h43 gfp_t gfp, const struct zpool_ops *ops);
51 int zpool_malloc(struct zpool *pool, size_t size, gfp_t gfp,
91 gfp_t gfp,
97 int (*malloc)(void *pool, size_t size, gfp_t gfp,
Dxarray.h584 void *entry, gfp_t gfp) in xa_store_bh() argument
589 curr = __xa_store(xa, index, entry, gfp); in xa_store_bh()
610 void *entry, gfp_t gfp) in xa_store_irq() argument
615 curr = __xa_store(xa, index, entry, gfp); in xa_store_irq()
685 void *old, void *entry, gfp_t gfp) in xa_cmpxchg() argument
690 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg()
712 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_bh() argument
717 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_bh()
739 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_irq() argument
744 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_irq()
[all …]
Didr.h273 static inline int ida_alloc(struct ida *ida, gfp_t gfp) in ida_alloc() argument
275 return ida_alloc_range(ida, 0, ~0, gfp); in ida_alloc()
291 static inline int ida_alloc_min(struct ida *ida, unsigned int min, gfp_t gfp) in ida_alloc_min() argument
293 return ida_alloc_range(ida, min, ~0, gfp); in ida_alloc_min()
309 static inline int ida_alloc_max(struct ida *ida, unsigned int max, gfp_t gfp) in ida_alloc_max() argument
311 return ida_alloc_range(ida, 0, max, gfp); in ida_alloc_max()
323 #define ida_simple_get(ida, start, end, gfp) \ argument
324 ida_alloc_range(ida, start, (end) - 1, gfp)
Dskb_array.h180 static inline int skb_array_init(struct skb_array *a, int size, gfp_t gfp) in skb_array_init() argument
182 return ptr_ring_init(&a->ring, size, gfp); in skb_array_init()
196 static inline int skb_array_resize(struct skb_array *a, int size, gfp_t gfp) in skb_array_resize() argument
198 return ptr_ring_resize(&a->ring, size, gfp, __skb_array_destroy_skb); in skb_array_resize()
203 gfp_t gfp) in skb_array_resize_multiple() argument
207 nrings, size, gfp, in skb_array_resize_multiple()
Dfscache.h574 gfp_t gfp) in fscache_read_or_alloc_page() argument
578 context, gfp); in fscache_read_or_alloc_page()
625 gfp_t gfp) in fscache_read_or_alloc_pages() argument
630 context, gfp); in fscache_read_or_alloc_pages()
656 gfp_t gfp) in fscache_alloc_page() argument
659 return __fscache_alloc_page(cookie, page, gfp); in fscache_alloc_page()
707 gfp_t gfp) in fscache_write_page() argument
710 return __fscache_write_page(cookie, page, object_size, gfp); in fscache_write_page()
793 gfp_t gfp) in fscache_maybe_release_page() argument
796 return __fscache_maybe_release_page(cookie, page, gfp); in fscache_maybe_release_page()
Dbtree-type.h32 gfp_t gfp) in BTREE_FN()
34 return btree_merge(&target->h, &victim->h, BTREE_TYPE_GEO, gfp); in BTREE_FN()
45 void *val, gfp_t gfp) in BTREE_FN()
48 return btree_insert(&head->h, BTREE_TYPE_GEO, &_key, val, gfp); in BTREE_FN()
88 void *val, gfp_t gfp) in BTREE_FN()
91 val, gfp); in BTREE_FN()
Ddma-mapping.h149 gfp_t gfp, unsigned long attrs);
167 enum dma_data_direction dir, gfp_t gfp, unsigned long attrs);
242 dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) in dmam_alloc_attrs() argument
295 size_t size, enum dma_data_direction dir, gfp_t gfp, in dma_alloc_noncontiguous() argument
320 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp);
327 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) in dma_alloc_noncoherent() argument
329 struct page *page = dma_alloc_pages(dev, size, dma_handle, dir, gfp); in dma_alloc_noncoherent()
433 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
435 return dma_alloc_attrs(dev, size, dma_handle, gfp, in dma_alloc_coherent()
436 (gfp & __GFP_NOWARN) ? DMA_ATTR_NO_WARN : 0); in dma_alloc_coherent()
[all …]
Dvirtio.h40 gfp_t gfp);
45 gfp_t gfp);
51 gfp_t gfp);
58 gfp_t gfp);
Dbtree-128.h42 void *val, gfp_t gfp) in btree_insert128() argument
46 (unsigned long *)&key, val, gfp); in btree_insert128()
79 gfp_t gfp) in btree_merge128() argument
81 return btree_merge(&target->h, &victim->h, &btree_geo128, gfp); in btree_merge128()
Dptr_ring.h467 static inline void **__ptr_ring_init_queue_alloc(unsigned int size, gfp_t gfp) in __ptr_ring_init_queue_alloc() argument
471 return kvmalloc_array(size, sizeof(void *), gfp | __GFP_ZERO); in __ptr_ring_init_queue_alloc()
487 static inline int ptr_ring_init(struct ptr_ring *r, int size, gfp_t gfp) in ptr_ring_init() argument
489 r->queue = __ptr_ring_init_queue_alloc(size, gfp); in ptr_ring_init()
559 int size, gfp_t gfp, in __ptr_ring_swap_queue() argument
590 static inline int ptr_ring_resize(struct ptr_ring *r, int size, gfp_t gfp, in ptr_ring_resize() argument
594 void **queue = __ptr_ring_init_queue_alloc(size, gfp); in ptr_ring_resize()
603 old = __ptr_ring_swap_queue(r, queue, size, gfp, destroy); in ptr_ring_resize()
622 gfp_t gfp, void (*destroy)(void *)) in ptr_ring_resize_multiple() argument
628 queues = kmalloc_array(nrings, sizeof(*queues), gfp); in ptr_ring_resize_multiple()
[all …]
Dpercpu.h134 extern void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp);
139 #define alloc_percpu_gfp(type, gfp) \ argument
141 __alignof__(type), gfp)
/include/net/sctp/
Dulpevent.h81 gfp_t gfp);
90 gfp_t gfp);
96 gfp_t gfp);
103 gfp_t gfp);
108 gfp_t gfp);
113 __u32 flags, gfp_t gfp);
116 const struct sctp_association *asoc, gfp_t gfp);
120 gfp_t gfp);
124 __u32 indication, gfp_t gfp);
127 const struct sctp_association *asoc, gfp_t gfp);
[all …]
Dstream_interleave.h25 int len, __u8 flags, gfp_t gfp);
29 struct sctp_chunk *chunk, gfp_t gfp);
33 struct sctp_chunk *chunk, gfp_t gfp);
34 void (*start_pd)(struct sctp_ulpq *ulpq, gfp_t gfp);
35 void (*abort_pd)(struct sctp_ulpq *ulpq, gfp_t gfp);
Dauth.h71 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp);
73 int sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp);
79 gfp_t gfp);
80 int sctp_auth_init_hmacs(struct sctp_endpoint *ep, gfp_t gfp);
94 struct sctp_shared_key *ep_key, gfp_t gfp);
110 int sctp_auth_init(struct sctp_endpoint *ep, gfp_t gfp);
Dstream_sched.h22 gfp_t gfp);
28 int (*init_sid)(struct sctp_stream *stream, __u16 sid, gfp_t gfp);
50 __u16 value, gfp_t gfp);
56 int sctp_sched_init_sid(struct sctp_stream *stream, __u16 sid, gfp_t gfp);
/include/asm-generic/
Dpgalloc.h59 static inline pgtable_t __pte_alloc_one(struct mm_struct *mm, gfp_t gfp) in __pte_alloc_one() argument
63 pte = alloc_page(gfp); in __pte_alloc_one()
122 gfp_t gfp = GFP_PGTABLE_USER; in pmd_alloc_one() local
125 gfp = GFP_PGTABLE_KERNEL; in pmd_alloc_one()
126 page = alloc_pages(gfp, 0); in pmd_alloc_one()
162 gfp_t gfp = GFP_PGTABLE_USER; in pud_alloc_one() local
165 gfp = GFP_PGTABLE_KERNEL; in pud_alloc_one()
166 return (pud_t *)get_zeroed_page(gfp); in pud_alloc_one()
/include/drm/
Ddrm_managed.h47 void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp) __malloc;
59 static inline void *drmm_kzalloc(struct drm_device *dev, size_t size, gfp_t gfp) in drmm_kzalloc() argument
61 return drmm_kmalloc(dev, size, gfp | __GFP_ZERO); in drmm_kzalloc()
103 char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp);
/include/linux/greybus/
Doperation.h149 gfp_t gfp);
154 size_t response_size, gfp_t gfp) in gb_operation_create() argument
157 response_size, 0, gfp); in gb_operation_create()
164 gfp_t gfp);
170 size_t response_size, gfp_t gfp);
175 gfp_t gfp);
/include/net/
Dpage_pool.h140 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp);
144 gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN); in page_pool_dev_alloc_pages() local
146 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages()
150 unsigned int size, gfp_t gfp);
156 gfp_t gfp = (GFP_ATOMIC | __GFP_NOWARN); in page_pool_dev_alloc_frag() local
158 return page_pool_alloc_frag(pool, offset, size, gfp); in page_pool_dev_alloc_frag()
Dhwbm.h21 int hwbm_pool_refill(struct hwbm_pool *bm_pool, gfp_t gfp);
26 static inline int hwbm_pool_refill(struct hwbm_pool *bm_pool, gfp_t gfp) in hwbm_pool_refill() argument
/include/net/netfilter/
Dnf_conntrack_timestamp.h27 struct nf_conn_tstamp *nf_ct_tstamp_ext_add(struct nf_conn *ct, gfp_t gfp) in nf_ct_tstamp_ext_add() argument
35 return nf_ct_ext_add(ct, NF_CT_EXT_TSTAMP, gfp); in nf_ct_tstamp_ext_add()

12345