Home
last modified time | relevance | path

Searched refs:pool (Results 1 – 25 of 45) sorted by relevance

12

/include/net/
Dxdp_sock_drv.h23 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries);
24 bool xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc);
25 u32 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 max);
26 void xsk_tx_release(struct xsk_buff_pool *pool);
29 void xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool);
30 void xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool);
31 void xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool);
32 void xsk_clear_tx_need_wakeup(struct xsk_buff_pool *pool);
33 bool xsk_uses_need_wakeup(struct xsk_buff_pool *pool);
35 static inline u32 xsk_pool_get_headroom(struct xsk_buff_pool *pool) in xsk_pool_get_headroom() argument
[all …]
Dxsk_buff_pool.h30 struct xsk_buff_pool *pool; member
105 int xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *dev,
107 int xp_assign_dev_shared(struct xsk_buff_pool *pool, struct xdp_sock *umem_xs,
109 int xp_alloc_tx_descs(struct xsk_buff_pool *pool, struct xdp_sock *xs);
110 void xp_destroy(struct xsk_buff_pool *pool);
111 void xp_get_pool(struct xsk_buff_pool *pool);
112 bool xp_put_pool(struct xsk_buff_pool *pool);
113 void xp_clear_dev(struct xsk_buff_pool *pool);
114 void xp_add_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs);
115 void xp_del_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs);
[all …]
Dnetdev_rx_queue.h24 struct xsk_buff_pool *pool; member
/include/net/page_pool/
Dhelpers.h67 bool page_pool_get_stats(const struct page_pool *pool,
92 static inline struct page *page_pool_dev_alloc_pages(struct page_pool *pool) in page_pool_dev_alloc_pages() argument
96 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages()
110 static inline struct page *page_pool_dev_alloc_frag(struct page_pool *pool, in page_pool_dev_alloc_frag() argument
116 return page_pool_alloc_frag(pool, offset, size, gfp); in page_pool_dev_alloc_frag()
119 static inline struct page *page_pool_alloc(struct page_pool *pool, in page_pool_alloc() argument
123 unsigned int max_size = PAGE_SIZE << pool->p.order; in page_pool_alloc()
129 return page_pool_alloc_pages(pool, gfp); in page_pool_alloc()
132 page = page_pool_alloc_frag(pool, offset, *size, gfp); in page_pool_alloc()
140 if (pool->frag_offset + *size > max_size) { in page_pool_alloc()
[all …]
Dtypes.h180 void (*disconnect)(void *pool);
246 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp);
247 netmem_ref page_pool_alloc_netmem(struct page_pool *pool, gfp_t gfp);
248 struct page *page_pool_alloc_frag(struct page_pool *pool, unsigned int *offset,
250 netmem_ref page_pool_alloc_frag_netmem(struct page_pool *pool,
260 void page_pool_disable_direct_recycling(struct page_pool *pool);
261 void page_pool_destroy(struct page_pool *pool);
262 void page_pool_use_xdp_mem(struct page_pool *pool, void (*disconnect)(void *),
264 void page_pool_put_page_bulk(struct page_pool *pool, void **data,
267 static inline void page_pool_destroy(struct page_pool *pool) in page_pool_destroy() argument
[all …]
/include/trace/events/
Dpage_pool.h16 TP_PROTO(const struct page_pool *pool,
19 TP_ARGS(pool, inflight, hold, release),
22 __field(const struct page_pool *, pool)
30 __entry->pool = pool;
34 __entry->cnt = pool->destroy_cnt;
38 __entry->pool, __entry->inflight, __entry->hold,
44 TP_PROTO(const struct page_pool *pool,
47 TP_ARGS(pool, netmem, release),
50 __field(const struct page_pool *, pool)
57 __entry->pool = pool;
[all …]
/include/linux/
Dgenalloc.h52 void *data, struct gen_pool *pool,
97 extern phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long);
101 static inline int gen_pool_add_virt(struct gen_pool *pool, unsigned long addr, in gen_pool_add_virt() argument
104 return gen_pool_add_owner(pool, addr, phys, size, nid, NULL); in gen_pool_add_virt()
119 static inline int gen_pool_add(struct gen_pool *pool, unsigned long addr, in gen_pool_add() argument
122 return gen_pool_add_virt(pool, addr, -1, size, nid); in gen_pool_add()
125 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size,
128 static inline unsigned long gen_pool_alloc_owner(struct gen_pool *pool, in gen_pool_alloc_owner() argument
131 return gen_pool_alloc_algo_owner(pool, size, pool->algo, pool->data, in gen_pool_alloc_owner()
135 static inline unsigned long gen_pool_alloc_algo(struct gen_pool *pool, in gen_pool_alloc_algo() argument
[all …]
Dzpool.h40 const char *zpool_get_type(struct zpool *pool);
42 void zpool_destroy_pool(struct zpool *pool);
44 bool zpool_malloc_support_movable(struct zpool *pool);
46 int zpool_malloc(struct zpool *pool, size_t size, gfp_t gfp,
49 void zpool_free(struct zpool *pool, unsigned long handle);
51 void *zpool_map_handle(struct zpool *pool, unsigned long handle,
54 void zpool_unmap_handle(struct zpool *pool, unsigned long handle);
56 u64 zpool_get_total_pages(struct zpool *pool);
82 void (*destroy)(void *pool);
85 int (*malloc)(void *pool, size_t size, gfp_t gfp,
[all …]
Dzsmalloc.h44 void zs_destroy_pool(struct zs_pool *pool);
46 unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t flags);
47 void zs_free(struct zs_pool *pool, unsigned long obj);
49 size_t zs_huge_class_size(struct zs_pool *pool);
51 void *zs_map_object(struct zs_pool *pool, unsigned long handle,
53 void zs_unmap_object(struct zs_pool *pool, unsigned long handle);
55 unsigned long zs_get_total_pages(struct zs_pool *pool);
56 unsigned long zs_compact(struct zs_pool *pool);
58 unsigned int zs_lookup_class_index(struct zs_pool *pool, unsigned int size);
60 void zs_pool_stats(struct zs_pool *pool, struct zs_pool_stats *stats);
Ddmapool.h24 void dma_pool_destroy(struct dma_pool *pool);
26 void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags,
28 void dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t addr);
35 void dmam_pool_destroy(struct dma_pool *pool);
41 static inline void dma_pool_destroy(struct dma_pool *pool) { } in dma_pool_destroy() argument
42 static inline void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_alloc() argument
44 static inline void dma_pool_free(struct dma_pool *pool, void *vaddr, in dma_pool_free() argument
49 static inline void dmam_pool_destroy(struct dma_pool *pool) { } in dmam_pool_destroy() argument
52 static inline void *dma_pool_zalloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_zalloc() argument
55 return dma_pool_alloc(pool, mem_flags | __GFP_ZERO, handle); in dma_pool_zalloc()
Dobjpool.h121 int objpool_init(struct objpool_head *pool, int nr_objs, int object_size,
126 static inline void *__objpool_try_get_slot(struct objpool_head *pool, int cpu) in __objpool_try_get_slot() argument
128 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_get_slot()
147 if (READ_ONCE(slot->last) - head - 1 >= pool->nr_objs) { in __objpool_try_get_slot()
169 static inline void *objpool_pop(struct objpool_head *pool) in objpool_pop() argument
179 for (i = 0; i < pool->nr_possible_cpus; i++) { in objpool_pop()
180 obj = __objpool_try_get_slot(pool, cpu); in objpool_pop()
192 __objpool_try_add_slot(void *obj, struct objpool_head *pool, int cpu) in __objpool_try_add_slot() argument
194 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_add_slot()
203 WARN_ON_ONCE(tail - head > pool->nr_objs); in __objpool_try_add_slot()
[all …]
Dmempool.h30 static inline bool mempool_initialized(mempool_t *pool) in mempool_initialized() argument
32 return pool->elements != NULL; in mempool_initialized()
35 static inline bool mempool_is_saturated(mempool_t *pool) in mempool_is_saturated() argument
37 return READ_ONCE(pool->curr_nr) >= pool->min_nr; in mempool_is_saturated()
40 void mempool_exit(mempool_t *pool);
41 int mempool_init_node(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn,
45 int mempool_init_noprof(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn,
63 extern int mempool_resize(mempool_t *pool, int new_min_nr);
64 extern void mempool_destroy(mempool_t *pool);
66 extern void *mempool_alloc_noprof(mempool_t *pool, gfp_t gfp_mask) __malloc;
[all …]
Dswiotlb.h247 unsigned long attrs, struct io_tlb_pool *pool);
252 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_tbl_unmap_single() local
254 if (unlikely(pool)) in swiotlb_tbl_unmap_single()
255 __swiotlb_tbl_unmap_single(dev, addr, size, dir, attrs, pool); in swiotlb_tbl_unmap_single()
260 struct io_tlb_pool *pool);
264 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_device() local
266 if (unlikely(pool)) in swiotlb_sync_single_for_device()
267 __swiotlb_sync_single_for_device(dev, addr, size, dir, pool); in swiotlb_sync_single_for_device()
272 struct io_tlb_pool *pool);
276 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_cpu() local
[all …]
Dtee_core.h62 struct tee_shm_pool *pool; member
133 struct tee_shm_pool *pool,
202 int (*alloc)(struct tee_shm_pool *pool, struct tee_shm *shm,
204 void (*free)(struct tee_shm_pool *pool, struct tee_shm *shm);
205 void (*destroy_pool)(struct tee_shm_pool *pool);
227 static inline void tee_shm_pool_free(struct tee_shm_pool *pool) in tee_shm_pool_free() argument
229 pool->ops->destroy_pool(pool); in tee_shm_pool_free()
Dsram.h9 void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, size_t size);
11 static inline void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, in sram_exec_copy() argument
Drethook.h37 struct objpool_head pool; member
Dagpgart.h102 struct agp_memory *pool; member
/include/drm/ttm/
Dttm_pool.h51 struct ttm_pool *pool; member
82 int ttm_pool_alloc(struct ttm_pool *pool, struct ttm_tt *tt,
84 void ttm_pool_free(struct ttm_pool *pool, struct ttm_tt *tt);
86 void ttm_pool_init(struct ttm_pool *pool, struct device *dev,
88 void ttm_pool_fini(struct ttm_pool *pool);
90 int ttm_pool_debugfs(struct ttm_pool *pool, struct seq_file *m);
/include/linux/ceph/
Dmsgpool.h13 mempool_t *pool; member
19 int ceph_msgpool_init(struct ceph_msgpool *pool, int type,
22 extern void ceph_msgpool_destroy(struct ceph_msgpool *pool);
23 struct ceph_msg *ceph_msgpool_get(struct ceph_msgpool *pool, int front_len,
Dosdmap.h23 uint64_t pool; member
63 static inline bool ceph_can_shift_osds(struct ceph_pg_pool_info *pool) in ceph_can_shift_osds() argument
65 switch (pool->type) { in ceph_can_shift_osds()
76 s64 pool; member
82 oloc->pool = -1; in ceph_oloc_init()
88 return oloc->pool == -1; in ceph_oloc_empty()
246 pgid->pool = ceph_decode_64(p); in ceph_decode_pgid()
/include/soc/fsl/
Dbman.h94 void bman_free_pool(struct bman_pool *pool);
103 int bman_get_bpid(const struct bman_pool *pool);
114 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num);
127 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num);
/include/linux/firmware/qcom/
Dqcom_tzmem.h44 void qcom_tzmem_pool_free(struct qcom_tzmem_pool *pool);
49 void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp);
/include/uapi/linux/netfilter_bridge/
Debt_among.h43 struct ebt_mac_wormhash_tuple pool[]; member
/include/uapi/linux/
Dhyperv.h352 __u8 pool; member
407 __u8 pool; member
/include/linux/i3c/
Dmaster.h681 void i3c_generic_ibi_free_pool(struct i3c_generic_ibi_pool *pool);
684 i3c_generic_ibi_get_free_slot(struct i3c_generic_ibi_pool *pool);
685 void i3c_generic_ibi_recycle_slot(struct i3c_generic_ibi_pool *pool,

12