| /include/net/ |
| D | xdp_sock_drv.h | 23 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries); 24 bool xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc); 25 u32 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 max); 26 void xsk_tx_release(struct xsk_buff_pool *pool); 29 void xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool); 30 void xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool); 31 void xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool); 32 void xsk_clear_tx_need_wakeup(struct xsk_buff_pool *pool); 33 bool xsk_uses_need_wakeup(struct xsk_buff_pool *pool); 35 static inline u32 xsk_pool_get_headroom(struct xsk_buff_pool *pool) in xsk_pool_get_headroom() argument [all …]
|
| D | xsk_buff_pool.h | 30 struct xsk_buff_pool *pool; member 105 int xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *dev, 107 int xp_assign_dev_shared(struct xsk_buff_pool *pool, struct xdp_sock *umem_xs, 109 int xp_alloc_tx_descs(struct xsk_buff_pool *pool, struct xdp_sock *xs); 110 void xp_destroy(struct xsk_buff_pool *pool); 111 void xp_get_pool(struct xsk_buff_pool *pool); 112 bool xp_put_pool(struct xsk_buff_pool *pool); 113 void xp_clear_dev(struct xsk_buff_pool *pool); 114 void xp_add_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs); 115 void xp_del_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs); [all …]
|
| D | netdev_rx_queue.h | 24 struct xsk_buff_pool *pool; member
|
| /include/net/page_pool/ |
| D | helpers.h | 67 bool page_pool_get_stats(const struct page_pool *pool, 92 static inline struct page *page_pool_dev_alloc_pages(struct page_pool *pool) in page_pool_dev_alloc_pages() argument 96 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages() 110 static inline struct page *page_pool_dev_alloc_frag(struct page_pool *pool, in page_pool_dev_alloc_frag() argument 116 return page_pool_alloc_frag(pool, offset, size, gfp); in page_pool_dev_alloc_frag() 119 static inline struct page *page_pool_alloc(struct page_pool *pool, in page_pool_alloc() argument 123 unsigned int max_size = PAGE_SIZE << pool->p.order; in page_pool_alloc() 129 return page_pool_alloc_pages(pool, gfp); in page_pool_alloc() 132 page = page_pool_alloc_frag(pool, offset, *size, gfp); in page_pool_alloc() 140 if (pool->frag_offset + *size > max_size) { in page_pool_alloc() [all …]
|
| D | types.h | 180 void (*disconnect)(void *pool); 246 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); 247 netmem_ref page_pool_alloc_netmem(struct page_pool *pool, gfp_t gfp); 248 struct page *page_pool_alloc_frag(struct page_pool *pool, unsigned int *offset, 250 netmem_ref page_pool_alloc_frag_netmem(struct page_pool *pool, 260 void page_pool_disable_direct_recycling(struct page_pool *pool); 261 void page_pool_destroy(struct page_pool *pool); 262 void page_pool_use_xdp_mem(struct page_pool *pool, void (*disconnect)(void *), 264 void page_pool_put_page_bulk(struct page_pool *pool, void **data, 267 static inline void page_pool_destroy(struct page_pool *pool) in page_pool_destroy() argument [all …]
|
| /include/trace/events/ |
| D | page_pool.h | 16 TP_PROTO(const struct page_pool *pool, 19 TP_ARGS(pool, inflight, hold, release), 22 __field(const struct page_pool *, pool) 30 __entry->pool = pool; 34 __entry->cnt = pool->destroy_cnt; 38 __entry->pool, __entry->inflight, __entry->hold, 44 TP_PROTO(const struct page_pool *pool, 47 TP_ARGS(pool, netmem, release), 50 __field(const struct page_pool *, pool) 57 __entry->pool = pool; [all …]
|
| /include/linux/ |
| D | genalloc.h | 52 void *data, struct gen_pool *pool, 97 extern phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long); 101 static inline int gen_pool_add_virt(struct gen_pool *pool, unsigned long addr, in gen_pool_add_virt() argument 104 return gen_pool_add_owner(pool, addr, phys, size, nid, NULL); in gen_pool_add_virt() 119 static inline int gen_pool_add(struct gen_pool *pool, unsigned long addr, in gen_pool_add() argument 122 return gen_pool_add_virt(pool, addr, -1, size, nid); in gen_pool_add() 125 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, 128 static inline unsigned long gen_pool_alloc_owner(struct gen_pool *pool, in gen_pool_alloc_owner() argument 131 return gen_pool_alloc_algo_owner(pool, size, pool->algo, pool->data, in gen_pool_alloc_owner() 135 static inline unsigned long gen_pool_alloc_algo(struct gen_pool *pool, in gen_pool_alloc_algo() argument [all …]
|
| D | zpool.h | 40 const char *zpool_get_type(struct zpool *pool); 42 void zpool_destroy_pool(struct zpool *pool); 44 bool zpool_malloc_support_movable(struct zpool *pool); 46 int zpool_malloc(struct zpool *pool, size_t size, gfp_t gfp, 49 void zpool_free(struct zpool *pool, unsigned long handle); 51 void *zpool_map_handle(struct zpool *pool, unsigned long handle, 54 void zpool_unmap_handle(struct zpool *pool, unsigned long handle); 56 u64 zpool_get_total_pages(struct zpool *pool); 82 void (*destroy)(void *pool); 85 int (*malloc)(void *pool, size_t size, gfp_t gfp, [all …]
|
| D | zsmalloc.h | 44 void zs_destroy_pool(struct zs_pool *pool); 46 unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t flags); 47 void zs_free(struct zs_pool *pool, unsigned long obj); 49 size_t zs_huge_class_size(struct zs_pool *pool); 51 void *zs_map_object(struct zs_pool *pool, unsigned long handle, 53 void zs_unmap_object(struct zs_pool *pool, unsigned long handle); 55 unsigned long zs_get_total_pages(struct zs_pool *pool); 56 unsigned long zs_compact(struct zs_pool *pool); 58 unsigned int zs_lookup_class_index(struct zs_pool *pool, unsigned int size); 60 void zs_pool_stats(struct zs_pool *pool, struct zs_pool_stats *stats);
|
| D | dmapool.h | 24 void dma_pool_destroy(struct dma_pool *pool); 26 void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, 28 void dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t addr); 35 void dmam_pool_destroy(struct dma_pool *pool); 41 static inline void dma_pool_destroy(struct dma_pool *pool) { } in dma_pool_destroy() argument 42 static inline void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_alloc() argument 44 static inline void dma_pool_free(struct dma_pool *pool, void *vaddr, in dma_pool_free() argument 49 static inline void dmam_pool_destroy(struct dma_pool *pool) { } in dmam_pool_destroy() argument 52 static inline void *dma_pool_zalloc(struct dma_pool *pool, gfp_t mem_flags, in dma_pool_zalloc() argument 55 return dma_pool_alloc(pool, mem_flags | __GFP_ZERO, handle); in dma_pool_zalloc()
|
| D | objpool.h | 121 int objpool_init(struct objpool_head *pool, int nr_objs, int object_size, 126 static inline void *__objpool_try_get_slot(struct objpool_head *pool, int cpu) in __objpool_try_get_slot() argument 128 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_get_slot() 147 if (READ_ONCE(slot->last) - head - 1 >= pool->nr_objs) { in __objpool_try_get_slot() 169 static inline void *objpool_pop(struct objpool_head *pool) in objpool_pop() argument 179 for (i = 0; i < pool->nr_possible_cpus; i++) { in objpool_pop() 180 obj = __objpool_try_get_slot(pool, cpu); in objpool_pop() 192 __objpool_try_add_slot(void *obj, struct objpool_head *pool, int cpu) in __objpool_try_add_slot() argument 194 struct objpool_slot *slot = pool->cpu_slots[cpu]; in __objpool_try_add_slot() 203 WARN_ON_ONCE(tail - head > pool->nr_objs); in __objpool_try_add_slot() [all …]
|
| D | mempool.h | 30 static inline bool mempool_initialized(mempool_t *pool) in mempool_initialized() argument 32 return pool->elements != NULL; in mempool_initialized() 35 static inline bool mempool_is_saturated(mempool_t *pool) in mempool_is_saturated() argument 37 return READ_ONCE(pool->curr_nr) >= pool->min_nr; in mempool_is_saturated() 40 void mempool_exit(mempool_t *pool); 41 int mempool_init_node(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn, 45 int mempool_init_noprof(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn, 63 extern int mempool_resize(mempool_t *pool, int new_min_nr); 64 extern void mempool_destroy(mempool_t *pool); 66 extern void *mempool_alloc_noprof(mempool_t *pool, gfp_t gfp_mask) __malloc; [all …]
|
| D | swiotlb.h | 247 unsigned long attrs, struct io_tlb_pool *pool); 252 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_tbl_unmap_single() local 254 if (unlikely(pool)) in swiotlb_tbl_unmap_single() 255 __swiotlb_tbl_unmap_single(dev, addr, size, dir, attrs, pool); in swiotlb_tbl_unmap_single() 260 struct io_tlb_pool *pool); 264 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_device() local 266 if (unlikely(pool)) in swiotlb_sync_single_for_device() 267 __swiotlb_sync_single_for_device(dev, addr, size, dir, pool); in swiotlb_sync_single_for_device() 272 struct io_tlb_pool *pool); 276 struct io_tlb_pool *pool = swiotlb_find_pool(dev, addr); in swiotlb_sync_single_for_cpu() local [all …]
|
| D | tee_core.h | 62 struct tee_shm_pool *pool; member 133 struct tee_shm_pool *pool, 202 int (*alloc)(struct tee_shm_pool *pool, struct tee_shm *shm, 204 void (*free)(struct tee_shm_pool *pool, struct tee_shm *shm); 205 void (*destroy_pool)(struct tee_shm_pool *pool); 227 static inline void tee_shm_pool_free(struct tee_shm_pool *pool) in tee_shm_pool_free() argument 229 pool->ops->destroy_pool(pool); in tee_shm_pool_free()
|
| D | sram.h | 9 void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, size_t size); 11 static inline void *sram_exec_copy(struct gen_pool *pool, void *dst, void *src, in sram_exec_copy() argument
|
| D | rethook.h | 37 struct objpool_head pool; member
|
| D | agpgart.h | 102 struct agp_memory *pool; member
|
| /include/drm/ttm/ |
| D | ttm_pool.h | 51 struct ttm_pool *pool; member 82 int ttm_pool_alloc(struct ttm_pool *pool, struct ttm_tt *tt, 84 void ttm_pool_free(struct ttm_pool *pool, struct ttm_tt *tt); 86 void ttm_pool_init(struct ttm_pool *pool, struct device *dev, 88 void ttm_pool_fini(struct ttm_pool *pool); 90 int ttm_pool_debugfs(struct ttm_pool *pool, struct seq_file *m);
|
| /include/linux/ceph/ |
| D | msgpool.h | 13 mempool_t *pool; member 19 int ceph_msgpool_init(struct ceph_msgpool *pool, int type, 22 extern void ceph_msgpool_destroy(struct ceph_msgpool *pool); 23 struct ceph_msg *ceph_msgpool_get(struct ceph_msgpool *pool, int front_len,
|
| D | osdmap.h | 23 uint64_t pool; member 63 static inline bool ceph_can_shift_osds(struct ceph_pg_pool_info *pool) in ceph_can_shift_osds() argument 65 switch (pool->type) { in ceph_can_shift_osds() 76 s64 pool; member 82 oloc->pool = -1; in ceph_oloc_init() 88 return oloc->pool == -1; in ceph_oloc_empty() 246 pgid->pool = ceph_decode_64(p); in ceph_decode_pgid()
|
| /include/soc/fsl/ |
| D | bman.h | 94 void bman_free_pool(struct bman_pool *pool); 103 int bman_get_bpid(const struct bman_pool *pool); 114 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num); 127 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num);
|
| /include/linux/firmware/qcom/ |
| D | qcom_tzmem.h | 44 void qcom_tzmem_pool_free(struct qcom_tzmem_pool *pool); 49 void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp);
|
| /include/uapi/linux/netfilter_bridge/ |
| D | ebt_among.h | 43 struct ebt_mac_wormhash_tuple pool[]; member
|
| /include/uapi/linux/ |
| D | hyperv.h | 352 __u8 pool; member 407 __u8 pool; member
|
| /include/linux/i3c/ |
| D | master.h | 681 void i3c_generic_ibi_free_pool(struct i3c_generic_ibi_pool *pool); 684 i3c_generic_ibi_get_free_slot(struct i3c_generic_ibi_pool *pool); 685 void i3c_generic_ibi_recycle_slot(struct i3c_generic_ibi_pool *pool,
|