/include/linux/ |
D | mbcache.h | 41 void mb_cache_destroy(struct mb_cache *cache); 43 int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, 45 void __mb_cache_entry_free(struct mb_cache *cache, 48 static inline void mb_cache_entry_put(struct mb_cache *cache, in mb_cache_entry_put() argument 58 __mb_cache_entry_free(cache, entry); in mb_cache_entry_put() 61 struct mb_cache_entry *mb_cache_entry_delete_or_get(struct mb_cache *cache, 63 void mb_cache_entry_delete(struct mb_cache *cache, u32 key, u64 value); 64 struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key, 66 struct mb_cache_entry *mb_cache_entry_find_first(struct mb_cache *cache, 68 struct mb_cache_entry *mb_cache_entry_find_next(struct mb_cache *cache, [all …]
|
D | kasan.h | 152 void __kasan_cache_create(struct kmem_cache *cache, unsigned int *size, 154 static __always_inline void kasan_cache_create(struct kmem_cache *cache, in kasan_cache_create() argument 158 __kasan_cache_create(cache, size, flags); in kasan_cache_create() 161 void __kasan_cache_create_kmalloc(struct kmem_cache *cache); 162 static __always_inline void kasan_cache_create_kmalloc(struct kmem_cache *cache) in kasan_cache_create_kmalloc() argument 165 __kasan_cache_create_kmalloc(cache); in kasan_cache_create_kmalloc() 168 size_t __kasan_metadata_size(struct kmem_cache *cache); 169 static __always_inline size_t kasan_metadata_size(struct kmem_cache *cache) in kasan_metadata_size() argument 172 return __kasan_metadata_size(cache); in kasan_metadata_size() 183 void __kasan_unpoison_object_data(struct kmem_cache *cache, void *object); [all …]
|
D | slab_def.h | 90 static inline void *nearest_obj(struct kmem_cache *cache, struct page *page, in nearest_obj() argument 93 void *object = x - (x - page->s_mem) % cache->size; in nearest_obj() 94 void *last_object = page->s_mem + (cache->num - 1) * cache->size; in nearest_obj() 108 static inline unsigned int obj_to_index(const struct kmem_cache *cache, in obj_to_index() argument 112 return reciprocal_divide(offset, cache->reciprocal_buffer_size); in obj_to_index() 115 static inline int objs_per_slab_page(const struct kmem_cache *cache, in objs_per_slab_page() argument 120 return cache->num; in objs_per_slab_page()
|
D | slub_def.h | 167 static inline void *nearest_obj(struct kmem_cache *cache, struct page *page, in nearest_obj() argument 169 void *object = x - (x - page_address(page)) % cache->size; in nearest_obj() 171 (page->objects - 1) * cache->size; in nearest_obj() 174 result = fixup_red_left(cache, result); in nearest_obj() 179 static inline unsigned int __obj_to_index(const struct kmem_cache *cache, in __obj_to_index() argument 183 cache->reciprocal_size); in __obj_to_index() 186 static inline unsigned int obj_to_index(const struct kmem_cache *cache, in obj_to_index() argument 191 return __obj_to_index(cache, page_address(page), obj); in obj_to_index() 194 static inline int objs_per_slab_page(const struct kmem_cache *cache, in objs_per_slab_page() argument
|
D | fscache-cache.h | 45 struct fscache_cache *cache; /* cache referred to by this tag */ member 230 struct fscache_object *(*alloc_object)(struct fscache_cache *cache, 270 void (*sync_cache)(struct fscache_cache *cache); 306 void (*dissociate_pages)(struct fscache_cache *cache); 377 struct fscache_cache *cache; /* cache that supplied this object */ member 415 return test_bit(FSCACHE_IOERROR, &object->cache->flags); in fscache_cache_is_broken() 431 static inline void fscache_object_destroyed(struct fscache_cache *cache) in fscache_object_destroyed() argument 433 if (atomic_dec_and_test(&cache->object_count)) in fscache_object_destroyed() 530 void fscache_init_cache(struct fscache_cache *cache, 534 extern int fscache_add_cache(struct fscache_cache *cache, [all …]
|
D | bpf_local_storage.h | 81 struct bpf_local_storage_data __rcu *cache[BPF_LOCAL_STORAGE_CACHE_SIZE]; member 115 u16 bpf_local_storage_cache_idx_get(struct bpf_local_storage_cache *cache); 116 void bpf_local_storage_cache_idx_free(struct bpf_local_storage_cache *cache,
|
D | bch.h | 52 int *cache; member
|
D | mroute_base.h | 382 struct list_head *cache; member 413 it->cache = NULL; in mr_mfc_seq_start() 426 if (it->cache == &mrt->mfc_unres_queue) in mr_mfc_seq_stop() 428 else if (it->cache == &mrt->mfc_cache_list) in mr_mfc_seq_stop()
|
D | lru_cache.h | 239 extern struct lru_cache *lc_create(const char *name, struct kmem_cache *cache,
|
D | kvm_host.h | 759 void kvm_release_pfn(kvm_pfn_t pfn, bool dirty, struct gfn_to_pfn_cache *cache); 837 struct gfn_to_pfn_cache *cache, bool atomic); 841 struct gfn_to_pfn_cache *cache, bool dirty, bool atomic);
|
/include/net/ |
D | netlabel.h | 198 struct netlbl_lsm_cache *cache; member 280 struct netlbl_lsm_cache *cache; in netlbl_secattr_cache_alloc() local 282 cache = kzalloc(sizeof(*cache), flags); in netlbl_secattr_cache_alloc() 283 if (cache) in netlbl_secattr_cache_alloc() 284 refcount_set(&cache->refcount, 1); in netlbl_secattr_cache_alloc() 285 return cache; in netlbl_secattr_cache_alloc() 296 static inline void netlbl_secattr_cache_free(struct netlbl_lsm_cache *cache) in netlbl_secattr_cache_free() argument 298 if (!refcount_dec_and_test(&cache->refcount)) in netlbl_secattr_cache_free() 301 if (cache->free) in netlbl_secattr_cache_free() 302 cache->free(cache->data); in netlbl_secattr_cache_free() [all …]
|
D | dst_cache.h | 12 struct dst_cache_pcpu __percpu *cache; member
|
D | page_pool.h | 69 void *cache[PP_ALLOC_CACHE_SIZE]; member
|
D | dst_metadata.h | 128 if (new_md->u.tun_info.dst_cache.cache) { in tun_dst_unclone()
|
/include/trace/hooks/ |
D | mm.h | 263 TP_PROTO(struct swap_slots_cache *cache, unsigned int type, 265 TP_ARGS(cache, type, free_slots, skip), 1); 267 TP_PROTO(struct swap_slots_cache *cache, unsigned int type, 269 TP_ARGS(cache, type, free_slots, skip)); 271 TP_PROTO(struct swap_slots_cache *cache, int *ret, bool *skip), 272 TP_ARGS(cache, ret, skip), 1); 274 TP_PROTO(struct swap_slots_cache *cache, int *ret, bool *skip), 275 TP_ARGS(cache, ret, skip)); 277 TP_PROTO(swp_entry_t entry, struct swap_slots_cache *cache, bool *skip), 278 TP_ARGS(entry, cache, skip), 1); [all …]
|
/include/acpi/platform/ |
D | aclinuxex.h | 65 static inline void *acpi_os_acquire_object(acpi_cache_t * cache) in acpi_os_acquire_object() argument 67 return kmem_cache_zalloc(cache, in acpi_os_acquire_object()
|
/include/xen/ |
D | grant_table.h | 227 void gnttab_page_cache_init(struct gnttab_page_cache *cache); 228 int gnttab_page_cache_get(struct gnttab_page_cache *cache, struct page **page); 229 void gnttab_page_cache_put(struct gnttab_page_cache *cache, struct page **page, 231 void gnttab_page_cache_shrink(struct gnttab_page_cache *cache,
|
/include/acpi/ |
D | acpiosxf.h | 207 acpi_status acpi_os_delete_cache(acpi_cache_t * cache); 211 acpi_status acpi_os_purge_cache(acpi_cache_t * cache); 215 void *acpi_os_acquire_object(acpi_cache_t * cache); 219 acpi_status acpi_os_release_object(acpi_cache_t * cache, void *object);
|
/include/linux/platform_data/ |
D | mlxreg.h | 96 u32 cache; member
|
/include/net/netfilter/ |
D | nf_conntrack_ecache.h | 23 unsigned long cache; /* bitops want long */ member 118 set_bit(event, &e->cache); in nf_conntrack_event_cache()
|
/include/linux/power/ |
D | bq27xxx_battery.h | 70 struct bq27xxx_reg_cache cache; member
|
/include/uapi/linux/ |
D | iommu.h | 264 __u8 cache; member
|
/include/trace/events/ |
D | bcache.h | 429 TP_PROTO(struct cache *ca, size_t bucket), 450 TP_PROTO(struct cache *ca, size_t bucket), 468 TP_PROTO(struct cache *ca, unsigned reserve),
|
/include/net/9p/ |
D | 9p.h | 549 struct kmem_cache *cache; member
|
/include/linux/mtd/ |
D | nand.h | 206 unsigned long *cache; member 932 return !!nand->bbt.cache; in nanddev_bbt_is_initialized()
|