Searched refs:cache (Results 1 – 9 of 9) sorted by relevance
/kernel/bpf/ |
D | memalloc.c | 109 struct bpf_mem_cache cache[NUM_CACHES]; member 387 ma->cache = pc; in bpf_mem_alloc_init() 404 c = &cc->cache[i]; in bpf_mem_alloc_init() 437 free_percpu(ma->cache); in free_mem_alloc_no_barrier() 439 ma->cache = NULL; in free_mem_alloc_no_barrier() 481 copy->cache = ma->cache; in destroy_mem_alloc() 482 ma->cache = NULL; in destroy_mem_alloc() 495 if (ma->cache) { in bpf_mem_alloc_destroy() 498 c = per_cpu_ptr(ma->cache, cpu); in bpf_mem_alloc_destroy() 522 c = &cc->cache[i]; in bpf_mem_alloc_destroy() [all …]
|
D | bpf_local_storage.c | 159 if (rcu_access_pointer(local_storage->cache[smap->cache_idx]) == in bpf_selem_unlink_storage_nolock() 161 RCU_INIT_POINTER(local_storage->cache[smap->cache_idx], NULL); in bpf_selem_unlink_storage_nolock() 255 sdata = rcu_dereference_check(local_storage->cache[smap->cache_idx], in bpf_local_storage_lookup() 280 rcu_assign_pointer(local_storage->cache[smap->cache_idx], in bpf_local_storage_lookup() 504 u16 bpf_local_storage_cache_idx_get(struct bpf_local_storage_cache *cache) in bpf_local_storage_cache_idx_get() argument 509 spin_lock(&cache->idx_lock); in bpf_local_storage_cache_idx_get() 512 if (cache->idx_usage_counts[i] < min_usage) { in bpf_local_storage_cache_idx_get() 513 min_usage = cache->idx_usage_counts[i]; in bpf_local_storage_cache_idx_get() 521 cache->idx_usage_counts[res]++; in bpf_local_storage_cache_idx_get() 523 spin_unlock(&cache->idx_lock); in bpf_local_storage_cache_idx_get() [all …]
|
D | btf.c | 7726 struct bpf_cand_cache **cache, in __print_cand_cache() argument 7733 cc = cache[i]; in __print_cand_cache() 7763 struct bpf_cand_cache **cache, in check_cand_cache() argument 7766 struct bpf_cand_cache *cc = cache[hash_cands(cands) % cache_size]; in check_cand_cache() 7780 struct bpf_cand_cache **cache, in populate_cand_cache() argument 7783 struct bpf_cand_cache **cc = &cache[hash_cands(cands) % cache_size], *new_cands; in populate_cand_cache() 7808 static void __purge_cand_cache(struct btf *btf, struct bpf_cand_cache **cache, in __purge_cand_cache() argument 7815 cc = cache[i]; in __purge_cand_cache() 7824 cache[i] = NULL; in __purge_cand_cache() 7833 cache[i] = NULL; in __purge_cand_cache()
|
/kernel/ |
D | scs.c | 96 void **cache = per_cpu_ptr(scs_cache, cpu); in scs_cleanup() local 99 vfree(cache[i]); in scs_cleanup() 100 cache[i] = NULL; in scs_cleanup()
|
D | seccomp.c | 226 struct action_cache cache; member 368 const struct action_cache *cache = &sfilter->cache; in seccomp_cache_check_allow() local 372 return seccomp_cache_check_allow_bitmap(cache->allow_native, in seccomp_cache_check_allow() 377 return seccomp_cache_check_allow_bitmap(cache->allow_native, in seccomp_cache_check_allow() 381 return seccomp_cache_check_allow_bitmap(cache->allow_compat, in seccomp_cache_check_allow() 837 struct action_cache *cache = &sfilter->cache; in seccomp_cache_prepare() local 839 sfilter->prev ? &sfilter->prev->cache : NULL; in seccomp_cache_prepare() 841 seccomp_cache_prepare_bitmap(sfilter, cache->allow_native, in seccomp_cache_prepare() 847 seccomp_cache_prepare_bitmap(sfilter, cache->allow_compat, in seccomp_cache_prepare() 2452 f->cache.allow_native, in proc_pid_seccomp_cache() [all …]
|
D | kprobes.c | 92 struct kprobe_insn_cache *cache; member 192 kip->cache = c; in __get_insn_slot() 223 kip->cache->sym); in collect_one_slot() 226 kip->cache->free(kip->insns); in collect_one_slot()
|
/kernel/rcu/ |
D | Kconfig | 174 implementations of RCU, and allows trading off cache misses 181 number of cache misses incurred during RCU's grace-period
|
/kernel/trace/ |
D | Kconfig | 835 "START". The second string records the cold cache time of the first 838 As it is a tight loop, it benchmarks as hot cache. That's fine because 839 we care most about hot paths that are probably in cache already.
|
/kernel/power/ |
D | Kconfig | 316 better performance thanks to cache locality; unfortunately,
|