/kernel/trace/ |
D | ftrace.c | 407 struct hlist_head *hash; member 584 memset(stat->hash, 0, in ftrace_profile_reset() 651 if (stat->hash) { in ftrace_profile_init_cpu() 663 stat->hash = kcalloc(size, sizeof(struct hlist_head), GFP_KERNEL); in ftrace_profile_init_cpu() 665 if (!stat->hash) in ftrace_profile_init_cpu() 670 kfree(stat->hash); in ftrace_profile_init_cpu() 671 stat->hash = NULL; in ftrace_profile_init_cpu() 701 hhd = &stat->hash[key]; in ftrace_find_profiled_func() 720 hlist_add_head_rcu(&rec->node, &stat->hash[key]); in ftrace_add_profile() 773 if (!stat->hash || !ftrace_profile_enabled) in function_profile_call() [all …]
|
D | trace_functions.c | 639 struct ftrace_hash *hash, char *glob, in ftrace_trace_probe_callback() argument 676 ftrace_trace_onoff_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_trace_onoff_callback() argument 690 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_trace_onoff_callback() 695 ftrace_stacktrace_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_stacktrace_callback() argument 705 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_stacktrace_callback() 710 ftrace_dump_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_dump_callback() argument 721 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_dump_callback() 726 ftrace_cpudump_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_cpudump_callback() argument 737 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_cpudump_callback()
|
D | trace.h | 887 ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip); 889 static __always_inline bool ftrace_hash_empty(struct ftrace_hash *hash) in ftrace_hash_empty() argument 891 return !hash || !(hash->count || (hash->flags & FTRACE_HASH_FL_MOD)); in ftrace_hash_empty() 1035 struct ftrace_hash *hash,
|
D | Kconfig | 654 to be aggregated into hash tables and dumped to stdout by
|
D | trace_events.c | 2693 event_enable_func(struct trace_array *tr, struct ftrace_hash *hash, in event_enable_func() argument
|
D | trace.c | 7713 ftrace_trace_snapshot_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_trace_snapshot_callback() argument
|
/kernel/bpf/ |
D | hashtab.c | 56 u32 hash; member 120 u32 hash) in prealloc_lru_pop() argument 122 struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash); in prealloc_lru_pop() 166 offsetof(struct htab_elem, hash) - in prealloc_init() 410 static inline struct bucket *__select_bucket(struct bpf_htab *htab, u32 hash) in __select_bucket() argument 412 return &htab->buckets[hash & (htab->n_buckets - 1)]; in __select_bucket() 415 static inline struct hlist_nulls_head *select_bucket(struct bpf_htab *htab, u32 hash) in select_bucket() argument 417 return &__select_bucket(htab, hash)->head; in select_bucket() 421 static struct htab_elem *lookup_elem_raw(struct hlist_nulls_head *head, u32 hash, in lookup_elem_raw() argument 428 if (l->hash == hash && !memcmp(&l->key, key, key_size)) in lookup_elem_raw() [all …]
|
D | bpf_lru_list.c | 351 u32 hash) in __local_list_add_pending() argument 353 *(u32 *)((void *)node + lru->hash_offset) = hash; in __local_list_add_pending() 400 u32 hash) in bpf_percpu_lru_pop_free() argument 421 *(u32 *)((void *)node + lru->hash_offset) = hash; in bpf_percpu_lru_pop_free() 432 u32 hash) in bpf_common_lru_pop_free() argument 452 __local_list_add_pending(lru, loc_l, cpu, node, hash); in bpf_common_lru_pop_free() 487 __local_list_add_pending(lru, loc_l, cpu, node, hash); in bpf_common_lru_pop_free() 494 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash) in bpf_lru_pop_free() argument 497 return bpf_percpu_lru_pop_free(lru, hash); in bpf_lru_pop_free() 499 return bpf_common_lru_pop_free(lru, hash); in bpf_lru_pop_free()
|
D | stackmap.c | 20 u32 hash; member 356 u32 hash, id, trace_nr, trace_len; in BPF_CALL_3() local 385 hash = jhash2((u32 *)ips, trace_len / sizeof(u32), 0); in BPF_CALL_3() 386 id = hash & (smap->n_buckets - 1); in BPF_CALL_3() 389 hash_matches = bucket && bucket->hash == hash; in BPF_CALL_3() 428 new_bucket->hash = hash; in BPF_CALL_3()
|
D | devmap.c | 94 struct hlist_head *hash; in dev_map_create_hash() local 96 hash = kmalloc_array(entries, sizeof(*hash), GFP_KERNEL); in dev_map_create_hash() 97 if (hash != NULL) in dev_map_create_hash() 99 INIT_HLIST_HEAD(&hash[i]); in dev_map_create_hash() 101 return hash; in dev_map_create_hash()
|
D | bpf_lru_list.h | 78 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
|
/kernel/locking/ |
D | qspinlock_paravirt.h | 207 #define for_each_hash_entry(he, offset, hash) \ argument 208 for (hash &= ~(PV_HE_PER_LINE - 1), he = &pv_lock_hash[hash], offset = 0; \ 210 offset++, he = &pv_lock_hash[(hash + offset) & ((1 << pv_lock_hash_bits) - 1)]) 214 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_hash() local 218 for_each_hash_entry(he, offset, hash) { in pv_hash() 241 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_unhash() local 245 for_each_hash_entry(he, offset, hash) { in pv_unhash()
|
D | lockdep.c | 461 u32 hash; member 475 return t1->hash == t2->hash && t1->nr_entries == t2->nr_entries && in traces_identical() 484 u32 hash; in save_trace() local 505 hash = jhash(trace->entries, trace->nr_entries * in save_trace() 507 trace->hash = hash; in save_trace() 508 hash_head = stack_trace_hash + (hash & (STACK_TRACE_HASH_SIZE - 1)); in save_trace() 1107 unsigned long hash = hash_long((uintptr_t)key, KEYHASH_BITS); in keyhashentry() local 1109 return lock_keys_hash + hash; in keyhashentry()
|
/kernel/ |
D | kprobes.c | 68 static raw_spinlock_t *kretprobe_table_lock_ptr(unsigned long hash) in kretprobe_table_lock_ptr() argument 70 return &(kretprobe_table_locks[hash].lock); in kretprobe_table_lock_ptr() 1179 unsigned long hash = hash_ptr(tsk, KPROBE_HASH_BITS); in kretprobe_hash_lock() local 1182 *head = &kretprobe_inst_table[hash]; in kretprobe_hash_lock() 1183 hlist_lock = kretprobe_table_lock_ptr(hash); in kretprobe_hash_lock() 1188 static void kretprobe_table_lock(unsigned long hash, in kretprobe_table_lock() argument 1192 raw_spinlock_t *hlist_lock = kretprobe_table_lock_ptr(hash); in kretprobe_table_lock() 1201 unsigned long hash = hash_ptr(tsk, KPROBE_HASH_BITS); in kretprobe_hash_unlock() local 1204 hlist_lock = kretprobe_table_lock_ptr(hash); in kretprobe_hash_unlock() 1209 static void kretprobe_table_unlock(unsigned long hash, in kretprobe_table_unlock() argument [all …]
|
D | audit_tree.c | 26 struct list_head hash; member 199 INIT_LIST_HEAD(&chunk->hash); in alloc_chunk() 240 list_add_rcu(&chunk->hash, list); in insert_hash() 250 list_for_each_entry_rcu(p, list, hash) { in audit_tree_lookup() 324 list_replace_rcu(&old->hash, &new->hash); in replace_chunk() 369 list_del_rcu(&chunk->hash); in untag_chunk() 1031 list_del_rcu(&chunk->hash); in evict_chunk()
|
D | module_signature.c | 34 ms->hash != 0 || in mod_check_sig()
|
D | sysctl_binary.c | 1373 u32 hash = FNV32_OFFSET; in warn_on_bintable() local 1376 hash = (hash ^ name[i]) * FNV32_PRIME; in warn_on_bintable() 1377 hash %= WARN_ONCE_HASH_SIZE; in warn_on_bintable() 1378 if (__test_and_set_bit(hash, warn_once_bitmap)) in warn_on_bintable()
|
D | workqueue.c | 3388 u32 hash = 0; in wqattrs_hash() local 3390 hash = jhash_1word(attrs->nice, hash); in wqattrs_hash() 3391 hash = jhash(cpumask_bits(attrs->cpumask), in wqattrs_hash() 3392 BITS_TO_LONGS(nr_cpumask_bits) * sizeof(long), hash); in wqattrs_hash() 3393 return hash; in wqattrs_hash() 3587 u32 hash = wqattrs_hash(attrs); in get_unbound_pool() local 3595 hash_for_each_possible(unbound_pool_hash, pool, hash_node, hash) { in get_unbound_pool() 3636 hash_add(unbound_pool_hash, &pool->hash_node, hash); in get_unbound_pool()
|
D | futex.c | 388 u32 hash = jhash2((u32*)&key->both.word, in hash_futex() local 391 return &futex_queues[hash & (futex_hashsize - 1)]; in hash_futex()
|
/kernel/time/ |
D | posix-timers.c | 112 static int hash(struct signal_struct *sig, unsigned int nr) in hash() function 133 struct hlist_head *head = &posix_timers_hashtable[hash(sig, id)]; in posix_timer_by_id() 147 head = &posix_timers_hashtable[hash(sig, sig->posix_timer_id)]; in posix_timer_add()
|
/kernel/power/ |
D | Kconfig | 242 dmesg -s 1000000 | grep 'hash matches'
|
/kernel/events/ |
D | core.c | 8366 u64 hash = swevent_hash(type, event_id); in __find_swevent_head() local 8368 return &hlist->heads[hash]; in __find_swevent_head()
|