Home
last modified time | relevance | path

Searched refs:hash (Results 1 – 21 of 21) sorted by relevance

/kernel/trace/
Dftrace.c403 struct hlist_head *hash; member
580 memset(stat->hash, 0, in ftrace_profile_reset()
647 if (stat->hash) { in ftrace_profile_init_cpu()
659 stat->hash = kcalloc(size, sizeof(struct hlist_head), GFP_KERNEL); in ftrace_profile_init_cpu()
661 if (!stat->hash) in ftrace_profile_init_cpu()
666 kfree(stat->hash); in ftrace_profile_init_cpu()
667 stat->hash = NULL; in ftrace_profile_init_cpu()
697 hhd = &stat->hash[key]; in ftrace_find_profiled_func()
716 hlist_add_head_rcu(&rec->node, &stat->hash[key]); in ftrace_add_profile()
769 if (!stat->hash || !ftrace_profile_enabled) in function_profile_call()
[all …]
Dtrace_functions.c645 struct ftrace_hash *hash, char *glob, in ftrace_trace_probe_callback() argument
682 ftrace_trace_onoff_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_trace_onoff_callback() argument
696 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_trace_onoff_callback()
701 ftrace_stacktrace_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_stacktrace_callback() argument
711 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_stacktrace_callback()
716 ftrace_dump_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_dump_callback() argument
727 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_dump_callback()
732 ftrace_cpudump_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_cpudump_callback() argument
743 return ftrace_trace_probe_callback(tr, ops, hash, glob, cmd, in ftrace_cpudump_callback()
Dtrace.h951 ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip);
953 static __always_inline bool ftrace_hash_empty(struct ftrace_hash *hash) in ftrace_hash_empty() argument
955 return !hash || !(hash->count || (hash->flags & FTRACE_HASH_FL_MOD)); in ftrace_hash_empty()
1006 struct ftrace_hash *hash; in ftrace_graph_addr() local
1016 hash = rcu_dereference_protected(ftrace_graph_hash, !preemptible()); in ftrace_graph_addr()
1018 if (ftrace_hash_empty(hash)) { in ftrace_graph_addr()
1023 if (ftrace_lookup_ip(hash, addr)) { in ftrace_graph_addr()
1122 struct ftrace_hash *hash,
DKconfig677 to be aggregated into hash tables and dumped to stdout by
Dtrace_events.c3016 event_enable_func(struct trace_array *tr, struct ftrace_hash *hash, in event_enable_func() argument
Dtrace.c8087 ftrace_trace_snapshot_callback(struct trace_array *tr, struct ftrace_hash *hash, in ftrace_trace_snapshot_callback() argument
/kernel/bpf/
Dhashtab.c121 u32 hash; member
236 u32 hash) in prealloc_lru_pop() argument
238 struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash); in prealloc_lru_pop()
282 offsetof(struct htab_elem, hash) - in prealloc_init()
523 static inline struct bucket *__select_bucket(struct bpf_htab *htab, u32 hash) in __select_bucket() argument
525 return &htab->buckets[hash & (htab->n_buckets - 1)]; in __select_bucket()
528 static inline struct hlist_nulls_head *select_bucket(struct bpf_htab *htab, u32 hash) in select_bucket() argument
530 return &__select_bucket(htab, hash)->head; in select_bucket()
534 static struct htab_elem *lookup_elem_raw(struct hlist_nulls_head *head, u32 hash, in lookup_elem_raw() argument
541 if (l->hash == hash && !memcmp(&l->key, key, key_size)) in lookup_elem_raw()
[all …]
Dbpf_lru_list.c356 u32 hash) in __local_list_add_pending() argument
358 *(u32 *)((void *)node + lru->hash_offset) = hash; in __local_list_add_pending()
405 u32 hash) in bpf_percpu_lru_pop_free() argument
426 *(u32 *)((void *)node + lru->hash_offset) = hash; in bpf_percpu_lru_pop_free()
437 u32 hash) in bpf_common_lru_pop_free() argument
457 __local_list_add_pending(lru, loc_l, cpu, node, hash); in bpf_common_lru_pop_free()
492 __local_list_add_pending(lru, loc_l, cpu, node, hash); in bpf_common_lru_pop_free()
499 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash) in bpf_lru_pop_free() argument
502 return bpf_percpu_lru_pop_free(lru, hash); in bpf_lru_pop_free()
504 return bpf_common_lru_pop_free(lru, hash); in bpf_lru_pop_free()
Dstackmap.c22 u32 hash; member
403 u32 hash, id, trace_nr, trace_len; in __bpf_get_stackid() local
415 hash = jhash2((u32 *)ips, trace_len / sizeof(u32), 0); in __bpf_get_stackid()
416 id = hash & (smap->n_buckets - 1); in __bpf_get_stackid()
419 hash_matches = bucket && bucket->hash == hash; in __bpf_get_stackid()
458 new_bucket->hash = hash; in __bpf_get_stackid()
Ddevmap.c93 struct hlist_head *hash; in dev_map_create_hash() local
95 hash = bpf_map_area_alloc((u64) entries * sizeof(*hash), numa_node); in dev_map_create_hash()
96 if (hash != NULL) in dev_map_create_hash()
98 INIT_HLIST_HEAD(&hash[i]); in dev_map_create_hash()
100 return hash; in dev_map_create_hash()
Dbpf_lru_list.h76 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
/kernel/locking/
Dqspinlock_paravirt.h207 #define for_each_hash_entry(he, offset, hash) \ argument
208 for (hash &= ~(PV_HE_PER_LINE - 1), he = &pv_lock_hash[hash], offset = 0; \
210 offset++, he = &pv_lock_hash[(hash + offset) & ((1 << pv_lock_hash_bits) - 1)])
214 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_hash() local
218 for_each_hash_entry(he, offset, hash) { in pv_hash()
241 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_unhash() local
245 for_each_hash_entry(he, offset, hash) { in pv_unhash()
Dlockdep.c508 u32 hash; member
522 return t1->hash == t2->hash && t1->nr_entries == t2->nr_entries && in traces_identical()
531 u32 hash; in save_trace() local
552 hash = jhash(trace->entries, trace->nr_entries * in save_trace()
554 trace->hash = hash; in save_trace()
555 hash_head = stack_trace_hash + (hash & (STACK_TRACE_HASH_SIZE - 1)); in save_trace()
1166 unsigned long hash = hash_long((uintptr_t)key, KEYHASH_BITS); in keyhashentry() local
1168 return lock_keys_hash + hash; in keyhashentry()
/kernel/
Dkprobes.c75 static raw_spinlock_t *kretprobe_table_lock_ptr(unsigned long hash) in kretprobe_table_lock_ptr() argument
77 return &(kretprobe_table_locks[hash].lock); in kretprobe_table_lock_ptr()
1244 unsigned long hash = hash_ptr(tsk, KPROBE_HASH_BITS); in kretprobe_hash_lock() local
1247 *head = &kretprobe_inst_table[hash]; in kretprobe_hash_lock()
1248 hlist_lock = kretprobe_table_lock_ptr(hash); in kretprobe_hash_lock()
1259 static void kretprobe_table_lock(unsigned long hash, in kretprobe_table_lock() argument
1263 raw_spinlock_t *hlist_lock = kretprobe_table_lock_ptr(hash); in kretprobe_table_lock()
1278 unsigned long hash = hash_ptr(tsk, KPROBE_HASH_BITS); in kretprobe_hash_unlock() local
1281 hlist_lock = kretprobe_table_lock_ptr(hash); in kretprobe_hash_unlock()
1286 static void kretprobe_table_unlock(unsigned long hash, in kretprobe_table_unlock() argument
[all …]
Daudit_tree.c26 struct list_head hash; member
197 INIT_LIST_HEAD(&chunk->hash); in alloc_chunk()
238 list_add_rcu(&chunk->hash, list); in insert_hash()
248 list_for_each_entry_rcu(p, list, hash) { in audit_tree_lookup()
322 list_replace_rcu(&old->hash, &new->hash); in replace_chunk()
367 list_del_rcu(&chunk->hash); in untag_chunk()
1029 list_del_rcu(&chunk->hash); in evict_chunk()
Dmodule_signature.c34 ms->hash != 0 || in mod_check_sig()
Dworkqueue.c3424 u32 hash = 0; in wqattrs_hash() local
3426 hash = jhash_1word(attrs->nice, hash); in wqattrs_hash()
3427 hash = jhash(cpumask_bits(attrs->cpumask), in wqattrs_hash()
3428 BITS_TO_LONGS(nr_cpumask_bits) * sizeof(long), hash); in wqattrs_hash()
3429 return hash; in wqattrs_hash()
3635 u32 hash = wqattrs_hash(attrs); in get_unbound_pool() local
3643 hash_for_each_possible(unbound_pool_hash, pool, hash_node, hash) { in get_unbound_pool()
3684 hash_add(unbound_pool_hash, &pool->hash_node, hash); in get_unbound_pool()
/kernel/time/
Dposix-timers.c113 static int hash(struct signal_struct *sig, unsigned int nr) in hash() function
135 struct hlist_head *head = &posix_timers_hashtable[hash(sig, id)]; in posix_timer_by_id()
149 head = &posix_timers_hashtable[hash(sig, sig->posix_timer_id)]; in posix_timer_add()
/kernel/futex/
Dcore.c364 u32 hash = jhash2((u32 *)key, offsetof(typeof(*key), both.offset) / 4, in hash_futex() local
367 return &futex_queues[hash & (futex_hashsize - 1)]; in hash_futex()
/kernel/power/
DKconfig254 dmesg -s 1000000 | grep 'hash matches'
/kernel/events/
Dcore.c9152 u64 hash = swevent_hash(type, event_id); in __find_swevent_head() local
9154 return &hlist->heads[hash]; in __find_swevent_head()