/kernel/bpf/ |
D | btf.c | 204 u32 *resolved_ids; 205 u32 *resolved_sizes; 209 u32 nr_types; 210 u32 types_size; 211 u32 data_size; 213 u32 id; 224 u32 type_id; 245 u32 off; 246 u32 len; 254 u32 log_type_id; [all …]
|
D | arraymap.c | 75 u32 elem_size, index_mask, max_entries; in array_map_alloc() 146 u32 index = *(u32 *)key; in array_map_lookup_elem() 155 u32 off) in array_map_direct_value_addr() 169 u32 *off) in array_map_direct_value_meta() 185 static u32 array_map_gen_lookup(struct bpf_map *map, struct bpf_insn *insn_buf) in array_map_gen_lookup() 189 u32 elem_size = round_up(map->value_size, 8); in array_map_gen_lookup() 218 u32 index = *(u32 *)key; in percpu_array_map_lookup_elem() 229 u32 index = *(u32 *)key; in bpf_percpu_array_copy() 232 u32 size; in bpf_percpu_array_copy() 256 u32 index = key ? *(u32 *)key : U32_MAX; in array_map_get_next_key() [all …]
|
D | core.c | 186 const u32 *insn_to_jit_off) in bpf_prog_fill_jited_linfo() 188 u32 linfo_idx, insn_start, insn_end, nr_linfo, i; in bpf_prog_fill_jited_linfo() 225 u32 pages, delta; in bpf_prog_realloc() 269 const u32 bits_offset = SHA_MESSAGE_BYTES - sizeof(__be64); in bpf_prog_calc_tag() 270 u32 raw_size = bpf_prog_tag_scratch_size(fp); in bpf_prog_calc_tag() 271 u32 digest[SHA_DIGEST_WORDS]; in bpf_prog_calc_tag() 272 u32 ws[SHA_WORKSPACE_WORDS]; in bpf_prog_calc_tag() 273 u32 i, bsize, psize, blocks; in bpf_prog_calc_tag() 340 static int bpf_adj_delta_to_imm(struct bpf_insn *insn, u32 pos, s32 end_old, in bpf_adj_delta_to_imm() 358 static int bpf_adj_delta_to_off(struct bpf_insn *insn, u32 pos, s32 end_old, in bpf_adj_delta_to_off() [all …]
|
D | reuseport_array.c | 42 if (attr->value_size != sizeof(u32) && in reuseport_array_alloc_check() 52 u32 index = *(u32 *)key; in reuseport_array_lookup_elem() 64 u32 index = *(u32 *)key; in reuseport_array_delete_elem() 97 u32 i; in reuseport_array_free() 208 u32 map_flags) in reuseport_array_update_check() 254 u32 index = *(u32 *)key; in bpf_fd_reuseport_array_update_elem() 338 u32 index = key ? *(u32 *)key : U32_MAX; in reuseport_array_get_next_key() 339 u32 *next = (u32 *)next_key; in reuseport_array_get_next_key()
|
D | devmap.c | 85 u32 n_buckets; 278 u32 index = key ? *(u32 *)key : U32_MAX; in dev_map_get_next_key() 279 u32 *next = next_key; in dev_map_get_next_key() 292 struct bpf_dtab_netdev *__dev_map_hash_lookup_elem(struct bpf_map *map, u32 key) in __dev_map_hash_lookup_elem() 310 u32 idx, *next = next_key; in dev_map_hash_get_next_key() 318 idx = *(u32 *)key; in dev_map_hash_get_next_key() 351 static int bq_xmit_all(struct xdp_bulk_queue *bq, u32 flags, in bq_xmit_all() 423 struct bpf_dtab_netdev *__dev_map_lookup_elem(struct bpf_map *map, u32 key) in __dev_map_lookup_elem() 500 struct bpf_dtab_netdev *obj = __dev_map_lookup_elem(map, *(u32 *)key); in dev_map_lookup_elem() 509 *(u32 *)key); in dev_map_hash_lookup_elem() [all …]
|
D | stackmap.c | 20 u32 hash; 21 u32 nr; 29 u32 n_buckets; 88 u32 value_size = attr->value_size; in stack_map_alloc() 289 u64 *ips, u32 trace_nr, bool user) in stack_map_get_build_id_offset() 358 u32 max_depth = map->value_size / stack_map_data_size(map); in BPF_CALL_3() 360 u32 init_nr = sysctl_perf_event_max_stack - max_depth; in BPF_CALL_3() 361 u32 skip = flags & BPF_F_SKIP_FIELD_MASK; in BPF_CALL_3() 362 u32 hash, id, trace_nr, trace_len; in BPF_CALL_3() 391 hash = jhash2((u32 *)ips, trace_len / sizeof(u32), 0); in BPF_CALL_3() [all …]
|
D | hashtab.c | 35 u32 n_buckets; /* number of hash buckets */ 36 u32 elem_size; /* size of each element in bytes */ 37 u32 hashrnd; 56 u32 hash; 79 static inline void htab_elem_set_ptr(struct htab_elem *l, u32 key_size, in htab_elem_set_ptr() 85 static inline void __percpu *htab_elem_get_ptr(struct htab_elem *l, u32 key_size) in htab_elem_get_ptr() 120 u32 hash) in prealloc_lru_pop() 136 u32 num_entries = htab->map.max_entries; in prealloc_init() 151 u32 size = round_up(htab->map.value_size, 8); in prealloc_init() 409 static inline u32 htab_map_hash(const void *key, u32 key_len, u32 hashrnd) in htab_map_hash() [all …]
|
D | cpumap.c | 53 u32 cpu; /* kthread CPU and map index */ 55 u32 qsize; /* Queue size placeholder for map lookup */ 337 static struct bpf_cpu_map_entry *__cpu_map_entry_alloc(u32 qsize, u32 cpu, in __cpu_map_entry_alloc() 446 u32 key_cpu, struct bpf_cpu_map_entry *rcpu) in __cpu_map_entry_replace() 461 u32 key_cpu = *(u32 *)key; in cpu_map_delete_elem() 478 u32 key_cpu = *(u32 *)key; in cpu_map_update_elem() 480 u32 qsize = *(u32 *)value; in cpu_map_update_elem() 514 u32 i; in cpu_map_free() 558 struct bpf_cpu_map_entry *__cpu_map_lookup_elem(struct bpf_map *map, u32 key) in __cpu_map_lookup_elem() 573 __cpu_map_lookup_elem(map, *(u32 *)key); in cpu_map_lookup_elem() [all …]
|
D | bpf_lru_list.c | 356 u32 hash) in __local_list_add_pending() 358 *(u32 *)((void *)node + lru->hash_offset) = hash; in __local_list_add_pending() 405 u32 hash) in bpf_percpu_lru_pop_free() 426 *(u32 *)((void *)node + lru->hash_offset) = hash; in bpf_percpu_lru_pop_free() 437 u32 hash) in bpf_common_lru_pop_free() 499 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash) in bpf_lru_pop_free() 565 u32 node_offset, u32 elem_size, in bpf_common_lru_populate() 566 u32 nr_elems) in bpf_common_lru_populate() 569 u32 i; in bpf_common_lru_populate() 583 u32 node_offset, u32 elem_size, in bpf_percpu_lru_populate() [all …]
|
D | syscall.c | 104 u32 type = attr->map_type; in find_and_alloc_map() 165 static u32 bpf_map_flags_retain_permanent(u32 flags) in bpf_map_flags_retain_permanent() 187 static int bpf_charge_memlock(struct user_struct *user, u32 pages) in bpf_charge_memlock() 198 static void bpf_uncharge_memlock(struct user_struct *user, u32 pages) in bpf_uncharge_memlock() 206 u32 pages = round_up(size, PAGE_SIZE) >> PAGE_SHIFT; in bpf_map_charge_init() 241 int bpf_map_charge_memlock(struct bpf_map *map, u32 pages) in bpf_map_charge_memlock() 252 void bpf_map_uncharge_memlock(struct bpf_map *map, u32 pages) in bpf_map_uncharge_memlock() 377 u32 owner_prog_type = 0; in bpf_map_show_fdinfo() 378 u32 owner_jited = 0; in bpf_map_show_fdinfo() 503 u32 btf_key_id, u32 btf_value_id) in map_check_btf() [all …]
|
D | bpf_lru_list.h | 70 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, 72 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, 73 u32 elem_size, u32 nr_elems); 75 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
|
/kernel/ |
D | futex.c | 240 u32 bitset; 389 u32 hash = jhash2((u32 *)key, offsetof(typeof(*key), both.offset) / 4, in hash_futex() 573 get_futex_key(u32 __user *uaddr, int fshared, union futex_key *key, enum futex_access rw) in get_futex_key() 585 if (unlikely((address % sizeof(u32)) != 0)) in get_futex_key() 589 if (unlikely(!access_ok(uaddr, sizeof(u32)))) in get_futex_key() 769 static int fault_in_user_writeable(u32 __user *uaddr) in fault_in_user_writeable() 801 static int cmpxchg_futex_value_locked(u32 *curval, u32 __user *uaddr, in cmpxchg_futex_value_locked() 802 u32 uval, u32 newval) in cmpxchg_futex_value_locked() 813 static int get_futex_value_locked(u32 *dest, u32 __user *from) in get_futex_value_locked() 1098 static int attach_to_pi_state(u32 __user *uaddr, u32 uval, in attach_to_pi_state() [all …]
|
D | test_kprobes.c | 16 static u32 rand1, preh_val, posth_val; 18 static u32 (*target)(u32 value); 19 static u32 (*target2)(u32 value); 21 static noinline u32 kprobe_target(u32 value) in kprobe_target() 82 static noinline u32 kprobe_target2(u32 value) in kprobe_target2() 157 static u32 krph_val;
|
D | audit.h | 80 u32 osid; 137 u32 target_sid; 154 u32 osid; 212 static inline int audit_hash_ino(u32 ino) in audit_hash_ino() 221 extern int audit_comparator(const u32 left, const u32 op, const u32 right); 222 extern int audit_uid_comparator(kuid_t left, u32 op, kuid_t right); 223 extern int audit_gid_comparator(kgid_t left, u32 op, kgid_t right); 262 u32 op); 284 extern int audit_make_tree(struct audit_krule *rule, char *pathname, u32 op); 336 extern u32 audit_sig_sid;
|
D | rseq.c | 86 u32 cpu_id = raw_smp_processor_id(); in rseq_update_cpu_id() 98 u32 cpu_id_start = 0, cpu_id = RSEQ_CPU_ID_UNINITIALIZED; in rseq_reset_rseq_cpu_id() 119 u32 __user *usig; in rseq_get_rseq_cs() 120 u32 sig; in rseq_get_rseq_cs() 147 usig = (u32 __user *)(unsigned long)(rseq_cs->abort_ip - sizeof(u32)); in rseq_get_rseq_cs() 161 static int rseq_need_restart(struct task_struct *t, u32 cs_flags) in rseq_need_restart() 163 u32 flags, event_mask; in rseq_need_restart() 307 SYSCALL_DEFINE4(rseq, struct rseq __user *, rseq, u32, rseq_len, in SYSCALL_DEFINE4() argument 308 int, flags, u32, sig) in SYSCALL_DEFINE4()
|
D | user_namespace.c | 215 u32 id; /* id to find */ 216 u32 count; /* == 0 unless used with map_id_range_down() */ 225 u32 first, last, id2; in cmp_map_id() 254 map_id_range_down_max(unsigned extents, struct uid_gid_map *map, u32 id, u32 count) in map_id_range_down_max() 272 map_id_range_down_base(unsigned extents, struct uid_gid_map *map, u32 id, u32 count) in map_id_range_down_base() 275 u32 first, last, id2; in map_id_range_down_base() 290 static u32 map_id_range_down(struct uid_gid_map *map, u32 id, u32 count) in map_id_range_down() 305 id = (u32) -1; in map_id_range_down() 310 static u32 map_id_down(struct uid_gid_map *map, u32 id) in map_id_down() 321 map_id_up_base(unsigned extents, struct uid_gid_map *map, u32 id) in map_id_up_base() [all …]
|
D | audit.c | 72 u32 audit_enabled = AUDIT_OFF; 78 static u32 audit_default = AUDIT_OFF; 81 static u32 audit_failure = AUDIT_FAIL_PRINTK; 107 u32 portid; 116 static u32 audit_rate_limit; 120 static u32 audit_backlog_limit = 64; 122 static u32 audit_backlog_wait_time = AUDIT_BACKLOG_WAIT_TIME; 127 u32 audit_sig_sid = 0; 380 static int audit_log_config_change(char *function_name, u32 new, u32 old, in audit_log_config_change() 399 static int audit_do_config_change(char *function_name, u32 *to_change, u32 new) in audit_do_config_change() [all …]
|
/kernel/locking/ |
D | qspinlock.c | 114 static inline __pure u32 encode_tail(int cpu, int idx) in encode_tail() 116 u32 tail; in encode_tail() 124 static inline __pure struct mcs_spinlock *decode_tail(u32 tail) in decode_tail() 175 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 181 return (u32)xchg_relaxed(&lock->tail, in xchg_tail() 219 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 221 u32 old, new, val = atomic_read(&lock->val); in xchg_tail() 248 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire() 276 static __always_inline u32 __pv_wait_head_or_lock(struct qspinlock *lock, in __pv_wait_head_or_lock() 314 void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in queued_spin_lock_slowpath() [all …]
|
/kernel/gcov/ |
D | clang.c | 63 u32 checksum; 71 u32 ident; 72 u32 checksum; 76 u32 cfg_checksum; 78 u32 num_counters; 114 u32 checksum) in llvm_gcda_start_file() 122 void llvm_gcda_start_file(const char *orig_filename, u32 version, u32 checksum) in llvm_gcda_start_file() 132 void llvm_gcda_emit_function(u32 ident, const char *function_name, in llvm_gcda_emit_function() 133 u32 func_checksum, u8 use_extra_checksum, u32 cfg_checksum) in llvm_gcda_emit_function() 151 void llvm_gcda_emit_function(u32 ident, u32 func_checksum, u32 cfg_checksum) in llvm_gcda_emit_function() [all …]
|
/kernel/trace/ |
D | trace_probe_tmpl.h | 17 *(u32 *)buf = (u32)val; in fetch_store_raw() 41 *(u32 *)buf <<= code->lshift; in fetch_apply_bitfield() 42 *(u32 *)buf >>= code->rshift; in fetch_apply_bitfield() 77 u32 loc = 0; in process_fetch_insn_bottom() 125 loc = *(u32 *)dest; in process_fetch_insn_bottom() 129 loc = *(u32 *)dest; in process_fetch_insn_bottom() 160 dest += sizeof(u32); in process_fetch_insn_bottom() 161 *(u32 *)dest = update_data_loc(loc, ret); in process_fetch_insn_bottom() 199 u32 *dl; /* Data location */ in store_trace_args()
|
D | trace_probe.h | 61 (((u32)(len) << 16) | ((u32)(offs) & 0xffff)) 62 #define get_loc_len(dl) ((u32)(dl) >> 16) 63 #define get_loc_offs(dl) ((u32)(dl) & 0xffff) 65 static nokprobe_inline void *get_loc_data(u32 *dl, void *ent) in get_loc_data() 70 static nokprobe_inline u32 update_data_loc(u32 loc, int consumed) in update_data_loc() 72 u32 maxlen = get_loc_len(loc); in update_data_loc() 73 u32 offset = get_loc_offs(loc); in update_data_loc() 143 typedef u32 string; 144 typedef u32 string_size; 156 DECLARE_BASIC_PRINT_TYPE_FUNC(u32);
|
/kernel/sched/ |
D | pelt.c | 63 static u32 __accumulate_pelt_segments(u64 periods, u32 d1, u32 d3) in __accumulate_pelt_segments() 65 u32 c1, c2, c3 = d3; /* y^0 == 1 */ in __accumulate_pelt_segments() 109 static __always_inline u32 113 u32 contrib = (u32)delta; /* p == 0 -> delta < 1024 */ in accumulate_sum() 229 u32 divider = LOAD_AVG_MAX - 1024 + sa->period_contrib; in ___update_load_avg()
|
D | psi.c | 238 enum psi_aggregators aggregator, u32 *times, in get_recent_times() 239 u32 *pchanged_states) in get_recent_times() 245 u32 state_mask; in get_recent_times() 260 u32 delta; in get_recent_times() 304 u32 *pchanged_states) in collect_percpu_times() 308 u32 changed_states = 0; in collect_percpu_times() 321 u32 times[NR_PSI_STATES]; in collect_percpu_times() 322 u32 nonidle; in collect_percpu_times() 323 u32 cpu_changed_states; in collect_percpu_times() 381 u32 sample; in update_averages() [all …]
|
/kernel/irq/ |
D | generic-chip.c | 40 u32 mask = d->mask; in irq_gc_mask_disable_reg() 59 u32 mask = d->mask; in irq_gc_mask_set_bit() 79 u32 mask = d->mask; in irq_gc_mask_clr_bit() 99 u32 mask = d->mask; in irq_gc_unmask_enable_reg() 115 u32 mask = d->mask; in irq_gc_ack_set_bit() 131 u32 mask = ~d->mask; in irq_gc_ack_clr_bit() 154 u32 mask = d->mask; in irq_gc_mask_disable_and_ack_set() 171 u32 mask = d->mask; in irq_gc_eoi() 190 u32 mask = d->mask; in irq_gc_set_wake() 204 static u32 irq_readl_be(void __iomem *addr) in irq_readl_be() [all …]
|
/kernel/printk/ |
D | printk.c | 381 u32 caller_id; /* thread id or processor id */ 428 static u32 syslog_idx; 434 static u32 log_first_idx; 438 static u32 log_next_idx; 442 static u32 console_idx; 447 static u32 clear_idx; 462 #define LOG_BUF_LEN_MAX (u32)(1 << 31) 465 static u32 log_buf_len = __LOG_BUF_LEN; 486 u32 log_buf_len_get(void) in log_buf_len_get() 504 static struct printk_log *log_from_idx(u32 idx) in log_from_idx() [all …]
|