Home
last modified time | relevance | path

Searched refs:u32 (Results 1 – 25 of 136) sorted by relevance

123456

/kernel/bpf/
Dbtf.c202 u32 *resolved_ids;
203 u32 *resolved_sizes;
207 u32 nr_types;
208 u32 types_size;
209 u32 data_size;
211 u32 id;
222 u32 type_id;
243 u32 off;
244 u32 len;
252 u32 log_type_id;
[all …]
Dcore.c192 const u32 *insn_to_jit_off) in bpf_prog_fill_jited_linfo()
194 u32 linfo_idx, insn_start, insn_end, nr_linfo, i; in bpf_prog_fill_jited_linfo()
231 u32 pages, delta; in bpf_prog_realloc()
276 const u32 bits_offset = SHA1_BLOCK_SIZE - sizeof(__be64); in bpf_prog_calc_tag()
277 u32 raw_size = bpf_prog_tag_scratch_size(fp); in bpf_prog_calc_tag()
278 u32 digest[SHA1_DIGEST_WORDS]; in bpf_prog_calc_tag()
279 u32 ws[SHA1_WORKSPACE_WORDS]; in bpf_prog_calc_tag()
280 u32 i, bsize, psize, blocks; in bpf_prog_calc_tag()
347 static int bpf_adj_delta_to_imm(struct bpf_insn *insn, u32 pos, s32 end_old, in bpf_adj_delta_to_imm()
365 static int bpf_adj_delta_to_off(struct bpf_insn *insn, u32 pos, s32 end_old, in bpf_adj_delta_to_off()
[all …]
Darraymap.c85 u32 elem_size, index_mask, max_entries; in array_map_alloc()
178 u32 index = *(u32 *)key; in array_map_lookup_elem()
187 u32 off) in array_map_direct_value_addr()
201 u32 *off) in array_map_direct_value_meta()
221 u32 elem_size = round_up(map->value_size, 8); in array_map_gen_lookup()
253 u32 index = *(u32 *)key; in percpu_array_map_lookup_elem()
264 u32 index = *(u32 *)key; in bpf_percpu_array_copy()
267 u32 size; in bpf_percpu_array_copy()
291 u32 index = key ? *(u32 *)key : U32_MAX; in array_map_get_next_key()
292 u32 *next = (u32 *)next_key; in array_map_get_next_key()
[all …]
Dreuseport_array.c45 if (attr->value_size != sizeof(u32) && in reuseport_array_alloc_check()
55 u32 index = *(u32 *)key; in reuseport_array_lookup_elem()
67 u32 index = *(u32 *)key; in reuseport_array_delete_elem()
100 u32 i; in reuseport_array_free()
209 u32 map_flags) in reuseport_array_update_check()
255 u32 index = *(u32 *)key; in bpf_fd_reuseport_array_update_elem()
337 u32 index = key ? *(u32 *)key : U32_MAX; in reuseport_array_get_next_key()
338 u32 *next = (u32 *)next_key; in reuseport_array_get_next_key()
Dstackmap.c22 u32 hash;
23 u32 nr;
31 u32 n_buckets;
92 u32 value_size = attr->value_size; in stack_map_alloc()
294 u64 *ips, u32 trace_nr, bool user) in stack_map_get_build_id_offset()
361 get_callchain_entry_for_task(struct task_struct *task, u32 max_depth) in get_callchain_entry_for_task()
402 u32 skip = flags & BPF_F_SKIP_FIELD_MASK; in __bpf_get_stackid()
403 u32 hash, id, trace_nr, trace_len; in __bpf_get_stackid()
415 hash = jhash2((u32 *)ips, trace_len / sizeof(u32), 0); in __bpf_get_stackid()
470 u32 max_depth = map->value_size / stack_map_data_size(map); in BPF_CALL_3()
[all …]
Dbpf_lru_list.c356 u32 hash) in __local_list_add_pending()
358 *(u32 *)((void *)node + lru->hash_offset) = hash; in __local_list_add_pending()
405 u32 hash) in bpf_percpu_lru_pop_free()
426 *(u32 *)((void *)node + lru->hash_offset) = hash; in bpf_percpu_lru_pop_free()
437 u32 hash) in bpf_common_lru_pop_free()
499 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash) in bpf_lru_pop_free()
565 u32 node_offset, u32 elem_size, in bpf_common_lru_populate()
566 u32 nr_elems) in bpf_common_lru_populate()
569 u32 i; in bpf_common_lru_populate()
583 u32 node_offset, u32 elem_size, in bpf_percpu_lru_populate()
[all …]
Dhashtab.c99 u32 n_buckets; /* number of hash buckets */
100 u32 elem_size; /* size of each element in bytes */
101 u32 hashrnd;
121 u32 hash;
184 static inline void htab_elem_set_ptr(struct htab_elem *l, u32 key_size, in htab_elem_set_ptr()
190 static inline void __percpu *htab_elem_get_ptr(struct htab_elem *l, u32 key_size) in htab_elem_get_ptr()
236 u32 hash) in prealloc_lru_pop()
252 u32 num_entries = htab->map.max_entries; in prealloc_init()
267 u32 size = round_up(htab->map.value_size, 8); in prealloc_init()
518 static inline u32 htab_map_hash(const void *key, u32 key_len, u32 hashrnd) in htab_map_hash()
[all …]
Ddevmap.c82 u32 n_buckets;
111 u32 valsize = attr->value_size; in dev_map_init_map()
261 u32 index = key ? *(u32 *)key : U32_MAX; in dev_map_get_next_key()
262 u32 *next = next_key; in dev_map_get_next_key()
275 struct bpf_dtab_netdev *__dev_map_hash_lookup_elem(struct bpf_map *map, u32 key) in __dev_map_hash_lookup_elem()
293 u32 idx, *next = next_key; in dev_map_hash_get_next_key()
301 idx = *(u32 *)key; in dev_map_hash_get_next_key()
344 static void bq_xmit_all(struct xdp_dev_bulk_queue *bq, u32 flags) in bq_xmit_all()
409 struct bpf_dtab_netdev *__dev_map_lookup_elem(struct bpf_map *map, u32 key) in __dev_map_lookup_elem()
472 u32 act; in dev_map_run_prog()
[all …]
Dcpumap.c53 u32 cpu; /* kthread CPU and map index */
84 u32 value_size = attr->value_size; in cpu_map_alloc()
246 u32 act; in cpu_map_bpf_prog_run_xdp()
415 __cpu_map_entry_alloc(struct bpf_cpumap_val *value, u32 cpu, int map_id) in __cpu_map_entry_alloc()
521 u32 key_cpu, struct bpf_cpu_map_entry *rcpu) in __cpu_map_entry_replace()
536 u32 key_cpu = *(u32 *)key; in cpu_map_delete_elem()
553 u32 key_cpu = *(u32 *)key; in cpu_map_update_elem()
588 u32 i; in cpu_map_free()
619 struct bpf_cpu_map_entry *__cpu_map_lookup_elem(struct bpf_map *map, u32 key) in __cpu_map_lookup_elem()
634 __cpu_map_lookup_elem(map, *(u32 *)key); in cpu_map_lookup_elem()
[all …]
Dbpf_lru_list.h71 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
73 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
74 u32 elem_size, u32 nr_elems);
76 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
Dsyscall.c107 u32 type = attr->map_type; in find_and_alloc_map()
148 static u32 bpf_map_value_size(struct bpf_map *map) in bpf_map_value_size()
156 return sizeof(u32); in bpf_map_value_size()
340 static u32 bpf_map_flags_retain_permanent(u32 flags) in bpf_map_flags_retain_permanent()
362 static int bpf_charge_memlock(struct user_struct *user, u32 pages) in bpf_charge_memlock()
373 static void bpf_uncharge_memlock(struct user_struct *user, u32 pages) in bpf_uncharge_memlock()
381 u32 pages = round_up(size, PAGE_SIZE) >> PAGE_SHIFT; in bpf_map_charge_init()
416 int bpf_map_charge_memlock(struct bpf_map *map, u32 pages) in bpf_map_charge_memlock()
427 void bpf_map_uncharge_memlock(struct bpf_map *map, u32 pages) in bpf_map_uncharge_memlock()
552 u32 type = 0, jited = 0; in bpf_map_show_fdinfo()
[all …]
/kernel/futex/
Dcore.c222 u32 bitset;
364 u32 hash = jhash2((u32 *)key, offsetof(typeof(*key), both.offset) / 4, in hash_futex()
487 static int get_futex_key(u32 __user *uaddr, bool fshared, union futex_key *key, in get_futex_key()
500 if (unlikely((address % sizeof(u32)) != 0)) in get_futex_key()
504 if (unlikely(!access_ok(uaddr, sizeof(u32)))) in get_futex_key()
686 static int fault_in_user_writeable(u32 __user *uaddr) in fault_in_user_writeable()
718 static int cmpxchg_futex_value_locked(u32 *curval, u32 __user *uaddr, in cmpxchg_futex_value_locked()
719 u32 uval, u32 newval) in cmpxchg_futex_value_locked()
730 static int get_futex_value_locked(u32 *dest, u32 __user *from) in get_futex_value_locked()
1015 static int attach_to_pi_state(u32 __user *uaddr, u32 uval, in attach_to_pi_state()
[all …]
/kernel/locking/
Dqspinlock.c115 static inline __pure u32 encode_tail(int cpu, int idx) in encode_tail()
117 u32 tail; in encode_tail()
125 static inline __pure struct mcs_spinlock *decode_tail(u32 tail) in decode_tail()
176 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail()
182 return (u32)xchg_relaxed(&lock->tail, in xchg_tail()
220 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail()
222 u32 old, new, val = atomic_read(&lock->val); in xchg_tail()
249 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire()
277 static __always_inline u32 __pv_wait_head_or_lock(struct qspinlock *lock, in __pv_wait_head_or_lock()
315 void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in queued_spin_lock_slowpath()
[all …]
/kernel/gcov/
Dclang.c63 u32 checksum;
71 u32 ident;
72 u32 checksum;
76 u32 cfg_checksum;
78 u32 num_counters;
114 u32 checksum) in llvm_gcda_start_file()
122 void llvm_gcda_start_file(const char *orig_filename, u32 version, u32 checksum) in llvm_gcda_start_file()
132 void llvm_gcda_emit_function(u32 ident, const char *function_name, in llvm_gcda_emit_function()
133 u32 func_checksum, u8 use_extra_checksum, u32 cfg_checksum) in llvm_gcda_emit_function()
151 void llvm_gcda_emit_function(u32 ident, u32 func_checksum, u32 cfg_checksum) in llvm_gcda_emit_function()
[all …]
/kernel/
Dtest_kprobes.c16 static u32 rand1, preh_val, posth_val;
18 static u32 (*target)(u32 value);
19 static u32 (*target2)(u32 value);
21 static noinline u32 kprobe_target(u32 value) in kprobe_target()
82 static noinline u32 kprobe_target2(u32 value) in kprobe_target2()
157 static u32 krph_val;
Drseq.c86 u32 cpu_id = raw_smp_processor_id(); in rseq_update_cpu_id()
98 u32 cpu_id_start = 0, cpu_id = RSEQ_CPU_ID_UNINITIALIZED; in rseq_reset_rseq_cpu_id()
119 u32 __user *usig; in rseq_get_rseq_cs()
120 u32 sig; in rseq_get_rseq_cs()
152 usig = (u32 __user *)(unsigned long)(rseq_cs->abort_ip - sizeof(u32)); in rseq_get_rseq_cs()
166 static int rseq_need_restart(struct task_struct *t, u32 cs_flags) in rseq_need_restart()
168 u32 flags, event_mask; in rseq_need_restart()
323 SYSCALL_DEFINE4(rseq, struct rseq __user *, rseq, u32, rseq_len, in SYSCALL_DEFINE4() argument
324 int, flags, u32, sig) in SYSCALL_DEFINE4()
Daudit.h80 u32 osid;
137 u32 target_sid;
154 u32 osid;
212 static inline int audit_hash_ino(u32 ino) in audit_hash_ino()
221 extern int audit_comparator(const u32 left, const u32 op, const u32 right);
222 extern int audit_uid_comparator(kuid_t left, u32 op, kuid_t right);
223 extern int audit_gid_comparator(kgid_t left, u32 op, kgid_t right);
262 u32 op);
284 extern int audit_make_tree(struct audit_krule *rule, char *pathname, u32 op);
Daudit.c72 u32 audit_enabled = AUDIT_OFF;
78 static u32 audit_default = AUDIT_OFF;
81 static u32 audit_failure = AUDIT_FAIL_PRINTK;
107 u32 portid;
117 static u32 audit_rate_limit;
121 static u32 audit_backlog_limit = 64;
123 static u32 audit_backlog_wait_time = AUDIT_BACKLOG_WAIT_TIME;
128 static u32 audit_sig_sid;
386 static int audit_log_config_change(char *function_name, u32 new, u32 old, in audit_log_config_change()
405 static int audit_do_config_change(char *function_name, u32 *to_change, u32 new) in audit_do_config_change()
[all …]
Duser_namespace.c216 u32 id; /* id to find */
217 u32 count; /* == 0 unless used with map_id_range_down() */
226 u32 first, last, id2; in cmp_map_id()
255 map_id_range_down_max(unsigned extents, struct uid_gid_map *map, u32 id, u32 count) in map_id_range_down_max()
273 map_id_range_down_base(unsigned extents, struct uid_gid_map *map, u32 id, u32 count) in map_id_range_down_base()
276 u32 first, last, id2; in map_id_range_down_base()
291 static u32 map_id_range_down(struct uid_gid_map *map, u32 id, u32 count) in map_id_range_down()
306 id = (u32) -1; in map_id_range_down()
311 static u32 map_id_down(struct uid_gid_map *map, u32 id) in map_id_down()
322 map_id_up_base(unsigned extents, struct uid_gid_map *map, u32 id) in map_id_up_base()
[all …]
/kernel/trace/
Dtrace_probe_tmpl.h17 *(u32 *)buf = (u32)val; in fetch_store_raw()
41 *(u32 *)buf <<= code->lshift; in fetch_apply_bitfield()
42 *(u32 *)buf >>= code->rshift; in fetch_apply_bitfield()
77 u32 loc = 0; in process_fetch_insn_bottom()
125 loc = *(u32 *)dest; in process_fetch_insn_bottom()
129 loc = *(u32 *)dest; in process_fetch_insn_bottom()
160 dest += sizeof(u32); in process_fetch_insn_bottom()
161 *(u32 *)dest = update_data_loc(loc, ret); in process_fetch_insn_bottom()
199 u32 *dl; /* Data location */ in store_trace_args()
Dbpf_trace.c74 static int bpf_btf_printf_prepare(struct btf_ptr *ptr, u32 btf_ptr_size,
150 bpf_probe_read_user_common(void *dst, u32 size, const void __user *unsafe_ptr) in bpf_probe_read_user_common()
160 BPF_CALL_3(bpf_probe_read_user, void *, dst, u32, size, in BPF_CALL_3() argument
176 bpf_probe_read_user_str_common(void *dst, u32 size, in bpf_probe_read_user_str_common()
197 BPF_CALL_3(bpf_probe_read_user_str, void *, dst, u32, size, in BPF_CALL_3() argument
213 bpf_probe_read_kernel_common(void *dst, u32 size, const void *unsafe_ptr) in bpf_probe_read_kernel_common()
223 BPF_CALL_3(bpf_probe_read_kernel, void *, dst, u32, size, in BPF_CALL_3() argument
239 bpf_probe_read_kernel_str_common(void *dst, u32 size, const void *unsafe_ptr) in bpf_probe_read_kernel_str_common()
258 BPF_CALL_3(bpf_probe_read_kernel_str, void *, dst, u32, size, in BPF_CALL_3() argument
274 BPF_CALL_3(bpf_probe_read_compat, void *, dst, u32, size, in BPF_CALL_3() argument
[all …]
Dtrace_probe.h60 (((u32)(len) << 16) | ((u32)(offs) & 0xffff))
61 #define get_loc_len(dl) ((u32)(dl) >> 16)
62 #define get_loc_offs(dl) ((u32)(dl) & 0xffff)
64 static nokprobe_inline void *get_loc_data(u32 *dl, void *ent) in get_loc_data()
69 static nokprobe_inline u32 update_data_loc(u32 loc, int consumed) in update_data_loc()
71 u32 maxlen = get_loc_len(loc); in update_data_loc()
72 u32 offset = get_loc_offs(loc); in update_data_loc()
142 typedef u32 string;
143 typedef u32 string_size;
155 DECLARE_BASIC_PRINT_TYPE_FUNC(u32);
/kernel/irq/
Dsettings.h42 __irq_settings_clr_and_set(struct irq_desc *desc, u32 clr, u32 set, u32 mask) in __irq_settings_clr_and_set()
49 irq_settings_clr_and_set(struct irq_desc *desc, u32 clr, u32 set) in irq_settings_clr_and_set()
79 static inline u32 irq_settings_get_trigger_mask(struct irq_desc *desc) in irq_settings_get_trigger_mask()
85 irq_settings_set_trigger_mask(struct irq_desc *desc, u32 mask) in irq_settings_set_trigger_mask()
Dgeneric-chip.c40 u32 mask = d->mask; in irq_gc_mask_disable_reg()
59 u32 mask = d->mask; in irq_gc_mask_set_bit()
79 u32 mask = d->mask; in irq_gc_mask_clr_bit()
99 u32 mask = d->mask; in irq_gc_unmask_enable_reg()
115 u32 mask = d->mask; in irq_gc_ack_set_bit()
131 u32 mask = ~d->mask; in irq_gc_ack_clr_bit()
154 u32 mask = d->mask; in irq_gc_mask_disable_and_ack_set()
171 u32 mask = d->mask; in irq_gc_eoi()
190 u32 mask = d->mask; in irq_gc_set_wake()
205 static u32 irq_readl_be(void __iomem *addr) in irq_readl_be()
[all …]
/kernel/sched/
Dpelt.c33 const u32 *pelt_runnable_avg_yN_inv = pelt32_runnable_avg_yN_inv;
120 static u32 __accumulate_pelt_segments(u64 periods, u32 d1, u32 d3) in __accumulate_pelt_segments()
122 u32 c1, c2, c3 = d3; /* y^0 == 1 */ in __accumulate_pelt_segments()
164 static __always_inline u32
168 u32 contrib = (u32)delta; /* p == 0 -> delta < 1024 */ in accumulate_sum()
322 u32 divider = get_pelt_divider(sa); in ___update_load_avg()

123456