Home
last modified time | relevance | path

Searched refs:b (Results 1 – 25 of 54) sorted by relevance

123

/kernel/bpf/
Dtnum.c62 struct tnum tnum_add(struct tnum a, struct tnum b) in tnum_add() argument
66 sm = a.mask + b.mask; in tnum_add()
67 sv = a.value + b.value; in tnum_add()
70 mu = chi | a.mask | b.mask; in tnum_add()
74 struct tnum tnum_sub(struct tnum a, struct tnum b) in tnum_sub() argument
78 dv = a.value - b.value; in tnum_sub()
80 beta = dv - b.mask; in tnum_sub()
82 mu = chi | a.mask | b.mask; in tnum_sub()
86 struct tnum tnum_and(struct tnum a, struct tnum b) in tnum_and() argument
91 beta = b.value | b.mask; in tnum_and()
[all …]
Dhashtab.c160 struct bucket *b, u32 hash, in htab_lock_bucket() argument
183 raw_spin_lock_irqsave(&b->raw_lock, flags); in htab_lock_bucket()
185 spin_lock_irqsave(&b->lock, flags); in htab_lock_bucket()
192 struct bucket *b, u32 hash, in htab_unlock_bucket() argument
199 raw_spin_unlock_irqrestore(&b->raw_lock, flags); in htab_unlock_bucket()
201 spin_unlock_irqrestore(&b->lock, flags); in htab_unlock_bucket()
755 struct bucket *b; in htab_lru_map_delete_node() local
759 b = __select_bucket(htab, tgt_l->hash); in htab_lru_map_delete_node()
760 head = &b->head; in htab_lru_map_delete_node()
762 ret = htab_lock_bucket(htab, b, tgt_l->hash, &flags); in htab_lru_map_delete_node()
[all …]
Dbpf_local_storage.c180 struct bpf_local_storage_map_bucket *b; in bpf_selem_unlink_map() local
188 b = select_bucket(smap, selem); in bpf_selem_unlink_map()
189 raw_spin_lock_irqsave(&b->lock, flags); in bpf_selem_unlink_map()
192 raw_spin_unlock_irqrestore(&b->lock, flags); in bpf_selem_unlink_map()
198 struct bpf_local_storage_map_bucket *b = select_bucket(smap, selem); in bpf_selem_link_map() local
201 raw_spin_lock_irqsave(&b->lock, flags); in bpf_selem_link_map()
203 hlist_add_head_rcu(&selem->map_node, &b->list); in bpf_selem_link_map()
204 raw_spin_unlock_irqrestore(&b->lock, flags); in bpf_selem_link_map()
491 struct bpf_local_storage_map_bucket *b; in bpf_local_storage_map_free() local
510 b = &smap->buckets[i]; in bpf_local_storage_map_free()
[all …]
/kernel/time/
Dtimeconst.bc5 define gcd(a,b) {
7 while (b) {
8 t = b;
9 b = a % b;
16 define fmul(b,n,d) {
17 return (2^b*n+d-1)/d;
22 define fadj(b,n,d) {
25 v = 2^b*(d-1)/d;
30 which brings the mul value into the range 2^b-1 <= x < 2^b. Such
33 define fmuls(b,n,d) {
[all …]
Dtimer.c1956 int b; in timers_prepare_cpu() local
1958 for (b = 0; b < NR_BASES; b++) { in timers_prepare_cpu()
1959 base = per_cpu_ptr(&timer_bases[b], cpu); in timers_prepare_cpu()
1972 int b, i; in timers_dead_cpu() local
1976 for (b = 0; b < NR_BASES; b++) { in timers_dead_cpu()
1977 old_base = per_cpu_ptr(&timer_bases[b], cpu); in timers_dead_cpu()
1978 new_base = get_cpu_ptr(&timer_bases[b]); in timers_dead_cpu()
/kernel/trace/
Dtrace_events_filter_test.h12 TP_PROTO(int a, int b, int c, int d, int e, int f, int g, int h),
14 TP_ARGS(a, b, c, d, e, f, g, h),
18 __field(int, b)
29 __entry->b = b;
39 __entry->a, __entry->b, __entry->c, __entry->d,
Dtracing_map.c131 char *b = val_b; in tracing_map_cmp_string() local
133 return strcmp(a, b); in tracing_map_cmp_string()
144 u64 b = atomic64_read((atomic64_t *)val_b); in tracing_map_cmp_atomic64() local
146 return (a > b) ? 1 : ((a < b) ? -1 : 0); in tracing_map_cmp_atomic64()
153 type b = (type)(*(u64 *)val_b); \
155 return (a > b) ? 1 : ((a < b) ? -1 : 0); \
847 const struct tracing_map_sort_entry *a, *b; in cmp_entries_dup() local
851 b = *(const struct tracing_map_sort_entry **)B; in cmp_entries_dup()
853 if (memcmp(a->key, b->key, a->elt->map->key_size)) in cmp_entries_dup()
862 const struct tracing_map_sort_entry *a, *b; in cmp_entries_sum() local
[all …]
Dtrace_branch.c338 const struct ftrace_branch_data *b = p2; in annotated_branch_stat_cmp() local
343 percent_b = get_incorrect_percent(b); in annotated_branch_stat_cmp()
350 if (a->incorrect < b->incorrect) in annotated_branch_stat_cmp()
352 if (a->incorrect > b->incorrect) in annotated_branch_stat_cmp()
360 if (a->correct > b->correct) in annotated_branch_stat_cmp()
362 if (a->correct < b->correct) in annotated_branch_stat_cmp()
Dtrace_probe.c17 #define C(a, b) b argument
1177 int trace_probe_compare_arg_type(struct trace_probe *a, struct trace_probe *b) in trace_probe_compare_arg_type() argument
1182 if (a->nr_args < b->nr_args) in trace_probe_compare_arg_type()
1184 if (a->nr_args > b->nr_args) in trace_probe_compare_arg_type()
1185 return b->nr_args + 1; in trace_probe_compare_arg_type()
1188 if ((b->nr_args <= i) || in trace_probe_compare_arg_type()
1189 ((a->args[i].type != b->args[i].type) || in trace_probe_compare_arg_type()
1190 (a->args[i].count != b->args[i].count) || in trace_probe_compare_arg_type()
1191 strcmp(a->args[i].name, b->args[i].name))) in trace_probe_compare_arg_type()
Dtrace_hwlat.c159 #define time_sub(a, b) ((a) - (b)) argument
160 #define init_time(a, b) (a = b) argument
Dtrace_events_filter.c37 #define C(a, b) a argument
42 #define C(a, b) b argument
74 #define C(a, b) FILT_ERR_##a argument
79 #define C(a, b) b argument
2179 .rec = { .a = va, .b = vb, .c = vc, .d = vd, \
/kernel/rcu/
Dtree_plugin.h289 __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); in rcu_qs()
291 WRITE_ONCE(current->rcu_read_unlock_special.b.need_qs, false); in rcu_qs()
318 !t->rcu_read_unlock_special.b.blocked) { in rcu_note_context_switch()
323 t->rcu_read_unlock_special.b.blocked = true; in rcu_note_context_switch()
402 WRITE_ONCE(current->rcu_read_unlock_special.b.need_qs, true); in __rcu_read_lock()
485 if (special.b.need_qs) { in rcu_preempt_deferred_qs_irqrestore()
504 if (special.b.blocked) { in rcu_preempt_deferred_qs_irqrestore()
734 __this_cpu_read(rcu_data.cpu_no_qs.b.norm) && in rcu_flavor_sched_clock_irq()
735 !t->rcu_read_unlock_special.b.need_qs && in rcu_flavor_sched_clock_irq()
737 t->rcu_read_unlock_special.b.need_qs = true; in rcu_flavor_sched_clock_irq()
[all …]
Dtasks.h881 int nq = READ_ONCE(t->trc_reader_special.b.need_qs); in rcu_read_unlock_trace_special()
884 t->trc_reader_special.b.need_mb) in rcu_read_unlock_trace_special()
888 WRITE_ONCE(t->trc_reader_special.b.need_qs, false); in rcu_read_unlock_trace_special()
939 WARN_ON_ONCE(READ_ONCE(t->trc_reader_special.b.need_qs)); in trc_read_check_handler()
940 WRITE_ONCE(t->trc_reader_special.b.need_qs, true); in trc_read_check_handler()
991 WARN_ON_ONCE(READ_ONCE(t->trc_reader_special.b.need_qs)); in trc_inspect_reader()
992 WRITE_ONCE(t->trc_reader_special.b.need_qs, true); in trc_inspect_reader()
1079 WRITE_ONCE(t->trc_reader_special.b.need_qs, false); in rcu_tasks_trace_pertask()
1122 " N"[!!READ_ONCE(t->trc_reader_special.b.need_qs)], in show_stalled_task_trace()
1208 if (READ_ONCE(t->trc_reader_special.b.need_qs)) in rcu_tasks_trace_postgp()
[all …]
Dtree_exp.h774 t->rcu_read_unlock_special.b.exp_hint = true; in rcu_exp_handler()
820 __this_cpu_write(rcu_data.cpu_no_qs.b.exp, true); in rcu_exp_need_qs()
834 __this_cpu_read(rcu_data.cpu_no_qs.b.exp)) in rcu_exp_handler()
857 rdp->cpu_no_qs.b.exp) { in sync_sched_exp_online_cleanup()
/kernel/
Dauditfilter.c696 static int audit_compare_rule(struct audit_krule *a, struct audit_krule *b) in audit_compare_rule() argument
700 if (a->flags != b->flags || in audit_compare_rule()
701 a->pflags != b->pflags || in audit_compare_rule()
702 a->listnr != b->listnr || in audit_compare_rule()
703 a->action != b->action || in audit_compare_rule()
704 a->field_count != b->field_count) in audit_compare_rule()
708 if (a->fields[i].type != b->fields[i].type || in audit_compare_rule()
709 a->fields[i].op != b->fields[i].op) in audit_compare_rule()
723 if (strcmp(a->fields[i].lsm_str, b->fields[i].lsm_str)) in audit_compare_rule()
728 audit_watch_path(b->watch))) in audit_compare_rule()
[all …]
Dstatic_call_inline.c68 const struct static_call_site *b = _b; in static_call_site_cmp() local
70 const struct static_call_key *key_b = static_call_key(b); in static_call_site_cmp()
85 struct static_call_site *b = _b; in static_call_site_swap() local
88 a->addr = b->addr - delta; in static_call_site_swap()
89 a->key = b->key - delta; in static_call_site_swap()
91 b->addr = tmp.addr + delta; in static_call_site_swap()
92 b->key = tmp.key + delta; in static_call_site_swap()
Dgroups.c79 kgid_t b = *(kgid_t *)_b; in gid_cmp() local
81 return gid_gt(a, b) - gid_lt(a, b); in gid_cmp()
Dcred.c632 int cred_fscmp(const struct cred *a, const struct cred *b) in cred_fscmp() argument
637 if (a == b) in cred_fscmp()
639 if (uid_lt(a->fsuid, b->fsuid)) in cred_fscmp()
641 if (uid_gt(a->fsuid, b->fsuid)) in cred_fscmp()
644 if (gid_lt(a->fsgid, b->fsgid)) in cred_fscmp()
646 if (gid_gt(a->fsgid, b->fsgid)) in cred_fscmp()
650 gb = b->group_info; in cred_fscmp()
Djump_label.c35 static int jump_label_cmp(const void *a, const void *b) in jump_label_cmp() argument
38 const struct jump_entry *jeb = b; in jump_label_cmp()
63 static void jump_label_swap(void *a, void *b, int size) in jump_label_swap() argument
65 long delta = (unsigned long)a - (unsigned long)b; in jump_label_swap()
67 struct jump_entry *jeb = b; in jump_label_swap()
Dsys.c81 # define SET_UNALIGN_CTL(a, b) (-EINVAL) argument
84 # define GET_UNALIGN_CTL(a, b) (-EINVAL) argument
87 # define SET_FPEMU_CTL(a, b) (-EINVAL) argument
90 # define GET_FPEMU_CTL(a, b) (-EINVAL) argument
93 # define SET_FPEXC_CTL(a, b) (-EINVAL) argument
96 # define GET_FPEXC_CTL(a, b) (-EINVAL) argument
99 # define GET_ENDIAN(a, b) (-EINVAL) argument
102 # define SET_ENDIAN(a, b) (-EINVAL) argument
114 # define SET_FP_MODE(a,b) (-EINVAL) argument
129 # define PAC_RESET_KEYS(a, b) (-EINVAL) argument
[all …]
Dmodule.c129 mod_tree_less(struct latch_tree_node *a, struct latch_tree_node *b) in mod_tree_less() argument
131 return __mod_tree_val(a) < __mod_tree_val(b); in mod_tree_less()
780 static int already_uses(struct module *a, struct module *b) in already_uses() argument
784 list_for_each_entry(use, &b->source_list, source_list) { in already_uses()
786 pr_debug("%s uses %s!\n", a->name, b->name); in already_uses()
790 pr_debug("%s does not use %s!\n", a->name, b->name); in already_uses()
801 static int add_module_usage(struct module *a, struct module *b) in add_module_usage() argument
811 use->target = b; in add_module_usage()
812 list_add(&use->source_list, &b->source_list); in add_module_usage()
818 static int ref_module(struct module *a, struct module *b) in ref_module() argument
[all …]
/kernel/locking/
Dww_mutex.h228 __ww_ctx_less(struct ww_acquire_ctx *a, struct ww_acquire_ctx *b) in __ww_ctx_less() argument
238 int b_prio = b->task->prio; in __ww_ctx_less()
251 if (dl_time_before(b->task->dl.deadline, in __ww_ctx_less()
256 b->task->dl.deadline)) in __ww_ctx_less()
265 return (signed long)(a->stamp - b->stamp) > 0; in __ww_ctx_less()
/kernel/kcsan/
Ddebugfs.c98 const unsigned long b = *(const unsigned long *)lhs; in cmp_filterlist_addrs() local
100 return a < b ? -1 : a == b ? 0 : 1; in cmp_filterlist_addrs()
/kernel/dma/
Ddebug.c270 static bool exact_match(struct dma_debug_entry *a, struct dma_debug_entry *b) in exact_match() argument
272 return ((a->dev_addr == b->dev_addr) && in exact_match()
273 (a->dev == b->dev)) ? true : false; in exact_match()
277 struct dma_debug_entry *b) in containing_match() argument
279 if (a->dev != b->dev) in containing_match()
282 if ((b->dev_addr <= a->dev_addr) && in containing_match()
283 ((b->dev_addr + b->size) >= (a->dev_addr + a->size))) in containing_match()
/kernel/cgroup/
Dcpuset.c707 static int cpusets_overlap(struct cpuset *a, struct cpuset *b) in cpusets_overlap() argument
709 return cpumask_intersects(a->effective_cpus, b->effective_cpus); in cpusets_overlap()
889 struct cpuset *b = csa[j]; in generate_sched_domains() local
890 int bpn = b->pn; in generate_sched_domains()
892 if (apn != bpn && cpusets_overlap(a, b)) { in generate_sched_domains()
946 struct cpuset *b = csa[j]; in generate_sched_domains() local
948 if (apn == b->pn) { in generate_sched_domains()
949 cpumask_or(dp, dp, b->effective_cpus); in generate_sched_domains()
952 update_domain_attr_tree(dattr + nslot, b); in generate_sched_domains()
955 b->pn = -1; in generate_sched_domains()

123