/kernel/bpf/ |
D | tnum.c | 62 struct tnum tnum_add(struct tnum a, struct tnum b) in tnum_add() argument 66 sm = a.mask + b.mask; in tnum_add() 67 sv = a.value + b.value; in tnum_add() 70 mu = chi | a.mask | b.mask; in tnum_add() 74 struct tnum tnum_sub(struct tnum a, struct tnum b) in tnum_sub() argument 78 dv = a.value - b.value; in tnum_sub() 80 beta = dv - b.mask; in tnum_sub() 82 mu = chi | a.mask | b.mask; in tnum_sub() 86 struct tnum tnum_and(struct tnum a, struct tnum b) in tnum_and() argument 91 beta = b.value | b.mask; in tnum_and() [all …]
|
D | hashtab.c | 577 struct bucket *b; in htab_lru_map_delete_node() local 580 b = __select_bucket(htab, tgt_l->hash); in htab_lru_map_delete_node() 581 head = &b->head; in htab_lru_map_delete_node() 583 raw_spin_lock_irqsave(&b->lock, flags); in htab_lru_map_delete_node() 591 raw_spin_unlock_irqrestore(&b->lock, flags); in htab_lru_map_delete_node() 826 struct bucket *b; in htab_map_update_elem() local 840 b = __select_bucket(htab, hash); in htab_map_update_elem() 841 head = &b->head; in htab_map_update_elem() 866 raw_spin_lock_irqsave(&b->lock, flags); in htab_map_update_elem() 907 raw_spin_unlock_irqrestore(&b->lock, flags); in htab_map_update_elem() [all …]
|
/kernel/time/ |
D | timeconst.bc | 5 define gcd(a,b) { 7 while (b) { 8 t = b; 9 b = a % b; 16 define fmul(b,n,d) { 17 return (2^b*n+d-1)/d; 22 define fadj(b,n,d) { 25 v = 2^b*(d-1)/d; 30 which brings the mul value into the range 2^b-1 <= x < 2^b. Such 33 define fmuls(b,n,d) { [all …]
|
D | timeconv.c | 44 static long math_div(long a, long b) in math_div() argument 46 return a / b - (a % b < 0); in math_div()
|
D | timer.c | 1961 int b; in timers_prepare_cpu() local 1963 for (b = 0; b < NR_BASES; b++) { in timers_prepare_cpu() 1964 base = per_cpu_ptr(&timer_bases[b], cpu); in timers_prepare_cpu() 1977 int b, i; in timers_dead_cpu() local 1981 for (b = 0; b < NR_BASES; b++) { in timers_dead_cpu() 1982 old_base = per_cpu_ptr(&timer_bases[b], cpu); in timers_dead_cpu() 1983 new_base = get_cpu_ptr(&timer_bases[b]); in timers_dead_cpu()
|
/kernel/trace/ |
D | trace_events_filter_test.h | 12 TP_PROTO(int a, int b, int c, int d, int e, int f, int g, int h), 14 TP_ARGS(a, b, c, d, e, f, g, h), 18 __field(int, b) 29 __entry->b = b; 39 __entry->a, __entry->b, __entry->c, __entry->d,
|
D | tracing_map.c | 130 char *b = val_b; in tracing_map_cmp_string() local 132 return strcmp(a, b); in tracing_map_cmp_string() 143 u64 b = atomic64_read((atomic64_t *)val_b); in tracing_map_cmp_atomic64() local 145 return (a > b) ? 1 : ((a < b) ? -1 : 0); in tracing_map_cmp_atomic64() 152 type b = (type)(*(u64 *)val_b); \ 154 return (a > b) ? 1 : ((a < b) ? -1 : 0); \ 838 const struct tracing_map_sort_entry **b) in cmp_entries_dup() argument 842 if (memcmp((*a)->key, (*b)->key, (*a)->elt->map->key_size)) in cmp_entries_dup() 849 const struct tracing_map_sort_entry **b) in cmp_entries_sum() argument 859 elt_b = (*b)->elt; in cmp_entries_sum() [all …]
|
D | trace_branch.c | 338 struct ftrace_branch_data *b = p2; in annotated_branch_stat_cmp() local 343 percent_b = get_incorrect_percent(b); in annotated_branch_stat_cmp() 350 if (a->incorrect < b->incorrect) in annotated_branch_stat_cmp() 352 if (a->incorrect > b->incorrect) in annotated_branch_stat_cmp() 360 if (a->correct > b->correct) in annotated_branch_stat_cmp() 362 if (a->correct < b->correct) in annotated_branch_stat_cmp()
|
D | trace_probe.c | 17 #define C(a, b) b argument 1094 int trace_probe_compare_arg_type(struct trace_probe *a, struct trace_probe *b) in trace_probe_compare_arg_type() argument 1099 if (a->nr_args < b->nr_args) in trace_probe_compare_arg_type() 1101 if (a->nr_args > b->nr_args) in trace_probe_compare_arg_type() 1102 return b->nr_args + 1; in trace_probe_compare_arg_type() 1105 if ((b->nr_args <= i) || in trace_probe_compare_arg_type() 1106 ((a->args[i].type != b->args[i].type) || in trace_probe_compare_arg_type() 1107 (a->args[i].count != b->args[i].count) || in trace_probe_compare_arg_type() 1108 strcmp(a->args[i].name, b->args[i].name))) in trace_probe_compare_arg_type()
|
D | trace_hwlat.c | 136 #define time_sub(a, b) ((a) - (b)) argument 137 #define init_time(a, b) (a = b) argument
|
D | trace_events_filter.c | 36 #define C(a, b) a argument 41 #define C(a, b) b argument 73 #define C(a, b) FILT_ERR_##a argument 78 #define C(a, b) b argument 2107 .rec = { .a = va, .b = vb, .c = vc, .d = vd, \
|
D | trace_probe.h | 336 int trace_probe_compare_arg_type(struct trace_probe *a, struct trace_probe *b); 444 #define C(a, b) TP_ERR_##a argument
|
D | trace.h | 436 #define TRACER_OPT(s, b) .name = #s, .bit = b argument 1265 #define C(a, b) TRACE_ITER_##a##_BIT argument 1278 #define C(a, b) TRACE_ITER_##a = (1 << TRACE_ITER_##a##_BIT) argument
|
/kernel/ |
D | auditfilter.c | 689 static int audit_compare_rule(struct audit_krule *a, struct audit_krule *b) in audit_compare_rule() argument 693 if (a->flags != b->flags || in audit_compare_rule() 694 a->pflags != b->pflags || in audit_compare_rule() 695 a->listnr != b->listnr || in audit_compare_rule() 696 a->action != b->action || in audit_compare_rule() 697 a->field_count != b->field_count) in audit_compare_rule() 701 if (a->fields[i].type != b->fields[i].type || in audit_compare_rule() 702 a->fields[i].op != b->fields[i].op) in audit_compare_rule() 716 if (strcmp(a->fields[i].lsm_str, b->fields[i].lsm_str)) in audit_compare_rule() 721 audit_watch_path(b->watch))) in audit_compare_rule() [all …]
|
D | cred.c | 613 int cred_fscmp(const struct cred *a, const struct cred *b) in cred_fscmp() argument 618 if (a == b) in cred_fscmp() 620 if (uid_lt(a->fsuid, b->fsuid)) in cred_fscmp() 622 if (uid_gt(a->fsuid, b->fsuid)) in cred_fscmp() 625 if (gid_lt(a->fsgid, b->fsgid)) in cred_fscmp() 627 if (gid_gt(a->fsgid, b->fsgid)) in cred_fscmp() 631 gb = b->group_info; in cred_fscmp()
|
D | groups.c | 84 kgid_t b = *(kgid_t *)_b; in gid_cmp() local 86 return gid_gt(a, b) - gid_lt(a, b); in gid_cmp()
|
D | jump_label.c | 35 static int jump_label_cmp(const void *a, const void *b) in jump_label_cmp() argument 38 const struct jump_entry *jeb = b; in jump_label_cmp() 63 static void jump_label_swap(void *a, void *b, int size) in jump_label_swap() argument 65 long delta = (unsigned long)a - (unsigned long)b; in jump_label_swap() 67 struct jump_entry *jeb = b; in jump_label_swap()
|
D | module.c | 120 mod_tree_less(struct latch_tree_node *a, struct latch_tree_node *b) in mod_tree_less() argument 122 return __mod_tree_val(a) < __mod_tree_val(b); in mod_tree_less() 828 static int already_uses(struct module *a, struct module *b) in already_uses() argument 832 list_for_each_entry(use, &b->source_list, source_list) { in already_uses() 834 pr_debug("%s uses %s!\n", a->name, b->name); in already_uses() 838 pr_debug("%s does not use %s!\n", a->name, b->name); in already_uses() 849 static int add_module_usage(struct module *a, struct module *b) in add_module_usage() argument 859 use->target = b; in add_module_usage() 860 list_add(&use->source_list, &b->source_list); in add_module_usage() 866 int ref_module(struct module *a, struct module *b) in ref_module() argument [all …]
|
D | params.c | 84 bool parameqn(const char *a, const char *b, size_t n) in parameqn() argument 89 if (dash2underscore(a[i]) != dash2underscore(b[i])) in parameqn() 95 bool parameq(const char *a, const char *b) in parameq() argument 97 return parameqn(a, b, strlen(a)+1); in parameq()
|
D | sys.c | 79 # define SET_UNALIGN_CTL(a, b) (-EINVAL) argument 82 # define GET_UNALIGN_CTL(a, b) (-EINVAL) argument 85 # define SET_FPEMU_CTL(a, b) (-EINVAL) argument 88 # define GET_FPEMU_CTL(a, b) (-EINVAL) argument 91 # define SET_FPEXC_CTL(a, b) (-EINVAL) argument 94 # define GET_FPEXC_CTL(a, b) (-EINVAL) argument 97 # define GET_ENDIAN(a, b) (-EINVAL) argument 100 # define SET_ENDIAN(a, b) (-EINVAL) argument 112 # define SET_FP_MODE(a,b) (-EINVAL) argument 121 # define PAC_RESET_KEYS(a, b) (-EINVAL) argument
|
D | user_namespace.c | 788 static int cmp_extents_forward(const void *a, const void *b) in cmp_extents_forward() argument 791 const struct uid_gid_extent *e2 = b; in cmp_extents_forward() 803 static int cmp_extents_reverse(const void *a, const void *b) in cmp_extents_reverse() argument 806 const struct uid_gid_extent *e2 = b; in cmp_extents_reverse()
|
/kernel/rcu/ |
D | tree_plugin.h | 266 __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); in rcu_qs() 268 WRITE_ONCE(current->rcu_read_unlock_special.b.need_qs, false); in rcu_qs() 295 !t->rcu_read_unlock_special.b.blocked) { in rcu_note_context_switch() 300 t->rcu_read_unlock_special.b.blocked = true; in rcu_note_context_switch() 447 t->rcu_read_unlock_special.b.deferred_qs = false; in rcu_preempt_deferred_qs_irqrestore() 448 if (special.b.need_qs) { in rcu_preempt_deferred_qs_irqrestore() 450 t->rcu_read_unlock_special.b.need_qs = false; in rcu_preempt_deferred_qs_irqrestore() 472 if (special.b.blocked) { in rcu_preempt_deferred_qs_irqrestore() 473 t->rcu_read_unlock_special.b.blocked = false; in rcu_preempt_deferred_qs_irqrestore() 613 t->rcu_read_unlock_special.b.exp_hint = false; in rcu_read_unlock_special() [all …]
|
D | tree_exp.h | 641 t->rcu_read_unlock_special.b.exp_hint = true; in rcu_exp_handler() 704 __this_cpu_write(rcu_data.cpu_no_qs.b.exp, true); in rcu_exp_need_qs() 720 __this_cpu_read(rcu_data.cpu_no_qs.b.exp)) in rcu_exp_handler() 743 __this_cpu_read(rcu_data.cpu_no_qs.b.exp)) { in sync_sched_exp_online_cleanup()
|
/kernel/dma/ |
D | debug.c | 266 static bool exact_match(struct dma_debug_entry *a, struct dma_debug_entry *b) in exact_match() argument 268 return ((a->dev_addr == b->dev_addr) && in exact_match() 269 (a->dev == b->dev)) ? true : false; in exact_match() 273 struct dma_debug_entry *b) in containing_match() argument 275 if (a->dev != b->dev) in containing_match() 278 if ((b->dev_addr <= a->dev_addr) && in containing_match() 279 ((b->dev_addr + b->size) >= (a->dev_addr + a->size))) in containing_match()
|
/kernel/cgroup/ |
D | cpuset.c | 640 static int cpusets_overlap(struct cpuset *a, struct cpuset *b) in cpusets_overlap() argument 642 return cpumask_intersects(a->effective_cpus, b->effective_cpus); in cpusets_overlap() 822 struct cpuset *b = csa[j]; in generate_sched_domains() local 823 int bpn = b->pn; in generate_sched_domains() 825 if (apn != bpn && cpusets_overlap(a, b)) { in generate_sched_domains() 879 struct cpuset *b = csa[j]; in generate_sched_domains() local 881 if (apn == b->pn) { in generate_sched_domains() 882 cpumask_or(dp, dp, b->effective_cpus); in generate_sched_domains() 885 update_domain_attr_tree(dattr + nslot, b); in generate_sched_domains() 888 b->pn = -1; in generate_sched_domains()
|