/kernel/locking/ |
D | qspinlock.c | 196 atomic_andnot(_Q_PENDING_VAL, &lock->val); in clear_pending() 207 atomic_add(-_Q_PENDING_VAL + _Q_LOCKED_VAL, &lock->val); in clear_pending_set_locked() 222 u32 old, new, val = atomic_read(&lock->val); in xchg_tail() local 225 new = (val & _Q_LOCKED_PENDING_MASK) | tail; in xchg_tail() 231 old = atomic_cmpxchg_relaxed(&lock->val, val, new); in xchg_tail() 232 if (old == val) in xchg_tail() 235 val = old; in xchg_tail() 251 return atomic_fetch_or_acquire(_Q_PENDING_VAL, &lock->val); in queued_fetch_set_pending_acquire() 315 void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in queued_spin_lock_slowpath() argument 335 if (val == _Q_PENDING_VAL) { in queued_spin_lock_slowpath() [all …]
|
D | qspinlock_paravirt.h | 88 int val = atomic_read(&lock->val); in pv_hybrid_queued_unfair_trylock() local 90 if (!(val & _Q_LOCKED_PENDING_MASK) && in pv_hybrid_queued_unfair_trylock() 95 if (!(val & _Q_TAIL_MASK) || (val & _Q_PENDING_MASK)) in pv_hybrid_queued_unfair_trylock() 128 atomic_or(_Q_PENDING_VAL, &lock->val); in set_pending() 133 int val = atomic_read(&lock->val); in trylock_clear_pending() local 138 if (val & _Q_LOCKED_MASK) in trylock_clear_pending() 144 old = val; in trylock_clear_pending() 145 new = (val & ~_Q_PENDING_MASK) | _Q_LOCKED_VAL; in trylock_clear_pending() 146 val = atomic_cmpxchg_acquire(&lock->val, old, new); in trylock_clear_pending() 148 if (val == old) in trylock_clear_pending() [all …]
|
/kernel/trace/ |
D | trace_probe_tmpl.h | 7 fetch_store_raw(unsigned long val, struct fetch_insn *code, void *buf) in fetch_store_raw() argument 11 *(u8 *)buf = (u8)val; in fetch_store_raw() 14 *(u16 *)buf = (u16)val; in fetch_store_raw() 17 *(u32 *)buf = (u32)val; in fetch_store_raw() 21 *(u64 *)buf = (u64)val; in fetch_store_raw() 24 *(unsigned long *)buf = val; in fetch_store_raw() 72 process_fetch_insn_bottom(struct fetch_insn *code, unsigned long val, in process_fetch_insn_bottom() argument 78 unsigned long lval = val; in process_fetch_insn_bottom() 84 lval = val; in process_fetch_insn_bottom() 85 ret = probe_mem_read(&val, (void *)val + code->offset, in process_fetch_insn_bottom() [all …]
|
D | trace_events_inject.c | 43 u64 val; in parse_field() local 95 ret = kstrtoll(num, 0, &val); in parse_field() 97 ret = kstrtoull(num, 0, &val); in parse_field() 102 *pv = val; in parse_field() 198 u64 val = 0; in parse_entry() local 210 while ((len = parse_field(str, call, &field, &val)) > 0) { in parse_entry() 215 char *addr = (char *)(unsigned long) val; in parse_entry() 244 u8 tmp = (u8) val; in parse_entry() 250 u16 tmp = (u16) val; in parse_entry() 256 u32 tmp = (u32) val; in parse_entry() [all …]
|
D | trace_hwlat.c | 425 u64 val; in hwlat_read() local 434 val = *entry; in hwlat_read() 436 len = snprintf(buf, sizeof(buf), "%llu\n", val); in hwlat_read() 460 u64 val; in hwlat_width_write() local 463 err = kstrtoull_from_user(ubuf, cnt, 10, &val); in hwlat_width_write() 468 if (val < hwlat_data.sample_window) in hwlat_width_write() 469 hwlat_data.sample_width = val; in hwlat_width_write() 499 u64 val; in hwlat_window_write() local 502 err = kstrtoull_from_user(ubuf, cnt, 10, &val); in hwlat_window_write() 507 if (hwlat_data.sample_width < val) in hwlat_window_write() [all …]
|
D | trace_events_synth.c | 290 int size, u64 val, char *space) in print_synth_event_num_val() argument 294 trace_seq_printf(s, print_fmt, name, (u8)val, space); in print_synth_event_num_val() 298 trace_seq_printf(s, print_fmt, name, (u16)val, space); in print_synth_event_num_val() 302 trace_seq_printf(s, print_fmt, name, (u32)val, space); in print_synth_event_num_val() 306 trace_seq_printf(s, print_fmt, name, val, space); in print_synth_event_num_val() 484 u64 val = var_ref_vals[val_idx]; in trace_event_raw_event_synth() local 488 *(u8 *)&entry->fields[n_u64] = (u8)val; in trace_event_raw_event_synth() 492 *(u16 *)&entry->fields[n_u64] = (u16)val; in trace_event_raw_event_synth() 496 *(u32 *)&entry->fields[n_u64] = (u32)val; in trace_event_raw_event_synth() 500 entry->fields[n_u64] = val; in trace_event_raw_event_synth() [all …]
|
D | trace_branch.c | 31 probe_likely_condition(struct ftrace_likely_data *f, int val, int expect) in probe_likely_condition() argument 83 entry->correct = val == expect; in probe_likely_condition() 94 void trace_likely_condition(struct ftrace_likely_data *f, int val, int expect) in trace_likely_condition() argument 99 probe_likely_condition(f, val, expect); in trace_likely_condition() 200 void trace_likely_condition(struct ftrace_likely_data *f, int val, int expect) in trace_likely_condition() argument 205 void ftrace_likely_update(struct ftrace_likely_data *f, int val, in ftrace_likely_update() argument 213 val = expect; in ftrace_likely_update() 221 trace_likely_condition(f, val, expect); in ftrace_likely_update() 224 if (val == expect) in ftrace_likely_update()
|
D | trace_readwrite.c | 18 void __log_write_mmio(u64 val, u8 width, volatile void __iomem *addr) in __log_write_mmio() argument 20 trace_rwmmio_write(CALLER_ADDR0, val, width, addr); in __log_write_mmio() 32 void __log_post_read_mmio(u64 val, u8 width, const volatile void __iomem *addr) in __log_post_read_mmio() argument 34 trace_rwmmio_post_read(CALLER_ADDR0, val, width, addr); in __log_post_read_mmio()
|
D | ring_buffer.c | 611 static inline int rb_time_cnt(unsigned long val) in rb_time_cnt() argument 613 return (val >> RB_TIME_SHIFT) & 3; in rb_time_cnt() 618 u64 val; in rb_time_val() local 620 val = top & RB_TIME_VAL_MASK; in rb_time_val() 621 val <<= RB_TIME_SHIFT; in rb_time_val() 622 val |= bottom & RB_TIME_VAL_MASK; in rb_time_val() 624 return val; in rb_time_val() 660 static inline unsigned long rb_time_val_cnt(unsigned long val, unsigned long cnt) in rb_time_val_cnt() argument 662 return (val & RB_TIME_VAL_MASK) | ((cnt & 3) << RB_TIME_SHIFT); in rb_time_val_cnt() 665 static inline void rb_time_split(u64 val, unsigned long *top, unsigned long *bottom) in rb_time_split() argument [all …]
|
D | trace_events_filter.c | 597 type val = (type)pred->val; \ 598 return *addr < val; \ 603 type val = (type)pred->val; \ 604 return *addr <= val; \ 609 type val = (type)pred->val; \ 610 return *addr > val; \ 615 type val = (type)pred->val; \ 616 return *addr >= val; \ 621 type val = (type)pred->val; \ 622 return !!(*addr & val); \ [all …]
|
D | tracing_map.c | 521 struct tracing_map_elt *val; in __tracing_map_insert() local 534 val = READ_ONCE(entry->val); in __tracing_map_insert() 535 if (val && in __tracing_map_insert() 536 keys_match(key, val->key, map->key_size)) { in __tracing_map_insert() 539 return val; in __tracing_map_insert() 540 } else if (unlikely(!val)) { in __tracing_map_insert() 582 WRITE_ONCE(entry->val, elt); in __tracing_map_insert() 585 return entry->val; in __tracing_map_insert() 1092 if (!entry->key || !entry->val) in tracing_map_sort_entries() 1095 entries[n_entries] = create_sort_entry(entry->val->key, in tracing_map_sort_entries() [all …]
|
/kernel/ |
D | params.c | 41 char val[]; member 58 return p->val; in kmalloc_parameter() 68 if (p->val == param) { in maybe_kfree_parameter() 116 char *val, in parse_one() argument 123 int (*handle_unknown)(char *param, char *val, in parse_one() 136 if (!val && in parse_one() 143 err = params[i].ops->set(val, ¶ms[i]); in parse_one() 152 pr_debug("doing %s: %s='%s'\n", doing, param, val); in parse_one() 153 return handle_unknown(param, val, doing, arg); in parse_one() 168 int (*unknown)(char *param, char *val, in parse_args() argument [all …]
|
D | notifier.c | 65 unsigned long val, void *v, in notifier_call_chain() argument 83 ret = nb->notifier_call(nb, val, v); in notifier_call_chain() 212 unsigned long val, void *v) in atomic_notifier_call_chain() argument 217 ret = notifier_call_chain(&nh->head, val, v, -1, NULL); in atomic_notifier_call_chain() 326 unsigned long val, void *v) in blocking_notifier_call_chain() argument 337 ret = notifier_call_chain(&nh->head, val, v, -1, NULL); in blocking_notifier_call_chain() 408 unsigned long val, void *v) in raw_notifier_call_chain() argument 410 return notifier_call_chain(&nh->head, val, v, -1, NULL); in raw_notifier_call_chain() 498 unsigned long val, void *v) in srcu_notifier_call_chain() argument 504 ret = notifier_call_chain(&nh->head, val, v, -1, NULL); in srcu_notifier_call_chain() [all …]
|
D | sysctl.c | 208 int val, ret; in bpf_stats_handler() local 210 .data = &val, in bpf_stats_handler() 211 .maxlen = sizeof(val), in bpf_stats_handler() 221 val = saved_val; in bpf_stats_handler() 223 if (write && !ret && val != saved_val) { in bpf_stats_handler() 224 if (val) in bpf_stats_handler() 228 saved_val = val; in bpf_stats_handler() 464 unsigned long *val, bool *neg, in proc_get_long() argument 488 if (strtoul_lenient(p, &p, 0, val)) in proc_get_long() 522 static void proc_put_long(void **buf, size_t *size, unsigned long val, bool neg) in proc_put_long() argument [all …]
|
D | fail_function.c | 107 static int fei_retval_set(void *data, u64 val) in fei_retval_set() argument 110 unsigned long retv = (unsigned long)val; in fei_retval_set() 126 val) != retv) in fei_retval_set() 130 attr->retval = val; in fei_retval_set() 137 static int fei_retval_get(void *data, u64 *val) in fei_retval_get() argument 147 *val = attr->retval; in fei_retval_get()
|
D | auditfilter.c | 221 switch(audit_classify_arch(arch->val)) { in audit_match_signal() 414 if ((f->val != 0) && (f->val != 1)) in audit_field_valid() 418 if (f->val & ~15) in audit_field_valid() 422 if (f->val & ~S_IFMT) in audit_field_valid() 426 if (f->val > AUDIT_MAX_FIELD_COMPARE) in audit_field_valid() 430 if (f->val >= AF_MAX) in audit_field_valid() 503 f->val = f_val; in audit_data_to_entry() 560 f->val = f_val; in audit_data_to_entry() 594 f->val = f_val; in audit_data_to_entry() 679 if (krule->pflags & AUDIT_LOGINUID_LEGACY && !f->val) { in audit_krule_to_data() [all …]
|
D | auditsc.c | 205 static int audit_match_filetype(struct audit_context *ctx, int val) in audit_match_filetype() argument 208 umode_t mode = (umode_t)val; in audit_match_filetype() 389 switch (f->val) { in audit_field_compare() 490 result = audit_comparator(pid, f->op, f->val); in audit_filter_rules() 496 result = audit_comparator(ctx->ppid, f->op, f->val); in audit_filter_rules() 544 result = audit_comparator(sessionid, f->op, f->val); in audit_filter_rules() 547 result = audit_comparator(tsk->personality, f->op, f->val); in audit_filter_rules() 551 result = audit_comparator(ctx->arch, f->op, f->val); in audit_filter_rules() 556 result = audit_comparator(ctx->return_code, f->op, f->val); in audit_filter_rules() 560 if (f->val) in audit_filter_rules() [all …]
|
/kernel/power/ |
D | main.c | 92 int pm_notifier_call_chain(unsigned long val) in pm_notifier_call_chain() argument 94 return blocking_notifier_call_chain(&pm_chain_head, val, NULL); in pm_notifier_call_chain() 109 unsigned long val; in pm_async_store() local 111 if (kstrtoul(buf, 10, &val)) in pm_async_store() 114 if (val > 1) in pm_async_store() 117 pm_async_enabled = val; in pm_async_store() 212 unsigned long val; in sync_on_suspend_store() local 214 if (kstrtoul(buf, 10, &val)) in sync_on_suspend_store() 217 if (val > 1) in sync_on_suspend_store() 220 sync_on_suspend_enabled = !!val; in sync_on_suspend_store() [all …]
|
/kernel/sched/ |
D | cpuacct.c | 192 u64 val) in cpuusage_write() argument 200 if (val) in cpuusage_write() 263 s64 val[CPUACCT_STAT_NSTATS]; in cpuacct_stats_show() local 267 memset(val, 0, sizeof(val)); in cpuacct_stats_show() 271 val[CPUACCT_STAT_USER] += cpustat[CPUTIME_USER]; in cpuacct_stats_show() 272 val[CPUACCT_STAT_USER] += cpustat[CPUTIME_NICE]; in cpuacct_stats_show() 273 val[CPUACCT_STAT_SYSTEM] += cpustat[CPUTIME_SYSTEM]; in cpuacct_stats_show() 274 val[CPUACCT_STAT_SYSTEM] += cpustat[CPUTIME_IRQ]; in cpuacct_stats_show() 275 val[CPUACCT_STAT_SYSTEM] += cpustat[CPUTIME_SOFTIRQ]; in cpuacct_stats_show() 281 (long long)nsec_to_clock_t(val[stat])); in cpuacct_stats_show() [all …]
|
D | stats.h | 38 #define __schedstat_set(var, val) do { var = (val); } while (0) argument 39 #define schedstat_set(var, val) do { if (schedstat_enabled()) { var = (val); } } while (0) argument 52 # define __schedstat_set(var, val) do { } while (0) argument 53 # define schedstat_set(var, val) do { } while (0) argument
|
D | clock.c | 302 u64 *ptr, old_val, val; in sched_clock_remote() local 344 val = this_clock; in sched_clock_remote() 351 val = remote_clock; in sched_clock_remote() 354 if (cmpxchg64(ptr, old_val, val) != old_val) in sched_clock_remote() 357 return val; in sched_clock_remote()
|
/kernel/debug/kdb/ |
D | kdb_main.c | 370 unsigned long val; in kdbgetintenv() local 373 diag = kdbgetulenv(match, &val); in kdbgetintenv() 375 *value = (int) val; in kdbgetintenv() 392 unsigned long val; in kdbgetularg() local 394 val = simple_strtoul(arg, &endp, 0); in kdbgetularg() 401 val = simple_strtoul(arg, &endp, 16); in kdbgetularg() 406 *value = val; in kdbgetularg() 414 u64 val; in kdbgetu64arg() local 416 val = simple_strtoull(arg, &endp, 0); in kdbgetu64arg() 420 val = simple_strtoull(arg, &endp, 16); in kdbgetu64arg() [all …]
|
/kernel/bpf/ |
D | devmap.c | 70 struct bpf_devmap_val val; member 532 return obj ? &obj->val : NULL; in dev_map_lookup_elem() 539 return obj ? &obj->val : NULL; in dev_map_hash_lookup_elem() 599 struct bpf_devmap_val *val, in __dev_map_alloc_node() argument 610 dev->dev = dev_get_by_index(net, val->ifindex); in __dev_map_alloc_node() 614 if (val->bpf_prog.fd > 0) { in __dev_map_alloc_node() 615 prog = bpf_prog_get_type_dev(val->bpf_prog.fd, in __dev_map_alloc_node() 627 dev->val.bpf_prog.id = prog->aux->id; in __dev_map_alloc_node() 630 dev->val.bpf_prog.id = 0; in __dev_map_alloc_node() 632 dev->val.ifindex = val->ifindex; in __dev_map_alloc_node() [all …]
|
/kernel/time/ |
D | itimer.c | 50 u64 val, interval; in get_cpu_itimer() local 55 val = it->expires; in get_cpu_itimer() 57 if (val) { in get_cpu_itimer() 63 if (val < t) in get_cpu_itimer() 65 val = TICK_NSEC; in get_cpu_itimer() 67 val -= t; in get_cpu_itimer() 72 value->it_value = ns_to_timespec64(val); in get_cpu_itimer()
|
/kernel/cgroup/ |
D | cpuset.c | 80 int val; /* most recent output value */ member 1938 static int update_relax_domain_level(struct cpuset *cs, s64 val) in update_relax_domain_level() argument 1941 if (val < -1 || val >= sched_domain_level_max) in update_relax_domain_level() 1945 if (val != cs->relax_domain_level) { in update_relax_domain_level() 1946 cs->relax_domain_level = val; in update_relax_domain_level() 2161 fmp->val = 0; in fmeter_init() 2180 fmp->val = (FM_COEF * fmp->val) / FM_SCALE; in fmeter_update() 2183 fmp->val += ((FM_SCALE - FM_COEF) * fmp->cnt) / FM_SCALE; in fmeter_update() 2199 int val; in fmeter_getrate() local 2203 val = fmp->val; in fmeter_getrate() [all …]
|