/kernel/ |
D | delayacct.c | 134 s64 tmp; in delayacct_add_tsk() local 137 tmp = (s64)d->cpu_run_real_total; in delayacct_add_tsk() 138 tmp += utime + stime; in delayacct_add_tsk() 139 d->cpu_run_real_total = (tmp < (s64)d->cpu_run_real_total) ? 0 : tmp; in delayacct_add_tsk() 142 tmp = (s64)d->cpu_scaled_run_real_total; in delayacct_add_tsk() 143 tmp += utimescaled + stimescaled; in delayacct_add_tsk() 145 (tmp < (s64)d->cpu_scaled_run_real_total) ? 0 : tmp; in delayacct_add_tsk() 157 tmp = (s64)d->cpu_delay_total + t2; in delayacct_add_tsk() 158 d->cpu_delay_total = (tmp < (s64)d->cpu_delay_total) ? 0 : tmp; in delayacct_add_tsk() 160 tmp = (s64)d->cpu_run_virtual_total + t3; in delayacct_add_tsk() [all …]
|
D | resource.c | 175 struct resource *tmp, **p; in __request_resource() local 185 tmp = *p; in __request_resource() 186 if (!tmp || tmp->start > end) { in __request_resource() 187 new->sibling = tmp; in __request_resource() 192 p = &tmp->sibling; in __request_resource() 193 if (tmp->end < start) in __request_resource() 195 return tmp; in __request_resource() 201 struct resource *tmp, **p, *chd; in __release_resource() local 205 tmp = *p; in __release_resource() 206 if (!tmp) in __release_resource() [all …]
|
D | sysctl.c | 345 char *p, tmp[TMPBUFLEN]; in proc_get_long() local 354 memcpy(tmp, *buf, len); in proc_get_long() 356 tmp[len] = 0; in proc_get_long() 357 p = tmp; in proc_get_long() 369 len = p - tmp; in proc_get_long() 403 char tmp[TMPBUFLEN], *p = tmp; in proc_put_long() local 406 len = strlen(tmp); in proc_put_long() 409 memcpy(*buf, tmp, len); in proc_put_long() 708 struct ctl_table tmp; in proc_dobool() local 716 tmp = *table; in proc_dobool() [all …]
|
D | pid.c | 168 struct pid_namespace *tmp; in alloc_pid() local 187 tmp = ns; in alloc_pid() 203 if (tid != 1 && !tmp->child_reaper) in alloc_pid() 206 if (!checkpoint_restore_ns_capable(tmp->user_ns)) in alloc_pid() 215 nr = idr_alloc(&tmp->idr, NULL, tid, in alloc_pid() 229 if (idr_get_cursor(&tmp->idr) > RESERVED_PIDS) in alloc_pid() 236 nr = idr_alloc_cyclic(&tmp->idr, NULL, pid_min, in alloc_pid() 248 pid->numbers[i].ns = tmp; in alloc_pid() 249 tmp = tmp->parent; in alloc_pid() 345 int tmp; in __change_pid() local [all …]
|
D | sys.c | 1028 struct tms tmp; in SYSCALL_DEFINE1() local 1030 do_sys_times(&tmp); in SYSCALL_DEFINE1() 1031 if (copy_to_user(tbuf, &tmp, sizeof(struct tms))) in SYSCALL_DEFINE1() 1048 struct compat_tms tmp; in COMPAT_SYSCALL_DEFINE1() local 1052 tmp.tms_utime = clock_t_to_compat_clock_t(tms.tms_utime); in COMPAT_SYSCALL_DEFINE1() 1053 tmp.tms_stime = clock_t_to_compat_clock_t(tms.tms_stime); in COMPAT_SYSCALL_DEFINE1() 1054 tmp.tms_cutime = clock_t_to_compat_clock_t(tms.tms_cutime); in COMPAT_SYSCALL_DEFINE1() 1055 tmp.tms_cstime = clock_t_to_compat_clock_t(tms.tms_cstime); in COMPAT_SYSCALL_DEFINE1() 1056 if (copy_to_user(tbuf, &tmp, sizeof(tmp))) in COMPAT_SYSCALL_DEFINE1() 1310 struct new_utsname tmp; in SYSCALL_DEFINE1() local [all …]
|
D | crash_core.c | 55 char *cur = cmdline, *tmp; in parse_crashkernel_mem() local 71 start = memparse(cur, &tmp); in parse_crashkernel_mem() 72 if (cur == tmp) { in parse_crashkernel_mem() 76 cur = tmp; in parse_crashkernel_mem() 85 end = memparse(cur, &tmp); in parse_crashkernel_mem() 86 if (cur == tmp) { in parse_crashkernel_mem() 90 cur = tmp; in parse_crashkernel_mem() 103 size = memparse(cur, &tmp); in parse_crashkernel_mem() 104 if (cur == tmp) { in parse_crashkernel_mem() 108 cur = tmp; in parse_crashkernel_mem() [all …]
|
/kernel/power/ |
D | console.c | 48 struct pm_vt_switch *entry, *tmp; in pm_vt_switch_required() local 51 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_required() 52 if (tmp->dev == dev) { in pm_vt_switch_required() 54 tmp->required = required; in pm_vt_switch_required() 80 struct pm_vt_switch *tmp; in pm_vt_switch_unregister() local 83 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_unregister() 84 if (tmp->dev == dev) { in pm_vt_switch_unregister() 85 list_del(&tmp->head); in pm_vt_switch_unregister() 86 kfree(tmp); in pm_vt_switch_unregister()
|
D | swap.c | 957 struct swap_map_page_list *tmp; in release_swap_reader() local 962 tmp = handle->maps; in release_swap_reader() 964 kfree(tmp); in release_swap_reader() 973 struct swap_map_page_list *tmp, *last; in get_swap_reader() local 985 tmp = kzalloc(sizeof(*handle->maps), GFP_KERNEL); in get_swap_reader() 986 if (!tmp) { in get_swap_reader() 991 handle->maps = tmp; in get_swap_reader() 993 last->next = tmp; in get_swap_reader() 994 last = tmp; in get_swap_reader() 996 tmp->map = (struct swap_map_page *) in get_swap_reader() [all …]
|
/kernel/cgroup/ |
D | cpuset.c | 604 static inline int alloc_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in alloc_cpumasks() argument 613 pmask1 = &tmp->new_cpus; in alloc_cpumasks() 614 pmask2 = &tmp->addmask; in alloc_cpumasks() 615 pmask3 = &tmp->delmask; in alloc_cpumasks() 646 static inline void free_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in free_cpumasks() argument 654 if (tmp) { in free_cpumasks() 655 free_cpumask_var(tmp->new_cpus); in free_cpumasks() 656 free_cpumask_var(tmp->addmask); in free_cpumasks() 657 free_cpumask_var(tmp->delmask); in free_cpumasks() 1308 struct tmpmasks *tmp); [all …]
|
/kernel/trace/rv/ |
D | rv.c | 345 struct dentry *tmp; in create_monitor_dir() local 352 tmp = rv_create_file("enable", RV_MODE_WRITE, mdef->root_d, mdef, &interface_enable_fops); in create_monitor_dir() 353 if (!tmp) { in create_monitor_dir() 358 tmp = rv_create_file("desc", RV_MODE_READ, mdef->root_d, mdef, &interface_desc_fops); in create_monitor_dir() 359 if (!tmp) { in create_monitor_dir() 762 struct dentry *tmp; in rv_init_interface() local 773 tmp = rv_create_file("available_monitors", RV_MODE_READ, rv_root.root_dir, NULL, in rv_init_interface() 775 if (!tmp) in rv_init_interface() 778 tmp = rv_create_file("enabled_monitors", RV_MODE_WRITE, rv_root.root_dir, NULL, in rv_init_interface() 780 if (!tmp) in rv_init_interface() [all …]
|
/kernel/sched/ |
D | topology.c | 271 struct perf_domain *tmp; in free_pd() local 274 tmp = pd->next; in free_pd() 276 pd = tmp; in free_pd() 379 struct perf_domain *pd = NULL, *tmp; in build_perf_domains() local 421 tmp = pd_init(i); in build_perf_domains() 422 if (!tmp) in build_perf_domains() 424 tmp->next = pd; in build_perf_domains() 425 pd = tmp; in build_perf_domains() 445 tmp = rd->pd; in build_perf_domains() 447 if (tmp) in build_perf_domains() [all …]
|
D | swait.c | 64 LIST_HEAD(tmp); in swake_up_all() 67 list_splice_init(&q->task_list, &tmp); in swake_up_all() 68 while (!list_empty(&tmp)) { in swake_up_all() 69 curr = list_first_entry(&tmp, typeof(*curr), task_list); in swake_up_all() 74 if (list_empty(&tmp)) in swake_up_all()
|
/kernel/locking/ |
D | rwsem.c | 275 long tmp = RWSEM_UNLOCKED_VALUE; in rwsem_write_trylock() local 277 if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) { in rwsem_write_trylock() 436 struct rwsem_waiter *waiter, *tmp; in rwsem_mark_wake() local 531 list_for_each_entry_safe(waiter, tmp, &sem->wait_list, list) { in rwsem_mark_wake() 571 list_for_each_entry_safe(waiter, tmp, &wlist, list) { in rwsem_mark_wake() 1348 long tmp; in __down_read_trylock() local 1353 tmp = atomic_long_read(&sem->count); in __down_read_trylock() 1354 while (!(tmp & RWSEM_READ_FAILED_MASK)) { in __down_read_trylock() 1355 if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, in __down_read_trylock() 1356 tmp + RWSEM_READER_BIAS)) { in __down_read_trylock() [all …]
|
D | mutex.c | 871 unsigned tmp; in ww_mutex_deadlock_injection() local 874 tmp = ctx->deadlock_inject_interval; in ww_mutex_deadlock_injection() 875 if (tmp > UINT_MAX/4) in ww_mutex_deadlock_injection() 876 tmp = UINT_MAX; in ww_mutex_deadlock_injection() 878 tmp = tmp*2 + tmp + tmp/2; in ww_mutex_deadlock_injection() 880 ctx->deadlock_inject_interval = tmp; in ww_mutex_deadlock_injection() 881 ctx->deadlock_inject_countdown = tmp; in ww_mutex_deadlock_injection()
|
/kernel/trace/ |
D | trace_probe.c | 222 char *tmp; in traceprobe_split_symbol_offset() local 228 tmp = strpbrk(symbol, "+-"); in traceprobe_split_symbol_offset() 229 if (tmp) { in traceprobe_split_symbol_offset() 230 ret = kstrtol(tmp, 0, offset); in traceprobe_split_symbol_offset() 233 *tmp = '\0'; in traceprobe_split_symbol_offset() 899 char *tmp; in parse_probe_arg() local 983 tmp = strchr(arg, '('); in parse_probe_arg() 984 if (!tmp) { in parse_probe_arg() 988 *tmp = '\0'; in parse_probe_arg() 994 ctx->offset += (tmp + 1 - arg) + (arg[0] != '-' ? 1 : 0); in parse_probe_arg() [all …]
|
D | trace_boot.c | 424 char *tmp; in trace_boot_init_histograms() local 432 tmp = kstrdup(buf, GFP_KERNEL); in trace_boot_init_histograms() 433 if (!tmp) in trace_boot_init_histograms() 436 pr_err("Failed to apply hist trigger: %s\n", tmp); in trace_boot_init_histograms() 437 kfree(tmp); in trace_boot_init_histograms() 443 tmp = kstrdup(buf, GFP_KERNEL); in trace_boot_init_histograms() 444 if (!tmp) in trace_boot_init_histograms() 447 pr_err("Failed to apply hist trigger: %s\n", tmp); in trace_boot_init_histograms() 448 kfree(tmp); in trace_boot_init_histograms()
|
D | trace_events_inject.c | 249 u8 tmp = (u8) val; in parse_entry() local 251 memcpy(entry + field->offset, &tmp, 1); in parse_entry() 255 u16 tmp = (u16) val; in parse_entry() local 257 memcpy(entry + field->offset, &tmp, 2); in parse_entry() 261 u32 tmp = (u32) val; in parse_entry() local 263 memcpy(entry + field->offset, &tmp, 4); in parse_entry()
|
D | trace_osnoise.c | 2693 struct dentry *tmp; in init_timerlat_stack_tracefs() local 2695 tmp = tracefs_create_file("print_stack", TRACE_MODE_WRITE, top_dir, in init_timerlat_stack_tracefs() 2697 if (!tmp) in init_timerlat_stack_tracefs() 2754 struct dentry *tmp; in init_timerlat_tracefs() local 2757 tmp = tracefs_create_file("timerlat_period_us", TRACE_MODE_WRITE, top_dir, in init_timerlat_tracefs() 2759 if (!tmp) in init_timerlat_tracefs() 2785 struct dentry *tmp; in init_tracefs() local 2796 tmp = tracefs_create_file("period_us", TRACE_MODE_WRITE, top_dir, in init_tracefs() 2798 if (!tmp) in init_tracefs() 2801 tmp = tracefs_create_file("runtime_us", TRACE_MODE_WRITE, top_dir, in init_tracefs() [all …]
|
/kernel/bpf/ |
D | disasm.c | 291 char tmp[64]; in print_bpf_insn() local 299 tmp, sizeof(tmp))); in print_bpf_insn() 308 char tmp[64]; in print_bpf_insn() local 314 tmp, sizeof(tmp))); in print_bpf_insn() 316 strcpy(tmp, "unknown"); in print_bpf_insn() 319 tmp, sizeof(tmp)), in print_bpf_insn()
|
D | dispatcher.c | 108 void *new, *tmp; in bpf_dispatcher_update() local 115 tmp = d->num_progs ? d->rw_image + noff : NULL; in bpf_dispatcher_update() 120 if (bpf_dispatcher_prepare(d, new, tmp)) in bpf_dispatcher_update() 122 if (IS_ERR(bpf_arch_text_copy(new, tmp, PAGE_SIZE / 2))) in bpf_dispatcher_update()
|
/kernel/time/ |
D | clocksource.c | 59 u64 tmp; in clocks_calc_mult_shift() local 66 tmp = ((u64)maxsec * from) >> 32; in clocks_calc_mult_shift() 67 while (tmp) { in clocks_calc_mult_shift() 68 tmp >>=1; in clocks_calc_mult_shift() 77 tmp = (u64) to << sft; in clocks_calc_mult_shift() 78 tmp += from / 2; in clocks_calc_mult_shift() 79 do_div(tmp, from); in clocks_calc_mult_shift() 80 if ((tmp >> sftacc) == 0) in clocks_calc_mult_shift() 83 *mult = tmp; in clocks_calc_mult_shift() 672 struct clocksource *cs, *tmp; in __clocksource_watchdog_kthread() local [all …]
|
D | timekeeping.c | 153 struct timespec64 tmp; in tk_set_wall_to_mono() local 159 set_normalized_timespec64(&tmp, -tk->wall_to_monotonic.tv_sec, in tk_set_wall_to_mono() 161 WARN_ON_ONCE(tk->offs_real != timespec64_to_ktime(tmp)); in tk_set_wall_to_mono() 163 set_normalized_timespec64(&tmp, -wtm.tv_sec, -wtm.tv_nsec); in tk_set_wall_to_mono() 164 tk->offs_real = timespec64_to_ktime(tmp); in tk_set_wall_to_mono() 312 u64 tmp, ntpinterval; in tk_setup_internals() local 326 tmp = NTP_INTERVAL_LENGTH; in tk_setup_internals() 327 tmp <<= clock->shift; in tk_setup_internals() 328 ntpinterval = tmp; in tk_setup_internals() 329 tmp += clock->mult/2; in tk_setup_internals() [all …]
|
/kernel/dma/ |
D | contiguous.c | 111 unsigned long tmp; in early_numa_cma() local 115 if (sscanf(s, "%lu%n", &tmp, &count) != 1) in early_numa_cma() 119 if (tmp >= MAX_NUMNODES) in early_numa_cma() 121 nid = array_index_nospec(tmp, MAX_NUMNODES); in early_numa_cma() 124 tmp = memparse(s, &s); in early_numa_cma() 125 numa_cma_size[nid] = tmp; in early_numa_cma()
|
/kernel/debug/ |
D | gdbstub.c | 238 char *tmp; in kgdb_mem2hex() local 245 tmp = buf + count; in kgdb_mem2hex() 247 err = copy_from_kernel_nofault(tmp, mem, count); in kgdb_mem2hex() 251 buf = hex_byte_pack(buf, *tmp); in kgdb_mem2hex() 252 tmp++; in kgdb_mem2hex() 957 int tmp; in gdb_serial_stub() local 1048 tmp = gdb_cmd_exception_pass(ks); in gdb_serial_stub() 1049 if (tmp > 0) in gdb_serial_stub() 1051 if (tmp == 0) in gdb_serial_stub()
|
/kernel/kcsan/ |
D | kcsan_test.c | 183 char tmp[2][64]; in __report_matches() local 187 scnprintf(tmp[0], sizeof(tmp[0]), "%pS", r->access[0].fn); in __report_matches() 188 scnprintf(tmp[1], sizeof(tmp[1]), "%pS", r->access[1].fn); in __report_matches() 189 cmp = strcmp(tmp[0], tmp[1]); in __report_matches() 1219 long tmp; in test_atomic_builtins() local 1229 tmp = 20L; in test_atomic_builtins() 1230 KUNIT_EXPECT_TRUE(test, __atomic_compare_exchange_n(&test_var, &tmp, 30L, in test_atomic_builtins() 1233 KUNIT_EXPECT_EQ(test, tmp, 20L); in test_atomic_builtins() 1235 KUNIT_EXPECT_FALSE(test, __atomic_compare_exchange_n(&test_var, &tmp, 40L, in test_atomic_builtins() 1238 KUNIT_EXPECT_EQ(test, tmp, 30L); in test_atomic_builtins()
|