/kernel/ |
D | delayacct.c | 90 s64 tmp; in __delayacct_add_tsk() local 93 tmp = (s64)d->cpu_run_real_total; in __delayacct_add_tsk() 94 tmp += utime + stime; in __delayacct_add_tsk() 95 d->cpu_run_real_total = (tmp < (s64)d->cpu_run_real_total) ? 0 : tmp; in __delayacct_add_tsk() 98 tmp = (s64)d->cpu_scaled_run_real_total; in __delayacct_add_tsk() 99 tmp += utimescaled + stimescaled; in __delayacct_add_tsk() 101 (tmp < (s64)d->cpu_scaled_run_real_total) ? 0 : tmp; in __delayacct_add_tsk() 113 tmp = (s64)d->cpu_delay_total + t2; in __delayacct_add_tsk() 114 d->cpu_delay_total = (tmp < (s64)d->cpu_delay_total) ? 0 : tmp; in __delayacct_add_tsk() 116 tmp = (s64)d->cpu_run_virtual_total + t3; in __delayacct_add_tsk() [all …]
|
D | resource.c | 189 struct resource *tmp, **p; in __request_resource() local 199 tmp = *p; in __request_resource() 200 if (!tmp || tmp->start > end) { in __request_resource() 201 new->sibling = tmp; in __request_resource() 206 p = &tmp->sibling; in __request_resource() 207 if (tmp->end < start) in __request_resource() 209 return tmp; in __request_resource() 215 struct resource *tmp, **p, *chd; in __release_resource() local 219 tmp = *p; in __release_resource() 220 if (!tmp) in __release_resource() [all …]
|
D | sysctl_binary.c | 1262 char *tmp, *result; in sysctl_getname() local 1265 tmp = __getname(); in sysctl_getname() 1266 if (tmp) { in sysctl_getname() 1267 const struct bin_table *table = get_sysctl(name, nlen, tmp); in sysctl_getname() 1268 result = tmp; in sysctl_getname() 1271 __putname(tmp); in sysctl_getname() 1404 struct __sysctl_args tmp; in SYSCALL_DEFINE1() local 1408 if (copy_from_user(&tmp, args, sizeof(tmp))) in SYSCALL_DEFINE1() 1411 if (tmp.oldval && !tmp.oldlenp) in SYSCALL_DEFINE1() 1414 if (tmp.oldlenp && get_user(oldlen, tmp.oldlenp)) in SYSCALL_DEFINE1() [all …]
|
D | crash_core.c | 41 char *cur = cmdline, *tmp; in parse_crashkernel_mem() local 48 start = memparse(cur, &tmp); in parse_crashkernel_mem() 49 if (cur == tmp) { in parse_crashkernel_mem() 53 cur = tmp; in parse_crashkernel_mem() 62 end = memparse(cur, &tmp); in parse_crashkernel_mem() 63 if (cur == tmp) { in parse_crashkernel_mem() 67 cur = tmp; in parse_crashkernel_mem() 80 size = memparse(cur, &tmp); in parse_crashkernel_mem() 81 if (cur == tmp) { in parse_crashkernel_mem() 85 cur = tmp; in parse_crashkernel_mem() [all …]
|
D | pid.c | 165 struct pid_namespace *tmp; in alloc_pid() local 173 tmp = ns; in alloc_pid() 186 if (idr_get_cursor(&tmp->idr) > RESERVED_PIDS) in alloc_pid() 193 nr = idr_alloc_cyclic(&tmp->idr, NULL, pid_min, in alloc_pid() 204 pid->numbers[i].ns = tmp; in alloc_pid() 205 tmp = tmp->parent; in alloc_pid() 294 int tmp; in __change_pid() local 301 for (tmp = PIDTYPE_MAX; --tmp >= 0; ) in __change_pid() 302 if (!hlist_empty(&pid->tasks[tmp])) in __change_pid()
|
D | sys.c | 961 struct tms tmp; in SYSCALL_DEFINE1() local 963 do_sys_times(&tmp); in SYSCALL_DEFINE1() 964 if (copy_to_user(tbuf, &tmp, sizeof(struct tms))) in SYSCALL_DEFINE1() 981 struct compat_tms tmp; in COMPAT_SYSCALL_DEFINE1() local 985 tmp.tms_utime = clock_t_to_compat_clock_t(tms.tms_utime); in COMPAT_SYSCALL_DEFINE1() 986 tmp.tms_stime = clock_t_to_compat_clock_t(tms.tms_stime); in COMPAT_SYSCALL_DEFINE1() 987 tmp.tms_cutime = clock_t_to_compat_clock_t(tms.tms_cutime); in COMPAT_SYSCALL_DEFINE1() 988 tmp.tms_cstime = clock_t_to_compat_clock_t(tms.tms_cstime); in COMPAT_SYSCALL_DEFINE1() 989 if (copy_to_user(tbuf, &tmp, sizeof(tmp))) in COMPAT_SYSCALL_DEFINE1() 1243 struct new_utsname tmp; in SYSCALL_DEFINE1() local [all …]
|
D | fork.c | 486 struct vm_area_struct *mpnt, *tmp, *prev, **pprev; in dup_mmap() local 546 tmp = vm_area_dup(mpnt); in dup_mmap() 547 if (!tmp) in dup_mmap() 549 retval = vma_dup_policy(mpnt, tmp); in dup_mmap() 552 tmp->vm_mm = mm; in dup_mmap() 553 retval = dup_userfaultfd(tmp, &uf); in dup_mmap() 556 if (tmp->vm_flags & VM_WIPEONFORK) { in dup_mmap() 558 tmp->anon_vma = NULL; in dup_mmap() 559 if (anon_vma_prepare(tmp)) in dup_mmap() 561 } else if (anon_vma_fork(tmp, mpnt)) in dup_mmap() [all …]
|
D | sysctl.c | 2140 char *tmp = skip_spaces(*buf); in proc_skip_spaces() local 2141 ret = tmp - *buf; in proc_skip_spaces() 2142 *buf = tmp; in proc_skip_spaces() 2213 char *p, tmp[TMPBUFLEN]; in proc_get_long() local 2222 memcpy(tmp, *buf, len); in proc_get_long() 2224 tmp[len] = 0; in proc_get_long() 2225 p = tmp; in proc_get_long() 2237 len = p - tmp; in proc_get_long() 2272 char tmp[TMPBUFLEN], *p = tmp; in proc_put_long() local 2275 len = strlen(tmp); in proc_put_long() [all …]
|
D | acct.c | 281 struct filename *tmp = getname(name); in SYSCALL_DEFINE1() local 283 if (IS_ERR(tmp)) in SYSCALL_DEFINE1() 284 return PTR_ERR(tmp); in SYSCALL_DEFINE1() 286 error = acct_on(tmp); in SYSCALL_DEFINE1() 288 putname(tmp); in SYSCALL_DEFINE1()
|
D | ptrace.c | 1157 unsigned long tmp = 0; in ptrace_request() local 1165 tmp = mm->context.exec_fdpic_loadmap; in ptrace_request() 1168 tmp = mm->context.interp_fdpic_loadmap; in ptrace_request() 1175 ret = put_user(tmp, datalp); in ptrace_request() 1290 unsigned long tmp; in generic_ptrace_peekdata() local 1293 copied = ptrace_access_vm(tsk, addr, &tmp, sizeof(tmp), FOLL_FORCE); in generic_ptrace_peekdata() 1294 if (copied != sizeof(tmp)) in generic_ptrace_peekdata() 1296 return put_user(tmp, (unsigned long __user *)data); in generic_ptrace_peekdata()
|
/kernel/power/ |
D | console.c | 48 struct pm_vt_switch *entry, *tmp; in pm_vt_switch_required() local 51 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_required() 52 if (tmp->dev == dev) { in pm_vt_switch_required() 54 tmp->required = required; in pm_vt_switch_required() 80 struct pm_vt_switch *tmp; in pm_vt_switch_unregister() local 83 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_unregister() 84 if (tmp->dev == dev) { in pm_vt_switch_unregister() 85 list_del(&tmp->head); in pm_vt_switch_unregister() 86 kfree(tmp); in pm_vt_switch_unregister()
|
D | swap.c | 949 struct swap_map_page_list *tmp; in release_swap_reader() local 954 tmp = handle->maps; in release_swap_reader() 956 kfree(tmp); in release_swap_reader() 965 struct swap_map_page_list *tmp, *last; in get_swap_reader() local 977 tmp = kzalloc(sizeof(*handle->maps), GFP_KERNEL); in get_swap_reader() 978 if (!tmp) { in get_swap_reader() 983 handle->maps = tmp; in get_swap_reader() 985 last->next = tmp; in get_swap_reader() 986 last = tmp; in get_swap_reader() 988 tmp->map = (struct swap_map_page *) in get_swap_reader() [all …]
|
/kernel/sched/ |
D | isolation.c | 83 cpumask_var_t tmp; in housekeeping_setup() local 94 alloc_bootmem_cpumask_var(&tmp); in housekeeping_setup() 100 cpumask_andnot(tmp, cpu_present_mask, non_housekeeping_mask); in housekeeping_setup() 101 if (cpumask_empty(tmp)) { in housekeeping_setup() 108 cpumask_andnot(tmp, cpu_present_mask, non_housekeeping_mask); in housekeeping_setup() 109 if (cpumask_empty(tmp)) in housekeeping_setup() 111 cpumask_andnot(tmp, cpu_possible_mask, non_housekeeping_mask); in housekeeping_setup() 112 if (!cpumask_equal(tmp, housekeeping_mask)) { in housekeeping_setup() 114 free_bootmem_cpumask_var(tmp); in housekeeping_setup() 119 free_bootmem_cpumask_var(tmp); in housekeeping_setup()
|
D | topology.c | 237 struct perf_domain *tmp; in free_pd() local 240 tmp = pd->next; in free_pd() 242 pd = tmp; in free_pd() 345 struct perf_domain *pd = NULL, *tmp; in build_perf_domains() local 382 tmp = pd_init(i); in build_perf_domains() 383 if (!tmp) in build_perf_domains() 385 tmp->next = pd; in build_perf_domains() 386 pd = tmp; in build_perf_domains() 406 tmp = rd->pd; in build_perf_domains() 408 if (tmp) in build_perf_domains() [all …]
|
D | swait.c | 52 LIST_HEAD(tmp); in swake_up_all() 55 list_splice_init(&q->task_list, &tmp); in swake_up_all() 56 while (!list_empty(&tmp)) { in swake_up_all() 57 curr = list_first_entry(&tmp, typeof(*curr), task_list); in swake_up_all() 62 if (list_empty(&tmp)) in swake_up_all()
|
/kernel/cgroup/ |
D | cpuset.c | 460 static inline int alloc_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in alloc_cpumasks() argument 469 pmask1 = &tmp->new_cpus; in alloc_cpumasks() 470 pmask2 = &tmp->addmask; in alloc_cpumasks() 471 pmask3 = &tmp->delmask; in alloc_cpumasks() 497 static inline void free_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in free_cpumasks() argument 504 if (tmp) { in free_cpumasks() 505 free_cpumask_var(tmp->new_cpus); in free_cpumasks() 506 free_cpumask_var(tmp->addmask); in free_cpumasks() 507 free_cpumask_var(tmp->delmask); in free_cpumasks() 1121 struct tmpmasks *tmp) in update_parent_subparts_cpumask() argument [all …]
|
/kernel/locking/ |
D | rwsem.c | 405 struct rwsem_waiter *waiter, *tmp; in rwsem_mark_wake() local 502 list_for_each_entry_safe(waiter, tmp, &sem->wait_list, list) { in rwsem_mark_wake() 534 list_for_each_entry_safe(waiter, tmp, &wlist, list) { in rwsem_mark_wake() 1365 long tmp; in __down_read_trylock() local 1372 tmp = RWSEM_UNLOCKED_VALUE; in __down_read_trylock() 1374 if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, in __down_read_trylock() 1375 tmp + RWSEM_READER_BIAS)) { in __down_read_trylock() 1379 } while (!(tmp & RWSEM_READ_FAILED_MASK)); in __down_read_trylock() 1388 long tmp = RWSEM_UNLOCKED_VALUE; in __down_write() local 1390 if (unlikely(!atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, in __down_write() [all …]
|
D | mutex.c | 1162 unsigned tmp; in ww_mutex_deadlock_injection() local 1165 tmp = ctx->deadlock_inject_interval; in ww_mutex_deadlock_injection() 1166 if (tmp > UINT_MAX/4) in ww_mutex_deadlock_injection() 1167 tmp = UINT_MAX; in ww_mutex_deadlock_injection() 1169 tmp = tmp*2 + tmp + tmp/2; in ww_mutex_deadlock_injection() 1171 ctx->deadlock_inject_interval = tmp; in ww_mutex_deadlock_injection() 1172 ctx->deadlock_inject_countdown = tmp; in ww_mutex_deadlock_injection()
|
/kernel/trace/ |
D | trace_probe.c | 210 char *tmp; in traceprobe_split_symbol_offset() local 216 tmp = strpbrk(symbol, "+-"); in traceprobe_split_symbol_offset() 217 if (tmp) { in traceprobe_split_symbol_offset() 218 ret = kstrtol(tmp, 0, offset); in traceprobe_split_symbol_offset() 221 *tmp = '\0'; in traceprobe_split_symbol_offset() 362 char *tmp; in parse_probe_arg() local 441 tmp = strchr(arg, '('); in parse_probe_arg() 442 if (!tmp) { in parse_probe_arg() 446 *tmp = '\0'; in parse_probe_arg() 452 offs += (tmp + 1 - arg) + (arg[0] != '-' ? 1 : 0); in parse_probe_arg() [all …]
|
/kernel/bpf/ |
D | disasm.c | 202 char tmp[64]; in print_bpf_insn() local 210 tmp, sizeof(tmp))); in print_bpf_insn() 219 char tmp[64]; in print_bpf_insn() local 225 tmp, sizeof(tmp))); in print_bpf_insn() 227 strcpy(tmp, "unknown"); in print_bpf_insn() 230 tmp, sizeof(tmp)), in print_bpf_insn()
|
D | xskmap.c | 69 struct xsk_map_node *n, *tmp; in xsk_map_sock_delete() local 72 list_for_each_entry_safe(n, tmp, &xs->map_list, node) { in xsk_map_sock_delete() 196 struct xdp_sock *xs, *tmp; in __xsk_map_flush() local 198 list_for_each_entry_safe(xs, tmp, flush_list, flush_node) { in __xsk_map_flush()
|
/kernel/time/ |
D | clocksource.c | 47 u64 tmp; in clocks_calc_mult_shift() local 54 tmp = ((u64)maxsec * from) >> 32; in clocks_calc_mult_shift() 55 while (tmp) { in clocks_calc_mult_shift() 56 tmp >>=1; in clocks_calc_mult_shift() 65 tmp = (u64) to << sft; in clocks_calc_mult_shift() 66 tmp += from / 2; in clocks_calc_mult_shift() 67 do_div(tmp, from); in clocks_calc_mult_shift() 68 if ((tmp >> sftacc) == 0) in clocks_calc_mult_shift() 71 *mult = tmp; in clocks_calc_mult_shift() 399 struct clocksource *cs, *tmp; in __clocksource_watchdog_kthread() local [all …]
|
D | timekeeping.c | 131 struct timespec64 tmp; in tk_set_wall_to_mono() local 137 set_normalized_timespec64(&tmp, -tk->wall_to_monotonic.tv_sec, in tk_set_wall_to_mono() 139 WARN_ON_ONCE(tk->offs_real != timespec64_to_ktime(tmp)); in tk_set_wall_to_mono() 141 set_normalized_timespec64(&tmp, -wtm.tv_sec, -wtm.tv_nsec); in tk_set_wall_to_mono() 142 tk->offs_real = timespec64_to_ktime(tmp); in tk_set_wall_to_mono() 290 u64 tmp, ntpinterval; in tk_setup_internals() local 304 tmp = NTP_INTERVAL_LENGTH; in tk_setup_internals() 305 tmp <<= clock->shift; in tk_setup_internals() 306 ntpinterval = tmp; in tk_setup_internals() 307 tmp += clock->mult/2; in tk_setup_internals() [all …]
|
/kernel/debug/ |
D | gdbstub.c | 241 char *tmp; in kgdb_mem2hex() local 248 tmp = buf + count; in kgdb_mem2hex() 250 err = probe_kernel_read(tmp, mem, count); in kgdb_mem2hex() 254 buf = hex_byte_pack(buf, *tmp); in kgdb_mem2hex() 255 tmp++; in kgdb_mem2hex() 947 int tmp; in gdb_serial_stub() local 1038 tmp = gdb_cmd_exception_pass(ks); in gdb_serial_stub() 1039 if (tmp > 0) in gdb_serial_stub() 1041 if (tmp == 0) in gdb_serial_stub()
|
/kernel/debug/kdb/ |
D | kdb_io.c | 211 char tmp; in kdb_read() local 251 tmp = *cp; in kdb_read() 255 *cp = tmp; in kdb_read() 273 tmp = *cp; in kdb_read() 277 *cp = tmp; in kdb_read() 387 tmp = *cp; in kdb_read() 391 *cp = tmp; in kdb_read()
|