| /kernel/ |
| D | delayacct.c | 133 s64 tmp; in delayacct_add_tsk() local 136 tmp = (s64)d->cpu_run_real_total; in delayacct_add_tsk() 137 tmp += utime + stime; in delayacct_add_tsk() 138 d->cpu_run_real_total = (tmp < (s64)d->cpu_run_real_total) ? 0 : tmp; in delayacct_add_tsk() 141 tmp = (s64)d->cpu_scaled_run_real_total; in delayacct_add_tsk() 142 tmp += utimescaled + stimescaled; in delayacct_add_tsk() 144 (tmp < (s64)d->cpu_scaled_run_real_total) ? 0 : tmp; in delayacct_add_tsk() 156 tmp = (s64)d->cpu_delay_total + t2; in delayacct_add_tsk() 157 d->cpu_delay_total = (tmp < (s64)d->cpu_delay_total) ? 0 : tmp; in delayacct_add_tsk() 159 tmp = (s64)d->cpu_run_virtual_total + t3; in delayacct_add_tsk() [all …]
|
| D | resource.c | 167 struct resource *tmp, **p; in __request_resource() local 177 tmp = *p; in __request_resource() 178 if (!tmp || tmp->start > end) { in __request_resource() 179 new->sibling = tmp; in __request_resource() 184 p = &tmp->sibling; in __request_resource() 185 if (tmp->end < start) in __request_resource() 187 return tmp; in __request_resource() 193 struct resource *tmp, **p, *chd; in __release_resource() local 197 tmp = *p; in __release_resource() 198 if (!tmp) in __release_resource() [all …]
|
| D | sysctl.c | 345 char *p, tmp[TMPBUFLEN]; in proc_get_long() local 354 memcpy(tmp, *buf, len); in proc_get_long() 356 tmp[len] = 0; in proc_get_long() 357 p = tmp; in proc_get_long() 369 len = p - tmp; in proc_get_long() 403 char tmp[TMPBUFLEN], *p = tmp; in proc_put_long() local 406 len = strlen(tmp); in proc_put_long() 409 memcpy(*buf, tmp, len); in proc_put_long() 708 struct ctl_table tmp; in proc_dobool() local 716 tmp = *table; in proc_dobool() [all …]
|
| D | crash_reserve.c | 61 char *cur = cmdline, *tmp; in parse_crashkernel_mem() local 77 start = memparse(cur, &tmp); in parse_crashkernel_mem() 78 if (cur == tmp) { in parse_crashkernel_mem() 82 cur = tmp; in parse_crashkernel_mem() 91 end = memparse(cur, &tmp); in parse_crashkernel_mem() 92 if (cur == tmp) { in parse_crashkernel_mem() 96 cur = tmp; in parse_crashkernel_mem() 109 size = memparse(cur, &tmp); in parse_crashkernel_mem() 110 if (cur == tmp) { in parse_crashkernel_mem() 114 cur = tmp; in parse_crashkernel_mem() [all …]
|
| D | pid.c | 172 struct pid_namespace *tmp; in alloc_pid() local 191 tmp = ns; in alloc_pid() 207 if (tid != 1 && !tmp->child_reaper) in alloc_pid() 210 if (!checkpoint_restore_ns_capable(tmp->user_ns)) in alloc_pid() 219 nr = idr_alloc(&tmp->idr, NULL, tid, in alloc_pid() 233 if (idr_get_cursor(&tmp->idr) > RESERVED_PIDS) in alloc_pid() 240 nr = idr_alloc_cyclic(&tmp->idr, NULL, pid_min, in alloc_pid() 252 pid->numbers[i].ns = tmp; in alloc_pid() 253 tmp = tmp->parent; in alloc_pid() 351 int tmp; in __change_pid() local [all …]
|
| D | sys.c | 1049 struct tms tmp; in SYSCALL_DEFINE1() local 1051 do_sys_times(&tmp); in SYSCALL_DEFINE1() 1052 if (copy_to_user(tbuf, &tmp, sizeof(struct tms))) in SYSCALL_DEFINE1() 1069 struct compat_tms tmp; in COMPAT_SYSCALL_DEFINE1() local 1073 tmp.tms_utime = clock_t_to_compat_clock_t(tms.tms_utime); in COMPAT_SYSCALL_DEFINE1() 1074 tmp.tms_stime = clock_t_to_compat_clock_t(tms.tms_stime); in COMPAT_SYSCALL_DEFINE1() 1075 tmp.tms_cutime = clock_t_to_compat_clock_t(tms.tms_cutime); in COMPAT_SYSCALL_DEFINE1() 1076 tmp.tms_cstime = clock_t_to_compat_clock_t(tms.tms_cstime); in COMPAT_SYSCALL_DEFINE1() 1077 if (copy_to_user(tbuf, &tmp, sizeof(tmp))) in COMPAT_SYSCALL_DEFINE1() 1331 struct new_utsname tmp; in SYSCALL_DEFINE1() local [all …]
|
| D | fork.c | 220 struct vm_struct *tmp = NULL; in try_release_thread_stack_to_cache() local 222 if (this_cpu_try_cmpxchg(cached_stacks[i], &tmp, vm)) in try_release_thread_stack_to_cache() 622 struct vm_area_struct *mpnt, *tmp; in dup_mmap() local 680 tmp = vm_area_dup(mpnt); in dup_mmap() 681 if (!tmp) in dup_mmap() 685 if (unlikely(tmp->vm_flags & VM_PFNMAP)) in dup_mmap() 686 untrack_pfn_clear(tmp); in dup_mmap() 688 retval = vma_dup_policy(mpnt, tmp); in dup_mmap() 691 tmp->vm_mm = mm; in dup_mmap() 692 retval = dup_userfaultfd(tmp, &uf); in dup_mmap() [all …]
|
| /kernel/power/ |
| D | console.c | 49 struct pm_vt_switch *entry, *tmp; in pm_vt_switch_required() local 52 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_required() 53 if (tmp->dev == dev) { in pm_vt_switch_required() 55 tmp->required = required; in pm_vt_switch_required() 81 struct pm_vt_switch *tmp; in pm_vt_switch_unregister() local 84 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_unregister() 85 if (tmp->dev == dev) { in pm_vt_switch_unregister() 86 list_del(&tmp->head); in pm_vt_switch_unregister() 87 kfree(tmp); in pm_vt_switch_unregister()
|
| D | swap.c | 1011 struct swap_map_page_list *tmp; in release_swap_reader() local 1016 tmp = handle->maps; in release_swap_reader() 1018 kfree(tmp); in release_swap_reader() 1027 struct swap_map_page_list *tmp, *last; in get_swap_reader() local 1039 tmp = kzalloc(sizeof(*handle->maps), GFP_KERNEL); in get_swap_reader() 1040 if (!tmp) { in get_swap_reader() 1045 handle->maps = tmp; in get_swap_reader() 1047 last->next = tmp; in get_swap_reader() 1048 last = tmp; in get_swap_reader() 1050 tmp->map = (struct swap_map_page *) in get_swap_reader() [all …]
|
| /kernel/cgroup/ |
| D | cpuset.c | 411 static inline int alloc_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in alloc_cpumasks() argument 421 pmask1 = &tmp->new_cpus; in alloc_cpumasks() 422 pmask2 = &tmp->addmask; in alloc_cpumasks() 423 pmask3 = &tmp->delmask; in alloc_cpumasks() 456 static inline void free_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in free_cpumasks() argument 464 if (tmp) { in free_cpumasks() 465 free_cpumask_var(tmp->new_cpus); in free_cpumasks() 466 free_cpumask_var(tmp->addmask); in free_cpumasks() 467 free_cpumask_var(tmp->delmask); in free_cpumasks() 1160 struct tmpmasks *tmp); [all …]
|
| /kernel/trace/rv/ |
| D | rv.c | 343 struct dentry *tmp; in create_monitor_dir() local 350 tmp = rv_create_file("enable", RV_MODE_WRITE, mdef->root_d, mdef, &interface_enable_fops); in create_monitor_dir() 351 if (!tmp) { in create_monitor_dir() 356 tmp = rv_create_file("desc", RV_MODE_READ, mdef->root_d, mdef, &interface_desc_fops); in create_monitor_dir() 357 if (!tmp) { in create_monitor_dir() 759 struct dentry *tmp; in rv_init_interface() local 770 tmp = rv_create_file("available_monitors", RV_MODE_READ, rv_root.root_dir, NULL, in rv_init_interface() 772 if (!tmp) in rv_init_interface() 775 tmp = rv_create_file("enabled_monitors", RV_MODE_WRITE, rv_root.root_dir, NULL, in rv_init_interface() 777 if (!tmp) in rv_init_interface() [all …]
|
| /kernel/trace/ |
| D | trace_probe.c | 232 char *tmp; in traceprobe_split_symbol_offset() local 238 tmp = strpbrk(symbol, "+-"); in traceprobe_split_symbol_offset() 239 if (tmp) { in traceprobe_split_symbol_offset() 240 ret = kstrtol(tmp, 0, offset); in traceprobe_split_symbol_offset() 243 *tmp = '\0'; in traceprobe_split_symbol_offset() 1063 char *tmp; in parse_probe_arg() local 1147 tmp = strchr(arg, '('); in parse_probe_arg() 1148 if (!tmp) { in parse_probe_arg() 1152 *tmp = '\0'; in parse_probe_arg() 1158 ctx->offset += (tmp + 1 - arg) + (arg[0] != '-' ? 1 : 0); in parse_probe_arg() [all …]
|
| D | trace_boot.c | 424 char *tmp; in trace_boot_init_histograms() local 432 tmp = kstrdup(buf, GFP_KERNEL); in trace_boot_init_histograms() 433 if (!tmp) in trace_boot_init_histograms() 436 pr_err("Failed to apply hist trigger: %s\n", tmp); in trace_boot_init_histograms() 437 kfree(tmp); in trace_boot_init_histograms() 443 tmp = kstrdup(buf, GFP_KERNEL); in trace_boot_init_histograms() 444 if (!tmp) in trace_boot_init_histograms() 447 pr_err("Failed to apply hist trigger: %s\n", tmp); in trace_boot_init_histograms() 448 kfree(tmp); in trace_boot_init_histograms()
|
| D | trace_events_inject.c | 249 u8 tmp = (u8) val; in parse_entry() local 251 memcpy(entry + field->offset, &tmp, 1); in parse_entry() 255 u16 tmp = (u16) val; in parse_entry() local 257 memcpy(entry + field->offset, &tmp, 2); in parse_entry() 261 u32 tmp = (u32) val; in parse_entry() local 263 memcpy(entry + field->offset, &tmp, 4); in parse_entry()
|
| D | trace_fprobe.c | 1006 char *tmp = strchr(argv[1], '%'); in parse_symbol_and_return() local 1009 if (tmp) { in parse_symbol_and_return() 1010 int len = tmp - argv[1]; in parse_symbol_and_return() 1012 if (!is_tracepoint && !strcmp(tmp, "%return")) { in parse_symbol_and_return() 1028 tmp = *symbol; in parse_symbol_and_return() 1029 while (*tmp && (isalnum(*tmp) || *tmp == '_')) in parse_symbol_and_return() 1030 tmp++; in parse_symbol_and_return() 1031 if (*tmp) { in parse_symbol_and_return() 1033 trace_probe_log_err(tmp - *symbol, BAD_TP_NAME); in parse_symbol_and_return() 1042 tmp = strstr(argv[i], "$retval"); in parse_symbol_and_return() [all …]
|
| D | trace_osnoise.c | 2725 struct dentry *tmp; in init_timerlat_stack_tracefs() local 2727 tmp = tracefs_create_file("print_stack", TRACE_MODE_WRITE, top_dir, in init_timerlat_stack_tracefs() 2729 if (!tmp) in init_timerlat_stack_tracefs() 2786 struct dentry *tmp; in init_timerlat_tracefs() local 2789 tmp = tracefs_create_file("timerlat_period_us", TRACE_MODE_WRITE, top_dir, in init_timerlat_tracefs() 2791 if (!tmp) in init_timerlat_tracefs() 2817 struct dentry *tmp; in init_tracefs() local 2828 tmp = tracefs_create_file("period_us", TRACE_MODE_WRITE, top_dir, in init_tracefs() 2830 if (!tmp) in init_tracefs() 2833 tmp = tracefs_create_file("runtime_us", TRACE_MODE_WRITE, top_dir, in init_tracefs() [all …]
|
| /kernel/sched/ |
| D | topology.c | 347 struct perf_domain *tmp; in free_pd() local 350 tmp = pd->next; in free_pd() 352 pd = tmp; in free_pd() 437 struct perf_domain *pd = NULL, *tmp; in build_perf_domains() local 453 tmp = pd_init(i); in build_perf_domains() 454 if (!tmp) in build_perf_domains() 456 tmp->next = pd; in build_perf_domains() 457 pd = tmp; in build_perf_domains() 463 tmp = rd->pd; in build_perf_domains() 465 if (tmp) in build_perf_domains() [all …]
|
| D | swait.c | 64 LIST_HEAD(tmp); in swake_up_all() 67 list_splice_init(&q->task_list, &tmp); in swake_up_all() 68 while (!list_empty(&tmp)) { in swake_up_all() 69 curr = list_first_entry(&tmp, typeof(*curr), task_list); in swake_up_all() 74 if (list_empty(&tmp)) in swake_up_all()
|
| /kernel/locking/ |
| D | rwsem.c | 282 long tmp = RWSEM_UNLOCKED_VALUE; in rwsem_write_trylock() local 284 if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) { in rwsem_write_trylock() 434 struct rwsem_waiter *waiter, *tmp; in rwsem_mark_wake() local 529 list_for_each_entry_safe(waiter, tmp, &sem->wait_list, list) { in rwsem_mark_wake() 569 list_for_each_entry_safe(waiter, tmp, &wlist, list) { in rwsem_mark_wake() 1346 long tmp; in __down_read_trylock() local 1351 tmp = atomic_long_read(&sem->count); in __down_read_trylock() 1352 while (!(tmp & RWSEM_READ_FAILED_MASK)) { in __down_read_trylock() 1353 if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, in __down_read_trylock() 1354 tmp + RWSEM_READER_BIAS)) { in __down_read_trylock() [all …]
|
| D | mutex.c | 899 unsigned tmp; in ww_mutex_deadlock_injection() local 902 tmp = ctx->deadlock_inject_interval; in ww_mutex_deadlock_injection() 903 if (tmp > UINT_MAX/4) in ww_mutex_deadlock_injection() 904 tmp = UINT_MAX; in ww_mutex_deadlock_injection() 906 tmp = tmp*2 + tmp + tmp/2; in ww_mutex_deadlock_injection() 908 ctx->deadlock_inject_interval = tmp; in ww_mutex_deadlock_injection() 909 ctx->deadlock_inject_countdown = tmp; in ww_mutex_deadlock_injection()
|
| /kernel/bpf/ |
| D | disasm.c | 315 char tmp[64]; in print_bpf_insn() local 323 tmp, sizeof(tmp))); in print_bpf_insn() 332 char tmp[64]; in print_bpf_insn() local 338 tmp, sizeof(tmp))); in print_bpf_insn() 340 strcpy(tmp, "unknown"); in print_bpf_insn() 343 tmp, sizeof(tmp)), in print_bpf_insn()
|
| D | dispatcher.c | 108 void *new, *tmp; in bpf_dispatcher_update() local 115 tmp = d->num_progs ? d->rw_image + noff : NULL; in bpf_dispatcher_update() 120 if (bpf_dispatcher_prepare(d, new, tmp)) in bpf_dispatcher_update() 122 if (IS_ERR(bpf_arch_text_copy(new, tmp, PAGE_SIZE / 2))) in bpf_dispatcher_update()
|
| /kernel/time/ |
| D | clocksource.c | 59 u64 tmp; in clocks_calc_mult_shift() local 66 tmp = ((u64)maxsec * from) >> 32; in clocks_calc_mult_shift() 67 while (tmp) { in clocks_calc_mult_shift() 68 tmp >>=1; in clocks_calc_mult_shift() 77 tmp = (u64) to << sft; in clocks_calc_mult_shift() 78 tmp += from / 2; in clocks_calc_mult_shift() 79 do_div(tmp, from); in clocks_calc_mult_shift() 80 if ((tmp >> sftacc) == 0) in clocks_calc_mult_shift() 83 *mult = tmp; in clocks_calc_mult_shift() 689 struct clocksource *cs, *tmp; in __clocksource_watchdog_kthread() local [all …]
|
| D | timekeeping.c | 153 struct timespec64 tmp; in tk_set_wall_to_mono() local 159 set_normalized_timespec64(&tmp, -tk->wall_to_monotonic.tv_sec, in tk_set_wall_to_mono() 161 WARN_ON_ONCE(tk->offs_real != timespec64_to_ktime(tmp)); in tk_set_wall_to_mono() 163 set_normalized_timespec64(&tmp, -wtm.tv_sec, -wtm.tv_nsec); in tk_set_wall_to_mono() 164 tk->offs_real = timespec64_to_ktime(tmp); in tk_set_wall_to_mono() 212 u64 tmp, ntpinterval; in tk_setup_internals() local 226 tmp = NTP_INTERVAL_LENGTH; in tk_setup_internals() 227 tmp <<= clock->shift; in tk_setup_internals() 228 ntpinterval = tmp; in tk_setup_internals() 229 tmp += clock->mult/2; in tk_setup_internals() [all …]
|
| /kernel/dma/ |
| D | contiguous.c | 104 unsigned long tmp; in early_numa_cma() local 108 if (sscanf(s, "%lu%n", &tmp, &count) != 1) in early_numa_cma() 112 if (tmp >= MAX_NUMNODES) in early_numa_cma() 114 nid = array_index_nospec(tmp, MAX_NUMNODES); in early_numa_cma() 117 tmp = memparse(s, &s); in early_numa_cma() 118 numa_cma_size[nid] = tmp; in early_numa_cma()
|