/kernel/ |
D | pid.c | 57 .nr = 0, 155 idr_remove(&ns->idr, upid->nr); in free_pid() 167 int i, nr; in alloc_pid() local 215 nr = idr_alloc(&tmp->idr, NULL, tid, in alloc_pid() 221 if (nr == -ENOSPC) in alloc_pid() 222 nr = -EEXIST; in alloc_pid() 236 nr = idr_alloc_cyclic(&tmp->idr, NULL, pid_min, in alloc_pid() 242 if (nr < 0) { in alloc_pid() 243 retval = (nr == -ENOSPC) ? -EAGAIN : nr; in alloc_pid() 247 pid->numbers[i].nr = nr; in alloc_pid() [all …]
|
D | softirq.c | 715 inline void raise_softirq_irqoff(unsigned int nr) in raise_softirq_irqoff() argument 717 __raise_softirq_irqoff(nr); in raise_softirq_irqoff() 732 void raise_softirq(unsigned int nr) in raise_softirq() argument 737 raise_softirq_irqoff(nr); in raise_softirq() 742 void __raise_softirq_irqoff(unsigned int nr) in __raise_softirq_irqoff() argument 745 trace_softirq_raise(nr); in __raise_softirq_irqoff() 746 or_softirq_pending(1UL << nr); in __raise_softirq_irqoff() 749 void open_softirq(int nr, void (*action)(struct softirq_action *)) in open_softirq() argument 751 softirq_vec[nr].action = action; in open_softirq()
|
D | seccomp.c | 256 sd->nr = syscall_get_nr(task, regs); in populate_seccomp_data() 372 int syscall_nr = sd->nr; in seccomp_cache_check_allow() 747 case offsetof(struct seccomp_data, nr): in seccomp_is_const_allow() 748 reg_value = sd->nr; in seccomp_is_const_allow() 808 int nr; in seccomp_cache_prepare_bitmap() local 818 for (nr = 0; nr < bitmap_size; nr++) { in seccomp_cache_prepare_bitmap() 820 if (!test_bit(nr, bitmap)) in seccomp_cache_prepare_bitmap() 823 sd.nr = nr; in seccomp_cache_prepare_bitmap() 834 __clear_bit(nr, bitmap); in seccomp_cache_prepare_bitmap() 1346 this_syscall = sd ? sd->nr : in __secure_computing() [all …]
|
D | notifier.c | 126 int ret, nr = 0; in notifier_call_chain_robust() local 128 ret = notifier_call_chain(nl, val_up, v, -1, &nr); in notifier_call_chain_robust() 130 notifier_call_chain(nl, val_down, v, nr-1, NULL); in notifier_call_chain_robust()
|
D | pid_namespace.c | 171 int nr; in zap_pid_ns_processes() local 204 nr = 2; in zap_pid_ns_processes() 205 idr_for_each_entry_continue(&pid_ns->idr, pid, nr) { in zap_pid_ns_processes()
|
D | fork.c | 2078 pid_t nr = -1; in pidfd_show_fdinfo() local 2082 nr = pid_nr_ns(pid, ns); in pidfd_show_fdinfo() 2085 seq_put_decimal_ll(m, "Pid:\t", nr); in pidfd_show_fdinfo() 2088 seq_put_decimal_ll(m, "\nNSpid:\t", nr); in pidfd_show_fdinfo() 2089 if (nr > 0) { in pidfd_show_fdinfo() 2098 seq_put_decimal_ll(m, "\t", pid->numbers[i].nr); in pidfd_show_fdinfo() 2911 pid_t nr; in kernel_clone() local 2960 nr = pid_vnr(pid); in kernel_clone() 2963 put_user(nr, args->parent_tid); in kernel_clone() 2990 return nr; in kernel_clone()
|
D | ptrace.c | 746 if (arg.nr < 0) in ptrace_peek_siginfo() 758 for (i = 0; i < arg.nr; ) { in ptrace_peek_siginfo() 945 info->entry.nr = syscall_get_nr(child, regs); in ptrace_get_syscall_info_entry()
|
/kernel/bpf/ |
D | stackmap.c | 22 u32 nr; member 188 entry->nr = stack_trace_save_tsk(task, (unsigned long *)entry->ip, in get_callchain_entry_for_task() 201 for (i = entry->nr - 1; i >= 0; i--) in get_callchain_entry_for_task() 224 if (trace->nr <= skip) in __bpf_get_stackid() 228 trace_nr = trace->nr - skip; in __bpf_get_stackid() 246 new_bucket->nr = trace_nr; in __bpf_get_stackid() 251 if (hash_matches && bucket->nr == trace_nr && in __bpf_get_stackid() 261 if (hash_matches && bucket->nr == trace_nr && in __bpf_get_stackid() 275 new_bucket->nr = trace_nr; in __bpf_get_stackid() 323 while (nr_kernel < trace->nr) { in count_kernel_ip() [all …]
|
/kernel/locking/ |
D | lockdep_proc.c | 412 unsigned long nl, nr; in lock_stat_cmp() local 414 nl = dl->stats.read_waittime.nr + dl->stats.write_waittime.nr; in lock_stat_cmp() 415 nr = dr->stats.read_waittime.nr + dr->stats.write_waittime.nr; in lock_stat_cmp() 417 return nr - nl; in lock_stat_cmp() 431 static void snprint_time(char *buf, size_t bufsiz, s64 nr) in snprint_time() argument 436 nr += 5; /* for display rounding */ in snprint_time() 437 div = div_s64_rem(nr, 1000, &rem); in snprint_time() 451 seq_printf(m, "%14lu", lt->nr); in seq_lock_time() 455 seq_time(m, lt->nr ? div64_u64(lt->total, lt->nr) : 0); in seq_lock_time() 505 if (stats->write_holdtime.nr) { in seq_stats() [all …]
|
D | lockdep.c | 272 if (time < lt->min || !lt->nr) in lock_time_inc() 276 lt->nr++; in lock_time_inc() 281 if (!src->nr) in lock_time_add() 287 if (src->min < dst->min || !dst->nr) in lock_time_add() 291 dst->nr += src->nr; in lock_time_add()
|
/kernel/trace/ |
D | trace_syscalls.c | 104 static struct syscall_metadata *syscall_nr_to_meta(int nr) in syscall_nr_to_meta() argument 107 return xa_load(&syscalls_metadata_sparse, (unsigned long)nr); in syscall_nr_to_meta() 109 if (!syscalls_metadata || nr >= NR_syscalls || nr < 0) in syscall_nr_to_meta() 112 return syscalls_metadata[nr]; in syscall_nr_to_meta() 138 syscall = trace->nr; in print_syscall_enter() 184 syscall = trace->nr; in print_syscall_exit() 325 entry->nr = syscall_nr; in ftrace_syscall_enter() 362 entry->nr = syscall_nr; in ftrace_syscall_exit() 504 unsigned long __init __weak arch_syscall_addr(int nr) in arch_syscall_addr() argument 506 return (unsigned long)sys_call_table[nr]; in arch_syscall_addr() [all …]
|
D | trace_btf.c | 43 const struct btf_param *btf_get_func_param(const struct btf_type *func_proto, s32 *nr) in btf_get_func_param() argument 48 *nr = btf_type_vlen(func_proto); in btf_get_func_param() 49 if (*nr > 0) in btf_get_func_param()
|
D | trace_btf.h | 7 s32 *nr);
|
D | trace_probe.c | 443 s32 nr; in query_btf_context() local 459 nr = 0; in query_btf_context() 460 param = btf_get_func_param(type, &nr); in query_btf_context() 464 nr--; in query_btf_context() 469 if (nr > 0) { in query_btf_context() 470 ctx->nr_params = nr; in query_btf_context()
|
D | trace.h | 138 int nr; member 144 int nr; member
|
/kernel/rcu/ |
D | rcuscale.c | 580 int nr; in compute_real() local 583 nr = n; in compute_real() 585 nr = num_online_cpus() + 1 + n; in compute_real() 586 if (nr <= 0) in compute_real() 587 nr = 1; in compute_real() 589 return nr; in compute_real()
|
/kernel/events/ |
D | hw_breakpoint.c | 391 int nr; in max_bp_pinned_slots() local 393 nr = info->cpu_pinned; in max_bp_pinned_slots() 395 nr += max_task_bp_pinned(cpu, type); in max_bp_pinned_slots() 397 nr += task_bp_pinned(cpu, bp, type); in max_bp_pinned_slots() 399 pinned_slots = max(nr, pinned_slots); in max_bp_pinned_slots()
|
D | callchain.c | 193 ctx.nr = entry->nr = init_nr; in get_perf_callchain()
|
D | ring_buffer.c | 895 int i, nr; in rb_free_work() local 898 nr = data_page_nr(rb); in rb_free_work() 902 for (i = 0; i <= nr; i++) in rb_free_work()
|
D | core.c | 1325 u32 nr; in perf_event_pid_type() local 1332 nr = __task_pid_nr_ns(p, type, event->ns); in perf_event_pid_type() 1334 if (!nr && !pid_alive(p)) in perf_event_pid_type() 1335 nr = -1; in perf_event_pid_type() 1336 return nr; in perf_event_pid_type() 1821 int nr = 1; in __perf_event_read_size() local 1836 nr += nr_siblings; in __perf_event_read_size() 1844 return size + nr * entry; in __perf_event_read_size() 3709 itrs[heap->nr] = event; in __heap_add() 3710 heap->nr++; in __heap_add() [all …]
|
/kernel/entry/ |
D | common.c | 267 unsigned long nr = syscall_get_nr(current, regs); in syscall_exit_to_user_mode_prepare() local 272 if (WARN(irqs_disabled(), "syscall %lu left IRQs disabled", nr)) in syscall_exit_to_user_mode_prepare()
|
/kernel/irq/ |
D | irqdesc.c | 539 static int irq_expand_nr_irqs(unsigned int nr) in irq_expand_nr_irqs() argument 541 if (nr > MAX_SPARSE_IRQS) in irq_expand_nr_irqs() 543 nr_irqs = nr; in irq_expand_nr_irqs() 642 static int irq_expand_nr_irqs(unsigned int nr) in irq_expand_nr_irqs() argument
|
/kernel/sched/ |
D | wait.c | 173 void __wake_up_locked(struct wait_queue_head *wq_head, unsigned int mode, int nr) in __wake_up_locked() argument 175 __wake_up_common(wq_head, mode, nr, 0, NULL, NULL); in __wake_up_locked()
|
D | fair.c | 1805 unsigned long rate_limit, int nr) in numa_promotion_rate_limit() argument 1811 mod_node_page_state(pgdat, PGPROMOTE_CANDIDATE, nr); in numa_promotion_rate_limit() 4488 if (cfs_rq->removed.nr) { in update_cfs_rq_load_avg() 4496 cfs_rq->removed.nr = 0; in update_cfs_rq_load_avg() 4720 ++cfs_rq->removed.nr; in remove_entity_load_avg() 7332 int i, cpu, idle_cpu = -1, nr = INT_MAX; in select_idle_cpu() local 7366 nr = div_u64(span_avg, avg_cost); in select_idle_cpu() 7368 nr = 4; in select_idle_cpu() 7377 nr = READ_ONCE(sd_share->nr_idle_scan) + 1; in select_idle_cpu() 7379 if (nr == 1) in select_idle_cpu() [all …]
|
/kernel/time/ |
D | posix-timers.c | 70 static int hash(struct signal_struct *sig, unsigned int nr) in hash() argument 72 return hash_32(hash32_ptr(sig) ^ nr, HASH_BITS(posix_timers_hashtable)); in hash()
|