/kernel/ |
D | signal.c | 3360 int copy_siginfo_to_user(siginfo_t __user *to, const kernel_siginfo_t *from) in copy_siginfo_to_user() argument 3363 if (copy_to_user(to, from , sizeof(struct kernel_siginfo))) in copy_siginfo_to_user() 3371 const siginfo_t __user *from) in post_copy_siginfo_from_user() argument 3374 char __user *expansion = si_expansion(from); in post_copy_siginfo_from_user() 3394 const siginfo_t __user *from) in __copy_siginfo_from_user() argument 3396 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in __copy_siginfo_from_user() 3399 return post_copy_siginfo_from_user(to, from); in __copy_siginfo_from_user() 3402 int copy_siginfo_from_user(kernel_siginfo_t *to, const siginfo_t __user *from) in copy_siginfo_from_user() argument 3404 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in copy_siginfo_from_user() 3406 return post_copy_siginfo_from_user(to, from); in copy_siginfo_from_user() [all …]
|
D | softirq.c | 1012 unsigned int __weak arch_dynirq_lower_bound(unsigned int from) in arch_dynirq_lower_bound() argument 1014 return from; in arch_dynirq_lower_bound()
|
D | relay.c | 1051 void *from; in relay_file_read() local 1062 from = buf->start + read_start; in relay_file_read() 1064 if (copy_to_user(buffer, from, avail)) in relay_file_read()
|
/kernel/irq/ |
D | devres.c | 150 unsigned int from; member 158 irq_free_descs(this->from, this->cnt); in devm_irq_desc_release() 178 int __devm_irq_alloc_descs(struct device *dev, int irq, unsigned int from, in __devm_irq_alloc_descs() argument 189 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs() 195 dr->from = base; in __devm_irq_alloc_descs()
|
D | irqdesc.c | 767 void irq_free_descs(unsigned int from, unsigned int cnt) in irq_free_descs() argument 771 if (from >= nr_irqs || (from + cnt) > nr_irqs) in irq_free_descs() 776 free_desc(from + i); in irq_free_descs() 778 bitmap_clear(allocated_irqs, from, cnt); in irq_free_descs() 797 __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, in __irq_alloc_descs() argument 806 if (from > irq) in __irq_alloc_descs() 808 from = irq; in __irq_alloc_descs() 815 from = arch_dynirq_lower_bound(from); in __irq_alloc_descs() 821 from, cnt, 0); in __irq_alloc_descs()
|
D | Kconfig | 29 # Support for delayed migration from interrupt context
|
/kernel/events/ |
D | ring_buffer.c | 572 unsigned long from, unsigned long to) in perf_output_copy_aux() argument 578 from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1; in perf_output_copy_aux() 582 tocopy = PAGE_SIZE - offset_in_page(from); in perf_output_copy_aux() 583 if (to > from) in perf_output_copy_aux() 584 tocopy = min(tocopy, to - from); in perf_output_copy_aux() 588 addr = rb->aux_pages[from >> PAGE_SHIFT]; in perf_output_copy_aux() 589 addr += offset_in_page(from); in perf_output_copy_aux() 596 from += tocopy; in perf_output_copy_aux() 597 from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1; in perf_output_copy_aux() 598 } while (to != from); in perf_output_copy_aux()
|
D | hw_breakpoint.c | 469 struct perf_event_attr *from) in hw_breakpoint_copy_attr() argument 471 to->bp_addr = from->bp_addr; in hw_breakpoint_copy_attr() 472 to->bp_type = from->bp_type; in hw_breakpoint_copy_attr() 473 to->bp_len = from->bp_len; in hw_breakpoint_copy_attr() 474 to->disabled = from->disabled; in hw_breakpoint_copy_attr()
|
/kernel/printk/ |
D | printk_ringbuffer.h | 355 #define prb_for_each_record(from, rb, s, r) \ argument 356 for ((s) = from; prb_read_valid(rb, s, r); (s) = (r)->info->seq + 1) 373 #define prb_for_each_info(from, rb, s, i, lc) \ argument 374 for ((s) = from; prb_read_valid_info(rb, s, i, lc); (s) = (i)->seq + 1)
|
/kernel/bpf/ |
D | core.c | 969 static int bpf_jit_blind_insn(const struct bpf_insn *from, in bpf_jit_blind_insn() argument 998 if (from->dst_reg == BPF_REG_AX || from->src_reg == BPF_REG_AX) in bpf_jit_blind_insn() 1001 if (from->imm == 0 && in bpf_jit_blind_insn() 1002 (from->code == (BPF_ALU | BPF_MOV | BPF_K) || in bpf_jit_blind_insn() 1003 from->code == (BPF_ALU64 | BPF_MOV | BPF_K))) { in bpf_jit_blind_insn() 1004 *to++ = BPF_ALU64_REG(BPF_XOR, from->dst_reg, from->dst_reg); in bpf_jit_blind_insn() 1008 switch (from->code) { in bpf_jit_blind_insn() 1018 *to++ = BPF_ALU32_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn() 1020 *to++ = BPF_ALU32_REG(from->code, from->dst_reg, BPF_REG_AX); in bpf_jit_blind_insn() 1032 *to++ = BPF_ALU64_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn() [all …]
|
D | dispatcher.c | 132 void bpf_dispatcher_change_prog(struct bpf_dispatcher *d, struct bpf_prog *from, in bpf_dispatcher_change_prog() argument 138 if (from == to) in bpf_dispatcher_change_prog() 150 changed |= bpf_dispatcher_remove_prog(d, from); in bpf_dispatcher_change_prog()
|
D | bpf_iter.c | 111 err = copy_to_user(buf, seq->buf + seq->from, n); in bpf_seq_read() 117 seq->from += n; in bpf_seq_read() 122 seq->from = 0; in bpf_seq_read() 222 seq->from = n; in bpf_seq_read()
|
D | Kconfig | 72 disable it by setting it to 1 (from which no other transition to
|
/kernel/time/ |
D | tick-common.c | 197 int from = tick_do_timer_boot_cpu; in tick_take_do_timer_from_boot() local 199 if (from >= 0 && from != cpu) in tick_take_do_timer_from_boot() 200 smp_call_function_single(from, giveup_do_timer, &cpu, 1); in tick_take_do_timer_from_boot()
|
D | Kconfig | 16 # Architecture has extra clocksource init called from registration 51 # Select to handle posix CPU timers from task_work 52 # and not from the timer interrupt context
|
D | clocksource.c | 47 clocks_calc_mult_shift(u32 *mult, u32 *shift, u32 from, u32 to, u32 maxsec) in clocks_calc_mult_shift() argument 56 tmp = ((u64)maxsec * from) >> 32; in clocks_calc_mult_shift() 68 tmp += from / 2; in clocks_calc_mult_shift() 69 do_div(tmp, from); in clocks_calc_mult_shift()
|
/kernel/rcu/ |
D | rcu_segcblist.c | 122 static void rcu_segcblist_move_seglen(struct rcu_segcblist *rsclp, int from, int to) in rcu_segcblist_move_seglen() argument 126 if (from == to) in rcu_segcblist_move_seglen() 129 len = rcu_segcblist_get_seglen(rsclp, from); in rcu_segcblist_move_seglen() 134 rcu_segcblist_set_seglen(rsclp, from, 0); in rcu_segcblist_move_seglen()
|
D | Kconfig | 35 designed for UP systems from which real-time response 46 side-effect of preventing "make oldconfig" from asking you all 179 they have RCU callbacks queued, and prevents RCU from waking 198 This option also prevents heavy loads from blocking RCU 232 bool "Offload RCU callback processing from boot-selected CPUs" 245 This option offloads callback invocation from the set of CPUs 251 from running on the specified CPUs, but (1) the kthreads may be
|
/kernel/power/ |
D | Kconfig | 29 of suspend, or they are content with invoking sync() from 55 from <http://suspend.sf.net>. 65 have it detect the saved image, restore memory state from it, and 90 Say N if no snapshotting from userspace is needed, this also 190 fields of device objects from user space. If you are not a kernel 232 functions from <linux/resume-trace.h> as well as the 270 and the Battery Powered Linux mini-HOWTO, available from 329 from which subsystems can access the energy models.
|
/kernel/gcov/ |
D | Kconfig | 28 To exclude files from being profiled even when CONFIG_GCOV_PROFILE_ALL 53 larger and run slower. Also be sure to exclude files from profiling
|
/kernel/cgroup/ |
D | cgroup-v1.c | 57 int cgroup_attach_task_all(struct task_struct *from, struct task_struct *tsk) in cgroup_attach_task_all() argument 72 from_cgrp = task_cgroup_from_root(from, root); in cgroup_attach_task_all() 100 int cgroup_transfer_tasks(struct cgroup *to, struct cgroup *from) in cgroup_transfer_tasks() argument 121 list_for_each_entry(link, &from->cset_links, cset_link) in cgroup_transfer_tasks() 134 css_task_iter_start(&from->self, 0, &it); in cgroup_transfer_tasks()
|
D | cpuset.c | 1703 nodemask_t from; member 1713 do_migrate_pages(mwork->mm, &mwork->from, &mwork->to, MPOL_MF_MOVE_ALL); in cpuset_migrate_mm_workfn() 1718 static void cpuset_migrate_mm(struct mm_struct *mm, const nodemask_t *from, in cpuset_migrate_mm() argument 1723 if (nodes_equal(*from, *to)) { in cpuset_migrate_mm() 1731 mwork->from = *from; in cpuset_migrate_mm()
|
/kernel/bpf/preload/ |
D | Kconfig | 10 # The dependency on !COMPILE_TEST prevents it from being enabled
|
/kernel/configs/ |
D | xen.config | 16 # pretty useless systems starting from allnoconfig
|
/kernel/trace/ |
D | Kconfig | 37 If this is set, then arguments and stack can be found from 326 to be scheduled in, starting from the point it has woken up. 384 without any interference from the operating system level, the 422 events can be used to trace the source of interference from NMI, 433 debugging and reverse engineering. It is called from the ioremap 549 conditions. This prevents the tracers themselves from being 580 support tools needed), fetch the blktrace tools from: 612 bool "Do NOT protect notrace function from kprobe events" 621 functions are protected from kprobe-events to prevent an infinite 719 used to combine data from other trace events or in fact any
|