/kernel/ |
D | signal.c | 3236 int copy_siginfo_to_user(siginfo_t __user *to, const kernel_siginfo_t *from) in copy_siginfo_to_user() argument 3239 if (copy_to_user(to, from , sizeof(struct kernel_siginfo))) in copy_siginfo_to_user() 3247 const siginfo_t __user *from) in post_copy_siginfo_from_user() argument 3250 char __user *expansion = si_expansion(from); in post_copy_siginfo_from_user() 3270 const siginfo_t __user *from) in __copy_siginfo_from_user() argument 3272 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in __copy_siginfo_from_user() 3275 return post_copy_siginfo_from_user(to, from); in __copy_siginfo_from_user() 3278 int copy_siginfo_from_user(kernel_siginfo_t *to, const siginfo_t __user *from) in copy_siginfo_from_user() argument 3280 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in copy_siginfo_from_user() 3282 return post_copy_siginfo_from_user(to, from); in copy_siginfo_from_user() [all …]
|
D | softirq.c | 774 unsigned int __weak arch_dynirq_lower_bound(unsigned int from) in arch_dynirq_lower_bound() argument 776 return from; in arch_dynirq_lower_bound()
|
D | relay.c | 1136 void *from; in relay_file_read() local 1147 from = buf->start + read_start; in relay_file_read() 1149 if (copy_to_user(buffer, from, avail)) in relay_file_read()
|
/kernel/irq/ |
D | irqdesc.c | 754 void irq_free_descs(unsigned int from, unsigned int cnt) in irq_free_descs() argument 758 if (from >= nr_irqs || (from + cnt) > nr_irqs) in irq_free_descs() 763 free_desc(from + i); in irq_free_descs() 765 bitmap_clear(allocated_irqs, from, cnt); in irq_free_descs() 784 __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, in __irq_alloc_descs() argument 793 if (from > irq) in __irq_alloc_descs() 795 from = irq; in __irq_alloc_descs() 802 from = arch_dynirq_lower_bound(from); in __irq_alloc_descs() 808 from, cnt, 0); in __irq_alloc_descs() 863 void irq_free_hwirqs(unsigned int from, int cnt) in irq_free_hwirqs() argument [all …]
|
D | devres.c | 150 unsigned int from; member 158 irq_free_descs(this->from, this->cnt); in devm_irq_desc_release() 178 int __devm_irq_alloc_descs(struct device *dev, int irq, unsigned int from, in __devm_irq_alloc_descs() argument 189 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs() 195 dr->from = base; in __devm_irq_alloc_descs()
|
D | Kconfig | 34 # Support for delayed migration from interrupt context
|
/kernel/events/ |
D | ring_buffer.c | 572 unsigned long from, unsigned long to) in perf_output_copy_aux() argument 578 from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1; in perf_output_copy_aux() 582 tocopy = PAGE_SIZE - offset_in_page(from); in perf_output_copy_aux() 583 if (to > from) in perf_output_copy_aux() 584 tocopy = min(tocopy, to - from); in perf_output_copy_aux() 588 addr = rb->aux_pages[from >> PAGE_SHIFT]; in perf_output_copy_aux() 589 addr += offset_in_page(from); in perf_output_copy_aux() 596 from += tocopy; in perf_output_copy_aux() 597 from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1; in perf_output_copy_aux() 598 } while (to != from); in perf_output_copy_aux()
|
D | hw_breakpoint.c | 468 struct perf_event_attr *from) in hw_breakpoint_copy_attr() argument 470 to->bp_addr = from->bp_addr; in hw_breakpoint_copy_attr() 471 to->bp_type = from->bp_type; in hw_breakpoint_copy_attr() 472 to->bp_len = from->bp_len; in hw_breakpoint_copy_attr() 473 to->disabled = from->disabled; in hw_breakpoint_copy_attr()
|
/kernel/printk/ |
D | printk_ringbuffer.h | 355 #define prb_for_each_record(from, rb, s, r) \ argument 356 for ((s) = from; prb_read_valid(rb, s, r); (s) = (r)->info->seq + 1) 373 #define prb_for_each_info(from, rb, s, i, lc) \ argument 374 for ((s) = from; prb_read_valid_info(rb, s, i, lc); (s) = (i)->seq + 1)
|
/kernel/bpf/ |
D | core.c | 970 static int bpf_jit_blind_insn(const struct bpf_insn *from, in bpf_jit_blind_insn() argument 999 if (from->dst_reg == BPF_REG_AX || from->src_reg == BPF_REG_AX) in bpf_jit_blind_insn() 1002 if (from->imm == 0 && in bpf_jit_blind_insn() 1003 (from->code == (BPF_ALU | BPF_MOV | BPF_K) || in bpf_jit_blind_insn() 1004 from->code == (BPF_ALU64 | BPF_MOV | BPF_K))) { in bpf_jit_blind_insn() 1005 *to++ = BPF_ALU64_REG(BPF_XOR, from->dst_reg, from->dst_reg); in bpf_jit_blind_insn() 1009 switch (from->code) { in bpf_jit_blind_insn() 1019 *to++ = BPF_ALU32_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn() 1021 *to++ = BPF_ALU32_REG(from->code, from->dst_reg, BPF_REG_AX); in bpf_jit_blind_insn() 1033 *to++ = BPF_ALU64_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn() [all …]
|
D | dispatcher.c | 132 void bpf_dispatcher_change_prog(struct bpf_dispatcher *d, struct bpf_prog *from, in bpf_dispatcher_change_prog() argument 138 if (from == to) in bpf_dispatcher_change_prog() 150 changed |= bpf_dispatcher_remove_prog(d, from); in bpf_dispatcher_change_prog()
|
D | bpf_iter.c | 101 err = copy_to_user(buf, seq->buf + seq->from, n); in bpf_seq_read() 107 seq->from += n; in bpf_seq_read() 112 seq->from = 0; in bpf_seq_read() 208 seq->from = n; in bpf_seq_read()
|
D | stackmap.c | 380 unsigned long *from = (unsigned long *) entry->ip; in get_callchain_entry_for_task() local 386 to[i] = (u64)(from[i]); in get_callchain_entry_for_task()
|
/kernel/time/ |
D | tick-common.c | 195 int from = tick_do_timer_boot_cpu; in tick_take_do_timer_from_boot() local 197 if (from >= 0 && from != cpu) in tick_take_do_timer_from_boot() 198 smp_call_function_single(from, giveup_do_timer, &cpu, 1); in tick_take_do_timer_from_boot()
|
D | Kconfig | 16 # Architecture has extra clocksource init called from registration 55 # Select to handle posix CPU timers from task_work 56 # and not from the timer interrupt context
|
D | clocksource.c | 45 clocks_calc_mult_shift(u32 *mult, u32 *shift, u32 from, u32 to, u32 maxsec) in clocks_calc_mult_shift() argument 54 tmp = ((u64)maxsec * from) >> 32; in clocks_calc_mult_shift() 66 tmp += from / 2; in clocks_calc_mult_shift() 67 do_div(tmp, from); in clocks_calc_mult_shift()
|
/kernel/rcu/ |
D | Kconfig | 35 designed for UP systems from which real-time response 46 side-effect of preventing "make oldconfig" from asking you all 178 they have RCU callbacks queued, and prevents RCU from waking 197 This option also prevents heavy loads from blocking RCU 217 bool "Offload RCU callback processing from boot-selected CPUs" 227 This option offloads callback invocation from the set of CPUs 232 (!PREEMPTION kernels). Nothing prevents this kthread from running
|
/kernel/power/ |
D | Kconfig | 29 of suspend, or they are content with invoking sync() from 55 from <http://suspend.sf.net>. 65 have it detect the saved image, restore memory state from it, and 90 Say N if no snapshotting from userspace is needed, this also 192 fields of device objects from user space. If you are not a kernel 234 functions from <linux/resume-trace.h> as well as the 272 and the Battery Powered Linux mini-HOWTO, available from 332 smarter decisions. This config option enables the framework from
|
/kernel/gcov/ |
D | Kconfig | 26 To exclude files from being profiled even when CONFIG_GCOV_PROFILE_ALL 51 larger and run slower. Also be sure to exclude files from profiling
|
/kernel/cgroup/ |
D | cgroup-v1.c | 55 int cgroup_attach_task_all(struct task_struct *from, struct task_struct *tsk) in cgroup_attach_task_all() argument 70 from_cgrp = task_cgroup_from_root(from, root); in cgroup_attach_task_all() 96 int cgroup_transfer_tasks(struct cgroup *to, struct cgroup *from) in cgroup_transfer_tasks() argument 117 list_for_each_entry(link, &from->cset_links, cset_link) in cgroup_transfer_tasks() 130 css_task_iter_start(&from->self, 0, &it); in cgroup_transfer_tasks()
|
D | cpuset.c | 1681 nodemask_t from; member 1691 do_migrate_pages(mwork->mm, &mwork->from, &mwork->to, MPOL_MF_MOVE_ALL); in cpuset_migrate_mm_workfn() 1696 static void cpuset_migrate_mm(struct mm_struct *mm, const nodemask_t *from, in cpuset_migrate_mm() argument 1704 mwork->from = *from; in cpuset_migrate_mm()
|
/kernel/bpf/preload/ |
D | Kconfig | 10 # The dependency on !COMPILE_TEST prevents it from being enabled
|
/kernel/configs/ |
D | xen.config | 16 # pretty useless systems starting from allnoconfig
|
/kernel/trace/ |
D | Kconfig | 328 to be scheduled in, starting from the point it has woken up. 371 debugging and reverse engineering. It is called from the ioremap 487 conditions. This prevents the tracers themselves from being 518 support tools needed), fetch the blktrace tools from: 550 bool "Do NOT protect notrace function from kprobe events" 559 functions are protected from kprobe-events to prevent an infinit 657 used to combine data from other trace events or in fact any
|
D | blktrace.c | 1054 dev_t dev, sector_t from) in blk_add_trace_bio_remap() argument 1068 r.sector_from = cpu_to_be64(from); in blk_add_trace_bio_remap() 1093 sector_t from) in blk_add_trace_rq_remap() argument 1107 r.sector_from = cpu_to_be64(from); in blk_add_trace_rq_remap()
|