/kernel/ |
D | signal.c | 3184 int copy_siginfo_to_user(siginfo_t __user *to, const kernel_siginfo_t *from) in copy_siginfo_to_user() argument 3187 if (copy_to_user(to, from , sizeof(struct kernel_siginfo))) in copy_siginfo_to_user() 3195 const siginfo_t __user *from) in post_copy_siginfo_from_user() argument 3198 char __user *expansion = si_expansion(from); in post_copy_siginfo_from_user() 3218 const siginfo_t __user *from) in __copy_siginfo_from_user() argument 3220 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in __copy_siginfo_from_user() 3223 return post_copy_siginfo_from_user(to, from); in __copy_siginfo_from_user() 3226 int copy_siginfo_from_user(kernel_siginfo_t *to, const siginfo_t __user *from) in copy_siginfo_from_user() argument 3228 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in copy_siginfo_from_user() 3230 return post_copy_siginfo_from_user(to, from); in copy_siginfo_from_user() [all …]
|
D | softirq.c | 710 unsigned int __weak arch_dynirq_lower_bound(unsigned int from) in arch_dynirq_lower_bound() argument 712 return from; in arch_dynirq_lower_bound()
|
D | relay.c | 1136 void *from; in relay_file_read() local 1147 from = buf->start + read_start; in relay_file_read() 1149 if (copy_to_user(buffer, from, avail)) in relay_file_read()
|
D | workqueue.c | 3391 const struct workqueue_attrs *from) in copy_workqueue_attrs() argument 3393 to->nice = from->nice; in copy_workqueue_attrs() 3394 cpumask_copy(to->cpumask, from->cpumask); in copy_workqueue_attrs() 3400 to->no_numa = from->no_numa; in copy_workqueue_attrs()
|
/kernel/irq/ |
D | irqdesc.c | 741 void irq_free_descs(unsigned int from, unsigned int cnt) in irq_free_descs() argument 745 if (from >= nr_irqs || (from + cnt) > nr_irqs) in irq_free_descs() 750 free_desc(from + i); in irq_free_descs() 752 bitmap_clear(allocated_irqs, from, cnt); in irq_free_descs() 771 __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, in __irq_alloc_descs() argument 780 if (from > irq) in __irq_alloc_descs() 782 from = irq; in __irq_alloc_descs() 789 from = arch_dynirq_lower_bound(from); in __irq_alloc_descs() 795 from, cnt, 0); in __irq_alloc_descs() 850 void irq_free_hwirqs(unsigned int from, int cnt) in irq_free_hwirqs() argument [all …]
|
D | devres.c | 150 unsigned int from; member 158 irq_free_descs(this->from, this->cnt); in devm_irq_desc_release() 178 int __devm_irq_alloc_descs(struct device *dev, int irq, unsigned int from, in __devm_irq_alloc_descs() argument 189 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs() 195 dr->from = base; in __devm_irq_alloc_descs()
|
D | Kconfig | 34 # Support for delayed migration from interrupt context
|
/kernel/bpf/ |
D | core.c | 906 static int bpf_jit_blind_insn(const struct bpf_insn *from, in bpf_jit_blind_insn() argument 935 if (from->dst_reg == BPF_REG_AX || from->src_reg == BPF_REG_AX) in bpf_jit_blind_insn() 938 if (from->imm == 0 && in bpf_jit_blind_insn() 939 (from->code == (BPF_ALU | BPF_MOV | BPF_K) || in bpf_jit_blind_insn() 940 from->code == (BPF_ALU64 | BPF_MOV | BPF_K))) { in bpf_jit_blind_insn() 941 *to++ = BPF_ALU64_REG(BPF_XOR, from->dst_reg, from->dst_reg); in bpf_jit_blind_insn() 945 switch (from->code) { in bpf_jit_blind_insn() 955 *to++ = BPF_ALU32_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn() 957 *to++ = BPF_ALU32_REG(from->code, from->dst_reg, BPF_REG_AX); in bpf_jit_blind_insn() 969 *to++ = BPF_ALU64_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn() [all …]
|
D | btf.c | 183 #define for_each_member_from(i, from, struct_type, member) \ argument 184 for (i = from, member = btf_type_member(struct_type) + from; \ 193 #define for_each_vsi_from(i, from, struct_type, member) \ argument 194 for (i = from, member = btf_type_var_secinfo(struct_type) + from; \
|
/kernel/rcu/ |
D | Kconfig | 34 designed for UP systems from which real-time response 45 side-effect of preventing "make oldconfig" from asking you all 150 they have RCU callbacks queued, and prevents RCU from waking 169 This option also prevents heavy loads from blocking RCU 189 bool "Offload RCU callback processing from boot-selected CPUs" 199 This option offloads callback invocation from the set of CPUs 204 (!PREEMPT kernels). Nothing prevents this kthread from running
|
/kernel/time/ |
D | tick-common.c | 193 int from = tick_do_timer_boot_cpu; in tick_take_do_timer_from_boot() local 195 if (from >= 0 && from != cpu) in tick_take_do_timer_from_boot() 196 smp_call_function_single(from, giveup_do_timer, &cpu, 1); in tick_take_do_timer_from_boot()
|
D | clocksource.c | 45 clocks_calc_mult_shift(u32 *mult, u32 *shift, u32 from, u32 to, u32 maxsec) in clocks_calc_mult_shift() argument 54 tmp = ((u64)maxsec * from) >> 32; in clocks_calc_mult_shift() 66 tmp += from / 2; in clocks_calc_mult_shift() 67 do_div(tmp, from); in clocks_calc_mult_shift()
|
D | Kconfig | 16 # Architecture has extra clocksource init called from registration
|
/kernel/events/ |
D | hw_breakpoint.c | 452 struct perf_event_attr *from) in hw_breakpoint_copy_attr() argument 454 to->bp_addr = from->bp_addr; in hw_breakpoint_copy_attr() 455 to->bp_type = from->bp_type; in hw_breakpoint_copy_attr() 456 to->bp_len = from->bp_len; in hw_breakpoint_copy_attr() 457 to->disabled = from->disabled; in hw_breakpoint_copy_attr()
|
/kernel/power/ |
D | Kconfig | 29 of suspend, or they are content with invoking sync() from 52 from <http://suspend.sf.net>. 62 have it detect the saved image, restore memory state from it, and 180 fields of device objects from user space. If you are not a kernel 222 functions from <linux/resume-trace.h> as well as the 260 and the Battery Powered Linux mini-HOWTO, available from 320 smarter decisions. This config option enables the framework from
|
/kernel/gcov/ |
D | Kconfig | 26 To exclude files from being profiled even when CONFIG_GCOV_PROFILE_ALL 51 larger and run slower. Also be sure to exclude files from profiling
|
/kernel/sched/ |
D | OWNERS | 1 # include OWNERS from the authoritative android-mainline branch
|
/kernel/cgroup/ |
D | cgroup-v1.c | 61 int cgroup_attach_task_all(struct task_struct *from, struct task_struct *tsk) in cgroup_attach_task_all() argument 76 from_cgrp = task_cgroup_from_root(from, root); in cgroup_attach_task_all() 102 int cgroup_transfer_tasks(struct cgroup *to, struct cgroup *from) in cgroup_transfer_tasks() argument 123 list_for_each_entry(link, &from->cset_links, cset_link) in cgroup_transfer_tasks() 136 css_task_iter_start(&from->self, 0, &it); in cgroup_transfer_tasks()
|
D | cpuset.c | 1600 nodemask_t from; member 1610 do_migrate_pages(mwork->mm, &mwork->from, &mwork->to, MPOL_MF_MOVE_ALL); in cpuset_migrate_mm_workfn() 1615 static void cpuset_migrate_mm(struct mm_struct *mm, const nodemask_t *from, in cpuset_migrate_mm() argument 1623 mwork->from = *from; in cpuset_migrate_mm()
|
D | rdma.c | 373 argstr.from = value; in parse_resource()
|
/kernel/configs/ |
D | xen.config | 16 # pretty useless systems starting from allnoconfig
|
/kernel/trace/ |
D | Kconfig | 245 to be scheduled in, starting from the point it has woken up. 391 conditions. This prevents the tracers themselves from being 446 support tools needed), fetch the blktrace tools from: 478 bool "Do NOT protect notrace function from kprobe events" 487 functions are protected from kprobe-events to prevent an infinit 627 debugging and reverse engineering. It is called from the ioremap
|
D | blktrace.c | 1056 dev_t dev, sector_t from) in blk_add_trace_bio_remap() argument 1070 r.sector_from = cpu_to_be64(from); in blk_add_trace_bio_remap() 1095 sector_t from) in blk_add_trace_rq_remap() argument 1109 r.sector_from = cpu_to_be64(from); in blk_add_trace_rq_remap()
|
/kernel/dma/ |
D | Kconfig | 188 unexpected behaviour from DMA API implementations if the scatterlist
|
/kernel/printk/ |
D | printk.c | 821 static ssize_t devkmsg_write(struct kiocb *iocb, struct iov_iter *from) in devkmsg_write() argument 828 size_t len = iov_iter_count(from); in devkmsg_write() 849 if (!copy_from_iter_full(buf, len, from)) { in devkmsg_write()
|