Home
last modified time | relevance | path

Searched refs:from (Results 1 – 25 of 35) sorted by relevance

12

/kernel/
Dsignal.c3236 int copy_siginfo_to_user(siginfo_t __user *to, const kernel_siginfo_t *from) in copy_siginfo_to_user() argument
3239 if (copy_to_user(to, from , sizeof(struct kernel_siginfo))) in copy_siginfo_to_user()
3247 const siginfo_t __user *from) in post_copy_siginfo_from_user() argument
3250 char __user *expansion = si_expansion(from); in post_copy_siginfo_from_user()
3270 const siginfo_t __user *from) in __copy_siginfo_from_user() argument
3272 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in __copy_siginfo_from_user()
3275 return post_copy_siginfo_from_user(to, from); in __copy_siginfo_from_user()
3278 int copy_siginfo_from_user(kernel_siginfo_t *to, const siginfo_t __user *from) in copy_siginfo_from_user() argument
3280 if (copy_from_user(to, from, sizeof(struct kernel_siginfo))) in copy_siginfo_from_user()
3282 return post_copy_siginfo_from_user(to, from); in copy_siginfo_from_user()
[all …]
Dsoftirq.c774 unsigned int __weak arch_dynirq_lower_bound(unsigned int from) in arch_dynirq_lower_bound() argument
776 return from; in arch_dynirq_lower_bound()
Drelay.c1136 void *from; in relay_file_read() local
1147 from = buf->start + read_start; in relay_file_read()
1149 if (copy_to_user(buffer, from, avail)) in relay_file_read()
/kernel/irq/
Dirqdesc.c754 void irq_free_descs(unsigned int from, unsigned int cnt) in irq_free_descs() argument
758 if (from >= nr_irqs || (from + cnt) > nr_irqs) in irq_free_descs()
763 free_desc(from + i); in irq_free_descs()
765 bitmap_clear(allocated_irqs, from, cnt); in irq_free_descs()
784 __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, in __irq_alloc_descs() argument
793 if (from > irq) in __irq_alloc_descs()
795 from = irq; in __irq_alloc_descs()
802 from = arch_dynirq_lower_bound(from); in __irq_alloc_descs()
808 from, cnt, 0); in __irq_alloc_descs()
863 void irq_free_hwirqs(unsigned int from, int cnt) in irq_free_hwirqs() argument
[all …]
Ddevres.c150 unsigned int from; member
158 irq_free_descs(this->from, this->cnt); in devm_irq_desc_release()
178 int __devm_irq_alloc_descs(struct device *dev, int irq, unsigned int from, in __devm_irq_alloc_descs() argument
189 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs()
195 dr->from = base; in __devm_irq_alloc_descs()
DKconfig34 # Support for delayed migration from interrupt context
/kernel/events/
Dring_buffer.c572 unsigned long from, unsigned long to) in perf_output_copy_aux() argument
578 from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1; in perf_output_copy_aux()
582 tocopy = PAGE_SIZE - offset_in_page(from); in perf_output_copy_aux()
583 if (to > from) in perf_output_copy_aux()
584 tocopy = min(tocopy, to - from); in perf_output_copy_aux()
588 addr = rb->aux_pages[from >> PAGE_SHIFT]; in perf_output_copy_aux()
589 addr += offset_in_page(from); in perf_output_copy_aux()
596 from += tocopy; in perf_output_copy_aux()
597 from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1; in perf_output_copy_aux()
598 } while (to != from); in perf_output_copy_aux()
Dhw_breakpoint.c468 struct perf_event_attr *from) in hw_breakpoint_copy_attr() argument
470 to->bp_addr = from->bp_addr; in hw_breakpoint_copy_attr()
471 to->bp_type = from->bp_type; in hw_breakpoint_copy_attr()
472 to->bp_len = from->bp_len; in hw_breakpoint_copy_attr()
473 to->disabled = from->disabled; in hw_breakpoint_copy_attr()
/kernel/printk/
Dprintk_ringbuffer.h355 #define prb_for_each_record(from, rb, s, r) \ argument
356 for ((s) = from; prb_read_valid(rb, s, r); (s) = (r)->info->seq + 1)
373 #define prb_for_each_info(from, rb, s, i, lc) \ argument
374 for ((s) = from; prb_read_valid_info(rb, s, i, lc); (s) = (i)->seq + 1)
/kernel/bpf/
Dcore.c970 static int bpf_jit_blind_insn(const struct bpf_insn *from, in bpf_jit_blind_insn() argument
999 if (from->dst_reg == BPF_REG_AX || from->src_reg == BPF_REG_AX) in bpf_jit_blind_insn()
1002 if (from->imm == 0 && in bpf_jit_blind_insn()
1003 (from->code == (BPF_ALU | BPF_MOV | BPF_K) || in bpf_jit_blind_insn()
1004 from->code == (BPF_ALU64 | BPF_MOV | BPF_K))) { in bpf_jit_blind_insn()
1005 *to++ = BPF_ALU64_REG(BPF_XOR, from->dst_reg, from->dst_reg); in bpf_jit_blind_insn()
1009 switch (from->code) { in bpf_jit_blind_insn()
1019 *to++ = BPF_ALU32_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn()
1021 *to++ = BPF_ALU32_REG(from->code, from->dst_reg, BPF_REG_AX); in bpf_jit_blind_insn()
1033 *to++ = BPF_ALU64_IMM(BPF_MOV, BPF_REG_AX, imm_rnd ^ from->imm); in bpf_jit_blind_insn()
[all …]
Ddispatcher.c132 void bpf_dispatcher_change_prog(struct bpf_dispatcher *d, struct bpf_prog *from, in bpf_dispatcher_change_prog() argument
138 if (from == to) in bpf_dispatcher_change_prog()
150 changed |= bpf_dispatcher_remove_prog(d, from); in bpf_dispatcher_change_prog()
Dbpf_iter.c101 err = copy_to_user(buf, seq->buf + seq->from, n); in bpf_seq_read()
107 seq->from += n; in bpf_seq_read()
112 seq->from = 0; in bpf_seq_read()
208 seq->from = n; in bpf_seq_read()
Dstackmap.c380 unsigned long *from = (unsigned long *) entry->ip; in get_callchain_entry_for_task() local
386 to[i] = (u64)(from[i]); in get_callchain_entry_for_task()
/kernel/time/
Dtick-common.c195 int from = tick_do_timer_boot_cpu; in tick_take_do_timer_from_boot() local
197 if (from >= 0 && from != cpu) in tick_take_do_timer_from_boot()
198 smp_call_function_single(from, giveup_do_timer, &cpu, 1); in tick_take_do_timer_from_boot()
DKconfig16 # Architecture has extra clocksource init called from registration
55 # Select to handle posix CPU timers from task_work
56 # and not from the timer interrupt context
Dclocksource.c45 clocks_calc_mult_shift(u32 *mult, u32 *shift, u32 from, u32 to, u32 maxsec) in clocks_calc_mult_shift() argument
54 tmp = ((u64)maxsec * from) >> 32; in clocks_calc_mult_shift()
66 tmp += from / 2; in clocks_calc_mult_shift()
67 do_div(tmp, from); in clocks_calc_mult_shift()
/kernel/rcu/
DKconfig35 designed for UP systems from which real-time response
46 side-effect of preventing "make oldconfig" from asking you all
178 they have RCU callbacks queued, and prevents RCU from waking
197 This option also prevents heavy loads from blocking RCU
217 bool "Offload RCU callback processing from boot-selected CPUs"
227 This option offloads callback invocation from the set of CPUs
232 (!PREEMPTION kernels). Nothing prevents this kthread from running
/kernel/power/
DKconfig29 of suspend, or they are content with invoking sync() from
55 from <http://suspend.sf.net>.
65 have it detect the saved image, restore memory state from it, and
90 Say N if no snapshotting from userspace is needed, this also
192 fields of device objects from user space. If you are not a kernel
234 functions from <linux/resume-trace.h> as well as the
272 and the Battery Powered Linux mini-HOWTO, available from
332 smarter decisions. This config option enables the framework from
/kernel/gcov/
DKconfig26 To exclude files from being profiled even when CONFIG_GCOV_PROFILE_ALL
51 larger and run slower. Also be sure to exclude files from profiling
/kernel/cgroup/
Dcgroup-v1.c55 int cgroup_attach_task_all(struct task_struct *from, struct task_struct *tsk) in cgroup_attach_task_all() argument
70 from_cgrp = task_cgroup_from_root(from, root); in cgroup_attach_task_all()
96 int cgroup_transfer_tasks(struct cgroup *to, struct cgroup *from) in cgroup_transfer_tasks() argument
117 list_for_each_entry(link, &from->cset_links, cset_link) in cgroup_transfer_tasks()
130 css_task_iter_start(&from->self, 0, &it); in cgroup_transfer_tasks()
Dcpuset.c1681 nodemask_t from; member
1691 do_migrate_pages(mwork->mm, &mwork->from, &mwork->to, MPOL_MF_MOVE_ALL); in cpuset_migrate_mm_workfn()
1696 static void cpuset_migrate_mm(struct mm_struct *mm, const nodemask_t *from, in cpuset_migrate_mm() argument
1704 mwork->from = *from; in cpuset_migrate_mm()
/kernel/bpf/preload/
DKconfig10 # The dependency on !COMPILE_TEST prevents it from being enabled
/kernel/configs/
Dxen.config16 # pretty useless systems starting from allnoconfig
/kernel/trace/
DKconfig328 to be scheduled in, starting from the point it has woken up.
371 debugging and reverse engineering. It is called from the ioremap
487 conditions. This prevents the tracers themselves from being
518 support tools needed), fetch the blktrace tools from:
550 bool "Do NOT protect notrace function from kprobe events"
559 functions are protected from kprobe-events to prevent an infinit
657 used to combine data from other trace events or in fact any
Dblktrace.c1054 dev_t dev, sector_t from) in blk_add_trace_bio_remap() argument
1068 r.sector_from = cpu_to_be64(from); in blk_add_trace_bio_remap()
1093 sector_t from) in blk_add_trace_rq_remap() argument
1107 r.sector_from = cpu_to_be64(from); in blk_add_trace_rq_remap()

12