/kernel/bpf/ |
D | core.c | 159 if (i < pos && i + insn->off + 1 > pos) in bpf_adj_branches() 160 insn->off += delta; in bpf_adj_branches() 161 else if (i > pos + delta && i + insn->off + 1 <= pos + delta) in bpf_adj_branches() 162 insn->off -= delta; in bpf_adj_branches() 166 struct bpf_prog *bpf_patch_insn_single(struct bpf_prog *prog, u32 off, in bpf_patch_insn_single() argument 174 memcpy(prog->insnsi + off, patch, sizeof(*patch)); in bpf_patch_insn_single() 199 insn_rest = insn_adj_cnt - off - len; in bpf_patch_insn_single() 201 memmove(prog_adj->insnsi + off + len, prog_adj->insnsi + off + 1, in bpf_patch_insn_single() 203 memcpy(prog_adj->insnsi + off, patch, sizeof(*patch) * len); in bpf_patch_insn_single() 205 bpf_adj_branches(prog_adj, off, insn_delta); in bpf_patch_insn_single() [all …]
|
D | verifier.c | 209 reg->id, reg->off, reg->range); in print_verifier_state() 308 insn->off, insn->src_reg); in print_bpf_insn() 313 insn->dst_reg, insn->off, in print_bpf_insn() 326 insn->off, insn->imm); in print_bpf_insn() 335 insn->src_reg, insn->off); in print_bpf_insn() 371 insn->code, insn->off); in print_bpf_insn() 378 insn->src_reg, insn->off); in print_bpf_insn() 383 insn->imm, insn->off); in print_bpf_insn() 543 static int check_stack_write(struct bpf_verifier_state *state, int off, in check_stack_write() argument 561 state->spilled_regs[(MAX_BPF_STACK + off) / BPF_REG_SIZE] = in check_stack_write() [all …]
|
D | hashtab.c | 484 int off = 0, cpu; in alloc_htab_elem() local 488 value + off, size); in alloc_htab_elem() 489 off += size; in alloc_htab_elem() 614 int off = 0, cpu; in __htab_percpu_map_update_elem() local 619 value + off, size); in __htab_percpu_map_update_elem() 620 off += size; in __htab_percpu_map_update_elem() 750 int cpu, off = 0; in bpf_percpu_hash_copy() local 764 bpf_long_memcpy(value + off, in bpf_percpu_hash_copy() 766 off += size; in bpf_percpu_hash_copy()
|
D | arraymap.c | 172 int cpu, off = 0; in bpf_percpu_array_copy() local 186 bpf_long_memcpy(value + off, per_cpu_ptr(pptr, cpu), size); in bpf_percpu_array_copy() 187 off += size; in bpf_percpu_array_copy() 247 int cpu, off = 0; in bpf_percpu_array_update() local 272 bpf_long_memcpy(per_cpu_ptr(pptr, cpu), value + off, size); in bpf_percpu_array_update() 273 off += size; in bpf_percpu_array_update()
|
/kernel/ |
D | kallsyms.c | 95 static unsigned int kallsyms_expand_symbol(unsigned int off, in kallsyms_expand_symbol() argument 102 data = &kallsyms_names[off]; in kallsyms_expand_symbol() 110 off += len + 1; in kallsyms_expand_symbol() 139 return off; in kallsyms_expand_symbol() 146 static char kallsyms_get_symbol_type(unsigned int off) in kallsyms_get_symbol_type() argument 152 return kallsyms_token_table[kallsyms_token_index[kallsyms_names[off + 1]]]; in kallsyms_get_symbol_type() 205 unsigned int off; in kallsyms_lookup_name() local 207 for (i = 0, off = 0; i < kallsyms_num_syms; i++) { in kallsyms_lookup_name() 208 off = kallsyms_expand_symbol(off, namebuf, ARRAY_SIZE(namebuf)); in kallsyms_lookup_name() 223 unsigned int off; in kallsyms_on_each_symbol() local [all …]
|
D | kcov.c | 274 unsigned long size, off; in kcov_mmap() local 292 for (off = 0; off < size; off += PAGE_SIZE) { in kcov_mmap() 293 page = vmalloc_to_page(kcov->area + off); in kcov_mmap() 294 if (vm_insert_page(vma, vma->vm_start + off, page)) in kcov_mmap()
|
D | params.c | 510 int i, off, ret; in param_array_get() local 514 for (i = off = 0; i < (arr->num ? *arr->num : arr->max); i++) { in param_array_get() 516 buffer[off++] = ','; in param_array_get() 519 ret = arr->ops->get(buffer + off, &p); in param_array_get() 522 off += ret; in param_array_get() 524 buffer[off] = '\0'; in param_array_get() 525 return off; in param_array_get()
|
D | pid.c | 56 struct pidmap *map, int off) in mk_pid() argument 58 return (map - pid_ns->pidmap)*BITS_PER_PAGE + off; in mk_pid() 61 #define find_next_offset(map, off) \ argument 62 find_next_zero_bit((map)->page, BITS_PER_PAGE, off)
|
D | ksysfs.c | 192 char *buf, loff_t off, size_t count) in notes_read() argument 194 memcpy(buf, &__start_notes + off, count); in notes_read()
|
D | ptrace.c | 710 s32 off = arg.off + i; in ptrace_peek_siginfo() local 714 if (!off--) { in ptrace_peek_siginfo() 721 if (off >= 0) /* beyond the end of the list */ in ptrace_peek_siginfo()
|
D | cgroup_pids.c | 258 size_t nbytes, loff_t off) in pids_max_write() argument
|
D | cgroup_freezer.c | 420 char *buf, size_t nbytes, loff_t off) in freezer_write() argument
|
D | cgroup.c | 2893 size_t nbytes, loff_t off, bool threadgroup) in __cgroup_procs_write() argument 2989 char *buf, size_t nbytes, loff_t off) in cgroup_tasks_write() argument 2991 return __cgroup_procs_write(of, buf, nbytes, off, false); in cgroup_tasks_write() 2995 char *buf, size_t nbytes, loff_t off) in cgroup_procs_write() argument 2997 return __cgroup_procs_write(of, buf, nbytes, off, true); in cgroup_procs_write() 3001 char *buf, size_t nbytes, loff_t off) in cgroup_release_agent_write() argument 3384 loff_t off) in cgroup_subtree_control_write() argument 3509 size_t nbytes, loff_t off) in cgroup_file_write() argument 3517 return cft->write(of, buf, nbytes, off); in cgroup_file_write()
|
D | cpuset.c | 1687 char *buf, size_t nbytes, loff_t off) in cpuset_write_resmask() argument
|
/kernel/trace/ |
D | bpf_trace.c | 477 static bool kprobe_prog_is_valid_access(int off, int size, enum bpf_access_type type, in kprobe_prog_is_valid_access() argument 480 if (off < 0 || off >= sizeof(struct pt_regs)) in kprobe_prog_is_valid_access() 484 if (off % size != 0) in kprobe_prog_is_valid_access() 558 static bool tp_prog_is_valid_access(int off, int size, enum bpf_access_type type, in tp_prog_is_valid_access() argument 561 if (off < sizeof(void *) || off >= PERF_MAX_TRACE_SIZE) in tp_prog_is_valid_access() 565 if (off % size != 0) in tp_prog_is_valid_access() 580 static bool pe_prog_is_valid_access(int off, int size, enum bpf_access_type type, in pe_prog_is_valid_access() argument 583 if (off < 0 || off >= sizeof(struct bpf_perf_event_data)) in pe_prog_is_valid_access() 587 if (off % size != 0) in pe_prog_is_valid_access() 589 if (off == offsetof(struct bpf_perf_event_data, sample_period)) { in pe_prog_is_valid_access()
|
D | Kconfig | 175 bool "Interrupts-off Latency Tracer" 186 This option measures the time spent in irqs-off critical 196 enabled. This option and the preempt-off timing option can be 200 bool "Preemption-off Latency Tracer" 210 This option measures the time spent in preemption-off critical 220 enabled. This option and the irqs-off timing option can be 616 When the tracepoint is enabled, it kicks off a kernel thread that 663 kernel boot sequence, the test will start that kicks off
|
/kernel/power/ |
D | swap.c | 679 size_t off; in save_image_lzo() local 775 for (off = 0; off < LZO_UNC_SIZE; off += PAGE_SIZE) { in save_image_lzo() 783 memcpy(data[thr].unc + off, in save_image_lzo() 793 if (!off) in save_image_lzo() 796 data[thr].unc_len = off; in save_image_lzo() 840 for (off = 0; in save_image_lzo() 841 off < LZO_HEADER + data[thr].cmp_len; in save_image_lzo() 842 off += PAGE_SIZE) { in save_image_lzo() 843 memcpy(page, data[thr].cmp + off, PAGE_SIZE); in save_image_lzo() 1175 size_t off; in load_image_lzo() local [all …]
|
D | Kconfig | 17 This allows you to turn off the freezer for suspend. If this is 45 system and powers it off; and restores that checkpoint on reboot. 254 manpage ("man 8 hdparm") for that), and it doesn't turn off
|
/kernel/gcov/ |
D | gcc_4_7.c | 380 static size_t store_gcov_u32(void *buffer, size_t off, u32 v) in store_gcov_u32() argument 385 data = buffer + off; in store_gcov_u32() 404 static size_t store_gcov_u64(void *buffer, size_t off, u64 v) in store_gcov_u64() argument 409 data = buffer + off; in store_gcov_u64()
|
/kernel/debug/kdb/ |
D | kdb_main.c | 497 unsigned long off = 0; in kdbgetaddrarg() local 610 diag = kdbgetularg(cp, &off); in kdbgetaddrarg() 615 off = -off; in kdbgetaddrarg() 618 *offset += off; in kdbgetaddrarg() 621 *value += off; in kdbgetaddrarg()
|
/kernel/irq/ |
D | Kconfig | 33 # Support for generic irq migrating off cpu before the cpu is offline.
|
/kernel/time/ |
D | timeconst.bc | 29 a shift value will be correct in the signed integer range and off
|
D | Kconfig | 113 dynamically off.
|
/kernel/events/ |
D | core.c | 6783 unsigned long off = vma->vm_pgoff << PAGE_SHIFT, flags; in __perf_addr_filters_adjust() local 6796 if (perf_addr_filter_match(filter, file, off, in __perf_addr_filters_adjust() 7893 int off = trace_event_get_offsets(event->tp_event); in perf_event_set_bpf_prog() local 7895 if (prog->aux->max_ctx_offset > off) { in perf_event_set_bpf_prog() 8025 unsigned long off = vma->vm_pgoff << PAGE_SHIFT; in perf_addr_filter_apply() local 8031 if (!perf_addr_filter_match(filter, file, off, vma_size)) in perf_addr_filter_apply()
|