Home
last modified time | relevance | path

Searched refs:entry (Results 1 – 25 of 65) sorted by relevance

123

/kernel/dma/
Ddebug.c166 static inline void dump_entry_trace(struct dma_debug_entry *entry) in dump_entry_trace() argument
169 if (entry) { in dump_entry_trace()
171 stack_trace_print(entry->stack_entries, entry->stack_len, 0); in dump_entry_trace()
217 #define err_printk(dev, entry, format, arg...) do { \ argument
224 dump_entry_trace(entry); \
236 static int hash_fn(struct dma_debug_entry *entry) in hash_fn() argument
242 return (entry->dev_addr >> HASH_FN_SHIFT) & HASH_FN_MASK; in hash_fn()
248 static struct hash_bucket *get_hash_bucket(struct dma_debug_entry *entry, in get_hash_bucket() argument
252 int idx = hash_fn(entry); in get_hash_bucket()
296 struct dma_debug_entry *entry, *ret = NULL; in __hash_bucket_find() local
[all …]
/kernel/
Dauditfilter.c106 struct audit_entry *entry; in audit_init_entry() local
109 entry = kzalloc(sizeof(*entry), GFP_KERNEL); in audit_init_entry()
110 if (unlikely(!entry)) in audit_init_entry()
115 kfree(entry); in audit_init_entry()
118 entry->rule.fields = fields; in audit_init_entry()
120 return entry; in audit_init_entry()
208 static int audit_match_signal(struct audit_entry *entry) in audit_match_signal() argument
210 struct audit_field *arch = entry->rule.arch_f; in audit_match_signal()
216 entry->rule.mask) && in audit_match_signal()
218 entry->rule.mask)); in audit_match_signal()
[all …]
Dasync.c111 struct async_entry *entry = in async_run_entry_fn() local
119 (long long)entry->cookie, in async_run_entry_fn()
120 entry->func, task_pid_nr(current)); in async_run_entry_fn()
123 entry->func(entry->data, entry->cookie); in async_run_entry_fn()
128 (long long)entry->cookie, in async_run_entry_fn()
129 entry->func, in async_run_entry_fn()
135 list_del_init(&entry->domain_list); in async_run_entry_fn()
136 list_del_init(&entry->global_list); in async_run_entry_fn()
139 kfree(entry); in async_run_entry_fn()
151 struct async_entry *entry) in __async_schedule_node_domain() argument
[all …]
Djump_label.c309 static int addr_conflict(struct jump_entry *entry, void *start, void *end) in addr_conflict() argument
311 if (jump_entry_code(entry) <= (unsigned long)end && in addr_conflict()
312 jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE > (unsigned long)start) in addr_conflict()
341 void __weak __init_or_module arch_jump_label_transform_static(struct jump_entry *entry, in arch_jump_label_transform_static() argument
344 arch_jump_label_transform(entry, type); in arch_jump_label_transform_static()
393 static enum jump_label_type jump_label_type(struct jump_entry *entry) in jump_label_type() argument
395 struct static_key *key = jump_entry_key(entry); in jump_label_type()
397 bool branch = jump_entry_is_branch(entry); in jump_label_type()
403 static bool jump_label_can_update(struct jump_entry *entry, bool init) in jump_label_can_update() argument
408 if (!init && jump_entry_is_init(entry)) in jump_label_can_update()
[all …]
Dkexec_core.c264 image->entry = &image->head; in do_kimage_alloc_init()
534 static int kimage_add_entry(struct kimage *image, kimage_entry_t entry) in kimage_add_entry() argument
536 if (*image->entry != 0) in kimage_add_entry()
537 image->entry++; in kimage_add_entry()
539 if (image->entry == image->last_entry) { in kimage_add_entry()
548 *image->entry = virt_to_boot_phys(ind_page) | IND_INDIRECTION; in kimage_add_entry()
549 image->entry = ind_page; in kimage_add_entry()
553 *image->entry = entry; in kimage_add_entry()
554 image->entry++; in kimage_add_entry()
555 *image->entry = 0; in kimage_add_entry()
[all …]
Dkexec.c39 static int kimage_alloc_init(struct kimage **rimage, unsigned long entry, in kimage_alloc_init() argument
50 if ((entry < phys_to_boot_phys(crashk_res.start)) || in kimage_alloc_init()
51 (entry > phys_to_boot_phys(crashk_res.end))) in kimage_alloc_init()
60 image->start = entry; in kimage_alloc_init()
106 static int do_kexec_load(unsigned long entry, unsigned long nr_segments, in do_kexec_load() argument
144 ret = kimage_alloc_init(&image, entry, nr_segments, segments, flags); in do_kexec_load()
247 SYSCALL_DEFINE4(kexec_load, unsigned long, entry, unsigned long, nr_segments, in SYSCALL_DEFINE4() argument
261 result = do_kexec_load(entry, nr_segments, segments, flags); in SYSCALL_DEFINE4()
267 COMPAT_SYSCALL_DEFINE4(kexec_load, compat_ulong_t, entry, in COMPAT_SYSCALL_DEFINE4() argument
302 result = do_kexec_load(entry, nr_segments, ksegments, flags); in COMPAT_SYSCALL_DEFINE4()
Dconfigs.c57 struct proc_dir_entry *entry; in ikconfig_init() local
60 entry = proc_create("config.gz", S_IFREG | S_IRUGO, NULL, in ikconfig_init()
62 if (!entry) in ikconfig_init()
65 proc_set_size(entry, &kernel_config_data_end - &kernel_config_data); in ikconfig_init()
/kernel/trace/
Dtrace_events_inject.c21 void *entry; in trace_inject_entry() local
24 entry = trace_event_buffer_reserve(&fbuffer, file, len); in trace_inject_entry()
25 if (entry) { in trace_inject_entry()
26 memcpy(entry, rec, len); in trace_inject_entry()
158 void *entry = NULL; in trace_alloc_entry() local
161 entry = kzalloc(entry_size + 1, GFP_KERNEL); in trace_alloc_entry()
162 if (!entry) in trace_alloc_entry()
175 str_item = (u32 *)(entry + field->offset); in trace_alloc_entry()
180 paddr = (char **)(entry + field->offset); in trace_alloc_entry()
186 return entry; in trace_alloc_entry()
[all …]
Dtrace_syscalls.c26 struct syscall_metadata *entry = call->data; in syscall_get_enter_fields() local
28 return &entry->enter_fields; in syscall_get_enter_fields()
117 struct syscall_metadata *entry; in get_syscall_name() local
119 entry = syscall_nr_to_meta(syscall); in get_syscall_name()
120 if (!entry) in get_syscall_name()
123 return entry->name; in get_syscall_name()
134 struct syscall_metadata *entry; in print_syscall_enter() local
139 entry = syscall_nr_to_meta(syscall); in print_syscall_enter()
141 if (!entry) in print_syscall_enter()
144 if (entry->enter_event->event.type != ent->type) { in print_syscall_enter()
[all …]
Dtrace_mmiotrace.c169 struct trace_entry *entry = iter->ent; in mmio_print_rw() local
177 trace_assign_type(field, entry); in mmio_print_rw()
214 struct trace_entry *entry = iter->ent; in mmio_print_map() local
222 trace_assign_type(field, entry); in mmio_print_map()
248 struct trace_entry *entry = iter->ent; in mmio_print_mark() local
249 struct print_entry *print = (struct print_entry *)entry; in mmio_print_mark()
302 struct trace_mmiotrace_rw *entry; in __trace_mmiotrace_rw() local
306 sizeof(*entry), 0, pc); in __trace_mmiotrace_rw()
311 entry = ring_buffer_event_data(event); in __trace_mmiotrace_rw()
312 entry->rw = *rw; in __trace_mmiotrace_rw()
[all …]
Dtrace_output.c28 struct trace_entry *entry = iter->ent; in trace_print_bputs_msg_only() local
31 trace_assign_type(field, entry); in trace_print_bputs_msg_only()
41 struct trace_entry *entry = iter->ent; in trace_print_bprintk_msg_only() local
44 trace_assign_type(field, entry); in trace_print_bprintk_msg_only()
54 struct trace_entry *entry = iter->ent; in trace_print_printk_msg_only() local
57 trace_assign_type(field, entry); in trace_print_printk_msg_only()
298 struct trace_entry *entry; in trace_raw_output_prep() local
301 entry = iter->ent; in trace_raw_output_prep()
303 if (entry->type != event->event.type) { in trace_raw_output_prep()
440 int trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry) in trace_print_lat_fmt() argument
[all …]
Dtrace_sched_wakeup.c385 struct ctx_switch_entry *entry; in tracing_sched_switch_trace() local
388 sizeof(*entry), flags, pc); in tracing_sched_switch_trace()
391 entry = ring_buffer_event_data(event); in tracing_sched_switch_trace()
392 entry->prev_pid = prev->pid; in tracing_sched_switch_trace()
393 entry->prev_prio = prev->prio; in tracing_sched_switch_trace()
394 entry->prev_state = task_state_index(prev); in tracing_sched_switch_trace()
395 entry->next_pid = next->pid; in tracing_sched_switch_trace()
396 entry->next_prio = next->prio; in tracing_sched_switch_trace()
397 entry->next_state = task_state_index(next); in tracing_sched_switch_trace()
398 entry->next_cpu = task_cpu(next); in tracing_sched_switch_trace()
[all …]
Dtrace_events_synth.c317 struct synth_trace_event *entry; in print_synth_event() local
323 entry = (struct synth_trace_event *)iter->ent; in print_synth_event()
346 offset = (u32)entry->fields[n_u64]; in print_synth_event()
349 str_field = (char *)entry + data_offset; in print_synth_event()
359 (char *)&entry->fields[n_u64], in print_synth_event()
371 entry->fields[n_u64], in print_synth_event()
377 entry->fields[n_u64], in print_synth_event()
394 static unsigned int trace_string(struct synth_trace_event *entry, in trace_string() argument
407 data_offset = offsetof(typeof(*entry), fields); in trace_string()
411 str_field = (char *)entry + data_offset; in trace_string()
[all …]
Dtrace_hwlat.c110 struct hwlat_entry *entry; in trace_hwlat_sample() local
117 event = trace_buffer_lock_reserve(buffer, TRACE_HWLAT, sizeof(*entry), in trace_hwlat_sample()
121 entry = ring_buffer_event_data(event); in trace_hwlat_sample()
122 entry->seqnum = sample->seqnum; in trace_hwlat_sample()
123 entry->duration = sample->duration; in trace_hwlat_sample()
124 entry->outer_duration = sample->outer_duration; in trace_hwlat_sample()
125 entry->timestamp = sample->timestamp; in trace_hwlat_sample()
126 entry->nmi_total_ts = sample->nmi_total_ts; in trace_hwlat_sample()
127 entry->nmi_count = sample->nmi_count; in trace_hwlat_sample()
128 entry->count = sample->count; in trace_hwlat_sample()
[all …]
Dftrace.c960 struct dentry *entry; in ftrace_profile_tracefs() local
991 entry = tracefs_create_file("function_profile_enabled", 0644, in ftrace_profile_tracefs()
993 if (!entry) in ftrace_profile_tracefs()
1117 struct ftrace_func_entry *entry; in __ftrace_lookup_ip() local
1123 hlist_for_each_entry_rcu_notrace(entry, hhd, hlist) { in __ftrace_lookup_ip()
1124 if (entry->ip == ip) in __ftrace_lookup_ip()
1125 return entry; in __ftrace_lookup_ip()
1150 struct ftrace_func_entry *entry) in __add_hash_entry() argument
1155 key = ftrace_hash_key(hash, entry->ip); in __add_hash_entry()
1157 hlist_add_head(&entry->hlist, hhd); in __add_hash_entry()
[all …]
Dring_buffer_benchmark.c88 int *entry; in read_event() local
95 entry = ring_buffer_event_data(event); in read_event()
96 if (*entry != cpu) { in read_event()
111 int *entry; in read_page() local
145 entry = ring_buffer_event_data(event); in read_page()
146 if (*entry != cpu) { in read_page()
158 entry = ring_buffer_event_data(event); in read_page()
159 if (*entry != cpu) { in read_page()
250 int *entry; in ring_buffer_producer() local
259 entry = ring_buffer_event_data(event); in ring_buffer_producer()
[all …]
Dtrace_branch.c38 struct trace_branch *entry; in probe_likely_condition() local
65 sizeof(*entry), flags, pc); in probe_likely_condition()
69 entry = ring_buffer_event_data(event); in probe_likely_condition()
77 strncpy(entry->func, f->data.func, TRACE_FUNC_SIZE); in probe_likely_condition()
78 strncpy(entry->file, p, TRACE_FILE_SIZE); in probe_likely_condition()
79 entry->func[TRACE_FUNC_SIZE] = 0; in probe_likely_condition()
80 entry->file[TRACE_FILE_SIZE] = 0; in probe_likely_condition()
81 entry->constant = f->constant; in probe_likely_condition()
82 entry->line = f->data.line; in probe_likely_condition()
83 entry->correct = val == expect; in probe_likely_condition()
[all …]
Dtrace_kprobe.c1387 struct kprobe_trace_entry_head *entry; in NOKPROBE_SYMBOL() local
1406 sizeof(*entry) + tk->tp.size + dsize, in NOKPROBE_SYMBOL()
1412 entry = fbuffer.entry = ring_buffer_event_data(fbuffer.event); in NOKPROBE_SYMBOL()
1413 entry->ip = (unsigned long)tk->rp.kp.addr; in NOKPROBE_SYMBOL()
1414 store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize); in NOKPROBE_SYMBOL()
1435 struct kretprobe_trace_entry_head *entry; in __kretprobe_trace_func() local
1453 sizeof(*entry) + tk->tp.size + dsize, in __kretprobe_trace_func()
1459 entry = fbuffer.entry = ring_buffer_event_data(fbuffer.event); in __kretprobe_trace_func()
1460 entry->func = (unsigned long)tk->rp.kp.addr; in __kretprobe_trace_func()
1461 entry->ret_ip = (unsigned long)ri->ret_addr; in __kretprobe_trace_func()
[all …]
Dtracing_map.c520 struct tracing_map_entry *entry; in __tracing_map_insert() local
530 entry = TRACING_MAP_ENTRY(map->map, idx); in __tracing_map_insert()
531 test_key = entry->key; in __tracing_map_insert()
534 val = READ_ONCE(entry->val); in __tracing_map_insert()
566 if (!cmpxchg(&entry->key, 0, key_hash)) { in __tracing_map_insert()
572 entry->key = 0; in __tracing_map_insert()
582 WRITE_ONCE(entry->val, elt); in __tracing_map_insert()
585 return entry->val; in __tracing_map_insert()
922 static void destroy_sort_entry(struct tracing_map_sort_entry *entry) in destroy_sort_entry() argument
924 if (!entry) in destroy_sort_entry()
[all …]
/kernel/bpf/
Ddispatcher.c45 struct bpf_dispatcher_prog *entry; in bpf_dispatcher_add_prog() local
50 entry = bpf_dispatcher_find_prog(d, prog); in bpf_dispatcher_add_prog()
51 if (entry) { in bpf_dispatcher_add_prog()
52 refcount_inc(&entry->users); in bpf_dispatcher_add_prog()
56 entry = bpf_dispatcher_find_free(d); in bpf_dispatcher_add_prog()
57 if (!entry) in bpf_dispatcher_add_prog()
61 entry->prog = prog; in bpf_dispatcher_add_prog()
62 refcount_set(&entry->users, 1); in bpf_dispatcher_add_prog()
70 struct bpf_dispatcher_prog *entry; in bpf_dispatcher_remove_prog() local
75 entry = bpf_dispatcher_find_prog(d, prog); in bpf_dispatcher_remove_prog()
[all …]
/kernel/power/
Dconsole.c48 struct pm_vt_switch *entry, *tmp; in pm_vt_switch_required() local
59 entry = kmalloc(sizeof(*entry), GFP_KERNEL); in pm_vt_switch_required()
60 if (!entry) in pm_vt_switch_required()
63 entry->required = required; in pm_vt_switch_required()
64 entry->dev = dev; in pm_vt_switch_required()
66 list_add(&entry->head, &pm_vt_switch_list); in pm_vt_switch_required()
109 struct pm_vt_switch *entry; in pm_vt_switch() local
119 list_for_each_entry(entry, &pm_vt_switch_list, head) { in pm_vt_switch()
120 if (entry->required) in pm_vt_switch()
/kernel/sched/
Dwait.c77 curr = list_next_entry(bookmark, entry); in __wake_up_common()
79 list_del(&bookmark->entry); in __wake_up_common()
82 curr = list_first_entry(&wq_head->head, wait_queue_entry_t, entry); in __wake_up_common()
84 if (&curr->entry == &wq_head->head) in __wake_up_common()
87 list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) { in __wake_up_common()
101 (&next->entry != &wq_head->head)) { in __wake_up_common()
103 list_add_tail(&bookmark->entry, &next->entry); in __wake_up_common()
120 INIT_LIST_HEAD(&bookmark.entry); in __wake_up_common_lock()
256 if (list_empty(&wq_entry->entry)) in prepare_to_wait()
272 if (list_empty(&wq_entry->entry)) { in prepare_to_wait_exclusive()
[all …]
Ddebug.c209 struct ctl_table *entry = in sd_alloc_ctl_entry() local
212 return entry; in sd_alloc_ctl_entry()
217 struct ctl_table *entry; in sd_free_ctl_entry() local
225 for (entry = *tablep; entry->mode; entry++) { in sd_free_ctl_entry()
226 if (entry->child) in sd_free_ctl_entry()
227 sd_free_ctl_entry(&entry->child); in sd_free_ctl_entry()
228 if (entry->proc_handler == NULL) in sd_free_ctl_entry()
229 kfree(entry->procname); in sd_free_ctl_entry()
237 set_table_entry(struct ctl_table *entry, in set_table_entry() argument
241 entry->procname = procname; in set_table_entry()
[all …]
/kernel/events/
Dcallchain.c38 __weak void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument
43 __weak void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument
183 struct perf_callchain_entry *entry; in get_perf_callchain() local
187 entry = get_callchain_entry(&rctx); in get_perf_callchain()
188 if (!entry) in get_perf_callchain()
191 ctx.entry = entry; in get_perf_callchain()
193 ctx.nr = entry->nr = init_nr; in get_perf_callchain()
229 return entry; in get_perf_callchain()
/kernel/locking/
Dlockdep.c980 list_for_each_entry(e, h, entry) { in class_lock_list_valid()
1067 hlist_for_each_entry_rcu(chain, head, entry) { in __check_data_structures()
1079 if (!in_any_class_list(&e->entry)) { in __check_data_structures()
1094 if (in_any_class_list(&e->entry)) { in __check_data_structures()
1373 struct lock_list *entry; in add_lock_to_list() local
1378 entry = alloc_list_entry(); in add_lock_to_list()
1379 if (!entry) in add_lock_to_list()
1382 entry->class = this; in add_lock_to_list()
1383 entry->links_to = links_to; in add_lock_to_list()
1384 entry->dep = dep; in add_lock_to_list()
[all …]

123