/kernel/dma/ |
D | debug.c | 160 static inline void dump_entry_trace(struct dma_debug_entry *entry) in dump_entry_trace() argument 163 if (entry) { in dump_entry_trace() 165 stack_trace_print(entry->stack_entries, entry->stack_len, 0); in dump_entry_trace() 211 #define err_printk(dev, entry, format, arg...) do { \ argument 218 dump_entry_trace(entry); \ 230 static int hash_fn(struct dma_debug_entry *entry) in hash_fn() argument 236 return (entry->dev_addr >> HASH_FN_SHIFT) & HASH_FN_MASK; in hash_fn() 242 static struct hash_bucket *get_hash_bucket(struct dma_debug_entry *entry, in get_hash_bucket() argument 246 int idx = hash_fn(entry); in get_hash_bucket() 292 struct dma_debug_entry *entry, *ret = NULL; in __hash_bucket_find() local [all …]
|
/kernel/ |
D | auditfilter.c | 106 struct audit_entry *entry; in audit_init_entry() local 109 entry = kzalloc(sizeof(*entry), GFP_KERNEL); in audit_init_entry() 110 if (unlikely(!entry)) in audit_init_entry() 115 kfree(entry); in audit_init_entry() 118 entry->rule.fields = fields; in audit_init_entry() 120 return entry; in audit_init_entry() 208 static int audit_match_signal(struct audit_entry *entry) in audit_match_signal() argument 210 struct audit_field *arch = entry->rule.arch_f; in audit_match_signal() 216 entry->rule.mask) && in audit_match_signal() 218 entry->rule.mask)); in audit_match_signal() [all …]
|
D | async.c | 111 struct async_entry *entry = in async_run_entry_fn() local 119 (long long)entry->cookie, in async_run_entry_fn() 120 entry->func, task_pid_nr(current)); in async_run_entry_fn() 123 entry->func(entry->data, entry->cookie); in async_run_entry_fn() 128 (long long)entry->cookie, in async_run_entry_fn() 129 entry->func, in async_run_entry_fn() 135 list_del_init(&entry->domain_list); in async_run_entry_fn() 136 list_del_init(&entry->global_list); in async_run_entry_fn() 139 kfree(entry); in async_run_entry_fn() 168 struct async_entry *entry; in async_schedule_node_domain() local [all …]
|
D | jump_label.c | 309 static int addr_conflict(struct jump_entry *entry, void *start, void *end) in addr_conflict() argument 311 if (jump_entry_code(entry) <= (unsigned long)end && in addr_conflict() 312 jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE > (unsigned long)start) in addr_conflict() 339 void __weak __init_or_module arch_jump_label_transform_static(struct jump_entry *entry, in arch_jump_label_transform_static() argument 342 arch_jump_label_transform(entry, type); in arch_jump_label_transform_static() 391 static enum jump_label_type jump_label_type(struct jump_entry *entry) in jump_label_type() argument 393 struct static_key *key = jump_entry_key(entry); in jump_label_type() 395 bool branch = jump_entry_is_branch(entry); in jump_label_type() 401 static bool jump_label_can_update(struct jump_entry *entry, bool init) in jump_label_can_update() argument 406 if (!init && jump_entry_is_init(entry)) in jump_label_can_update() [all …]
|
D | kexec_core.c | 264 image->entry = &image->head; in do_kimage_alloc_init() 534 static int kimage_add_entry(struct kimage *image, kimage_entry_t entry) in kimage_add_entry() argument 536 if (*image->entry != 0) in kimage_add_entry() 537 image->entry++; in kimage_add_entry() 539 if (image->entry == image->last_entry) { in kimage_add_entry() 548 *image->entry = virt_to_boot_phys(ind_page) | IND_INDIRECTION; in kimage_add_entry() 549 image->entry = ind_page; in kimage_add_entry() 553 *image->entry = entry; in kimage_add_entry() 554 image->entry++; in kimage_add_entry() 555 *image->entry = 0; in kimage_add_entry() [all …]
|
D | kexec.c | 39 static int kimage_alloc_init(struct kimage **rimage, unsigned long entry, in kimage_alloc_init() argument 50 if ((entry < phys_to_boot_phys(crashk_res.start)) || in kimage_alloc_init() 51 (entry > phys_to_boot_phys(crashk_res.end))) in kimage_alloc_init() 60 image->start = entry; in kimage_alloc_init() 106 static int do_kexec_load(unsigned long entry, unsigned long nr_segments, in do_kexec_load() argument 135 ret = kimage_alloc_init(&image, entry, nr_segments, segments, flags); in do_kexec_load() 232 SYSCALL_DEFINE4(kexec_load, unsigned long, entry, unsigned long, nr_segments, in SYSCALL_DEFINE4() argument 257 result = do_kexec_load(entry, nr_segments, segments, flags); in SYSCALL_DEFINE4() 265 COMPAT_SYSCALL_DEFINE4(kexec_load, compat_ulong_t, entry, in COMPAT_SYSCALL_DEFINE4() argument 311 result = do_kexec_load(entry, nr_segments, ksegments, flags); in COMPAT_SYSCALL_DEFINE4()
|
D | configs.c | 58 struct proc_dir_entry *entry; in ikconfig_init() local 61 entry = proc_create("config.gz", S_IFREG | S_IRUGO, NULL, in ikconfig_init() 63 if (!entry) in ikconfig_init() 66 proc_set_size(entry, &kernel_config_data_end - &kernel_config_data); in ikconfig_init()
|
/kernel/trace/ |
D | trace_syscalls.c | 25 struct syscall_metadata *entry = call->data; in syscall_get_enter_fields() local 27 return &entry->enter_fields; in syscall_get_enter_fields() 112 struct syscall_metadata *entry; in get_syscall_name() local 114 entry = syscall_nr_to_meta(syscall); in get_syscall_name() 115 if (!entry) in get_syscall_name() 118 return entry->name; in get_syscall_name() 129 struct syscall_metadata *entry; in print_syscall_enter() local 134 entry = syscall_nr_to_meta(syscall); in print_syscall_enter() 136 if (!entry) in print_syscall_enter() 139 if (entry->enter_event->event.type != ent->type) { in print_syscall_enter() [all …]
|
D | trace_output.c | 28 struct trace_entry *entry = iter->ent; in trace_print_bputs_msg_only() local 31 trace_assign_type(field, entry); in trace_print_bputs_msg_only() 41 struct trace_entry *entry = iter->ent; in trace_print_bprintk_msg_only() local 44 trace_assign_type(field, entry); in trace_print_bprintk_msg_only() 54 struct trace_entry *entry = iter->ent; in trace_print_printk_msg_only() local 57 trace_assign_type(field, entry); in trace_print_printk_msg_only() 283 struct trace_entry *entry; in trace_raw_output_prep() local 286 entry = iter->ent; in trace_raw_output_prep() 288 if (entry->type != event->event.type) { in trace_raw_output_prep() 425 int trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry) in trace_print_lat_fmt() argument [all …]
|
D | trace_mmiotrace.c | 169 struct trace_entry *entry = iter->ent; in mmio_print_rw() local 177 trace_assign_type(field, entry); in mmio_print_rw() 214 struct trace_entry *entry = iter->ent; in mmio_print_map() local 222 trace_assign_type(field, entry); in mmio_print_map() 248 struct trace_entry *entry = iter->ent; in mmio_print_mark() local 249 struct print_entry *print = (struct print_entry *)entry; in mmio_print_mark() 302 struct trace_mmiotrace_rw *entry; in __trace_mmiotrace_rw() local 306 sizeof(*entry), 0, pc); in __trace_mmiotrace_rw() 311 entry = ring_buffer_event_data(event); in __trace_mmiotrace_rw() 312 entry->rw = *rw; in __trace_mmiotrace_rw() [all …]
|
D | trace_sched_wakeup.c | 383 struct ctx_switch_entry *entry; in tracing_sched_switch_trace() local 386 sizeof(*entry), flags, pc); in tracing_sched_switch_trace() 389 entry = ring_buffer_event_data(event); in tracing_sched_switch_trace() 390 entry->prev_pid = prev->pid; in tracing_sched_switch_trace() 391 entry->prev_prio = prev->prio; in tracing_sched_switch_trace() 392 entry->prev_state = task_state_index(prev); in tracing_sched_switch_trace() 393 entry->next_pid = next->pid; in tracing_sched_switch_trace() 394 entry->next_prio = next->prio; in tracing_sched_switch_trace() 395 entry->next_state = task_state_index(next); in tracing_sched_switch_trace() 396 entry->next_cpu = task_cpu(next); in tracing_sched_switch_trace() [all …]
|
D | trace_hwlat.c | 109 struct hwlat_entry *entry; in trace_hwlat_sample() local 116 event = trace_buffer_lock_reserve(buffer, TRACE_HWLAT, sizeof(*entry), in trace_hwlat_sample() 120 entry = ring_buffer_event_data(event); in trace_hwlat_sample() 121 entry->seqnum = sample->seqnum; in trace_hwlat_sample() 122 entry->duration = sample->duration; in trace_hwlat_sample() 123 entry->outer_duration = sample->outer_duration; in trace_hwlat_sample() 124 entry->timestamp = sample->timestamp; in trace_hwlat_sample() 125 entry->nmi_total_ts = sample->nmi_total_ts; in trace_hwlat_sample() 126 entry->nmi_count = sample->nmi_count; in trace_hwlat_sample() 128 if (!call_filter_check_discard(call, entry, buffer, event)) in trace_hwlat_sample() [all …]
|
D | ring_buffer_benchmark.c | 88 int *entry; in read_event() local 95 entry = ring_buffer_event_data(event); in read_event() 96 if (*entry != cpu) { in read_event() 111 int *entry; in read_page() local 145 entry = ring_buffer_event_data(event); in read_page() 146 if (*entry != cpu) { in read_page() 158 entry = ring_buffer_event_data(event); in read_page() 159 if (*entry != cpu) { in read_page() 250 int *entry; in ring_buffer_producer() local 259 entry = ring_buffer_event_data(event); in ring_buffer_producer() [all …]
|
D | trace_branch.c | 37 struct trace_branch *entry; in probe_likely_condition() local 65 sizeof(*entry), flags, pc); in probe_likely_condition() 69 entry = ring_buffer_event_data(event); in probe_likely_condition() 77 strncpy(entry->func, f->data.func, TRACE_FUNC_SIZE); in probe_likely_condition() 78 strncpy(entry->file, p, TRACE_FILE_SIZE); in probe_likely_condition() 79 entry->func[TRACE_FUNC_SIZE] = 0; in probe_likely_condition() 80 entry->file[TRACE_FILE_SIZE] = 0; in probe_likely_condition() 81 entry->constant = f->constant; in probe_likely_condition() 82 entry->line = f->data.line; in probe_likely_condition() 83 entry->correct = val == expect; in probe_likely_condition() [all …]
|
D | trace_kprobe.c | 1177 struct kprobe_trace_entry_head *entry; in NOKPROBE_SYMBOL() local 1193 size = sizeof(*entry) + tk->tp.size + dsize; in NOKPROBE_SYMBOL() 1201 entry = ring_buffer_event_data(event); in NOKPROBE_SYMBOL() 1202 entry->ip = (unsigned long)tk->rp.kp.addr; in NOKPROBE_SYMBOL() 1203 store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize); in NOKPROBE_SYMBOL() 1206 entry, irq_flags, pc, regs); in NOKPROBE_SYMBOL() 1225 struct kretprobe_trace_entry_head *entry; in __kretprobe_trace_func() local 1241 size = sizeof(*entry) + tk->tp.size + dsize; in __kretprobe_trace_func() 1249 entry = ring_buffer_event_data(event); in __kretprobe_trace_func() 1250 entry->func = (unsigned long)tk->rp.kp.addr; in __kretprobe_trace_func() [all …]
|
D | tracing_map.c | 517 struct tracing_map_entry *entry; in __tracing_map_insert() local 527 entry = TRACING_MAP_ENTRY(map->map, idx); in __tracing_map_insert() 528 test_key = entry->key; in __tracing_map_insert() 531 val = READ_ONCE(entry->val); in __tracing_map_insert() 563 if (!cmpxchg(&entry->key, 0, key_hash)) { in __tracing_map_insert() 569 entry->key = 0; in __tracing_map_insert() 574 entry->val = elt; in __tracing_map_insert() 577 return entry->val; in __tracing_map_insert() 905 static void destroy_sort_entry(struct tracing_map_sort_entry *entry) in destroy_sort_entry() argument 907 if (!entry) in destroy_sort_entry() [all …]
|
D | ftrace.c | 964 struct dentry *entry; in ftrace_profile_tracefs() local 995 entry = tracefs_create_file("function_profile_enabled", 0644, in ftrace_profile_tracefs() 997 if (!entry) in ftrace_profile_tracefs() 1129 struct ftrace_func_entry *entry; in __ftrace_lookup_ip() local 1135 hlist_for_each_entry_rcu_notrace(entry, hhd, hlist) { in __ftrace_lookup_ip() 1136 if (entry->ip == ip) in __ftrace_lookup_ip() 1137 return entry; in __ftrace_lookup_ip() 1162 struct ftrace_func_entry *entry) in __add_hash_entry() argument 1167 key = ftrace_hash_key(hash, entry->ip); in __add_hash_entry() 1169 hlist_add_head(&entry->hlist, hhd); in __add_hash_entry() [all …]
|
D | trace_functions_graph.c | 105 struct ftrace_graph_ent_entry *entry; in __trace_graph_entry() local 108 sizeof(*entry), flags, pc); in __trace_graph_entry() 111 entry = ring_buffer_event_data(event); in __trace_graph_entry() 112 entry->graph_ent = *trace; in __trace_graph_entry() 113 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_graph_entry() 225 struct ftrace_graph_ret_entry *entry; in __trace_graph_return() local 228 sizeof(*entry), flags, pc); in __trace_graph_return() 231 entry = ring_buffer_event_data(event); in __trace_graph_return() 232 entry->ret = *trace; in __trace_graph_return() 233 if (!call_filter_check_discard(call, entry, buffer, event)) in __trace_graph_return() [all …]
|
D | trace_uprobe.c | 34 #define DATAOF_TRACE_ENTRY(entry, is_return) \ argument 35 ((void*)(entry) + SIZEOF_TRACE_ENTRY(is_return)) 948 struct uprobe_trace_entry_head *entry; in __uprobe_trace_func() local 970 entry = ring_buffer_event_data(event); in __uprobe_trace_func() 972 entry->vaddr[0] = func; in __uprobe_trace_func() 973 entry->vaddr[1] = instruction_pointer(regs); in __uprobe_trace_func() 974 data = DATAOF_TRACE_ENTRY(entry, true); in __uprobe_trace_func() 976 entry->vaddr[0] = instruction_pointer(regs); in __uprobe_trace_func() 977 data = DATAOF_TRACE_ENTRY(entry, false); in __uprobe_trace_func() 982 event_trigger_unlock_commit(trace_file, buffer, event, entry, 0, 0); in __uprobe_trace_func() [all …]
|
D | trace.c | 837 struct print_entry *entry; in __trace_puts() local 850 alloc = sizeof(*entry) + size + 2; /* possible \n added */ in __trace_puts() 859 entry = ring_buffer_event_data(event); in __trace_puts() 860 entry->ip = ip; in __trace_puts() 862 memcpy(&entry->buf, str, size); in __trace_puts() 865 if (entry->buf[size - 1] != '\n') { in __trace_puts() 866 entry->buf[size] = '\n'; in __trace_puts() 867 entry->buf[size + 1] = '\0'; in __trace_puts() 869 entry->buf[size] = '\0'; in __trace_puts() 887 struct bputs_entry *entry; in __trace_bputs() local [all …]
|
/kernel/power/ |
D | console.c | 48 struct pm_vt_switch *entry, *tmp; in pm_vt_switch_required() local 59 entry = kmalloc(sizeof(*entry), GFP_KERNEL); in pm_vt_switch_required() 60 if (!entry) in pm_vt_switch_required() 63 entry->required = required; in pm_vt_switch_required() 64 entry->dev = dev; in pm_vt_switch_required() 66 list_add(&entry->head, &pm_vt_switch_list); in pm_vt_switch_required() 109 struct pm_vt_switch *entry; in pm_vt_switch() local 119 list_for_each_entry(entry, &pm_vt_switch_list, head) { in pm_vt_switch() 120 if (entry->required) in pm_vt_switch()
|
/kernel/sched/ |
D | wait.c | 76 curr = list_next_entry(bookmark, entry); in __wake_up_common() 78 list_del(&bookmark->entry); in __wake_up_common() 81 curr = list_first_entry(&wq_head->head, wait_queue_entry_t, entry); in __wake_up_common() 83 if (&curr->entry == &wq_head->head) in __wake_up_common() 86 list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) { in __wake_up_common() 100 (&next->entry != &wq_head->head)) { in __wake_up_common() 102 list_add_tail(&bookmark->entry, &next->entry); in __wake_up_common() 119 INIT_LIST_HEAD(&bookmark.entry); in __wake_up_common_lock() 228 if (list_empty(&wq_entry->entry)) in prepare_to_wait() 242 if (list_empty(&wq_entry->entry)) in prepare_to_wait_exclusive() [all …]
|
D | debug.c | 209 struct ctl_table *entry = in sd_alloc_ctl_entry() local 212 return entry; in sd_alloc_ctl_entry() 217 struct ctl_table *entry; in sd_free_ctl_entry() local 225 for (entry = *tablep; entry->mode; entry++) { in sd_free_ctl_entry() 226 if (entry->child) in sd_free_ctl_entry() 227 sd_free_ctl_entry(&entry->child); in sd_free_ctl_entry() 228 if (entry->proc_handler == NULL) in sd_free_ctl_entry() 229 kfree(entry->procname); in sd_free_ctl_entry() 237 set_table_entry(struct ctl_table *entry, in set_table_entry() argument 241 entry->procname = procname; in set_table_entry() [all …]
|
/kernel/events/ |
D | callchain.c | 38 __weak void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument 43 __weak void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 181 struct perf_callchain_entry *entry; in get_perf_callchain() local 185 entry = get_callchain_entry(&rctx); in get_perf_callchain() 189 if (!entry) in get_perf_callchain() 192 ctx.entry = entry; in get_perf_callchain() 194 ctx.nr = entry->nr = init_nr; in get_perf_callchain() 231 return entry; in get_perf_callchain()
|
/kernel/locking/ |
D | lockdep.c | 924 list_for_each_entry(e, h, entry) { in class_lock_list_valid() 1011 hlist_for_each_entry_rcu(chain, head, entry) { in __check_data_structures() 1023 if (!in_any_class_list(&e->entry)) { in __check_data_structures() 1038 if (in_any_class_list(&e->entry)) { in __check_data_structures() 1307 struct lock_list *entry; in add_lock_to_list() local 1312 entry = alloc_list_entry(); in add_lock_to_list() 1313 if (!entry) in add_lock_to_list() 1316 entry->class = this; in add_lock_to_list() 1317 entry->links_to = links_to; in add_lock_to_list() 1318 entry->distance = distance; in add_lock_to_list() [all …]
|