Lines Matching refs:event
274 struct ring_buffer_event *event, int flag) in trace_process_export() argument
280 entry = ring_buffer_event_data(event); in trace_process_export()
281 size = ring_buffer_event_length(event); in trace_process_export()
318 static void ftrace_exports(struct ring_buffer_event *event, int flag) in ftrace_exports() argument
326 trace_process_export(export, event, flag); in ftrace_exports()
499 struct ring_buffer_event *event) in call_filter_check_discard() argument
503 __trace_event_discard_commit(buffer, event); in call_filter_check_discard()
937 trace_event_setup(struct ring_buffer_event *event, in trace_event_setup() argument
940 struct trace_entry *ent = ring_buffer_event_data(event); in trace_event_setup()
951 struct ring_buffer_event *event; in __trace_buffer_lock_reserve() local
953 event = ring_buffer_lock_reserve(buffer, len); in __trace_buffer_lock_reserve()
954 if (event != NULL) in __trace_buffer_lock_reserve()
955 trace_event_setup(event, type, flags, pc); in __trace_buffer_lock_reserve()
957 return event; in __trace_buffer_lock_reserve()
991 __buffer_unlock_commit(struct trace_buffer *buffer, struct ring_buffer_event *event) in __buffer_unlock_commit() argument
996 if (this_cpu_read(trace_buffered_event) == event) { in __buffer_unlock_commit()
998 ring_buffer_write(buffer, event->array[0], &event->array[1]); in __buffer_unlock_commit()
1002 ring_buffer_unlock_commit(buffer, event); in __buffer_unlock_commit()
1013 struct ring_buffer_event *event; in __trace_puts() local
1033 event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, alloc, in __trace_puts()
1035 if (!event) { in __trace_puts()
1040 entry = ring_buffer_event_data(event); in __trace_puts()
1052 __buffer_unlock_commit(buffer, event); in __trace_puts()
1067 struct ring_buffer_event *event; in __trace_bputs() local
1087 event = __trace_buffer_lock_reserve(buffer, TRACE_BPUTS, size, in __trace_bputs()
1089 if (!event) in __trace_bputs()
1092 entry = ring_buffer_event_data(event); in __trace_bputs()
1096 __buffer_unlock_commit(buffer, event); in __trace_bputs()
2665 struct ring_buffer_event *event; in trace_buffered_event_enable() local
2683 event = page_address(page); in trace_buffered_event_enable()
2684 memset(event, 0, sizeof(*event)); in trace_buffered_event_enable()
2686 per_cpu(trace_buffered_event, cpu) = event; in trace_buffered_event_enable()
2805 struct trace_event *event; in output_printk() local
2814 if (!event_call || !event_call->event.funcs || in output_printk()
2815 !event_call->event.funcs->trace) in output_printk()
2824 event = &fbuffer->trace_file->event_call->event; in output_printk()
2829 event_call->event.funcs->trace(iter, 0, event); in output_printk()
2875 ftrace_exports(fbuffer->event, TRACE_EXPORT_EVENT); in trace_event_buffer_commit()
2877 fbuffer->event, fbuffer->entry, in trace_event_buffer_commit()
2893 struct ring_buffer_event *event, in trace_buffer_unlock_commit_regs() argument
2897 __buffer_unlock_commit(buffer, event); in trace_buffer_unlock_commit_regs()
2914 struct ring_buffer_event *event) in trace_buffer_unlock_commit_nostack() argument
2916 __buffer_unlock_commit(buffer, event); in trace_buffer_unlock_commit_nostack()
2926 struct ring_buffer_event *event; in trace_function() local
2929 event = __trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry), in trace_function()
2931 if (!event) in trace_function()
2933 entry = ring_buffer_event_data(event); in trace_function()
2937 if (!call_filter_check_discard(call, entry, buffer, event)) { in trace_function()
2939 ftrace_exports(event, TRACE_EXPORT_FUNCTION); in trace_function()
2940 __buffer_unlock_commit(buffer, event); in trace_function()
2968 struct ring_buffer_event *event; in __ftrace_trace_stack() local
3011 event = __trace_buffer_lock_reserve(buffer, TRACE_STACK, in __ftrace_trace_stack()
3014 if (!event) in __ftrace_trace_stack()
3016 entry = ring_buffer_event_data(event); in __ftrace_trace_stack()
3021 if (!call_filter_check_discard(call, entry, buffer, event)) in __ftrace_trace_stack()
3022 __buffer_unlock_commit(buffer, event); in __ftrace_trace_stack()
3097 struct ring_buffer_event *event; in ftrace_trace_userstack() local
3120 event = __trace_buffer_lock_reserve(buffer, TRACE_USER_STACK, in ftrace_trace_userstack()
3122 if (!event) in ftrace_trace_userstack()
3124 entry = ring_buffer_event_data(event); in ftrace_trace_userstack()
3130 if (!call_filter_check_discard(call, entry, buffer, event)) in ftrace_trace_userstack()
3131 __buffer_unlock_commit(buffer, event); in ftrace_trace_userstack()
3267 struct ring_buffer_event *event; in trace_vbprintk() local
3299 event = __trace_buffer_lock_reserve(buffer, TRACE_BPRINT, size, in trace_vbprintk()
3301 if (!event) in trace_vbprintk()
3303 entry = ring_buffer_event_data(event); in trace_vbprintk()
3308 if (!call_filter_check_discard(call, entry, buffer, event)) { in trace_vbprintk()
3309 __buffer_unlock_commit(buffer, event); in trace_vbprintk()
3332 struct ring_buffer_event *event; in __trace_array_vprintk() local
3359 event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size, in __trace_array_vprintk()
3361 if (!event) in __trace_array_vprintk()
3363 entry = ring_buffer_event_data(event); in __trace_array_vprintk()
3367 if (!call_filter_check_discard(call, entry, buffer, event)) { in __trace_array_vprintk()
3368 __buffer_unlock_commit(buffer, event); in __trace_array_vprintk()
3491 struct ring_buffer_event *event; in peek_next_entry() local
3495 event = ring_buffer_iter_peek(buf_iter, ts); in peek_next_entry()
3500 event = ring_buffer_peek(iter->array_buffer->buffer, cpu, ts, in peek_next_entry()
3504 if (event) { in peek_next_entry()
3505 iter->ent_size = ring_buffer_event_length(event); in peek_next_entry()
3506 return ring_buffer_event_data(event); in peek_next_entry()
3989 struct trace_event *event; in print_trace_fmt() local
3995 event = ftrace_find_event(entry->type); in print_trace_fmt()
4007 if (event) in print_trace_fmt()
4008 return event->funcs->trace(iter, sym_flags, event); in print_trace_fmt()
4020 struct trace_event *event; in print_raw_fmt() local
4031 event = ftrace_find_event(entry->type); in print_raw_fmt()
4032 if (event) in print_raw_fmt()
4033 return event->funcs->raw(iter, 0, event); in print_raw_fmt()
4046 struct trace_event *event; in print_hex_fmt() local
4058 event = ftrace_find_event(entry->type); in print_hex_fmt()
4059 if (event) { in print_hex_fmt()
4060 enum print_line_t ret = event->funcs->hex(iter, 0, event); in print_hex_fmt()
4075 struct trace_event *event; in print_bin_fmt() local
4087 event = ftrace_find_event(entry->type); in print_bin_fmt()
4088 return event ? event->funcs->binary(iter, 0, event) : in print_bin_fmt()
6737 struct ring_buffer_event *event; in tracing_mark_write() local
6769 event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size, in tracing_mark_write()
6771 if (unlikely(!event)) in tracing_mark_write()
6775 entry = ring_buffer_event_data(event); in tracing_mark_write()
6789 tt = event_triggers_call(tr->trace_marker_file, entry, event); in tracing_mark_write()
6799 ftrace_exports(event, TRACE_EXPORT_MARKER); in tracing_mark_write()
6800 __buffer_unlock_commit(buffer, event); in tracing_mark_write()
6819 struct ring_buffer_event *event; in tracing_mark_raw_write() local
6850 event = __trace_buffer_lock_reserve(buffer, TRACE_RAW_DATA, size, in tracing_mark_raw_write()
6852 if (!event) in tracing_mark_raw_write()
6856 entry = ring_buffer_event_data(event); in tracing_mark_raw_write()
6866 __buffer_unlock_commit(buffer, event); in tracing_mark_raw_write()
9326 unsigned long event, void *unused) in trace_panic_handler() argument