/kernel/trace/ |
D | trace_seq.c | 31 #define TRACE_SEQ_BUF_LEFT(s) seq_buf_buffer_left(&(s)->seq) argument 36 static inline void __trace_seq_init(struct trace_seq *s) in __trace_seq_init() argument 38 if (unlikely(!s->seq.size)) in __trace_seq_init() 39 trace_seq_init(s); in __trace_seq_init() 51 int trace_print_seq(struct seq_file *m, struct trace_seq *s) in trace_print_seq() argument 55 __trace_seq_init(s); in trace_print_seq() 57 ret = seq_buf_print_seq(m, &s->seq); in trace_print_seq() 65 trace_seq_init(s); in trace_print_seq() 81 void trace_seq_printf(struct trace_seq *s, const char *fmt, ...) in trace_seq_printf() argument 83 unsigned int save_len = s->seq.len; in trace_seq_printf() [all …]
|
D | trace_output.c | 27 struct trace_seq *s = &iter->seq; in trace_print_bputs_msg_only() local 33 trace_seq_puts(s, field->str); in trace_print_bputs_msg_only() 35 return trace_handle_return(s); in trace_print_bputs_msg_only() 40 struct trace_seq *s = &iter->seq; in trace_print_bprintk_msg_only() local 46 trace_seq_bprintf(s, field->fmt, field->buf); in trace_print_bprintk_msg_only() 48 return trace_handle_return(s); in trace_print_bprintk_msg_only() 53 struct trace_seq *s = &iter->seq; in trace_print_printk_msg_only() local 59 trace_seq_puts(s, field->buf); in trace_print_printk_msg_only() 61 return trace_handle_return(s); in trace_print_printk_msg_only() 296 struct trace_seq *s = &iter->seq; in trace_raw_output_prep() local [all …]
|
D | trace_functions_graph.c | 95 struct trace_seq *s, u32 flags); 331 static void print_graph_cpu(struct trace_seq *s, int cpu) in print_graph_cpu() argument 338 trace_seq_printf(s, " %*d) ", max_bytes_for_cpu, cpu); in print_graph_cpu() 343 static void print_graph_proc(struct trace_seq *s, pid_t pid) in print_graph_proc() argument 364 trace_seq_putc(s, ' '); in print_graph_proc() 366 trace_seq_printf(s, "%s-%s", comm, pid_str); in print_graph_proc() 370 trace_seq_putc(s, ' '); in print_graph_proc() 374 static void print_graph_lat_fmt(struct trace_seq *s, struct trace_entry *entry) in print_graph_lat_fmt() argument 376 trace_seq_putc(s, ' '); in print_graph_lat_fmt() 377 trace_print_lat_fmt(s, entry); in print_graph_lat_fmt() [all …]
|
D | trace_osnoise.c | 260 static void print_osnoise_headers(struct seq_file *s) in print_osnoise_headers() argument 263 seq_puts(s, "# osnoise is tainted!\n"); in print_osnoise_headers() 265 seq_puts(s, "# _-------=> irqs-off\n"); in print_osnoise_headers() 266 seq_puts(s, "# / _------=> need-resched\n"); in print_osnoise_headers() 267 seq_puts(s, "# | / _-----=> need-resched-lazy\n"); in print_osnoise_headers() 268 seq_puts(s, "# || / _----=> hardirq/softirq\n"); in print_osnoise_headers() 269 seq_puts(s, "# ||| / _---=> preempt-depth\n"); in print_osnoise_headers() 270 seq_puts(s, "# |||| / _--=> preempt-lazy-depth\n"); in print_osnoise_headers() 271 seq_puts(s, "# ||||| / _-=> migrate-disable\n"); in print_osnoise_headers() 273 seq_puts(s, "# |||||| / "); in print_osnoise_headers() [all …]
|
D | trace_mmiotrace.c | 61 static void mmio_print_pcidev(struct trace_seq *s, const struct pci_dev *dev) in mmio_print_pcidev() argument 67 trace_seq_printf(s, "PCIDEV %02x%02x %04x%04x %x", in mmio_print_pcidev() 72 trace_seq_printf(s, " %llx", in mmio_print_pcidev() 79 trace_seq_printf(s, " %llx", in mmio_print_pcidev() 84 trace_seq_printf(s, " %s\n", drv->name); in mmio_print_pcidev() 86 trace_seq_puts(s, " \n"); in mmio_print_pcidev() 100 struct trace_seq *s = &iter->seq; in mmio_pipe_open() local 102 trace_seq_puts(s, "VERSION 20070824\n"); in mmio_pipe_open() 136 struct trace_seq *s = &iter->seq; in mmio_read() local 142 trace_seq_printf(s, "MARK 0.000000 Lost %lu events.\n", n); in mmio_read() [all …]
|
D | trace_output.h | 16 seq_print_ip_sym(struct trace_seq *s, unsigned long ip, 19 extern void trace_seq_print_sym(struct trace_seq *s, unsigned long address, bool offset); 30 trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry); 36 #define SEQ_PUT_FIELD(s, x) \ argument 37 trace_seq_putmem(s, &(x), sizeof(x)) 39 #define SEQ_PUT_HEX_FIELD(s, x) \ argument 40 trace_seq_putmem_hex(s, &(x), sizeof(x))
|
D | trace_hwlat.c | 205 struct hwlat_sample s; in get_sample() local 259 ktime_get_real_ts64(&s.timestamp); in get_sample() 291 s.seqnum = hwlat_data.count; in get_sample() 292 s.duration = sample; in get_sample() 293 s.outer_duration = outer_sample; in get_sample() 294 s.nmi_total_ts = kdata->nmi_total_ts; in get_sample() 295 s.nmi_count = kdata->nmi_count; in get_sample() 296 s.count = count; in get_sample() 297 trace_hwlat_sample(&s); in get_sample() 606 static void *s_mode_start(struct seq_file *s, loff_t *pos) in s_mode_start() argument [all …]
|
D | blktrace.c | 1306 static void blk_log_dump_pdu(struct trace_seq *s, in blk_log_dump_pdu() argument 1325 trace_seq_putc(s, '('); in blk_log_dump_pdu() 1329 trace_seq_printf(s, "%s%02x", in blk_log_dump_pdu() 1337 trace_seq_puts(s, " ..) "); in blk_log_dump_pdu() 1342 trace_seq_puts(s, ") "); in blk_log_dump_pdu() 1345 static void blk_log_generic(struct trace_seq *s, const struct trace_entry *ent, bool has_cg) in blk_log_generic() argument 1352 trace_seq_printf(s, "%u ", t_bytes(ent)); in blk_log_generic() 1353 blk_log_dump_pdu(s, ent, has_cg); in blk_log_generic() 1354 trace_seq_printf(s, "[%s]\n", cmd); in blk_log_generic() 1357 trace_seq_printf(s, "%llu + %u [%s]\n", in blk_log_generic() [all …]
|
/kernel/bpf/preload/iterators/ |
D | iterators.skel.h | 106 struct bpf_object_skeleton *s; in iterators_bpf__create_skeleton() local 108 s = (struct bpf_object_skeleton *)calloc(1, sizeof(*s)); in iterators_bpf__create_skeleton() 109 if (!s) in iterators_bpf__create_skeleton() 111 obj->skeleton = s; in iterators_bpf__create_skeleton() 113 s->sz = sizeof(*s); in iterators_bpf__create_skeleton() 114 s->name = "iterators_bpf"; in iterators_bpf__create_skeleton() 115 s->obj = &obj->obj; in iterators_bpf__create_skeleton() 118 s->map_cnt = 1; in iterators_bpf__create_skeleton() 119 s->map_skel_sz = sizeof(*s->maps); in iterators_bpf__create_skeleton() 120 s->maps = (struct bpf_map_skeleton *)calloc(s->map_cnt, s->map_skel_sz); in iterators_bpf__create_skeleton() [all …]
|
/kernel/bpf/ |
D | percpu_freelist.c | 6 int pcpu_freelist_init(struct pcpu_freelist *s) in pcpu_freelist_init() argument 10 s->freelist = alloc_percpu(struct pcpu_freelist_head); in pcpu_freelist_init() 11 if (!s->freelist) in pcpu_freelist_init() 15 struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_init() 20 raw_spin_lock_init(&s->extralist.lock); in pcpu_freelist_init() 21 s->extralist.first = NULL; in pcpu_freelist_init() 25 void pcpu_freelist_destroy(struct pcpu_freelist *s) in pcpu_freelist_destroy() argument 27 free_percpu(s->freelist); in pcpu_freelist_destroy() 45 static inline bool pcpu_freelist_try_push_extra(struct pcpu_freelist *s, in pcpu_freelist_try_push_extra() argument 48 if (!raw_spin_trylock(&s->extralist.lock)) in pcpu_freelist_try_push_extra() [all …]
|
/kernel/ |
D | scs.c | 19 static void __scs_account(void *s, int account) in __scs_account() argument 21 struct page *scs_page = vmalloc_to_page(s); in __scs_account() 34 void *s; in __scs_alloc() local 37 s = this_cpu_xchg(scs_cache[i], NULL); in __scs_alloc() 38 if (s) { in __scs_alloc() 39 s = kasan_unpoison_vmalloc(s, SCS_SIZE, in __scs_alloc() 41 memset(s, 0, SCS_SIZE); in __scs_alloc() 46 s = __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END, in __scs_alloc() 51 return kasan_reset_tag(s); in __scs_alloc() 56 void *s; in scs_alloc() local [all …]
|
D | cfi.c | 71 static inline int ptr_to_shadow(const struct cfi_shadow *s, unsigned long ptr) in ptr_to_shadow() argument 76 if (unlikely(page < s->base)) in ptr_to_shadow() 79 index = page - s->base; in ptr_to_shadow() 88 static inline unsigned long shadow_to_ptr(const struct cfi_shadow *s, in shadow_to_ptr() argument 94 return (s->base + index) << PAGE_SHIFT; in shadow_to_ptr() 98 static inline unsigned long shadow_to_check_fn(const struct cfi_shadow *s, in shadow_to_check_fn() argument 104 if (unlikely(s->shadow[index] == SHADOW_INVALID)) in shadow_to_check_fn() 108 return (s->base + s->shadow[index]) << PAGE_SHIFT; in shadow_to_check_fn() 146 static void add_module_to_shadow(struct cfi_shadow *s, struct module *mod, in add_module_to_shadow() argument 158 check_index = ptr_to_shadow(s, check); in add_module_to_shadow() [all …]
|
D | audit.c | 1240 struct audit_status s; in audit_receive_msg() local 1241 memset(&s, 0, sizeof(s)); in audit_receive_msg() 1242 s.enabled = audit_enabled; in audit_receive_msg() 1243 s.failure = audit_failure; in audit_receive_msg() 1246 s.pid = auditd_pid_vnr(); in audit_receive_msg() 1247 s.rate_limit = audit_rate_limit; in audit_receive_msg() 1248 s.backlog_limit = audit_backlog_limit; in audit_receive_msg() 1249 s.lost = atomic_read(&audit_lost); in audit_receive_msg() 1250 s.backlog = skb_queue_len(&audit_queue); in audit_receive_msg() 1251 s.feature_bitmap = AUDIT_FEATURE_BITMAP_ALL; in audit_receive_msg() [all …]
|
D | taskstats.c | 113 struct listener *s, *tmp; in send_cpu_listeners() local 122 list_for_each_entry(s, &listeners->list, list) { in send_cpu_listeners() 124 if (!list_is_last(&s->list, &listeners->list)) { in send_cpu_listeners() 129 rc = genlmsg_unicast(&init_net, skb_cur, s->pid); in send_cpu_listeners() 131 s->valid = 0; in send_cpu_listeners() 146 list_for_each_entry_safe(s, tmp, &listeners->list, list) { in send_cpu_listeners() 147 if (!s->valid) { in send_cpu_listeners() 148 list_del(&s->list); in send_cpu_listeners() 149 kfree(s); in send_cpu_listeners() 277 struct listener *s, *tmp, *s2; in add_del_listener() local [all …]
|
D | panic.c | 481 char *s; in print_tainted() local 484 s = buf + sprintf(buf, "Tainted: "); in print_tainted() 487 *s++ = test_bit(i, &tainted_mask) ? in print_tainted() 490 *s = 0; in print_tainted() 767 static int __init oops_setup(char *s) in oops_setup() argument 769 if (!s) in oops_setup() 771 if (!strcmp(s, "panic")) in oops_setup() 777 static int __init panic_on_taint_setup(char *s) in panic_on_taint_setup() argument 781 if (!s) in panic_on_taint_setup() 784 taint_str = strsep(&s, ","); in panic_on_taint_setup() [all …]
|
/kernel/printk/ |
D | index.c | 44 static void *pi_next(struct seq_file *s, void *v, loff_t *pos) in pi_next() argument 46 const struct module *mod = s->file->f_inode->i_private; in pi_next() 54 static void *pi_start(struct seq_file *s, loff_t *pos) in pi_start() argument 63 return pi_next(s, NULL, pos); in pi_start() 71 #define seq_escape_printf_format(s, src) \ argument 72 seq_escape_str(s, src, ESCAPE_ANY | ESCAPE_NAP | ESCAPE_APPEND, "\"\\") 74 static int pi_show(struct seq_file *s, void *v) in pi_show() argument 82 seq_puts(s, "# <level/flags> filename:line function \"format\"\n"); in pi_show() 102 seq_puts(s, "<c>"); in pi_show() 104 seq_printf(s, "<%d,c>", level); in pi_show() [all …]
|
/kernel/time/ |
D | timeconst.bc | 34 auto s, m; 35 for (s = 0; 1; s++) { 36 m = fmul(s,n,d); 38 return s; 61 s=fmuls(32,1000,hz) 63 print "#define HZ_TO_MSEC_MUL32\tU64_C(0x", fmul(s,1000,hz), ")\n" 64 print "#define HZ_TO_MSEC_ADJ32\tU64_C(0x", fadj(s,1000,hz), ")\n" 66 print "#define HZ_TO_MSEC_SHR32\t", s, "\n" 68 s=fmuls(32,hz,1000) 70 print "#define MSEC_TO_HZ_MUL32\tU64_C(0x", fmul(s,hz,1000), ")\n" [all …]
|
D | test_udelay.c | 27 static int udelay_test_single(struct seq_file *s, int usecs, uint32_t iters) in udelay_test_single() argument 57 seq_printf(s, "%d usecs x %d: exp=%d allowed=%d min=%d avg=%lld max=%d", in udelay_test_single() 61 seq_printf(s, " FAIL=%d", fail_count); in udelay_test_single() 62 seq_puts(s, "\n"); in udelay_test_single() 67 static int udelay_test_show(struct seq_file *s, void *v) in udelay_test_show() argument 79 return udelay_test_single(s, usecs, iters); in udelay_test_show() 84 seq_printf(s, "udelay() test (lpj=%ld kt=%lld.%09ld)\n", in udelay_test_show() 86 seq_puts(s, "usage:\n"); in udelay_test_show() 87 seq_puts(s, "echo USECS [ITERS] > " DEBUGFS_FILENAME "\n"); in udelay_test_show() 88 seq_puts(s, "cat " DEBUGFS_FILENAME "\n"); in udelay_test_show()
|
/kernel/rcu/ |
D | tree_exp.h | 48 unsigned long s; in rcu_exp_gp_seq_snap() local 51 s = rcu_seq_snap(&rcu_state.expedited_sequence); in rcu_exp_gp_seq_snap() 52 trace_rcu_exp_grace_period(rcu_state.name, s, TPS("snap")); in rcu_exp_gp_seq_snap() 53 return s; in rcu_exp_gp_seq_snap() 61 static bool rcu_exp_gp_seq_done(unsigned long s) in rcu_exp_gp_seq_done() argument 63 return rcu_seq_done(&rcu_state.expedited_sequence, s); in rcu_exp_gp_seq_done() 263 static bool sync_exp_work_done(unsigned long s) in sync_exp_work_done() argument 265 if (rcu_exp_gp_seq_done(s)) { in sync_exp_work_done() 266 trace_rcu_exp_grace_period(rcu_state.name, s, TPS("done")); in sync_exp_work_done() 280 static bool exp_funnel_lock(unsigned long s) in exp_funnel_lock() argument [all …]
|
D | srcutree.c | 596 unsigned long s) in srcu_funnel_exp_start() argument 601 if (rcu_seq_done(&ssp->srcu_gp_seq, s) || in srcu_funnel_exp_start() 602 ULONG_CMP_GE(READ_ONCE(snp->srcu_gp_seq_needed_exp), s)) in srcu_funnel_exp_start() 605 if (ULONG_CMP_GE(snp->srcu_gp_seq_needed_exp, s)) { in srcu_funnel_exp_start() 609 WRITE_ONCE(snp->srcu_gp_seq_needed_exp, s); in srcu_funnel_exp_start() 613 if (ULONG_CMP_LT(ssp->srcu_gp_seq_needed_exp, s)) in srcu_funnel_exp_start() 614 WRITE_ONCE(ssp->srcu_gp_seq_needed_exp, s); in srcu_funnel_exp_start() 629 unsigned long s, bool do_norm) in srcu_funnel_gp_start() argument 632 int idx = rcu_seq_ctr(s) % ARRAY_SIZE(sdp->mynode->srcu_have_cbs); in srcu_funnel_gp_start() 638 if (rcu_seq_done(&ssp->srcu_gp_seq, s) && snp != sdp->mynode) in srcu_funnel_gp_start() [all …]
|
D | rcu.h | 30 static inline unsigned long rcu_seq_ctr(unsigned long s) in rcu_seq_ctr() argument 32 return s >> RCU_SEQ_CTR_SHIFT; in rcu_seq_ctr() 39 static inline int rcu_seq_state(unsigned long s) in rcu_seq_state() argument 41 return s & RCU_SEQ_STATE_MASK; in rcu_seq_state() 89 unsigned long s; in rcu_seq_snap() local 91 s = (READ_ONCE(*sp) + 2 * RCU_SEQ_STATE_MASK + 1) & ~RCU_SEQ_STATE_MASK; in rcu_seq_snap() 93 return s; in rcu_seq_snap() 106 static inline bool rcu_seq_started(unsigned long *sp, unsigned long s) in rcu_seq_started() argument 108 return ULONG_CMP_LT((s - 1) & ~RCU_SEQ_STATE_MASK, READ_ONCE(*sp)); in rcu_seq_started() 115 static inline bool rcu_seq_done(unsigned long *sp, unsigned long s) in rcu_seq_done() argument [all …]
|
/kernel/power/ |
D | main.c | 127 char *s = buf; in mem_sleep_show() local 135 s += sprintf(s, "[%s] ", label); in mem_sleep_show() 137 s += sprintf(s, "%s ", label); in mem_sleep_show() 141 if (s != buf) in mem_sleep_show() 142 *(s-1) = '\n'; in mem_sleep_show() 144 return (s - buf); in mem_sleep_show() 242 char *s = buf; in pm_test_show() local 248 s += sprintf(s, "[%s] ", pm_tests[level]); in pm_test_show() 250 s += sprintf(s, "%s ", pm_tests[level]); in pm_test_show() 253 if (s != buf) in pm_test_show() [all …]
|
/kernel/irq/ |
D | timings.c | 540 struct irqt_stat __percpu *s; in irq_timings_next_event() local 570 s = idr_find(&irqt_stats, irq); in irq_timings_next_event() 571 if (s) in irq_timings_next_event() 572 irq_timings_store(irq, this_cpu_ptr(s), ts); in irq_timings_next_event() 579 idr_for_each_entry(&irqt_stats, s, i) { in irq_timings_next_event() 581 irqs = this_cpu_ptr(s); in irq_timings_next_event() 596 struct irqt_stat __percpu *s; in irq_timings_free() local 598 s = idr_find(&irqt_stats, irq); in irq_timings_free() 599 if (s) { in irq_timings_free() 600 free_percpu(s); in irq_timings_free() [all …]
|
/kernel/sched/ |
D | psi.c | 263 enum psi_states s; in get_recent_times() local 279 for (s = 0; s < NR_PSI_STATES; s++) { in get_recent_times() 290 if (state_mask & (1 << s)) in get_recent_times() 291 times[s] += now - state_start; in get_recent_times() 293 delta = times[s] - groupc->times_prev[aggregator][s]; in get_recent_times() 294 groupc->times_prev[aggregator][s] = times[s]; in get_recent_times() 296 times[s] = delta; in get_recent_times() 298 *pchanged_states |= (1 << s); in get_recent_times() 330 int s; in collect_percpu_times() local 352 for (s = 0; s < PSI_NONIDLE; s++) in collect_percpu_times() [all …]
|
/kernel/dma/ |
D | debug.c | 1298 struct scatterlist *s; in debug_dma_map_sg() local 1304 for_each_sg(sg, s, nents, i) { in debug_dma_map_sg() 1305 check_for_stack(dev, sg_page(s), s->offset); in debug_dma_map_sg() 1306 if (!PageHighMem(sg_page(s))) in debug_dma_map_sg() 1307 check_for_illegal_area(dev, sg_virt(s), s->length); in debug_dma_map_sg() 1310 for_each_sg(sg, s, mapped_ents, i) { in debug_dma_map_sg() 1317 entry->pfn = page_to_pfn(sg_page(s)); in debug_dma_map_sg() 1318 entry->offset = s->offset; in debug_dma_map_sg() 1319 entry->size = sg_dma_len(s); in debug_dma_map_sg() 1320 entry->dev_addr = sg_dma_address(s); in debug_dma_map_sg() [all …]
|