/kernel/ |
D | delayacct.c | 45 static void delayacct_end(raw_spinlock_t *lock, u64 *start, u64 *total, in delayacct_end() argument 53 *total += ns; in delayacct_end() 71 u64 *total; in __delayacct_blkio_end() local 75 total = &delays->swapin_delay; in __delayacct_blkio_end() 78 total = &delays->blkio_delay; in __delayacct_blkio_end() 82 delayacct_end(&delays->lock, &delays->blkio_start, total, count); in __delayacct_blkio_end()
|
D | stacktrace.c | 49 unsigned int generated, i, total = 0; in stack_trace_snprint() local 58 total += generated; in stack_trace_snprint() 68 return total; in stack_trace_snprint()
|
D | kprobes.c | 2451 unsigned int i, total = 0, errors = 0; in arm_all_kprobes() local 2477 total++; in arm_all_kprobes() 2484 errors, total); in arm_all_kprobes() 2497 unsigned int i, total = 0, errors = 0; in disarm_all_kprobes() local 2520 total++; in disarm_all_kprobes() 2527 errors, total); in disarm_all_kprobes()
|
D | fork.c | 154 int total = 0; in nr_processes() local 157 total += per_cpu(process_counts, cpu); in nr_processes() 159 return total; in nr_processes()
|
/kernel/sched/ |
D | psi.c | 350 group->total[aggregator][s] += in collect_percpu_times() 383 sample = group->total[PSI_AVGS][s] - group->avg_total[s]; in update_averages() 497 group->total[PSI_POLL][t->state], 0); in init_triggers() 498 memcpy(group->polling_total, group->total[PSI_POLL], in init_triggers() 507 u64 *total = group->total[PSI_POLL]; in update_triggers() local 517 if (group->polling_total[t->state] == total[t->state]) in update_triggers() 529 growth = window_update(&t->win, now, total[t->state]); in update_triggers() 544 memcpy(group->polling_total, total, in update_triggers() 958 u64 total; in psi_show() local 963 total = div_u64(group->total[PSI_AVGS][res * 2 + full], in psi_show() [all …]
|
D | cputime.c | 42 irqtime->total += delta; in irqtime_account_delta() 538 static u64 scale_stime(u64 stime, u64 rtime, u64 total) in scale_stime() argument 548 if (total >> 32) in scale_stime() 567 total >>= 1; in scale_stime() 574 scaled = div_u64((u64) (u32) stime * (u64) (u32) rtime, (u32)total); in scale_stime()
|
D | sched.h | 2264 u64 total; member 2281 u64 total; in irq_time_read() local 2285 total = irqtime->total; in irq_time_read() 2288 return total; in irq_time_read()
|
D | rt.c | 2484 unsigned long total, sum = 0; in tg_rt_schedulable() local 2507 total = to_ratio(period, runtime); in tg_rt_schedulable() 2512 if (total > to_ratio(global_rt_period(), global_rt_runtime())) in tg_rt_schedulable() 2530 if (sum > total) in tg_rt_schedulable()
|
/kernel/trace/ |
D | trace_hwlat.c | 171 s64 diff, total, last_total = 0; in get_sample() local 208 total = time_to_us(time_sub(t2, start)); /* sample width */ in get_sample() 211 if (total < last_total) { in get_sample() 215 last_total = total; in get_sample() 229 } while (total <= hwlat_data.sample_width); in get_sample()
|
D | trace_probe_tmpl.h | 76 int total = 0, ret = 0, i = 0; in process_fetch_insn_bottom() local 146 total += ret; in process_fetch_insn_bottom() 164 ret = total; in process_fetch_insn_bottom()
|
D | trace.c | 3518 get_total_entries_cpu(struct trace_buffer *buf, unsigned long *total, in get_total_entries_cpu() argument 3532 *total = count; in get_total_entries_cpu() 3534 *total = count + in get_total_entries_cpu() 3541 unsigned long *total, unsigned long *entries) in get_total_entries() argument 3546 *total = 0; in get_total_entries() 3551 *total += t; in get_total_entries() 3558 unsigned long total, entries; in trace_total_entries_cpu() local 3563 get_total_entries_cpu(&tr->trace_buffer, &total, &entries, cpu); in trace_total_entries_cpu() 3570 unsigned long total, entries; in trace_total_entries() local 3575 get_total_entries(&tr->trace_buffer, &total, &entries); in trace_total_entries() [all …]
|
/kernel/dma/ |
D | Kconfig | 109 int "Percentage of total memory" 115 Allocator as a percentage of the total memory in the system.
|
/kernel/locking/ |
D | lockdep_proc.c | 402 seq_time(m, lt->total); in seq_lock_time() 403 seq_time(m, lt->nr ? div_s64(lt->total, lt->nr) : 0); in seq_lock_time()
|
D | lockdep.c | 210 lt->total += time; in lock_time_inc() 225 dst->total += src->total; in lock_time_add()
|
/kernel/bpf/ |
D | btf.c | 3167 u32 total, expected_total, i; in btf_check_sec_info() local 3183 total = 0; in btf_check_sec_info() 3190 if (total < secs[i].off) { in btf_check_sec_info() 3195 if (total > secs[i].off) { in btf_check_sec_info() 3199 if (expected_total - total < secs[i].len) { in btf_check_sec_info() 3204 total += secs[i].len; in btf_check_sec_info() 3208 if (expected_total != total) { in btf_check_sec_info()
|
/kernel/power/ |
D | snapshot.c | 1576 unsigned long total) in preallocate_highmem_fraction() argument 1578 unsigned long alloc = __fraction(nr_pages, highmem, total); in preallocate_highmem_fraction() 1590 unsigned long total) in preallocate_highmem_fraction() argument
|
/kernel/events/ |
D | core.c | 4796 u64 total = 0; in __perf_event_read_value() local 4804 total += perf_event_count(event); in __perf_event_read_value() 4813 total += perf_event_count(child); in __perf_event_read_value() 4819 return total; in __perf_event_read_value()
|