/kernel/bpf/ |
D | mprog.c | 131 int total = bpf_mprog_total(entry); in bpf_mprog_insert() local 138 if (idx == total) in bpf_mprog_insert() 155 int total = bpf_mprog_total(entry); in bpf_mprog_delete() local 162 else if (idx == total) in bpf_mprog_delete() 163 idx = total - 1; in bpf_mprog_delete() 300 int total = bpf_mprog_total(entry); in bpf_mprog_fetch() local 308 else if (idx == total) in bpf_mprog_fetch() 309 idx = total - 1; in bpf_mprog_fetch()
|
D | trampoline.c | 229 bpf_trampoline_get_progs(const struct bpf_trampoline *tr, int *total, bool *ip_arg) in bpf_trampoline_get_progs() argument 236 *total = 0; in bpf_trampoline_get_progs() 243 *total += tr->progs_cnt[kind]; in bpf_trampoline_get_progs() 399 int err, total; in bpf_trampoline_update() local 401 tlinks = bpf_trampoline_get_progs(tr, &total, &ip_arg); in bpf_trampoline_update() 405 if (total == 0) { in bpf_trampoline_update() 449 WARN_ON(tr->cur_image && total == 0); in bpf_trampoline_update()
|
D | hashtab.c | 1686 u32 bucket_cnt, total, key_size, value_size, roundup_key_size; in __htab_map_lookup_and_delete_batch() local 1731 total = 0; in __htab_map_lookup_and_delete_batch() 1775 if (bucket_cnt > (max_count - total)) { in __htab_map_lookup_and_delete_batch() 1776 if (total == 0) in __htab_map_lookup_and_delete_batch() 1874 if (bucket_cnt && (copy_to_user(ukeys + total * key_size, keys, in __htab_map_lookup_and_delete_batch() 1876 copy_to_user(uvalues + total * value_size, values, in __htab_map_lookup_and_delete_batch() 1882 total += bucket_cnt; in __htab_map_lookup_and_delete_batch() 1897 put_user(total, &uattr->batch.count)) in __htab_map_lookup_and_delete_batch()
|
D | btf.c | 527 u32 total = 0; in btf_nr_types() local 530 total += btf->nr_types; in btf_nr_types() 534 return total; in btf_nr_types() 541 u32 i, total; in btf_find_by_name_kind() local 543 total = btf_nr_types(btf); in btf_find_by_name_kind() 544 for (i = 1; i < total; i++) { in btf_find_by_name_kind() 5194 u32 total, expected_total, i; in btf_check_sec_info() local 5210 total = 0; in btf_check_sec_info() 5217 if (total < secs[i].off) { in btf_check_sec_info() 5222 if (total > secs[i].off) { in btf_check_sec_info() [all …]
|
/kernel/sched/ |
D | psi.c | 385 group->total[aggregator][s] += in collect_percpu_times() 442 u64 *total = group->total[aggregator]; in update_triggers() local 463 new_stall = aggregator_total[t->state] != total[t->state]; in update_triggers() 484 growth = window_update(&t->win, now, total[t->state]); in update_triggers() 542 sample = group->total[PSI_AVGS][s] - group->avg_total[s]; in update_averages() 611 group->total[PSI_POLL][t->state], 0); in init_rtpoll_triggers() 612 memcpy(group->rtpoll_total, group->total[PSI_POLL], in init_rtpoll_triggers() 717 memcpy(group->rtpoll_total, group->total[PSI_POLL], in psi_rtpoll_work() 1275 u64 total = 0; in psi_show() local 1282 total = div_u64(group->total[PSI_AVGS][res * 2 + full], in psi_show() [all …]
|
D | sched.h | 2958 u64 total; member 2975 u64 total; in irq_time_read() local 2979 total = irqtime->total; in irq_time_read() 2982 return total; in irq_time_read()
|
D | rt.c | 2885 unsigned long total, sum = 0; in tg_rt_schedulable() local 2909 total = to_ratio(period, runtime); in tg_rt_schedulable() 2914 if (total > to_ratio(global_rt_period(), global_rt_runtime())) in tg_rt_schedulable() 2932 if (sum > total) in tg_rt_schedulable()
|
D | cputime.c | 47 irqtime->total += delta; in irqtime_account_delta()
|
D | core.c | 11414 u64 total = 0; in throttled_time_self() local 11417 total += READ_ONCE(tg->cfs_rq[i]->throttled_clock_self_time); in throttled_time_self() 11420 return total; in throttled_time_self()
|
/kernel/kcsan/ |
D | selftest.c | 244 int total = 0; in kcsan_selftest() local 248 ++total; \ in kcsan_selftest() 259 pr_info("selftest: %d/%d tests passed\n", passed, total); in kcsan_selftest() 260 if (passed != total) in kcsan_selftest()
|
/kernel/ |
D | stacktrace.c | 50 unsigned int generated, i, total = 0; in stack_trace_snprint() local 59 total += generated; in stack_trace_snprint() 69 return total; in stack_trace_snprint()
|
D | delayacct.c | 99 static void delayacct_end(raw_spinlock_t *lock, u64 *start, u64 *total, u32 *count) in delayacct_end() argument 106 *total += ns; in delayacct_end()
|
D | kprobes.c | 2906 unsigned int i, total = 0, errors = 0; in arm_all_kprobes() local 2932 total++; in arm_all_kprobes() 2939 errors, total); in arm_all_kprobes() 2952 unsigned int i, total = 0, errors = 0; in disarm_all_kprobes() local 2975 total++; in disarm_all_kprobes() 2982 errors, total); in disarm_all_kprobes()
|
D | fork.c | 163 int total = 0; in nr_processes() local 166 total += per_cpu(process_counts, cpu); in nr_processes() 168 return total; in nr_processes()
|
/kernel/trace/ |
D | trace_probe_tmpl.h | 127 int total = 0, ret = 0, i = 0; in process_fetch_insn_bottom() local 209 total += ret; in process_fetch_insn_bottom() 227 ret = total; in process_fetch_insn_bottom()
|
D | trace_hwlat.c | 207 s64 diff, outer_diff, total, last_total = 0; in get_sample() local 245 total = time_to_us(time_sub(t2, start)); /* sample width */ in get_sample() 248 if (total < last_total) { in get_sample() 252 last_total = total; in get_sample() 272 } while (total <= hwlat_data.sample_width); in get_sample()
|
D | trace_osnoise.c | 1421 s64 total, last_total = 0; in run_osnoise() local 1502 total = time_sub(sample, start); in run_osnoise() 1507 if (total < last_total) { in run_osnoise() 1512 last_total = total; in run_osnoise() 1566 } while (total < runtime && !kthread_should_stop()); in run_osnoise() 1593 s.runtime = time_to_us(total); in run_osnoise()
|
D | trace.c | 4241 get_total_entries_cpu(struct array_buffer *buf, unsigned long *total, in get_total_entries_cpu() argument 4255 *total = count; in get_total_entries_cpu() 4257 *total = count + in get_total_entries_cpu() 4264 unsigned long *total, unsigned long *entries) in get_total_entries() argument 4269 *total = 0; in get_total_entries() 4274 *total += t; in get_total_entries() 4281 unsigned long total, entries; in trace_total_entries_cpu() local 4286 get_total_entries_cpu(&tr->array_buffer, &total, &entries, cpu); in trace_total_entries_cpu() 4293 unsigned long total, entries; in trace_total_entries() local 4298 get_total_entries(&tr->array_buffer, &total, &entries); in trace_total_entries() [all …]
|
D | trace_events_hist.c | 5430 static inline unsigned int __get_percentage(u64 val, u64 total) in __get_percentage() argument 5432 if (!total) in __get_percentage() 5436 return (unsigned int)div64_ul(val * 10000, total); in __get_percentage() 5438 total = div64_u64(total, 10000); in __get_percentage() 5439 if (!total) in __get_percentage() 5442 return (unsigned int)div64_ul(val, total); in __get_percentage() 5471 u64 total; member 5484 pc = __get_percentage(val, stats[idx].total); in hist_trigger_print_val() 5561 stats[j].total += val; in print_entries()
|
/kernel/dma/ |
D | Kconfig | 186 int "Percentage of total memory" 192 Allocator as a percentage of the total memory in the system.
|
/kernel/locking/ |
D | lockdep_proc.c | 454 seq_time(m, lt->total); in seq_lock_time() 455 seq_time(m, lt->nr ? div64_u64(lt->total, lt->nr) : 0); in seq_lock_time()
|
D | lockdep.c | 275 lt->total += time; in lock_time_inc() 290 dst->total += src->total; in lock_time_add()
|
/kernel/power/ |
D | snapshot.c | 1708 unsigned long total) in preallocate_highmem_fraction() argument 1710 unsigned long alloc = __fraction(nr_pages, highmem, total); in preallocate_highmem_fraction() 1722 unsigned long total) in preallocate_highmem_fraction() argument
|
/kernel/events/ |
D | core.c | 5450 u64 total = 0; in __perf_event_read_value() local 5458 total += perf_event_count(event); in __perf_event_read_value() 5467 total += perf_event_count(child); in __perf_event_read_value() 5473 return total; in __perf_event_read_value()
|