Home
last modified time | relevance | path

Searched refs:start (Results 1 – 25 of 93) sorted by relevance

1234

/kernel/
Dresource.c35 .start = 0,
43 .start = 0,
100 unsigned long long start, end; in r_show() local
109 start = r->start; in r_show()
112 start = end = 0; in r_show()
117 width, start, in r_show()
124 .start = r_start,
161 resource_size_t start = new->start; in __request_resource() local
165 if (end < start) in __request_resource()
167 if (start < root->start) in __request_resource()
[all …]
Drange.c12 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) in add_range() argument
14 if (start >= end) in add_range()
21 range[nr_range].start = start; in add_range()
30 u64 start, u64 end) in add_range_with_merge() argument
34 if (start >= end) in add_range_with_merge()
44 common_start = max(range[i].start, start); in add_range_with_merge()
50 start = min(range[i].start, start); in add_range_with_merge()
55 range[nr_range - 1].start = 0; in add_range_with_merge()
62 return add_range(range, az, nr_range, start, end); in add_range_with_merge()
65 void subtract_range(struct range *range, int az, u64 start, u64 end) in subtract_range() argument
[all …]
Dresource_kunit.c22 static struct resource r0 = { .start = R0_START, .end = R0_END };
23 static struct resource r1 = { .start = R1_START, .end = R1_END };
24 static struct resource r2 = { .start = R2_START, .end = R2_END };
25 static struct resource r3 = { .start = R3_START, .end = R3_END };
26 static struct resource r4 = { .start = R4_START, .end = R4_END };
37 .r1 = &r1, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true,
39 .r1 = &r2, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true,
41 .r1 = &r3, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true,
43 .r1 = &r4, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true,
49 .r1 = &r4, .r2 = &r1, .r.start = R1_START, .r.end = R4_END, .ret = true,
[all …]
Dstatic_call_inline.c95 static inline void static_call_sort_entries(struct static_call_site *start, in static_call_sort_entries() argument
98 sort(start, stop - start, sizeof(struct static_call_site), in static_call_sort_entries()
210 struct static_call_site *start, in __static_call_init() argument
217 if (start == stop) in __static_call_init()
220 static_call_sort_entries(start, stop); in __static_call_init()
222 for (site = start; site < stop; site++) { in __static_call_init()
282 static int addr_conflict(struct static_call_site *site, void *start, void *end) in addr_conflict() argument
287 addr + CALL_INSN_SIZE > (unsigned long)start) in addr_conflict()
295 void *start, void *end, bool init) in __static_call_text_reserved() argument
301 if (addr_conflict(iter, start, end)) in __static_call_text_reserved()
[all …]
Dkexec_file.c289 image->control_page = crashk_res.start; in kimage_file_alloc_init()
423 static int locate_mem_hole_top_down(unsigned long start, unsigned long end, in locate_mem_hole_top_down() argument
436 if (temp_start < start || temp_start < kbuf->buf_min) in locate_mem_hole_top_down()
461 static int locate_mem_hole_bottom_up(unsigned long start, unsigned long end, in locate_mem_hole_bottom_up() argument
467 temp_start = max(start, kbuf->buf_min); in locate_mem_hole_bottom_up()
498 u64 start = res->start, end = res->end; in locate_mem_hole_callback() local
499 unsigned long sz = end - start + 1; in locate_mem_hole_callback()
510 if (end < kbuf->buf_min || start > kbuf->buf_max) in locate_mem_hole_callback()
518 return locate_mem_hole_top_down(start, end, kbuf); in locate_mem_hole_callback()
519 return locate_mem_hole_bottom_up(start, end, kbuf); in locate_mem_hole_callback()
[all …]
Djump_label.c80 jump_label_sort_entries(struct jump_entry *start, struct jump_entry *stop) in jump_label_sort_entries() argument
88 size = (((unsigned long)stop - (unsigned long)start) in jump_label_sort_entries()
90 sort(start, size, sizeof(struct jump_entry), jump_label_cmp, swapfn); in jump_label_sort_entries()
309 static int addr_conflict(struct jump_entry *entry, void *start, void *end) in addr_conflict() argument
312 jump_entry_code(entry) + jump_entry_size(entry) > (unsigned long)start) in addr_conflict()
319 struct jump_entry *iter_stop, void *start, void *end, bool init) in __jump_label_text_reserved() argument
326 if (addr_conflict(iter, start, end)) in __jump_label_text_reserved()
550 static int __jump_label_mod_text_reserved(void *start, void *end) in __jump_label_mod_text_reserved() argument
556 mod = __module_text_address((unsigned long)start); in __jump_label_mod_text_reserved()
567 start, end, mod->state == MODULE_STATE_COMING); in __jump_label_mod_text_reserved()
[all …]
Dkexec_core.c61 .start = 0,
68 .start = 0,
246 if ((mstart < phys_to_boot_phys(crashk_res.start)) || in sanity_check_segment_list()
283 unsigned long start, in kimage_is_destination_range() argument
293 if ((end > mstart) && (start < mend)) in kimage_is_destination_range()
999 if (crashk_res.end != crashk_res.start) in crash_get_memory_size()
1018 unsigned long start, end; in crash_shrink_memory() local
1029 start = crashk_res.start; in crash_shrink_memory()
1031 old_size = (end == 0) ? 0 : end - start + 1; in crash_shrink_memory()
1044 end = start + new_size; in crash_shrink_memory()
[all …]
Dkprobes.c2265 int kprobe_add_area_blacklist(unsigned long start, unsigned long end) in kprobe_add_area_blacklist() argument
2270 for (entry = start; entry < end; entry += ret) { in kprobe_add_area_blacklist()
2281 static void kprobe_remove_area_blacklist(unsigned long start, unsigned long end) in kprobe_remove_area_blacklist() argument
2286 if (ent->start_addr < start || ent->start_addr >= end) in kprobe_remove_area_blacklist()
2333 static int __init populate_kprobe_blacklist(unsigned long *start, in populate_kprobe_blacklist() argument
2340 for (iter = start; iter < end; iter++) { in populate_kprobe_blacklist()
2364 unsigned long start, end; in add_module_kprobe_blacklist() local
2372 start = (unsigned long)mod->kprobes_text_start; in add_module_kprobe_blacklist()
2373 if (start) { in add_module_kprobe_blacklist()
2374 end = start + mod->kprobes_text_size; in add_module_kprobe_blacklist()
[all …]
Dpadata.c443 unsigned long start, size, end; in padata_mt_helper() local
445 start = job->start; in padata_mt_helper()
447 size = roundup(start + 1, ps->chunk_size) - start; in padata_mt_helper()
449 end = start + size; in padata_mt_helper()
451 job->start = end; in padata_mt_helper()
455 job->thread_fn(start, end, job->fn_arg); in padata_mt_helper()
491 job->thread_fn(job->start, job->start + job->size, job->fn_arg); in padata_do_multithreaded()
Dkexec.c33 if ((entry < phys_to_boot_phys(crashk_res.start)) || in kimage_alloc_init()
43 image->start = entry; in kimage_alloc_init()
49 image->control_page = crashk_res.start; in kimage_alloc_init()
Drelay.c42 page = vmalloc_to_page(buf->start + (pgoff << PAGE_SHIFT)); in relay_buf_fault()
165 buf->start = relay_alloc_buf(buf, &chan->alloc_size); in relay_create_buf()
166 if (!buf->start) in relay_create_buf()
201 if (likely(buf->start)) { in relay_destroy_buf()
202 vunmap(buf->start); in relay_destroy_buf()
302 buf->data = buf->start; in __relay_reset()
696 new = buf->start + new_subbuf * buf->chan->subbuf_size; in relay_switch_subbuf()
966 write_subbuf = (buf->data - buf->start) / subbuf_size; in relay_file_read_subbuf_avail()
1062 from = buf->start + read_start; in relay_file_read()
/kernel/bpf/
Dtrampoline.c47 ksym->start = (unsigned long) data; in bpf_image_ksym_add()
48 ksym->end = ksym->start + PAGE_SIZE; in bpf_image_ksym_add()
50 perf_event_ksymbol(PERF_RECORD_KSYMBOL_TYPE_BPF, ksym->start, in bpf_image_ksym_add()
57 perf_event_ksymbol(PERF_RECORD_KSYMBOL_TYPE_BPF, ksym->start, in bpf_image_ksym_del()
537 u64 start = NO_START_TIME; in bpf_prog_start_time() local
540 start = sched_clock(); in bpf_prog_start_time()
541 if (unlikely(!start)) in bpf_prog_start_time()
542 start = NO_START_TIME; in bpf_prog_start_time()
544 return start; in bpf_prog_start_time()
584 u64 start) in update_prog_stats() argument
[all …]
/kernel/dma/
Dswiotlb.c144 pr_info("mapped [mem %pa-%pa] (%luMB)\n", &mem->start, &mem->end, in swiotlb_print_info()
172 vaddr = phys_to_virt(mem->start); in swiotlb_update_mem_attributes()
178 static void swiotlb_init_io_tlb_mem(struct io_tlb_mem *mem, phys_addr_t start, in swiotlb_init_io_tlb_mem() argument
181 void *vaddr = phys_to_virt(start); in swiotlb_init_io_tlb_mem()
185 mem->start = start; in swiotlb_init_io_tlb_mem()
186 mem->end = mem->start + bytes; in swiotlb_init_io_tlb_mem()
339 tbl_vaddr = (unsigned long)phys_to_virt(mem->start); in swiotlb_exit()
340 tbl_size = PAGE_ALIGN(mem->end - mem->start); in swiotlb_exit()
348 memblock_free_late(mem->start, tbl_size); in swiotlb_exit()
370 int index = (tlb_addr - mem->start) >> IO_TLB_SHIFT; in swiotlb_bounce()
[all …]
/kernel/irq/
Dirqdesc.c470 static int alloc_descs(unsigned int start, unsigned int cnt, int node, in alloc_descs() argument
499 desc = alloc_desc(start + i, node, flags, mask, owner); in alloc_descs()
502 irq_insert_desc(start + i, desc); in alloc_descs()
503 irq_sysfs_add(start + i, desc); in alloc_descs()
504 irq_add_debugfs_entry(start + i, desc); in alloc_descs()
506 bitmap_set(allocated_irqs, start, cnt); in alloc_descs()
507 return start; in alloc_descs()
511 free_desc(start + i); in alloc_descs()
602 static inline int alloc_descs(unsigned int start, unsigned int cnt, int node, in alloc_descs() argument
609 struct irq_desc *desc = irq_to_desc(start + i); in alloc_descs()
[all …]
Dtimings.c384 int index, i, period_max, count, start, min = INT_MAX; in __irq_timings_next_event() local
412 start = irqs->count < IRQ_TIMINGS_SIZE ? in __irq_timings_next_event()
422 int index = (start + i) & IRQ_TIMINGS_MASK; in __irq_timings_next_event()
716 int index, start, i, count, period_max; in irq_timings_test_next_index() local
735 start = count < IRQ_TIMINGS_SIZE ? 0 : in irq_timings_test_next_index()
741 int index = (start + i) & IRQ_TIMINGS_MASK; in irq_timings_test_next_index()
839 int start = count >= IRQ_TIMINGS_SIZE ? count - IRQ_TIMINGS_SIZE : 0; in irq_timings_test_irqts() local
857 ots += start; in irq_timings_test_irqts()
858 oirq += start; in irq_timings_test_irqts()
/kernel/trace/
Dtrace_printk.c53 void hold_module_trace_bprintk_format(const char **start, const char **end) in hold_module_trace_bprintk_format() argument
59 if (start != end) in hold_module_trace_bprintk_format()
63 for (iter = start; iter < end; iter++) { in hold_module_trace_bprintk_format()
93 const char **start = mod->trace_bprintk_fmt_start; in module_trace_bprintk_format_notify() local
94 const char **end = start + mod->num_trace_bprintk_fmt; in module_trace_bprintk_format_notify()
97 hold_module_trace_bprintk_format(start, end); in module_trace_bprintk_format_notify()
354 .start = t_start,
Dtrace_mmiotrace.c64 resource_size_t start, end; in mmio_print_pcidev() local
71 start = dev->resource[i].start; in mmio_print_pcidev()
73 (unsigned long long)(start | in mmio_print_pcidev()
77 start = dev->resource[i].start; in mmio_print_pcidev()
80 dev->resource[i].start < dev->resource[i].end ? in mmio_print_pcidev()
81 (unsigned long long)(end - start) + 1 : 0); in mmio_print_pcidev()
279 .start = mmio_trace_start,
Dtrace_stack.c157 unsigned long this_size, flags; unsigned long *p, *top, *start; in check_stack() local
212 start = stack; in check_stack()
214 (((unsigned long)start & ~(THREAD_SIZE-1)) + THREAD_SIZE); in check_stack()
227 p = start; in check_stack()
240 start = p + 1; in check_stack()
470 .start = t_start,
Dtrace_benchmark.c39 u64 start; in trace_do_benchmark() local
53 start = trace_clock_local(); in trace_do_benchmark()
60 delta = stop - start; in trace_do_benchmark()
Dfgraph.c373 int start = 0, end = FTRACE_RETSTACK_ALLOC_SIZE; in alloc_retstack_tasklist() local
382 start = 0; in alloc_retstack_tasklist()
391 if (start == end) { in alloc_retstack_tasklist()
402 t->ret_stack = ret_stack_list[start++]; in alloc_retstack_tasklist()
409 for (i = start; i < end; i++) in alloc_retstack_tasklist()
Dpreemptirq_delay_test.c41 u64 start, end; in busy_wait() local
43 start = trace_clock_local(); in busy_wait()
49 } while ((end - start) < (time * 1000)); in busy_wait()
Dtrace_syscalls.c85 struct syscall_metadata **start; in find_syscall_meta() local
90 start = __start_syscalls_metadata; in find_syscall_meta()
97 for ( ; start < stop; start++) { in find_syscall_meta()
98 if ((*start)->name && arch_syscall_match_sym_name(str, (*start)->name)) in find_syscall_meta()
99 return *start; in find_syscall_meta()
/kernel/power/
Dswap.c124 unsigned long start; member
140 if (swap_offset < ext->start) { in swsusp_extents_insert()
142 if (swap_offset == ext->start - 1) { in swsusp_extents_insert()
143 ext->start--; in swsusp_extents_insert()
164 ext->start = swap_offset; in swsusp_extents_insert()
206 for (offset = ext->start; offset <= ext->end; offset++) in free_all_swap_pages()
547 ktime_t start; in save_image() local
558 start = ktime_get(); in save_image()
578 swsusp_show_speed(start, stop, nr_to_write, "Wrote"); in save_image()
684 ktime_t start; in save_image_lzo() local
[all …]
Dprocess.c39 ktime_t start, end, elapsed; in try_to_freeze_tasks() local
44 start = ktime_get_boottime(); in try_to_freeze_tasks()
87 elapsed = ktime_sub(end, start); in try_to_freeze_tasks()
/kernel/locking/
Dqspinlock_stat.h110 u64 start = sched_clock(); in __pv_kick() local
112 per_cpu(pv_kick_time, cpu) = start; in __pv_kick()
114 this_cpu_add(EVENT_COUNT(pv_latency_kick), sched_clock() - start); in __pv_kick()

1234