/kernel/ |
D | resource.c | 32 .start = 0, 40 .start = 0, 109 unsigned long long start, end; in r_show() local 118 start = r->start; in r_show() 121 start = end = 0; in r_show() 126 width, start, in r_show() 133 .start = r_start, 187 resource_size_t start = new->start; in __request_resource() local 191 if (end < start) in __request_resource() 193 if (start < root->start) in __request_resource() [all …]
|
D | range.c | 11 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) in add_range() argument 13 if (start >= end) in add_range() 20 range[nr_range].start = start; in add_range() 29 u64 start, u64 end) in add_range_with_merge() argument 33 if (start >= end) in add_range_with_merge() 43 common_start = max(range[i].start, start); in add_range_with_merge() 49 start = min(range[i].start, start); in add_range_with_merge() 54 range[nr_range - 1].start = 0; in add_range_with_merge() 61 return add_range(range, az, nr_range, start, end); in add_range_with_merge() 64 void subtract_range(struct range *range, int az, u64 start, u64 end) in subtract_range() argument [all …]
|
D | kexec_file.c | 301 image->control_page = crashk_res.start; in kimage_file_alloc_init() 431 static int locate_mem_hole_top_down(unsigned long start, unsigned long end, in locate_mem_hole_top_down() argument 444 if (temp_start < start || temp_start < kbuf->buf_min) in locate_mem_hole_top_down() 469 static int locate_mem_hole_bottom_up(unsigned long start, unsigned long end, in locate_mem_hole_bottom_up() argument 475 temp_start = max(start, kbuf->buf_min); in locate_mem_hole_bottom_up() 506 u64 start = res->start, end = res->end; in locate_mem_hole_callback() local 507 unsigned long sz = end - start + 1; in locate_mem_hole_callback() 513 if (end < kbuf->buf_min || start > kbuf->buf_max) in locate_mem_hole_callback() 521 return locate_mem_hole_top_down(start, end, kbuf); in locate_mem_hole_callback() 522 return locate_mem_hole_bottom_up(start, end, kbuf); in locate_mem_hole_callback() [all …]
|
D | jump_label.c | 80 jump_label_sort_entries(struct jump_entry *start, struct jump_entry *stop) in jump_label_sort_entries() argument 88 size = (((unsigned long)stop - (unsigned long)start) in jump_label_sort_entries() 90 sort(start, size, sizeof(struct jump_entry), jump_label_cmp, swapfn); in jump_label_sort_entries() 309 static int addr_conflict(struct jump_entry *entry, void *start, void *end) in addr_conflict() argument 312 jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE > (unsigned long)start) in addr_conflict() 319 struct jump_entry *iter_stop, void *start, void *end) in __jump_label_text_reserved() argument 325 if (addr_conflict(iter, start, end)) in __jump_label_text_reserved() 539 static int __jump_label_mod_text_reserved(void *start, void *end) in __jump_label_mod_text_reserved() argument 544 mod = __module_text_address((unsigned long)start); in __jump_label_mod_text_reserved() 554 start, end); in __jump_label_mod_text_reserved() [all …]
|
D | kexec_core.c | 60 .start = 0, 67 .start = 0, 245 if ((mstart < phys_to_boot_phys(crashk_res.start)) || in sanity_check_segment_list() 282 unsigned long start, in kimage_is_destination_range() argument 292 if ((end > mstart) && (start < mend)) in kimage_is_destination_range() 991 if (crashk_res.end != crashk_res.start) in crash_get_memory_size() 1009 unsigned long start, end; in crash_shrink_memory() local 1019 start = crashk_res.start; in crash_shrink_memory() 1021 old_size = (end == 0) ? 0 : end - start + 1; in crash_shrink_memory() 1034 end = start + new_size; in crash_shrink_memory() [all …]
|
D | sys.c | 2279 unsigned long start, unsigned long end, in prctl_update_vma_anon_name() argument 2291 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); in prctl_update_vma_anon_name() 2292 *prev = vma_merge(mm, *prev, start, end, vma->vm_flags, vma->anon_vma, in prctl_update_vma_anon_name() 2302 if (start != vma->vm_start) { in prctl_update_vma_anon_name() 2303 error = split_vma(mm, vma, start, 1); in prctl_update_vma_anon_name() 2324 static int prctl_set_vma_anon_name(unsigned long start, unsigned long end, in prctl_set_vma_anon_name() argument 2337 vma = find_vma_prev(current->mm, start, &prev); in prctl_set_vma_anon_name() 2338 if (vma && start > vma->vm_start) in prctl_set_vma_anon_name() 2348 if (start < vma->vm_start) { in prctl_set_vma_anon_name() 2350 start = vma->vm_start; in prctl_set_vma_anon_name() [all …]
|
D | kexec.c | 50 if ((entry < phys_to_boot_phys(crashk_res.start)) || in kimage_alloc_init() 60 image->start = entry; in kimage_alloc_init() 68 image->control_page = crashk_res.start; in kimage_alloc_init()
|
D | crash_core.c | 45 unsigned long long start, end = ULLONG_MAX, size; in parse_crashkernel_mem() local 48 start = memparse(cur, &tmp); in parse_crashkernel_mem() 68 if (end <= start) { in parse_crashkernel_mem() 92 if (system_ram >= start && system_ram < end) { in parse_crashkernel_mem()
|
D | relay.c | 51 page = vmalloc_to_page(buf->start + (pgoff << PAGE_SHIFT)); in relay_buf_fault() 177 buf->start = relay_alloc_buf(buf, &chan->alloc_size); in relay_create_buf() 178 if (!buf->start) in relay_create_buf() 213 if (likely(buf->start)) { in relay_destroy_buf() 214 vunmap(buf->start); in relay_destroy_buf() 368 buf->data = buf->start; in __relay_reset() 781 new = buf->start + new_subbuf * buf->chan->subbuf_size; in relay_switch_subbuf() 1051 write_subbuf = (buf->data - buf->start) / subbuf_size; in relay_file_read_subbuf_avail() 1147 from = buf->start + read_start; in relay_file_read()
|
/kernel/printk/ |
D | printk_safe.c | 123 static int printk_safe_flush_buffer(const char *start, size_t len) in printk_safe_flush_buffer() argument 128 c = start; in printk_safe_flush_buffer() 129 end = start + len; in printk_safe_flush_buffer() 135 printk_safe_flush_line(start, c - start + 1); in printk_safe_flush_buffer() 136 start = ++c; in printk_safe_flush_buffer() 148 printk_safe_flush_line(start, c - start); in printk_safe_flush_buffer() 149 start = c++; in printk_safe_flush_buffer() 159 if (start < end && !header) { in printk_safe_flush_buffer() 162 printk_safe_flush_line(start, end - start); in printk_safe_flush_buffer()
|
/kernel/irq/ |
D | irqdesc.c | 470 static int alloc_descs(unsigned int start, unsigned int cnt, int node, in alloc_descs() argument 499 desc = alloc_desc(start + i, node, flags, mask, owner); in alloc_descs() 502 irq_insert_desc(start + i, desc); in alloc_descs() 503 irq_sysfs_add(start + i, desc); in alloc_descs() 504 irq_add_debugfs_entry(start + i, desc); in alloc_descs() 506 bitmap_set(allocated_irqs, start, cnt); in alloc_descs() 507 return start; in alloc_descs() 511 free_desc(start + i); in alloc_descs() 602 static inline int alloc_descs(unsigned int start, unsigned int cnt, int node, in alloc_descs() argument 609 struct irq_desc *desc = irq_to_desc(start + i); in alloc_descs() [all …]
|
D | timings.c | 384 int index, i, period_max, count, start, min = INT_MAX; in __irq_timings_next_event() local 412 start = irqs->count < IRQ_TIMINGS_SIZE ? in __irq_timings_next_event() 422 int index = (start + i) & IRQ_TIMINGS_MASK; in __irq_timings_next_event() 716 int index, start, i, count, period_max; in irq_timings_test_next_index() local 735 start = count < IRQ_TIMINGS_SIZE ? 0 : in irq_timings_test_next_index() 741 int index = (start + i) & IRQ_TIMINGS_MASK; in irq_timings_test_next_index() 839 int start = count >= IRQ_TIMINGS_SIZE ? count - IRQ_TIMINGS_SIZE : 0; in irq_timings_test_irqts() local 857 ots += start; in irq_timings_test_irqts() 858 oirq += start; in irq_timings_test_irqts()
|
/kernel/trace/ |
D | trace_printk.c | 53 void hold_module_trace_bprintk_format(const char **start, const char **end) in hold_module_trace_bprintk_format() argument 59 if (start != end) in hold_module_trace_bprintk_format() 63 for (iter = start; iter < end; iter++) { in hold_module_trace_bprintk_format() 93 const char **start = mod->trace_bprintk_fmt_start; in module_trace_bprintk_format_notify() local 94 const char **end = start + mod->num_trace_bprintk_fmt; in module_trace_bprintk_format_notify() 97 hold_module_trace_bprintk_format(start, end); in module_trace_bprintk_format_notify() 343 .start = t_start,
|
D | trace_mmiotrace.c | 66 resource_size_t start, end; in mmio_print_pcidev() local 73 start = dev->resource[i].start; in mmio_print_pcidev() 75 (unsigned long long)(start | in mmio_print_pcidev() 79 start = dev->resource[i].start; in mmio_print_pcidev() 82 dev->resource[i].start < dev->resource[i].end ? in mmio_print_pcidev() 83 (unsigned long long)(end - start) + 1 : 0); in mmio_print_pcidev() 281 .start = mmio_trace_start,
|
D | preemptirq_delay_test.c | 28 u64 start, end; in busy_wait() local 29 start = trace_clock_local(); in busy_wait() 34 } while ((end - start) < (time * 1000)); in busy_wait()
|
D | trace_stack.c | 157 unsigned long this_size, flags; unsigned long *p, *top, *start; in check_stack() local 212 start = stack; in check_stack() 214 (((unsigned long)start & ~(THREAD_SIZE-1)) + THREAD_SIZE); in check_stack() 227 p = start; in check_stack() 240 start = p + 1; in check_stack() 471 .start = t_start,
|
D | trace_benchmark.c | 39 u64 start; in trace_do_benchmark() local 53 start = trace_clock_local(); in trace_do_benchmark() 60 delta = stop - start; in trace_do_benchmark()
|
D | fgraph.c | 346 int start = 0, end = FTRACE_RETSTACK_ALLOC_SIZE; in alloc_retstack_tasklist() local 355 start = 0; in alloc_retstack_tasklist() 364 if (start == end) { in alloc_retstack_tasklist() 375 t->ret_stack = ret_stack_list[start++]; in alloc_retstack_tasklist() 382 for (i = start; i < end; i++) in alloc_retstack_tasklist()
|
D | trace_syscalls.c | 83 struct syscall_metadata **start; in find_syscall_meta() local 88 start = __start_syscalls_metadata; in find_syscall_meta() 95 for ( ; start < stop; start++) { in find_syscall_meta() 96 if ((*start)->name && arch_syscall_match_sym_name(str, (*start)->name)) in find_syscall_meta() 97 return *start; in find_syscall_meta()
|
D | trace_hwlat.c | 170 time_type start, t1, t2, last_t2; in get_sample() local 188 start = time_get(); /* start timestamp */ in get_sample() 208 total = time_to_us(time_sub(t2, start)); /* sample width */ in get_sample() 618 .start = hwlat_tracer_start,
|
/kernel/power/ |
D | swap.c | 124 unsigned long start; member 140 if (swap_offset < ext->start) { in swsusp_extents_insert() 142 if (swap_offset == ext->start - 1) { in swsusp_extents_insert() 143 ext->start--; in swsusp_extents_insert() 164 ext->start = swap_offset; in swsusp_extents_insert() 206 for (offset = ext->start; offset <= ext->end; offset++) in free_all_swap_pages() 539 ktime_t start; in save_image() local 550 start = ktime_get(); in save_image() 569 swsusp_show_speed(start, stop, nr_to_write, "Wrote"); in save_image() 675 ktime_t start; in save_image_lzo() local [all …]
|
D | process.c | 37 ktime_t start, end, elapsed; in try_to_freeze_tasks() local 42 start = ktime_get_boottime(); in try_to_freeze_tasks() 85 elapsed = ktime_sub(end, start); in try_to_freeze_tasks()
|
D | snapshot.c | 501 unsigned long start, in create_zone_bm_rtree() argument 508 pages = end - start; in create_zone_bm_rtree() 515 zone->start_pfn = start; in create_zone_bm_rtree() 562 unsigned long start; member 604 if (&ext->hook == list || zone_end < ext->start) { in create_mem_extents() 613 new_ext->start = zone_start; in create_mem_extents() 620 if (zone_start < ext->start) in create_mem_extents() 621 ext->start = zone_start; in create_mem_extents() 628 if (zone_end < cur->start) in create_mem_extents() 662 ext->start, ext->end); in memory_bm_create() [all …]
|
/kernel/dma/ |
D | remap.c | 176 bool dma_in_atomic_pool(void *start, size_t size) in dma_in_atomic_pool() argument 181 return addr_in_gen_pool(atomic_pool, (unsigned long)start, size); in dma_in_atomic_pool() 206 bool dma_free_from_pool(void *start, size_t size) in dma_free_from_pool() argument 208 if (!dma_in_atomic_pool(start, size)) in dma_free_from_pool() 210 gen_pool_free(atomic_pool, (unsigned long)start, size); in dma_free_from_pool()
|
/kernel/locking/ |
D | qspinlock_stat.h | 110 u64 start = sched_clock(); in __pv_kick() local 112 per_cpu(pv_kick_time, cpu) = start; in __pv_kick() 114 this_cpu_add(EVENT_COUNT(pv_latency_kick), sched_clock() - start); in __pv_kick()
|