/arch/sh/kernel/ |
D | stacktrace.c | 41 if (trace->nr_entries < trace->max_entries) in save_stack_address() 42 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 55 if (trace->nr_entries < trace->max_entries) in save_stack_trace() 56 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace() 76 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 77 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched() 90 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 91 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_tsk()
|
/arch/s390/kernel/ |
D | stacktrace.c | 24 if (trace->nr_entries < trace->max_entries) { in __save_address() 25 trace->entries[trace->nr_entries++] = address; in __save_address() 47 if (trace->nr_entries < trace->max_entries) in save_stack_trace() 48 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace() 60 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 61 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_tsk() 71 if (trace->nr_entries < trace->max_entries) in save_stack_trace_regs() 72 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_regs()
|
/arch/parisc/kernel/ |
D | stacktrace.c | 35 trace->nr_entries = 0; in dump_trace() 36 while (trace->nr_entries < trace->max_entries) { in dump_trace() 41 trace->entries[trace->nr_entries++] = info.ip; in dump_trace() 52 if (trace->nr_entries < trace->max_entries) in save_stack_trace() 53 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace() 60 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 61 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_tsk()
|
/arch/x86/kernel/ |
D | stacktrace.c | 26 if (trace->nr_entries >= trace->max_entries) in save_stack_address() 29 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 50 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 51 trace->entries[trace->nr_entries++] = ULONG_MAX; in __save_stack_trace() 146 if (trace->nr_entries < trace->max_entries) in __save_stack_trace_reliable() 147 trace->entries[trace->nr_entries++] = ULONG_MAX; in __save_stack_trace_reliable() 207 if (trace->nr_entries < trace->max_entries) in __save_stack_trace_user() 208 trace->entries[trace->nr_entries++] = regs->ip; in __save_stack_trace_user() 210 while (trace->nr_entries < trace->max_entries) { in __save_stack_trace_user() 220 trace->entries[trace->nr_entries++] = in __save_stack_trace_user() [all …]
|
D | e820.c | 80 for (i = 0; i < e820_table->nr_entries; i++) { in e820__mapped_any() 104 for (i = 0; i < e820_table->nr_entries; i++) { in __e820__mapped_all() 155 int x = table->nr_entries; in __e820__range_add() 165 table->nr_entries++; in __e820__range_add() 192 for (i = 0; i < e820_table->nr_entries; i++) { in e820__print_table() 302 if (table->nr_entries < 2) in e820__update_table() 305 BUG_ON(table->nr_entries > max_nr_entries); in e820__update_table() 308 for (i = 0; i < table->nr_entries; i++) { in e820__update_table() 314 for (i = 0; i < 2 * table->nr_entries; i++) in e820__update_table() 322 for (i = 0; i < table->nr_entries; i++) { in e820__update_table() [all …]
|
D | ldt.c | 99 new_ldt->nr_entries = num_entries; in alloc_ldt_struct() 136 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in map_ldt_struct() 204 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in unmap_ldt_struct() 241 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct() 262 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct() 263 if (ldt->nr_entries * LDT_ENTRY_SIZE > PAGE_SIZE) in free_ldt_struct() 286 new_ldt = alloc_ldt_struct(old_mm->context.ldt->nr_entries); in ldt_dup_context() 293 new_ldt->nr_entries * LDT_ENTRY_SIZE); in ldt_dup_context() 341 entries_size = mm->context.ldt->nr_entries * LDT_ENTRY_SIZE; in read_ldt() 424 old_nr_entries = old_ldt ? old_ldt->nr_entries : 0; in write_ldt()
|
D | resource.c | 31 for (i = 0; i < e820_table->nr_entries; i++) { in remove_e820_regions()
|
/arch/arm/kernel/ |
D | stacktrace.c | 85 trace->entries[trace->nr_entries++] = addr; in save_trace() 87 if (trace->nr_entries >= trace->max_entries) in save_trace() 104 trace->entries[trace->nr_entries++] = regs->ARM_pc; in save_trace() 106 return trace->nr_entries >= trace->max_entries; in save_trace() 128 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 129 trace->entries[trace->nr_entries++] = ULONG_MAX; in __save_stack_trace() 147 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 148 trace->entries[trace->nr_entries++] = ULONG_MAX; in __save_stack_trace() 166 if (trace->nr_entries < trace->max_entries) in save_stack_trace_regs() 167 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_regs()
|
/arch/mips/kernel/ |
D | stacktrace.c | 29 trace->entries[trace->nr_entries++] = addr; in save_raw_context_stack() 30 if (trace->nr_entries >= trace->max_entries) in save_raw_context_stack() 57 trace->entries[trace->nr_entries++] = pc; in save_context_stack() 58 if (trace->nr_entries >= trace->max_entries) in save_context_stack() 82 WARN_ON(trace->nr_entries || !trace->max_entries); in save_stack_trace_tsk()
|
/arch/ia64/kernel/ |
D | stacktrace.c | 19 trace->nr_entries = 0; in ia64_do_save_stack() 25 trace->entries[trace->nr_entries++] = ip; in ia64_do_save_stack() 26 if (trace->nr_entries == trace->max_entries) in ia64_do_save_stack()
|
/arch/cris/kernel/ |
D | stacktrace.c | 44 trace->entries[trace->nr_entries++] = addr; in save_trace() 46 return trace->nr_entries >= trace->max_entries; in save_trace() 66 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 67 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_tsk()
|
/arch/um/kernel/ |
D | stacktrace.c | 53 if (trace->nr_entries >= trace->max_entries) in save_addr() 56 trace->entries[trace->nr_entries++] = address; in save_addr() 66 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 67 trace->entries[trace->nr_entries++] = ULONG_MAX; in __save_stack_trace()
|
/arch/metag/kernel/ |
D | stacktrace.c | 137 trace->entries[trace->nr_entries++] = addr; in save_trace() 139 return trace->nr_entries >= trace->max_entries; in save_trace() 157 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 158 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_tsk() 178 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 179 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_tsk()
|
/arch/arm64/kernel/ |
D | stacktrace.c | 126 trace->entries[trace->nr_entries++] = addr; in save_trace() 128 return trace->nr_entries >= trace->max_entries; in save_trace() 147 if (trace->nr_entries < trace->max_entries) in save_stack_trace_regs() 148 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_regs() 178 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 179 trace->entries[trace->nr_entries++] = ULONG_MAX; in __save_stack_trace()
|
/arch/sparc/kernel/ |
D | stacktrace.c | 57 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 63 if (trace->nr_entries < in __save_stack_trace() 65 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 71 } while (trace->nr_entries < trace->max_entries); in __save_stack_trace()
|
/arch/unicore32/kernel/ |
D | stacktrace.c | 93 trace->entries[trace->nr_entries++] = addr; in save_trace() 95 return trace->nr_entries >= trace->max_entries; in save_trace() 123 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 124 trace->entries[trace->nr_entries++] = ULONG_MAX; in save_stack_trace_tsk()
|
/arch/blackfin/kernel/ |
D | stacktrace.c | 41 trace->entries[trace->nr_entries++] = frame->rets; in save_stack_trace() 42 if (trace->nr_entries >= trace->max_entries) in save_stack_trace()
|
/arch/hexagon/kernel/ |
D | stacktrace.c | 52 trace->entries[trace->nr_entries++] = frame->rets; in save_stack_trace() 53 if (trace->nr_entries >= trace->max_entries) in save_stack_trace()
|
/arch/x86/xen/ |
D | setup.c | 205 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_find_pfn_range() 472 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_foreach_remap_area() 474 if (entry->type == E820_TYPE_RAM || i == xen_e820_table.nr_entries - 1) { in xen_foreach_remap_area() 602 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_ignore_unusable() 620 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++) { in xen_is_e820_reserved() 645 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++, entry++) { in xen_find_free_area() 748 memmap.nr_entries = ARRAY_SIZE(xen_e820_table.entries); in xen_memory_setup() 757 memmap.nr_entries = 1; in xen_memory_setup() 766 BUG_ON(memmap.nr_entries == 0); in xen_memory_setup() 767 xen_e820_table.nr_entries = memmap.nr_entries; in xen_memory_setup() [all …]
|
D | enlighten_pvh.c | 42 memmap.nr_entries = ARRAY_SIZE(pvh_bootparams.e820_table); in init_pvh_bootparams() 49 pvh_bootparams.e820_entries = memmap.nr_entries; in init_pvh_bootparams()
|
/arch/arc/kernel/ |
D | stacktrace.c | 182 trace->entries[trace->nr_entries++] = address; in __collect_all() 184 if (trace->nr_entries >= trace->max_entries) in __collect_all() 200 trace->entries[trace->nr_entries++] = address; in __collect_all_but_sched() 202 if (trace->nr_entries >= trace->max_entries) in __collect_all_but_sched()
|
/arch/powerpc/kernel/ |
D | stacktrace.c | 38 trace->entries[trace->nr_entries++] = ip; in save_context_stack() 43 if (trace->nr_entries >= trace->max_entries) in save_context_stack()
|
/arch/x86/include/asm/ |
D | mmu_context.h | 54 unsigned int nr_entries; member 143 set_ldt(ldt_slot_va(ldt->slot), ldt->nr_entries); in load_mm_ldt() 145 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt()
|
/arch/x86/include/asm/e820/ |
D | types.h | 87 __u32 nr_entries; member
|
/arch/microblaze/kernel/ |
D | unwind.c | 243 trace->entries[trace->nr_entries++] = pc; in microblaze_unwind_inner() 245 if (trace->nr_entries >= trace->max_entries) in microblaze_unwind_inner()
|