/arch/nios2/mm/ |
D | ioremap.c | 22 static inline void remap_area_pte(pte_t *pte, unsigned long address, in remap_area_pte() argument 31 address &= ~PMD_MASK; in remap_area_pte() 32 end = address + size; in remap_area_pte() 35 if (address >= end) in remap_area_pte() 44 address += PAGE_SIZE; in remap_area_pte() 47 } while (address && (address < end)); in remap_area_pte() 50 static inline int remap_area_pmd(pmd_t *pmd, unsigned long address, in remap_area_pmd() argument 56 address &= ~PGDIR_MASK; in remap_area_pmd() 57 end = address + size; in remap_area_pmd() 60 phys_addr -= address; in remap_area_pmd() [all …]
|
D | fault.c | 43 unsigned long address) in do_page_fault() argument 66 if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) { in do_page_fault() 73 if (unlikely(address >= TASK_SIZE)) in do_page_fault() 93 vma = find_vma(mm, address); in do_page_fault() 96 if (vma->vm_start <= address) in do_page_fault() 100 if (expand_stack(vma, address)) in do_page_fault() 134 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 190 "cause %ld\n", current->comm, SIGSEGV, address, cause); in do_page_fault() 193 _exception(SIGSEGV, regs, code, address); in do_page_fault() 209 address < PAGE_SIZE ? "NULL pointer dereference" : in do_page_fault() [all …]
|
/arch/x86/mm/ |
D | fault.c | 225 force_sig_info_fault(int si_signo, int si_code, unsigned long address, in force_sig_info_fault() argument 234 info.si_addr = (void __user *)address; in force_sig_info_fault() 250 static inline pmd_t *vmalloc_sync_one(pgd_t *pgd, unsigned long address) in vmalloc_sync_one() argument 252 unsigned index = pgd_index(address); in vmalloc_sync_one() 268 pud = pud_offset(pgd, address); in vmalloc_sync_one() 269 pud_k = pud_offset(pgd_k, address); in vmalloc_sync_one() 273 pmd = pmd_offset(pud, address); in vmalloc_sync_one() 274 pmd_k = pmd_offset(pud_k, address); in vmalloc_sync_one() 288 unsigned long address; in vmalloc_sync_all() local 293 for (address = VMALLOC_START & PMD_MASK; in vmalloc_sync_all() [all …]
|
/arch/mips/mm/ |
D | ioremap.c | 19 static inline void remap_area_pte(pte_t * pte, unsigned long address, in remap_area_pte() argument 27 address &= ~PMD_MASK; in remap_area_pte() 28 end = address + size; in remap_area_pte() 31 BUG_ON(address >= end); in remap_area_pte() 39 address += PAGE_SIZE; in remap_area_pte() 42 } while (address && (address < end)); in remap_area_pte() 45 static inline int remap_area_pmd(pmd_t * pmd, unsigned long address, in remap_area_pmd() argument 50 address &= ~PGDIR_MASK; in remap_area_pmd() 51 end = address + size; in remap_area_pmd() 54 phys_addr -= address; in remap_area_pmd() [all …]
|
D | fault.c | 39 unsigned long address) in __do_page_fault() argument 53 current->comm, current->pid, field, address, write, in __do_page_fault() 83 if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) in __do_page_fault() 86 if (unlikely(address >= MODULE_START && address < MODULE_END)) in __do_page_fault() 101 vma = find_vma(mm, address); in __do_page_fault() 104 if (vma->vm_start <= address) in __do_page_fault() 108 if (expand_stack(vma, address)) in __do_page_fault() 123 if (address == regs->cp0_epc && !(vma->vm_flags & VM_EXEC)) { in __do_page_fault() 128 field, address, write, in __do_page_fault() 134 exception_epc(regs) != address) { in __do_page_fault() [all …]
|
/arch/sh/mm/ |
D | fault.c | 41 force_sig_info_fault(int si_signo, int si_code, unsigned long address, in force_sig_info_fault() argument 49 info.si_addr = (void __user *)address; in force_sig_info_fault() 127 static inline pmd_t *vmalloc_sync_one(pgd_t *pgd, unsigned long address) in vmalloc_sync_one() argument 129 unsigned index = pgd_index(address); in vmalloc_sync_one() 140 pud = pud_offset(pgd, address); in vmalloc_sync_one() 141 pud_k = pud_offset(pgd_k, address); in vmalloc_sync_one() 148 pmd = pmd_offset(pud, address); in vmalloc_sync_one() 149 pmd_k = pmd_offset(pud_k, address); in vmalloc_sync_one() 177 static noinline int vmalloc_fault(unsigned long address) in vmalloc_fault() argument 184 if (!(address >= VMALLOC_START && address < __FAULT_ADDR_LIMIT)) in vmalloc_fault() [all …]
|
D | tlbex_32.c | 23 unsigned long address) in handle_tlbmiss() argument 36 if (address >= P3SEG && address < P3_ADDR_MAX) { in handle_tlbmiss() 37 pgd = pgd_offset_k(address); in handle_tlbmiss() 39 if (unlikely(address >= TASK_SIZE || !current->mm)) in handle_tlbmiss() 42 pgd = pgd_offset(current->mm, address); in handle_tlbmiss() 45 pud = pud_offset(pgd, address); in handle_tlbmiss() 48 pmd = pmd_offset(pud, address); in handle_tlbmiss() 51 pte = pte_offset_kernel(pmd, address); in handle_tlbmiss() 71 local_flush_tlb_one(get_asid(), address & PAGE_MASK); in handle_tlbmiss() 75 update_mmu_cache(NULL, address, pte); in handle_tlbmiss()
|
D | tlbex_64.c | 44 unsigned long address) in handle_tlbmiss() argument 52 if (is_vmalloc_addr((void *)address)) { in handle_tlbmiss() 53 pgd = pgd_offset_k(address); in handle_tlbmiss() 55 if (unlikely(address >= TASK_SIZE || !current->mm)) in handle_tlbmiss() 58 pgd = pgd_offset(current->mm, address); in handle_tlbmiss() 61 pud = pud_offset(pgd, address); in handle_tlbmiss() 65 pmd = pmd_offset(pud, address); in handle_tlbmiss() 69 pte = pte_offset_kernel(pmd, address); in handle_tlbmiss() 83 update_mmu_cache(NULL, address, pte); in handle_tlbmiss() 132 unsigned long address) in do_fast_page_fault() argument [all …]
|
/arch/mips/include/asm/octeon/ |
D | cvmx-scratch.h | 51 static inline uint8_t cvmx_scratch_read8(uint64_t address) in cvmx_scratch_read8() argument 53 return *CASTPTR(volatile uint8_t, CVMX_SCRATCH_BASE + address); in cvmx_scratch_read8() 63 static inline uint16_t cvmx_scratch_read16(uint64_t address) in cvmx_scratch_read16() argument 65 return *CASTPTR(volatile uint16_t, CVMX_SCRATCH_BASE + address); in cvmx_scratch_read16() 75 static inline uint32_t cvmx_scratch_read32(uint64_t address) in cvmx_scratch_read32() argument 77 return *CASTPTR(volatile uint32_t, CVMX_SCRATCH_BASE + address); in cvmx_scratch_read32() 87 static inline uint64_t cvmx_scratch_read64(uint64_t address) in cvmx_scratch_read64() argument 89 return *CASTPTR(volatile uint64_t, CVMX_SCRATCH_BASE + address); in cvmx_scratch_read64() 98 static inline void cvmx_scratch_write8(uint64_t address, uint64_t value) in cvmx_scratch_write8() argument 100 *CASTPTR(volatile uint8_t, CVMX_SCRATCH_BASE + address) = in cvmx_scratch_write8() [all …]
|
D | cvmx-asm.h | 93 #define CVMX_PREPARE_FOR_STORE(address, offset) \ argument 95 [rbase] "d" (address)) 101 #define CVMX_DONT_WRITE_BACK(address, offset) \ argument 103 [rbase] "d" (address)) 117 #define CVMX_CACHE(op, address, offset) \ argument 119 : : [rbase] "d" (address) ) 121 #define CVMX_CACHE_LCKL2(address, offset) CVMX_CACHE(31, address, offset) argument 123 #define CVMX_CACHE_WBIL2(address, offset) CVMX_CACHE(23, address, offset) argument 125 #define CVMX_CACHE_WBIL2I(address, offset) CVMX_CACHE(3, address, offset) argument 127 #define CVMX_CACHE_LTGL2I(address, offset) CVMX_CACHE(7, address, offset) argument
|
/arch/x86/math-emu/ |
D | get_address.c | 122 static unsigned long vm86_segment(u_char segment, struct address *addr) in vm86_segment() 137 struct address *addr, long offset) in pm_address() 140 unsigned long base_address, limit, address, seg_top; in pm_address() local 163 address = base_address + offset; in pm_address() 178 (address <= limit) || (address >= seg_top) ? 0 : in pm_address() 179 ((seg_top - address) >= 255 ? 255 : seg_top - address); in pm_address() 182 (address > limit) || (address < base_address) ? 0 : in pm_address() 183 ((limit - address) >= 254 ? 255 : limit - address + 1); in pm_address() 189 return address; in pm_address() 209 struct address *addr, fpu_addr_modes addr_modes) in FPU_get_address() [all …]
|
/arch/s390/mm/ |
D | vmem.c | 85 unsigned long address = start; in vmem_add_mem() local 93 while (address < end) { in vmem_add_mem() 94 pg_dir = pgd_offset_k(address); in vmem_add_mem() 101 pu_dir = pud_offset(pg_dir, address); in vmem_add_mem() 102 if (MACHINE_HAS_EDAT2 && pud_none(*pu_dir) && address && in vmem_add_mem() 103 !(address & ~PUD_MASK) && (address + PUD_SIZE <= end) && in vmem_add_mem() 105 pud_val(*pu_dir) = address | pgprot_val(REGION3_KERNEL); in vmem_add_mem() 106 address += PUD_SIZE; in vmem_add_mem() 116 pm_dir = pmd_offset(pu_dir, address); in vmem_add_mem() 117 if (MACHINE_HAS_EDAT1 && pmd_none(*pm_dir) && address && in vmem_add_mem() [all …]
|
/arch/tile/mm/ |
D | fault.c | 47 int si_code, unsigned long address, in force_sig_info_fault() argument 56 si_signo, si_code & 0xffff, address, in force_sig_info_fault() 63 info.si_addr = (void __user *)address; in force_sig_info_fault() 65 trace_unhandled_signal(type, regs, address, si_signo); in force_sig_info_fault() 74 SYSCALL_DEFINE1(cmpxchg_badaddr, unsigned long, address) in SYSCALL_DEFINE1() argument 78 if (address >= PAGE_OFFSET) in SYSCALL_DEFINE1() 80 address, INT_DTLB_MISS, current, regs); in SYSCALL_DEFINE1() 83 BUS_ADRALN, address, in SYSCALL_DEFINE1() 104 static inline pmd_t *vmalloc_sync_one(pgd_t *pgd, unsigned long address) in vmalloc_sync_one() argument 106 unsigned index = pgd_index(address); in vmalloc_sync_one() [all …]
|
/arch/xtensa/mm/ |
D | fault.c | 43 unsigned int address = regs->excvaddr; in do_page_fault() local 55 if (address >= TASK_SIZE && !user_mode(regs)) in do_page_fault() 62 bad_page_fault(regs, address, SIGSEGV); in do_page_fault() 73 address, exccause, regs->pc, is_write? "w":"", is_exec? "x":""); in do_page_fault() 80 vma = find_vma(mm, address); in do_page_fault() 84 if (vma->vm_start <= address) in do_page_fault() 88 if (expand_stack(vma, address)) in do_page_fault() 113 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 146 perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); in do_page_fault() 148 perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, regs, address); in do_page_fault() [all …]
|
/arch/arc/mm/ |
D | fault.c | 28 noinline static int handle_kernel_vaddr_fault(unsigned long address) in handle_kernel_vaddr_fault() argument 38 pgd = pgd_offset_fast(current->active_mm, address); in handle_kernel_vaddr_fault() 39 pgd_k = pgd_offset_k(address); in handle_kernel_vaddr_fault() 44 pud = pud_offset(pgd, address); in handle_kernel_vaddr_fault() 45 pud_k = pud_offset(pgd_k, address); in handle_kernel_vaddr_fault() 49 pmd = pmd_offset(pud, address); in handle_kernel_vaddr_fault() 50 pmd_k = pmd_offset(pud_k, address); in handle_kernel_vaddr_fault() 63 void do_page_fault(unsigned long address, struct pt_regs *regs) in do_page_fault() argument 82 if (address >= VMALLOC_START) { in do_page_fault() 83 ret = handle_kernel_vaddr_fault(address); in do_page_fault() [all …]
|
/arch/arc/kernel/ |
D | traps.c | 31 void die(const char *str, struct pt_regs *regs, unsigned long address) in die() argument 33 show_kernel_fault_diag(str, regs, address); in die() 66 int name(unsigned long address, struct pt_regs *regs) \ 72 .si_addr = (void __user *)address, \ 90 int do_misaligned_access(unsigned long address, struct pt_regs *regs, in do_misaligned_access() argument 94 if (misaligned_fixup(address, regs, cregs) != 0) in do_misaligned_access() 95 return do_misaligned_error(address, regs); in do_misaligned_access() 104 void do_machine_check_fault(unsigned long address, struct pt_regs *regs) in do_machine_check_fault() argument 106 die("Machine Check Exception", regs, address); in do_machine_check_fault() 119 void do_non_swi_trap(unsigned long address, struct pt_regs *regs) in do_non_swi_trap() argument [all …]
|
D | stacktrace.c | 117 unsigned int address; in arc_unwind_core() local 123 address = UNW_PC(&frame_info); in arc_unwind_core() 125 if (!address || !__kernel_text_address(address)) in arc_unwind_core() 128 if (consumer_fn(address, arg) == -1) in arc_unwind_core() 138 return address; /* return the last address it saw */ in arc_unwind_core() 162 static int __print_sym(unsigned int address, void *unused) in __print_sym() argument 164 __print_symbol(" %s\n", address); in __print_sym() 173 static int __collect_all(unsigned int address, void *arg) in __collect_all() argument 180 trace->entries[trace->nr_entries++] = address; in __collect_all() 188 static int __collect_all_but_sched(unsigned int address, void *arg) in __collect_all_but_sched() argument [all …]
|
/arch/score/mm/ |
D | fault.c | 45 unsigned long address) in do_page_fault() argument 66 if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) in do_page_fault() 69 if (unlikely(address >= MODULE_START && address < MODULE_END)) in do_page_fault() 84 vma = find_vma(mm, address); in do_page_fault() 87 if (vma->vm_start <= address) in do_page_fault() 91 if (expand_stack(vma, address)) in do_page_fault() 114 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 142 tsk->thread.cp0_badvaddr = address; in do_page_fault() 147 info.si_addr = (void __user *) address; in do_page_fault() 155 current->thread.cp0_baduaddr = address; in do_page_fault() [all …]
|
/arch/ia64/mm/ |
D | fault.c | 48 mapped_kernel_page_is_present (unsigned long address) in mapped_kernel_page_is_present() argument 55 pgd = pgd_offset_k(address); in mapped_kernel_page_is_present() 59 pud = pud_offset(pgd, address); in mapped_kernel_page_is_present() 63 pmd = pmd_offset(pud, address); in mapped_kernel_page_is_present() 67 ptep = pte_offset_kernel(pmd, address); in mapped_kernel_page_is_present() 80 ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *regs) in ia64_do_page_fault() argument 110 if ((REGION_NUMBER(address) == 5) && !user_mode(regs)) in ia64_do_page_fault() 127 vma = find_vma_prev(mm, address, &prev_vma); in ia64_do_page_fault() 138 if (( !vma && prev_vma ) || (address < vma->vm_start) ) in ia64_do_page_fault() 162 fault = handle_mm_fault(vma, address, flags); in ia64_do_page_fault() [all …]
|
/arch/mn10300/mm/ |
D | fault.c | 68 static void print_pagetable_entries(pgd_t *pgdir, unsigned long address) 74 pgd = pgdir + __pgd_offset(address); 82 pmd = pmd_offset(pgd, address); 90 pte = pte_offset(pmd, address); 119 unsigned long address) in do_page_fault() argument 141 fault_code & 0xffff, address); in do_page_fault() 158 if (address >= VMALLOC_START && address < VMALLOC_END && in do_page_fault() 179 vma = find_vma(mm, address); in do_page_fault() 182 if (vma->vm_start <= address) in do_page_fault() 190 if ((address & PAGE_MASK) + 2 * PAGE_SIZE < regs->sp) { in do_page_fault() [all …]
|
/arch/mips/include/asm/ |
D | pgtable-64.h | 240 #define __pgd_offset(address) pgd_index(address) argument 241 #define __pud_offset(address) (((address) >> PUD_SHIFT) & (PTRS_PER_PUD-1)) argument 242 #define __pmd_offset(address) pmd_index(address) argument 245 #define pgd_offset_k(address) pgd_offset(&init_mm, address) argument 247 #define pgd_index(address) (((address) >> PGDIR_SHIFT) & (PTRS_PER_PGD-1)) argument 248 #define pmd_index(address) (((address) >> PMD_SHIFT) & (PTRS_PER_PMD-1)) argument 262 static inline pmd_t *pmd_offset(pud_t * pud, unsigned long address) in pmd_offset() argument 264 return (pmd_t *) pud_page_vaddr(*pud) + pmd_index(address); in pmd_offset() 269 #define __pte_offset(address) \ argument 270 (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) [all …]
|
/arch/x86/mm/kmemcheck/ |
D | shadow.c | 18 void *kmemcheck_shadow_lookup(unsigned long address) in kmemcheck_shadow_lookup() argument 23 if (!virt_addr_valid(address)) in kmemcheck_shadow_lookup() 26 pte = kmemcheck_pte_lookup(address); in kmemcheck_shadow_lookup() 30 page = virt_to_page(address); in kmemcheck_shadow_lookup() 33 return page->shadow + (address & (PAGE_SIZE - 1)); in kmemcheck_shadow_lookup() 36 static void mark_shadow(void *address, unsigned int n, in mark_shadow() argument 39 unsigned long addr = (unsigned long) address; in mark_shadow() 77 void kmemcheck_mark_unallocated(void *address, unsigned int n) in kmemcheck_mark_unallocated() argument 79 mark_shadow(address, n, KMEMCHECK_SHADOW_UNALLOCATED); in kmemcheck_mark_unallocated() 82 void kmemcheck_mark_uninitialized(void *address, unsigned int n) in kmemcheck_mark_uninitialized() argument [all …]
|
/arch/sparc/mm/ |
D | fault_32.c | 38 static void __noreturn unhandled_fault(unsigned long address, in unhandled_fault() argument 42 if ((unsigned long) address < PAGE_SIZE) { in unhandled_fault() 47 address); in unhandled_fault() 58 unsigned long address) in lookup_fault() argument 99 unhandled_fault(address, current, ®s); in lookup_fault() 107 unsigned long address, struct task_struct *tsk) in show_signal_msg() argument 117 tsk->comm, task_pid_nr(tsk), address, in show_signal_msg() 168 unsigned long address) in do_sparc_fault() argument 180 address = regs->pc; in do_sparc_fault() 192 if (address >= TASK_SIZE) in do_sparc_fault() [all …]
|
/arch/cris/mm/ |
D | fault.c | 53 do_page_fault(unsigned long address, struct pt_regs *regs, in do_page_fault() argument 65 address, smp_processor_id(), instruction_pointer(regs), in do_page_fault() 89 if (address >= VMALLOC_START && in do_page_fault() 101 address == cris_signal_return_page && in do_page_fault() 123 vma = find_vma(mm, address); in do_page_fault() 126 if (vma->vm_start <= address) in do_page_fault() 137 if (address + PAGE_SIZE < rdusp()) in do_page_fault() 140 if (expand_stack(vma, address)) in do_page_fault() 171 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 228 address, instruction_pointer(regs)); in do_page_fault() [all …]
|
/arch/x86/include/asm/trace/ |
D | exceptions.h | 14 TP_PROTO(unsigned long address, struct pt_regs *regs, 17 TP_ARGS(address, regs, error_code), 20 __field( unsigned long, address ) 26 __entry->address = address; 32 (void *)__entry->address, (void *)__entry->ip, 37 TP_PROTO(unsigned long address, struct pt_regs *regs, \ 39 TP_ARGS(address, regs, error_code), \
|