Lines Matching refs:address
48 mapped_kernel_page_is_present (unsigned long address) in mapped_kernel_page_is_present() argument
55 pgd = pgd_offset_k(address); in mapped_kernel_page_is_present()
59 pud = pud_offset(pgd, address); in mapped_kernel_page_is_present()
63 pmd = pmd_offset(pud, address); in mapped_kernel_page_is_present()
67 ptep = pte_offset_kernel(pmd, address); in mapped_kernel_page_is_present()
80 ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *regs) in ia64_do_page_fault() argument
110 if ((REGION_NUMBER(address) == 5) && !user_mode(regs)) in ia64_do_page_fault()
127 vma = find_vma_prev(mm, address, &prev_vma); in ia64_do_page_fault()
138 if (( !vma && prev_vma ) || (address < vma->vm_start) ) in ia64_do_page_fault()
162 fault = handle_mm_fault(vma, address, flags); in ia64_do_page_fault()
206 if (!(prev_vma && (prev_vma->vm_flags & VM_GROWSUP) && (address == prev_vma->vm_end))) { in ia64_do_page_fault()
211 if (REGION_NUMBER(address) != REGION_NUMBER(vma->vm_start) in ia64_do_page_fault()
212 || REGION_OFFSET(address) >= RGN_MAP_LIMIT) in ia64_do_page_fault()
214 if (expand_stack(vma, address)) in ia64_do_page_fault()
218 if (REGION_NUMBER(address) != REGION_NUMBER(vma->vm_start) in ia64_do_page_fault()
219 || REGION_OFFSET(address) >= RGN_MAP_LIMIT) in ia64_do_page_fault()
225 if (address > vma->vm_end + PAGE_SIZE - sizeof(long)) in ia64_do_page_fault()
227 if (expand_upwards(vma, address)) in ia64_do_page_fault()
252 si.si_addr = (void __user *) address; in ia64_do_page_fault()
279 if (REGION_NUMBER(address) == 5 && mapped_kernel_page_is_present(address)) in ia64_do_page_fault()
291 if (address < PAGE_SIZE) in ia64_do_page_fault()
292 printk(KERN_ALERT "Unable to handle kernel NULL pointer dereference (address %016lx)\n", address); in ia64_do_page_fault()
295 "virtual address %016lx\n", address); in ia64_do_page_fault()