/arch/cris/mm/ |
D | ioremap.c | 34 unsigned long offset, last_addr; in __ioremap_prot() local 37 last_addr = phys_addr + size - 1; in __ioremap_prot() 38 if (!size || last_addr < phys_addr) in __ioremap_prot() 46 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_prot()
|
/arch/metag/mm/ |
D | ioremap.c | 33 unsigned long offset, last_addr; in __ioremap() local 37 last_addr = phys_addr + size - 1; in __ioremap() 38 if (!size || last_addr < phys_addr) in __ioremap() 51 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap()
|
/arch/avr32/mm/ |
D | ioremap.c | 27 unsigned long offset, last_addr; in __ioremap() local 39 last_addr = phys_addr + size - 1; in __ioremap() 40 if (!size || last_addr < phys_addr) in __ioremap() 54 size = PAGE_ALIGN(last_addr + 1) - phys_addr; in __ioremap()
|
/arch/sh/mm/ |
D | ioremap.c | 42 unsigned long offset, last_addr, addr, orig_addr; in __ioremap_caller() local 46 last_addr = phys_addr + size - 1; in __ioremap_caller() 47 if (!size || last_addr < phys_addr) in __ioremap_caller() 69 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_caller()
|
D | pmb.c | 414 phys_addr_t offset, last_addr; in pmb_remap_caller() local 435 last_addr = phys + size; in pmb_remap_caller() 439 aligned = ALIGN(last_addr, pmb_sizes[i].size) - phys; in pmb_remap_caller()
|
/arch/m32r/mm/ |
D | ioremap.c | 47 unsigned long offset, last_addr; in __ioremap() local 51 last_addr = phys_addr + size - 1; in __ioremap() 52 if (!size || last_addr < phys_addr) in __ioremap() 85 size = PAGE_ALIGN(last_addr + 1) - phys_addr; in __ioremap()
|
/arch/parisc/mm/ |
D | ioremap.c | 31 unsigned long offset, last_addr; in __ioremap() local 45 last_addr = phys_addr + size - 1; in __ioremap() 46 if (!size || last_addr < phys_addr) in __ioremap() 74 size = PAGE_ALIGN(last_addr + 1) - phys_addr; in __ioremap()
|
/arch/hexagon/mm/ |
D | ioremap.c | 26 unsigned long last_addr, addr; in ioremap_nocache() local 33 last_addr = phys_addr + size - 1; in ioremap_nocache() 36 if (!size || (last_addr < phys_addr)) in ioremap_nocache()
|
/arch/openrisc/mm/ |
D | ioremap.c | 46 unsigned long offset, last_addr; in __ioremap() local 50 last_addr = addr + size - 1; in __ioremap() 51 if (!size || last_addr < addr) in __ioremap() 59 size = PAGE_ALIGN(last_addr + 1) - p; in __ioremap()
|
/arch/mips/mm/ |
D | ioremap.c | 120 phys_t last_addr; in __ioremap() local 126 last_addr = phys_addr + size - 1; in __ioremap() 127 if (!size || last_addr < phys_addr) in __ioremap() 134 if (IS_LOW512(phys_addr) && IS_LOW512(last_addr) && in __ioremap() 158 size = PAGE_ALIGN(last_addr + 1) - phys_addr; in __ioremap()
|
/arch/arm64/mm/ |
D | ioremap.c | 35 unsigned long last_addr; in __ioremap_caller() local 51 last_addr = phys_addr + size - 1; in __ioremap_caller() 52 if (!size || last_addr < phys_addr || (last_addr & ~PHYS_MASK)) in __ioremap_caller()
|
/arch/x86/mm/ |
D | ioremap.c | 66 resource_size_t pfn, last_pfn, last_addr; in __ioremap_caller() local 76 last_addr = phys_addr + size - 1; in __ioremap_caller() 77 if (!size || last_addr < phys_addr) in __ioremap_caller() 90 if (is_ISA_range(phys_addr, last_addr)) in __ioremap_caller() 96 last_pfn = last_addr >> PAGE_SHIFT; in __ioremap_caller() 110 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_caller() 488 resource_size_t last_addr; in __early_ioremap() local 517 last_addr = phys_addr + size - 1; in __early_ioremap() 518 if (!size || last_addr < phys_addr) { in __early_ioremap() 529 size = PAGE_ALIGN(last_addr + 1) - phys_addr; in __early_ioremap()
|
/arch/powerpc/boot/ |
D | devtree.c | 235 u32 last_addr[MAX_ADDR_CELLS]; in dt_xlate() local 257 copy_val(last_addr, prop_buf + offset, naddr); in dt_xlate() 283 offset = find_range(last_addr, prop_buf, prev_naddr, in dt_xlate() 291 if (!sub_reg(last_addr, this_addr)) in dt_xlate() 296 if (!add_reg(last_addr, this_addr, naddr)) in dt_xlate() 303 ret_addr = ((u64)last_addr[2] << 32) | last_addr[3]; in dt_xlate()
|
/arch/unicore32/mm/ |
D | ioremap.c | 187 unsigned long last_addr; in __uc32_ioremap_caller() local 194 last_addr = phys_addr + size - 1; in __uc32_ioremap_caller() 195 if (!size || last_addr < phys_addr) in __uc32_ioremap_caller()
|
/arch/x86/mm/kmemcheck/ |
D | shadow.c | 40 unsigned long last_addr = addr + n - 1; in mark_shadow() local 42 unsigned long last_page = last_addr & PAGE_MASK; in mark_shadow()
|
/arch/mips/include/asm/ |
D | io.h | 203 phys_t phys_addr, last_addr; in __ioremap_mode() local 208 last_addr = phys_addr + size - 1; in __ioremap_mode() 209 if (!size || last_addr < phys_addr) in __ioremap_mode() 216 if (__IS_LOW512(phys_addr) && __IS_LOW512(last_addr) && in __ioremap_mode()
|
/arch/arm/mm/ |
D | ioremap.c | 337 unsigned long last_addr; in __arm_ioremap_caller() local 344 last_addr = phys_addr + size - 1; in __arm_ioremap_caller() 345 if (!size || last_addr < phys_addr) in __arm_ioremap_caller()
|
/arch/tile/mm/ |
D | pgtable.c | 541 unsigned long offset, last_addr; in ioremap_prot() local 545 last_addr = phys_addr + size - 1; in ioremap_prot() 546 if (!size || last_addr < phys_addr) in ioremap_prot() 559 size = PAGE_ALIGN(last_addr+1) - phys_addr; in ioremap_prot()
|
/arch/sh/include/asm/ |
D | io.h | 283 phys_addr_t last_addr = offset + size - 1; in __ioremap_29bit() local 291 if (likely(PXSEG(offset) < P3SEG && PXSEG(last_addr) < P3SEG)) { in __ioremap_29bit()
|
/arch/x86/kernel/ |
D | e820.c | 263 unsigned long long last_addr; in sanitize_e820_map() local 307 last_addr = 0; /* start with 0 as last starting address */ in sanitize_e820_map() 349 change_point[chgidx]->addr - last_addr; in sanitize_e820_map() 366 last_addr = change_point[chgidx]->addr; in sanitize_e820_map()
|
/arch/blackfin/kernel/ |
D | setup.c | 301 unsigned long long last_addr; in sanitize_memmap() local 395 last_addr = 0; /* start with 0 as last starting address */ in sanitize_memmap() 420 change_point[chgidx]->addr - last_addr; in sanitize_memmap() 429 last_addr = change_point[chgidx]->addr; in sanitize_memmap()
|