/arch/ia64/mm/ |
D | ioremap.c | 19 __ioremap_uc(unsigned long phys_addr) in __ioremap_uc() argument 21 return (void __iomem *) (__IA64_UNCACHED_OFFSET | phys_addr); in __ioremap_uc() 25 early_ioremap (unsigned long phys_addr, unsigned long size) in early_ioremap() argument 28 attr = kern_mem_attribute(phys_addr, size); in early_ioremap() 30 return (void __iomem *) phys_to_virt(phys_addr); in early_ioremap() 31 return __ioremap_uc(phys_addr); in early_ioremap() 35 ioremap (unsigned long phys_addr, unsigned long size) in ioremap() argument 50 attr = kern_mem_attribute(phys_addr, size); in ioremap() 52 return (void __iomem *) phys_to_virt(phys_addr); in ioremap() 54 return __ioremap_uc(phys_addr); in ioremap() [all …]
|
/arch/avr32/mm/ |
D | ioremap.c | 22 void __iomem *__ioremap(unsigned long phys_addr, size_t size, in __ioremap() argument 35 if ((phys_addr >= P4SEG) && (flags == 0)) in __ioremap() 36 return (void __iomem *)phys_addr; in __ioremap() 39 last_addr = phys_addr + size - 1; in __ioremap() 40 if (!size || last_addr < phys_addr) in __ioremap() 48 if (PHYSADDR(P2SEGADDR(phys_addr)) == phys_addr) in __ioremap() 49 return (void __iomem *)P2SEGADDR(phys_addr); in __ioremap() 52 offset = phys_addr & ~PAGE_MASK; in __ioremap() 53 phys_addr &= PAGE_MASK; in __ioremap() 54 size = PAGE_ALIGN(last_addr + 1) - phys_addr; in __ioremap() [all …]
|
/arch/arm64/mm/ |
D | ioremap.c | 32 static void __iomem *__ioremap_caller(phys_addr_t phys_addr, size_t size, in __ioremap_caller() argument 36 unsigned long offset = phys_addr & ~PAGE_MASK; in __ioremap_caller() 45 phys_addr &= PAGE_MASK; in __ioremap_caller() 51 last_addr = phys_addr + size - 1; in __ioremap_caller() 52 if (!size || last_addr < phys_addr || (last_addr & ~PHYS_MASK)) in __ioremap_caller() 58 if (WARN_ON(pfn_valid(__phys_to_pfn(phys_addr)))) in __ioremap_caller() 65 area->phys_addr = phys_addr; in __ioremap_caller() 67 err = ioremap_page_range(addr, addr + size, phys_addr, prot); in __ioremap_caller() 76 void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot) in __ioremap() argument 78 return __ioremap_caller(phys_addr, size, prot, in __ioremap() [all …]
|
/arch/m32r/mm/ |
D | ioremap.c | 43 __ioremap(unsigned long phys_addr, unsigned long size, unsigned long flags) in __ioremap() argument 51 last_addr = phys_addr + size - 1; in __ioremap() 52 if (!size || last_addr < phys_addr) in __ioremap() 59 if (IS_LOW512(phys_addr) && IS_LOW512(phys_addr + size - 1)) in __ioremap() 60 return (void *) KSEG1ADDR(phys_addr); in __ioremap() 65 if (phys_addr < virt_to_phys(high_memory)) { in __ioremap() 69 t_addr = __va(phys_addr); in __ioremap() 83 offset = phys_addr & ~PAGE_MASK; in __ioremap() 84 phys_addr &= PAGE_MASK; in __ioremap() 85 size = PAGE_ALIGN(last_addr + 1) - phys_addr; in __ioremap() [all …]
|
/arch/parisc/mm/ |
D | ioremap.c | 27 void __iomem * __ioremap(unsigned long phys_addr, unsigned long size, unsigned long flags) in __ioremap() argument 35 unsigned long end = phys_addr + size - 1; in __ioremap() 37 if ((phys_addr >= 0x00080000 && end < 0x000fffff) || in __ioremap() 38 (phys_addr >= 0x00500000 && end < 0x03bfffff)) { in __ioremap() 39 phys_addr |= F_EXTEND(0xfc000000); in __ioremap() 45 last_addr = phys_addr + size - 1; in __ioremap() 46 if (!size || last_addr < phys_addr) in __ioremap() 52 if (phys_addr < virt_to_phys(high_memory)) { in __ioremap() 56 t_addr = __va(phys_addr); in __ioremap() 72 offset = phys_addr & ~PAGE_MASK; in __ioremap() [all …]
|
/arch/x86/mm/ |
D | ioremap.c | 82 static void __iomem *__ioremap_caller(resource_size_t phys_addr, in __ioremap_caller() argument 87 const resource_size_t unaligned_phys_addr = phys_addr; in __ioremap_caller() 96 last_addr = phys_addr + size - 1; in __ioremap_caller() 97 if (!size || last_addr < phys_addr) in __ioremap_caller() 100 if (!phys_addr_valid(phys_addr)) { in __ioremap_caller() 102 (unsigned long long)phys_addr); in __ioremap_caller() 110 if (is_ISA_range(phys_addr, last_addr)) in __ioremap_caller() 111 return (__force void __iomem *)phys_to_virt(phys_addr); in __ioremap_caller() 116 pfn = phys_addr >> PAGE_SHIFT; in __ioremap_caller() 121 &phys_addr, &last_addr); in __ioremap_caller() [all …]
|
D | physaddr.c | 72 unsigned long phys_addr = x - PAGE_OFFSET; in __phys_addr() local 78 VIRTUAL_BUG_ON((phys_addr >> PAGE_SHIFT) > max_low_pfn); in __phys_addr() 79 BUG_ON(slow_virt_to_phys((void *)x) != phys_addr); in __phys_addr() 81 return phys_addr; in __phys_addr()
|
/arch/metag/mm/ |
D | ioremap.c | 28 void __iomem *__ioremap(unsigned long phys_addr, size_t size, in __ioremap() argument 37 last_addr = phys_addr + size - 1; in __ioremap() 38 if (!size || last_addr < phys_addr) in __ioremap() 42 if (phys_addr >= LINSYSCUSTOM_BASE && in __ioremap() 43 phys_addr < (LINSYSCUSTOM_BASE + LINSYSCUSTOM_LIMIT)) in __ioremap() 44 return (__force void __iomem *) phys_addr; in __ioremap() 49 offset = phys_addr & ~PAGE_MASK; in __ioremap() 50 phys_addr &= PAGE_MASK; in __ioremap() 51 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap() 62 area->phys_addr = phys_addr; in __ioremap() [all …]
|
/arch/cris/mm/ |
D | ioremap.c | 30 void __iomem * __ioremap_prot(unsigned long phys_addr, unsigned long size, pgprot_t prot) in __ioremap_prot() argument 37 last_addr = phys_addr + size - 1; in __ioremap_prot() 38 if (!size || last_addr < phys_addr) in __ioremap_prot() 44 offset = phys_addr & ~PAGE_MASK; in __ioremap_prot() 45 phys_addr &= PAGE_MASK; in __ioremap_prot() 46 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_prot() 56 phys_addr, prot)) { in __ioremap_prot() 63 void __iomem * __ioremap(unsigned long phys_addr, unsigned long size, unsigned long flags) in __ioremap() argument 65 return __ioremap_prot(phys_addr, size, in __ioremap() 79 void __iomem *ioremap_nocache(unsigned long phys_addr, unsigned long size) in ioremap_nocache() argument [all …]
|
/arch/mips/mm/ |
D | ioremap.c | 20 phys_addr_t size, phys_addr_t phys_addr, unsigned long flags) in remap_area_pte() argument 32 pfn = phys_addr >> PAGE_SHIFT; in remap_area_pte() 46 phys_addr_t size, phys_addr_t phys_addr, unsigned long flags) in remap_area_pmd() argument 54 phys_addr -= address; in remap_area_pmd() 60 remap_area_pte(pte, address, end - address, address + phys_addr, flags); in remap_area_pmd() 67 static int remap_area_pages(unsigned long address, phys_addr_t phys_addr, in remap_area_pages() argument 74 phys_addr -= address; in remap_area_pages() 90 phys_addr + address, flags)) in remap_area_pages() 116 void __iomem * __ioremap(phys_addr_t phys_addr, phys_addr_t size, unsigned long flags) in __ioremap() argument 123 phys_addr = fixup_bigphys_addr(phys_addr, size); in __ioremap() [all …]
|
/arch/nios2/mm/ |
D | ioremap.c | 23 unsigned long size, unsigned long phys_addr, in remap_area_pte() argument 37 pfn = PFN_DOWN(phys_addr); in remap_area_pte() 51 unsigned long size, unsigned long phys_addr, in remap_area_pmd() argument 60 phys_addr -= address; in remap_area_pmd() 68 remap_area_pte(pte, address, end - address, address + phys_addr, in remap_area_pmd() 76 static int remap_area_pages(unsigned long address, unsigned long phys_addr, in remap_area_pages() argument 83 phys_addr -= address; in remap_area_pages() 100 phys_addr + address, flags)) in remap_area_pages() 115 void __iomem *__ioremap(unsigned long phys_addr, unsigned long size, in __ioremap() argument 124 last_addr = phys_addr + size - 1; in __ioremap() [all …]
|
/arch/sh/mm/ |
D | ioremap.c | 38 __ioremap_caller(phys_addr_t phys_addr, unsigned long size, in __ioremap_caller() argument 46 last_addr = phys_addr + size - 1; in __ioremap_caller() 47 if (!size || last_addr < phys_addr) in __ioremap_caller() 54 return ioremap_fixed(phys_addr, size, pgprot); in __ioremap_caller() 60 mapped = pmb_remap_caller(phys_addr, size, pgprot, caller); in __ioremap_caller() 67 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller() 68 phys_addr &= PAGE_MASK; in __ioremap_caller() 69 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_caller() 77 area->phys_addr = phys_addr; in __ioremap_caller() 80 if (ioremap_page_range(addr, addr + size, phys_addr, pgprot)) { in __ioremap_caller()
|
D | ioremap_fixed.c | 47 ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) in ioremap_fixed() argument 58 offset = phys_addr & ~PAGE_MASK; in ioremap_fixed() 59 phys_addr &= PAGE_MASK; in ioremap_fixed() 60 size = PAGE_ALIGN(phys_addr + size) - phys_addr; in ioremap_fixed() 89 __set_fixmap(idx, phys_addr, prot); in ioremap_fixed() 90 phys_addr += PAGE_SIZE; in ioremap_fixed()
|
/arch/mips/alchemy/common/ |
D | setup.c | 77 phys_addr_t __fixup_bigphys_addr(phys_addr_t phys_addr, phys_addr_t size) in __fixup_bigphys_addr() argument 83 if ((phys_addr >> 32) != 0) in __fixup_bigphys_addr() 84 return phys_addr; in __fixup_bigphys_addr() 87 if (phys_addr >= start && (phys_addr + size - 1) <= end) in __fixup_bigphys_addr() 88 return (phys_addr_t)(AU1500_PCI_MEM_PHYS_ADDR + phys_addr); in __fixup_bigphys_addr() 91 return phys_addr; in __fixup_bigphys_addr()
|
/arch/ia64/kernel/ |
D | efi.c | 289 return (md->phys_addr + efi_md_size(md)); in efi_md_end() 367 "dropped @ %llx\n", md->phys_addr); in efi_get_pal_addr() 375 vaddr = PAGE_OFFSET + md->phys_addr; in efi_get_pal_addr() 404 smp_processor_id(), md->phys_addr, in efi_get_pal_addr() 405 md->phys_addr + efi_md_size(md), in efi_get_pal_addr() 408 return __va(md->phys_addr); in efi_get_pal_addr() 431 static void __init handle_palo(unsigned long phys_addr) in handle_palo() argument 433 struct palo_table *palo = __va(phys_addr); in handle_palo() 590 md->phys_addr, in efi_init() 591 md->phys_addr + efi_md_size(md), size, unit); in efi_init() [all …]
|
/arch/hexagon/mm/ |
D | ioremap.c | 25 void __iomem *ioremap_nocache(unsigned long phys_addr, unsigned long size) in ioremap_nocache() argument 28 unsigned long offset = phys_addr & ~PAGE_MASK; in ioremap_nocache() 34 last_addr = phys_addr + size - 1; in ioremap_nocache() 37 if (!size || (last_addr < phys_addr)) in ioremap_nocache() 46 if (ioremap_page_range(addr, addr+size, phys_addr, prot)) { in ioremap_nocache()
|
/arch/mips/include/asm/mach-au1x00/ |
D | ioremap.h | 17 static inline phys_addr_t __fixup_bigphys_addr(phys_addr_t phys_addr, phys_addr_t size) in __fixup_bigphys_addr() argument 19 return phys_addr; in __fixup_bigphys_addr() 26 static inline phys_addr_t fixup_bigphys_addr(phys_addr_t phys_addr, phys_addr_t size) in fixup_bigphys_addr() argument 28 return __fixup_bigphys_addr(phys_addr, size); in fixup_bigphys_addr()
|
/arch/unicore32/mm/ |
D | ioremap.c | 184 void __iomem *__uc32_ioremap_caller(unsigned long phys_addr, size_t size, in __uc32_ioremap_caller() argument 188 unsigned long offset = phys_addr & ~PAGE_MASK; in __uc32_ioremap_caller() 189 unsigned long pfn = __phys_to_pfn(phys_addr); in __uc32_ioremap_caller() 194 last_addr = phys_addr + size - 1; in __uc32_ioremap_caller() 195 if (!size || last_addr < phys_addr) in __uc32_ioremap_caller() 220 __uc32_ioremap(unsigned long phys_addr, size_t size) in __uc32_ioremap() argument 222 return __uc32_ioremap_caller(phys_addr, size, MT_DEVICE, in __uc32_ioremap() 228 __uc32_ioremap_cached(unsigned long phys_addr, size_t size) in __uc32_ioremap_cached() argument 230 return __uc32_ioremap_caller(phys_addr, size, MT_DEVICE_CACHED, in __uc32_ioremap_cached()
|
/arch/ia64/include/asm/ |
D | early_ioremap.h | 4 extern void __iomem * early_ioremap (unsigned long phys_addr, unsigned long size); 5 #define early_memremap(phys_addr, size) early_ioremap(phys_addr, size) argument
|
/arch/arm/mm/ |
D | ioremap.c | 59 if (vm->phys_addr > paddr || in find_static_vm_paddr() 60 paddr + size - 1 > vm->phys_addr + vm->size - 1) in find_static_vm_paddr() 294 addr += paddr - svm->vm.phys_addr; in __arm_ioremap_pfn_caller() 310 area->phys_addr = paddr; in __arm_ioremap_pfn_caller() 336 void __iomem *__arm_ioremap_caller(phys_addr_t phys_addr, size_t size, in __arm_ioremap_caller() argument 340 unsigned long offset = phys_addr & ~PAGE_MASK; in __arm_ioremap_caller() 341 unsigned long pfn = __phys_to_pfn(phys_addr); in __arm_ioremap_caller() 346 last_addr = phys_addr + size - 1; in __arm_ioremap_caller() 347 if (!size || last_addr < phys_addr) in __arm_ioremap_caller() 409 __arm_ioremap_exec(phys_addr_t phys_addr, size_t size, bool cached) in __arm_ioremap_exec() argument [all …]
|
/arch/powerpc/boot/ |
D | cuboot-pq2.c | 36 u32 phys_addr; member 181 if (mem->phys_addr + mem->size[1] == mmio->phys_addr) in fixup_pci() 183 else if (mmio->phys_addr + mmio->size[1] == mem->phys_addr) in fixup_pci() 188 out_be32(&pci_regs[1][0], mem_base->phys_addr | 1); in fixup_pci() 191 out_be32(&pci_regs[1][1], io->phys_addr | 1); in fixup_pci() 195 out_le32(&pci_regs[0][2], mem->phys_addr >> 12); in fixup_pci() 199 out_le32(&pci_regs[0][8], mmio->phys_addr >> 12); in fixup_pci() 203 out_le32(&pci_regs[0][14], io->phys_addr >> 12); in fixup_pci()
|
/arch/sparc/kernel/ |
D | auxio_32.c | 67 r.start = auxregs[0].phys_addr; in auxio_probe() 68 r.end = auxregs[0].phys_addr + auxregs[0].reg_size - 1; in auxio_probe() 71 if ((((unsigned long) auxregs[0].phys_addr) & 3) == 3) in auxio_probe() 131 r.start = regs.phys_addr; in auxio_power_probe() 132 r.end = regs.phys_addr + regs.reg_size - 1; in auxio_power_probe()
|
/arch/x86/platform/efi/ |
D | efi_64.c | 324 pfn = md->phys_addr >> PAGE_SHIFT; in __map_region() 327 md->phys_addr, va); in __map_region() 333 u64 pa = md->phys_addr; in efi_map_region() 343 __map_region(md, md->phys_addr); in efi_map_region() 351 md->virt_addr = md->phys_addr; in efi_map_region() 388 __map_region(md, md->phys_addr); in efi_map_region_fixed() 392 void __iomem *__init efi_ioremap(unsigned long phys_addr, unsigned long size, in efi_ioremap() argument 398 return ioremap(phys_addr, size); in efi_ioremap() 400 last_map_pfn = init_memory_mapping(phys_addr, phys_addr + size); in efi_ioremap() 401 if ((last_map_pfn << PAGE_SHIFT) < phys_addr + size) { in efi_ioremap() [all …]
|
D | efi.c | 106 unsigned long long start = md->phys_addr; in efi_find_mirror() 131 unsigned long long start = md->phys_addr; in do_add_efi_memmap() 219 u64 end = (md->num_pages << EFI_PAGE_SHIFT) + md->phys_addr - 1; in efi_memmap_entry_valid() 227 (md->phys_addr >> EFI_PAGE_SHIFT)) { in efi_memmap_entry_valid() 231 if ((md->phys_addr & U64_HIGH_BIT) && !(end & U64_HIGH_BIT)) in efi_memmap_entry_valid() 242 md->phys_addr, end_hi, end); in efi_memmap_entry_valid() 246 md->phys_addr, end); in efi_memmap_entry_valid() 287 md->phys_addr, in efi_print_memmap() 288 md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1, in efi_print_memmap() 594 start_pfn = PFN_DOWN(md->phys_addr); in old_map_region() [all …]
|
/arch/ia64/sn/pci/ |
D | pci_dma.c | 83 unsigned long phys_addr; in sn_dma_alloc_coherent() local 111 phys_addr = __pa(cpuaddr); in sn_dma_alloc_coherent() 119 *dma_handle = provider->dma_map_consistent(pdev, phys_addr, size, in sn_dma_alloc_coherent() 183 unsigned long phys_addr; in sn_dma_map_page() local 189 phys_addr = __pa(cpu_addr); in sn_dma_map_page() 191 dma_addr = provider->dma_map_consistent(pdev, phys_addr, in sn_dma_map_page() 194 dma_addr = provider->dma_map(pdev, phys_addr, size, in sn_dma_map_page() 275 unsigned long phys_addr; in sn_dma_map_sg() local 288 phys_addr = SG_ENT_PHYS_ADDRESS(sg); in sn_dma_map_sg() 291 phys_addr, in sn_dma_map_sg() [all …]
|