/arch/xtensa/mm/ |
D | cache.c | 87 unsigned long phys = page_to_phys(page); in flush_dcache_page() local 89 unsigned long alias = !(DCACHE_ALIAS_EQ(temp, phys)); in flush_dcache_page() 107 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_page() 110 __invalidate_icache_page_alias(virt, phys); in flush_dcache_page() 140 unsigned long phys = page_to_phys(pfn_to_page(pfn)); in flush_cache_page() local 143 __flush_invalidate_dcache_page_alias(virt, phys); in flush_cache_page() 144 __invalidate_icache_page_alias(virt, phys); in flush_cache_page() 171 unsigned long phys = page_to_phys(page); in update_mmu_cache() local 175 __flush_invalidate_dcache_page_alias(vaddr, phys); in update_mmu_cache() 176 __invalidate_icache_page_alias(vaddr, phys); in update_mmu_cache() [all …]
|
/arch/m68k/mm/ |
D | sun3kmap.c | 30 static inline void do_page_mapin(unsigned long phys, unsigned long virt, in do_page_mapin() argument 36 ptep = pfn_pte(phys >> PAGE_SHIFT, PAGE_KERNEL); in do_page_mapin() 48 static inline void do_pmeg_mapin(unsigned long phys, unsigned long virt, in do_pmeg_mapin() argument 56 do_page_mapin(phys, virt, type); in do_pmeg_mapin() 57 phys += PAGE_SIZE; in do_pmeg_mapin() 63 void __iomem *sun3_ioremap(unsigned long phys, unsigned long size, in sun3_ioremap() argument 74 offset = phys & (PAGE_SIZE-1); in sun3_ioremap() 75 phys &= ~(PAGE_SIZE-1); in sun3_ioremap() 98 do_pmeg_mapin(phys, virt, type, seg_pages); in sun3_ioremap() 101 phys += seg_pages * PAGE_SIZE; in sun3_ioremap() [all …]
|
/arch/um/kernel/ |
D | physmem.c | 58 void map_memory(unsigned long virt, unsigned long phys, unsigned long len, in map_memory() argument 64 fd = phys_mapping(phys, &offset); in map_memory() 110 int phys_mapping(unsigned long phys, unsigned long long *offset_out) in phys_mapping() argument 114 if (phys < physmem_size) { in phys_mapping() 116 *offset_out = phys; in phys_mapping() 118 else if (phys < __pa(end_iomem)) { in phys_mapping() 122 if ((phys >= region->phys) && in phys_mapping() 123 (phys < region->phys + region->size)) { in phys_mapping() 125 *offset_out = phys - region->phys; in phys_mapping() 131 else if (phys < __pa(end_iomem) + highmem) { in phys_mapping() [all …]
|
/arch/x86/include/asm/ |
D | fixmap.h | 18 unsigned long phys, pgprot_t flags); 22 unsigned long phys, pgprot_t flags) in __set_fixmap() argument 24 native_set_fixmap(idx, phys, flags); in __set_fixmap() 28 #define set_fixmap(idx, phys) \ argument 29 __set_fixmap(idx, phys, PAGE_KERNEL) 34 #define set_fixmap_nocache(idx, phys) \ argument 35 __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE)
|
/arch/sparc/include/asm/ |
D | fixmap.h | 58 unsigned long phys, pgprot_t flags); 60 #define set_fixmap(idx, phys) \ argument 61 __set_fixmap(idx, phys, PAGE_KERNEL) 65 #define set_fixmap_nocache(idx, phys) \ argument 66 __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE)
|
/arch/um/include/asm/ |
D | fixmap.h | 45 unsigned long phys, pgprot_t flags); 47 #define set_fixmap(idx, phys) \ argument 48 __set_fixmap(idx, phys, PAGE_KERNEL) 52 #define set_fixmap_nocache(idx, phys) \ argument 53 __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE)
|
D | page.h | 48 #define pte_set_val(pte, phys, prot) \ argument 49 ({ (pte).pte_high = (phys) >> 32; \ 50 (pte).pte_low = (phys) | pgprot_val(prot); }) 77 #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot)) argument 110 #define __va(phys) to_virt((unsigned long) (phys)) argument
|
/arch/powerpc/include/asm/ |
D | fixmap.h | 56 phys_addr_t phys, pgprot_t flags); 58 #define set_fixmap(idx, phys) \ argument 59 __set_fixmap(idx, phys, PAGE_KERNEL) 63 #define set_fixmap_nocache(idx, phys) \ argument 64 __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE)
|
D | smp.h | 65 static inline void set_hard_smp_processor_id(int cpu, int phys) in set_hard_smp_processor_id() argument 67 smp_hw_index[cpu] = phys; in set_hard_smp_processor_id() 114 static inline void set_hard_smp_processor_id(int cpu, int phys) in set_hard_smp_processor_id() argument 116 paca[cpu].hw_cpu_id = phys; in set_hard_smp_processor_id() 130 static inline void set_hard_smp_processor_id(int cpu, int phys) in set_hard_smp_processor_id() argument 132 boot_cpuid_phys = phys; in set_hard_smp_processor_id()
|
D | dma.h | 279 static __inline__ void set_dma_addr(unsigned int dmanr, unsigned int phys) in set_dma_addr() argument 282 dma_outb(phys & 0xff, in set_dma_addr() 284 dma_outb((phys >> 8) & 0xff, in set_dma_addr() 287 dma_outb((phys >> 1) & 0xff, in set_dma_addr() 289 dma_outb((phys >> 9) & 0xff, in set_dma_addr() 292 set_dma_page(dmanr, phys >> 16); in set_dma_addr()
|
/arch/sh/include/asm/ |
D | fixmap.h | 61 unsigned long phys, pgprot_t flags); 63 #define set_fixmap(idx, phys) \ argument 64 __set_fixmap(idx, phys, PAGE_KERNEL) 68 #define set_fixmap_nocache(idx, phys) \ argument 69 __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE)
|
/arch/um/include/shared/ |
D | mem.h | 9 extern int phys_mapping(unsigned long phys, unsigned long long *offset_out); 17 static inline void *to_virt(unsigned long phys) in to_virt() argument 19 return((void *) uml_physmem + phys); in to_virt()
|
D | mem_user.h | 40 unsigned long phys; member 58 extern unsigned long phys_offset(unsigned long phys); 59 extern void map_memory(unsigned long virt, unsigned long phys,
|
/arch/powerpc/mm/ |
D | ppc_mmu_32.c | 46 phys_addr_t phys; member 57 return bat_addrs[b].phys + (va - bat_addrs[b].start); in v_mapped_by_bats() 68 if (pa >= bat_addrs[b].phys in p_mapped_by_bats() 70 +bat_addrs[b].phys) in p_mapped_by_bats() 71 return bat_addrs[b].start+(pa-bat_addrs[b].phys); in p_mapped_by_bats() 119 void __init setbat(int index, unsigned long virt, phys_addr_t phys, in setbat() argument 138 bat[1].batl = BAT_PHYS_ADDR(phys) | wimgxpp; in setbat() 159 bat->batl = phys | bl | 0x40; /* V=1 */ in setbat() 164 bat_addrs[index].phys = phys; in setbat()
|
D | fsl_booke_mmu.c | 68 phys_addr_t phys; member 81 return tlbcam_addrs[b].phys + (va - tlbcam_addrs[b].start); in v_mapped_by_tlbcam() 92 if (pa >= tlbcam_addrs[b].phys in p_mapped_by_tlbcam() 94 +tlbcam_addrs[b].phys) in p_mapped_by_tlbcam() 95 return tlbcam_addrs[b].start+(pa-tlbcam_addrs[b].phys); in p_mapped_by_tlbcam() 104 void settlbcam(int index, unsigned long virt, phys_addr_t phys, in settlbcam() argument 127 TLBCAM[index].MAS3 = (phys & PAGE_MASK) | MAS3_SX | MAS3_SR; in settlbcam() 142 tlbcam_addrs[index].phys = phys; in settlbcam()
|
D | 44x_mmu.c | 64 static void __init ppc44x_pin_tlb(unsigned int virt, unsigned int phys) in ppc44x_pin_tlb() argument 76 "r" (phys), in ppc44x_pin_tlb()
|
/arch/mips/cavium-octeon/ |
D | flash_setup.c | 55 flash_map.phys = region_cfg.s.base << 16; in flash_init() 56 flash_map.size = 0x1fc00000 - flash_map.phys; in flash_init() 58 flash_map.virt = ioremap(flash_map.phys, flash_map.size); in flash_init() 60 "0x%08lx\n", flash_map.size >> 20, flash_map.phys); in flash_init()
|
/arch/sh/mm/ |
D | cache-sh4.c | 35 static void __flush_cache_4096(unsigned long addr, unsigned long phys, 225 unsigned long phys) in flush_cache_4096() argument 239 P1SEGADDR(phys), exec_offset); in flush_cache_4096() 250 unsigned long phys = PHYSADDR(page_address(page)); in flush_dcache_page() local 257 flush_cache_4096(addr, phys); in flush_dcache_page() 331 unsigned long phys; in __flush_cache_mm() local 340 phys = pte_val(entry) & PTE_PHYS_MASK; in __flush_cache_mm() 342 if ((p ^ phys) & alias_mask) { in __flush_cache_mm() 344 d |= 1 << ((phys & alias_mask) >> PAGE_SHIFT); in __flush_cache_mm() 423 unsigned long phys = pfn << PAGE_SHIFT; in flush_cache_page() local [all …]
|
D | cache-sh7705.c | 74 static void __uses_jump_to_uncached __flush_dcache_page(unsigned long phys) in __flush_dcache_page() argument 79 phys |= SH_CACHE_VALID; in __flush_dcache_page() 112 if (data == phys) { in __flush_dcache_page()
|
/arch/mips/pmc-sierra/yosemite/ |
D | py-console.c | 56 static unsigned char readb_outer_space(unsigned long long phys) in readb_outer_space() argument 58 unsigned long long vaddr = IO_BASE_64 | phys; in readb_outer_space() 80 static void writeb_outer_space(unsigned long long phys, unsigned char c) in writeb_outer_space() argument 82 unsigned long long vaddr = IO_BASE_64 | phys; in writeb_outer_space()
|
/arch/x86/mm/ |
D | init_64.c | 229 static void __init __init_extra_mapping(unsigned long phys, unsigned long size, in __init_extra_mapping() argument 236 BUG_ON((phys & ~PMD_MASK) || (size & ~PMD_MASK)); in __init_extra_mapping() 237 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping() 238 pgd = pgd_offset_k((unsigned long)__va(phys)); in __init_extra_mapping() 244 pud = pud_offset(pgd, (unsigned long)__va(phys)); in __init_extra_mapping() 250 pmd = pmd_offset(pud, phys); in __init_extra_mapping() 252 set_pmd(pmd, __pmd(phys | pgprot_val(prot))); in __init_extra_mapping() 256 void __init init_extra_mapping_wb(unsigned long phys, unsigned long size) in init_extra_mapping_wb() argument 258 __init_extra_mapping(phys, size, PAGE_KERNEL_LARGE); in init_extra_mapping_wb() 261 void __init init_extra_mapping_uc(unsigned long phys, unsigned long size) in init_extra_mapping_uc() argument [all …]
|
D | ioremap.c | 457 void *xlate_dev_mem_ptr(unsigned long phys) in xlate_dev_mem_ptr() argument 460 unsigned long start = phys & PAGE_MASK; in xlate_dev_mem_ptr() 464 return __va(phys); in xlate_dev_mem_ptr() 468 addr = (void *)((unsigned long)addr | (phys & ~PAGE_MASK)); in xlate_dev_mem_ptr() 473 void unxlate_dev_mem_ptr(unsigned long phys, void *addr) in unxlate_dev_mem_ptr() argument 475 if (page_is_ram(phys >> PAGE_SHIFT)) in unxlate_dev_mem_ptr() 547 unsigned long phys, pgprot_t flags) in __early_set_fixmap() argument 559 set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags)); in __early_set_fixmap() 566 unsigned long phys, pgprot_t prot) in early_set_fixmap() argument 569 __set_fixmap(idx, phys, prot); in early_set_fixmap() [all …]
|
/arch/avr32/mm/ |
D | dma-coherent.c | 126 dma_addr_t phys; in dma_alloc_writecombine() local 132 phys = page_to_phys(page); in dma_alloc_writecombine() 133 *handle = phys; in dma_alloc_writecombine() 136 return __ioremap(phys, size, _PAGE_BUFFER); in dma_alloc_writecombine()
|
/arch/sh/kernel/cpu/sh4/ |
D | sq.c | 144 unsigned long sq_remap(unsigned long phys, unsigned int size, in sq_remap() argument 153 end = phys + size - 1; in sq_remap() 154 if (unlikely(!size || end < phys)) in sq_remap() 157 if (unlikely(phys < virt_to_phys(high_memory))) in sq_remap() 160 phys &= PAGE_MASK; in sq_remap() 161 size = PAGE_ALIGN(end + 1) - phys; in sq_remap() 167 map->addr = phys; in sq_remap()
|
/arch/alpha/include/asm/ |
D | io.h | 74 unsigned long phys = (unsigned long)address; in virt_to_phys() local 77 phys <<= (64 - 41); in virt_to_phys() 78 phys = (long)phys >> (64 - 41); in virt_to_phys() 81 phys &= (1ul << hwrpb->pa_bits) - 1; in virt_to_phys() 83 return phys; in virt_to_phys() 115 unsigned long phys = virt_to_phys(address); in virt_to_bus() local 116 unsigned long bus = phys + __direct_map_base; in virt_to_bus() 117 return phys <= __direct_map_size ? bus : 0; in virt_to_bus()
|