/arch/xtensa/mm/ |
D | cache.c | 152 unsigned long phys = page_to_phys(page); in flush_dcache_page() local 154 unsigned long alias = !(DCACHE_ALIAS_EQ(temp, phys)); in flush_dcache_page() 168 virt = TLBTEMP_BASE_1 + (phys & DCACHE_ALIAS_MASK); in flush_dcache_page() 169 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_page() 174 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_page() 177 __invalidate_icache_page_alias(virt, phys); in flush_dcache_page() 208 unsigned long phys = page_to_phys(pfn_to_page(pfn)); in local_flush_cache_page() local 212 __flush_invalidate_dcache_page_alias(virt, phys); in local_flush_cache_page() 213 __invalidate_icache_page_alias(virt, phys); in local_flush_cache_page() 237 unsigned long phys = page_to_phys(page); in update_mmu_cache() local [all …]
|
/arch/m68k/mm/ |
D | sun3kmap.c | 30 static inline void do_page_mapin(unsigned long phys, unsigned long virt, in do_page_mapin() argument 36 ptep = pfn_pte(phys >> PAGE_SHIFT, PAGE_KERNEL); in do_page_mapin() 48 static inline void do_pmeg_mapin(unsigned long phys, unsigned long virt, in do_pmeg_mapin() argument 56 do_page_mapin(phys, virt, type); in do_pmeg_mapin() 57 phys += PAGE_SIZE; in do_pmeg_mapin() 63 void __iomem *sun3_ioremap(unsigned long phys, unsigned long size, in sun3_ioremap() argument 74 offset = phys & (PAGE_SIZE-1); in sun3_ioremap() 75 phys &= ~(PAGE_SIZE-1); in sun3_ioremap() 98 do_pmeg_mapin(phys, virt, type, seg_pages); in sun3_ioremap() 101 phys += seg_pages * PAGE_SIZE; in sun3_ioremap() [all …]
|
/arch/arm/mach-realview/include/mach/ |
D | memory.h | 52 #define __phys_to_virt(phys) \ argument 53 ((phys) >= 0x80000000 ? (phys) - 0x80000000 + PAGE_OFFSET2 : \ 54 (phys) >= 0x20000000 ? (phys) - 0x20000000 + PAGE_OFFSET1 : \ 55 (phys) + PAGE_OFFSET)
|
/arch/um/kernel/ |
D | physmem.c | 41 void map_memory(unsigned long virt, unsigned long phys, unsigned long len, in map_memory() argument 47 fd = phys_mapping(phys, &offset); in map_memory() 121 int phys_mapping(unsigned long phys, unsigned long long *offset_out) in phys_mapping() argument 125 if (phys < physmem_size) { in phys_mapping() 127 *offset_out = phys; in phys_mapping() 129 else if (phys < __pa(end_iomem)) { in phys_mapping() 133 if ((phys >= region->phys) && in phys_mapping() 134 (phys < region->phys + region->size)) { in phys_mapping() 136 *offset_out = phys - region->phys; in phys_mapping() 142 else if (phys < __pa(end_iomem) + highmem) { in phys_mapping() [all …]
|
/arch/s390/numa/ |
D | mode_emu.c | 208 static void toptree_to_numa_single(struct toptree *numa, struct toptree *phys, in toptree_to_numa_single() argument 213 toptree_for_each_safe(core, tmp, phys, CORE) { in toptree_to_numa_single() 225 static void move_level_to_numa_node(struct toptree *node, struct toptree *phys, in move_level_to_numa_node() argument 231 toptree_for_each_safe(cur, tmp, phys, level) { in move_level_to_numa_node() 248 static void move_level_to_numa(struct toptree *numa, struct toptree *phys, in move_level_to_numa() argument 254 move_level_to_numa_node(node, phys, level, perfect); in move_level_to_numa() 260 static void toptree_to_numa_first(struct toptree *numa, struct toptree *phys) in toptree_to_numa_first() argument 265 move_level_to_numa(numa, phys, BOOK, true); in toptree_to_numa_first() 266 move_level_to_numa(numa, phys, BOOK, false); in toptree_to_numa_first() 267 move_level_to_numa(numa, phys, MC, true); in toptree_to_numa_first() [all …]
|
/arch/powerpc/mm/ |
D | fsl_booke_mmu.c | 67 phys_addr_t phys; member 83 return tlbcam_addrs[b].phys + (va - tlbcam_addrs[b].start); in v_mapped_by_tlbcam() 94 if (pa >= tlbcam_addrs[b].phys in p_mapped_by_tlbcam() 96 +tlbcam_addrs[b].phys) in p_mapped_by_tlbcam() 97 return tlbcam_addrs[b].start+(pa-tlbcam_addrs[b].phys); in p_mapped_by_tlbcam() 108 static void settlbcam(int index, unsigned long virt, phys_addr_t phys, in settlbcam() argument 130 TLBCAM[index].MAS3 = (phys & MAS3_RPN) | MAS3_SX | MAS3_SR; in settlbcam() 133 TLBCAM[index].MAS7 = (u64)phys >> 32; in settlbcam() 143 tlbcam_addrs[index].phys = phys; in settlbcam() 147 phys_addr_t phys) in calc_cam_sz() argument [all …]
|
D | ppc_mmu_32.c | 46 phys_addr_t phys; member 57 return bat_addrs[b].phys + (va - bat_addrs[b].start); in v_mapped_by_bats() 68 if (pa >= bat_addrs[b].phys in p_mapped_by_bats() 70 +bat_addrs[b].phys) in p_mapped_by_bats() 71 return bat_addrs[b].start+(pa-bat_addrs[b].phys); in p_mapped_by_bats() 115 void __init setbat(int index, unsigned long virt, phys_addr_t phys, in setbat() argument 135 bat[1].batl = BAT_PHYS_ADDR(phys) | wimgxpp; in setbat() 154 bat->batl = phys | bl | 0x40; /* V=1 */ in setbat() 159 bat_addrs[index].phys = phys; in setbat()
|
D | init_64.c | 197 unsigned long phys) in vmemmap_create_mapping() argument 214 BUG_ON(map_kernel_page(start + i, phys, flags)); in vmemmap_create_mapping() 226 unsigned long phys) in vmemmap_create_mapping() argument 228 int mapped = htab_bolt_mapping(start, start + page_size, phys, in vmemmap_create_mapping() 280 static __meminit void vmemmap_list_populate(unsigned long phys, in vmemmap_list_populate() argument 292 vmem_back->phys = phys; in vmemmap_list_populate() 359 return vmem_back->phys; in vmemmap_list_free() 442 page = (struct page *) (vmem_back->phys + pg_va - in realmode_pfn_to_page()
|
D | 44x_mmu.c | 67 static void __init ppc44x_pin_tlb(unsigned int virt, unsigned int phys) in ppc44x_pin_tlb() argument 81 "r" (phys), in ppc44x_pin_tlb() 137 static void ppc47x_pin_tlb(unsigned int virt, unsigned int phys) in ppc47x_pin_tlb() argument 153 virt, phys, bolted); in ppc47x_pin_tlb() 168 "r" (phys), in ppc47x_pin_tlb()
|
/arch/arm64/mm/ |
D | mmu.c | 77 phys_addr_t phys; in early_pgtable_alloc() local 80 phys = memblock_alloc(PAGE_SIZE, PAGE_SIZE); in early_pgtable_alloc() 81 BUG_ON(!phys); in early_pgtable_alloc() 88 ptr = pte_set_fixmap(phys); in early_pgtable_alloc() 98 return phys; in early_pgtable_alloc() 181 phys_addr_t phys, pgprot_t prot, in alloc_init_pmd() argument 212 if (((addr | next | phys) & ~SECTION_MASK) == 0 && in alloc_init_pmd() 215 pmd_set_huge(pmd, phys, prot); in alloc_init_pmd() 229 alloc_init_pte(pmd, addr, next, __phys_to_pfn(phys), in alloc_init_pmd() 232 phys += next - addr; in alloc_init_pmd() [all …]
|
/arch/sh/boards/mach-sdk7786/ |
D | sram.c | 23 unsigned long phys; in fpga_sram_init() local 47 phys = (area << 26) + SZ_64M - SZ_4K; in fpga_sram_init() 53 vaddr = ioremap(phys, SZ_2K); in fpga_sram_init() 61 SZ_2K >> 10, phys, phys + SZ_2K - 1, area); in fpga_sram_init()
|
/arch/sh/include/asm/ |
D | mmu.h | 61 int pmb_bolt_mapping(unsigned long virt, phys_addr_t phys, 63 void __iomem *pmb_remap_caller(phys_addr_t phys, unsigned long size, 70 pmb_bolt_mapping(unsigned long virt, phys_addr_t phys, in pmb_bolt_mapping() argument 77 pmb_remap_caller(phys_addr_t phys, unsigned long size, in pmb_remap_caller() argument 99 pmb_remap(phys_addr_t phys, unsigned long size, pgprot_t prot) in pmb_remap() argument 101 return pmb_remap_caller(phys, size, prot, __builtin_return_address(0)); in pmb_remap()
|
/arch/x86/include/asm/ |
D | fixmap.h | 151 phys_addr_t phys, pgprot_t flags); 155 phys_addr_t phys, pgprot_t flags) in __set_fixmap() argument 157 native_set_fixmap(idx, phys, flags); in __set_fixmap() 163 #define __late_set_fixmap(idx, phys, flags) __set_fixmap(idx, phys, flags) argument 167 phys_addr_t phys, pgprot_t flags);
|
/arch/powerpc/sysdev/ |
D | ppc4xx_ocm.c | 47 phys_addr_t phys; member 62 phys_addr_t phys; member 140 ocm->phys = rsrc.start; in ocm_init_node() 152 if (!request_mem_region(ocm->phys, ocm->memtotal, "ppc4xx_ocm")) { in ocm_init_node() 160 ocm->nc.phys = ocm->phys; in ocm_init_node() 164 ocm->c.phys = ocm->phys + ocm->nc.memtotal; in ocm_init_node() 169 ocm->nc.phys = 0; in ocm_init_node() 172 ocm->c.phys = 0; in ocm_init_node() 183 ocm->nc.virt = __ioremap(ocm->nc.phys, ocm->nc.memtotal, in ocm_init_node() 199 ocm->c.virt = __ioremap(ocm->c.phys, ocm->c.memtotal, in ocm_init_node() [all …]
|
/arch/um/include/shared/ |
D | mem.h | 9 extern int phys_mapping(unsigned long phys, unsigned long long *offset_out); 17 static inline void *to_virt(unsigned long phys) in to_virt() argument 19 return((void *) uml_physmem + phys); in to_virt()
|
D | mem_user.h | 40 unsigned long phys; member 58 extern unsigned long phys_offset(unsigned long phys); 59 extern void map_memory(unsigned long virt, unsigned long phys,
|
/arch/x86/kernel/cpu/ |
D | perf_event_intel_bts.c | 161 struct bts_phys *phys = &buf->buf[buf->cur_buf]; in bts_config_buffer() local 162 unsigned long index, thresh = 0, end = phys->size; in bts_config_buffer() 163 struct page *page = phys->page; in bts_config_buffer() 168 if (buf->end < phys->offset + buf_size(page)) in bts_config_buffer() 169 end = buf->end - phys->offset - phys->displacement; in bts_config_buffer() 171 index -= phys->offset + phys->displacement; in bts_config_buffer() 181 ds->bts_buffer_base = (u64)(long)page_address(page) + phys->displacement; in bts_config_buffer() 189 static void bts_buffer_pad_out(struct bts_phys *phys, unsigned long head) in bts_buffer_pad_out() argument 191 unsigned long index = head - phys->offset; in bts_buffer_pad_out() 193 memset(page_address(phys->page) + index, 0, phys->size - index); in bts_buffer_pad_out() [all …]
|
/arch/arm/mach-davinci/include/mach/ |
D | uncompress.h | 52 static inline void set_uart_info(u32 phys) in set_uart_info() argument 54 uart = (u32 *)phys; in set_uart_info() 57 #define _DEBUG_LL_ENTRY(machine, phys) \ argument 60 set_uart_info(phys); \
|
/arch/powerpc/include/asm/ |
D | smp.h | 87 static inline void set_hard_smp_processor_id(int cpu, int phys) in set_hard_smp_processor_id() argument 89 smp_hw_index[cpu] = phys; in set_hard_smp_processor_id() 152 static inline void set_hard_smp_processor_id(int cpu, int phys) in set_hard_smp_processor_id() argument 154 paca[cpu].hw_cpu_id = phys; in set_hard_smp_processor_id() 168 static inline void set_hard_smp_processor_id(int cpu, int phys) in set_hard_smp_processor_id() argument 170 boot_cpuid_phys = phys; in set_hard_smp_processor_id()
|
D | dma.h | 271 static __inline__ void set_dma_addr(unsigned int dmanr, unsigned int phys) in set_dma_addr() argument 274 dma_outb(phys & 0xff, in set_dma_addr() 276 dma_outb((phys >> 8) & 0xff, in set_dma_addr() 279 dma_outb((phys >> 1) & 0xff, in set_dma_addr() 281 dma_outb((phys >> 9) & 0xff, in set_dma_addr() 284 set_dma_page(dmanr, phys >> 16); in set_dma_addr()
|
/arch/um/include/asm/ |
D | page.h | 48 #define pte_set_val(pte, phys, prot) \ argument 49 ({ (pte).pte_high = (phys) >> 32; \ 50 (pte).pte_low = (phys) | pgprot_val(prot); }) 77 #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot)) argument 110 #define __va(phys) to_virt((unsigned long) (phys)) argument
|
/arch/arm64/include/asm/ |
D | acpi.h | 32 static inline void __iomem *acpi_os_ioremap(acpi_physical_address phys, in acpi_os_ioremap() argument 35 if (!page_is_ram(phys >> PAGE_SHIFT)) in acpi_os_ioremap() 36 return ioremap(phys, size); in acpi_os_ioremap() 38 return ioremap_cache(phys, size); in acpi_os_ioremap()
|
/arch/metag/mm/ |
D | mmu-meta1.c | 26 static unsigned long map_addr(unsigned long phys) in map_addr() argument 31 offset = phys - dm_base; in map_addr() 36 dm_base = phys & ~(MMCU_DIRECTMAPn_ADDR_SCALE - 1); in map_addr() 42 offset = phys - dm_base; in map_addr()
|
/arch/arm/mach-davinci/ |
D | sram.c | 55 phys_addr_t phys = davinci_soc_info.sram_dma; in sram_init() local 68 addr = ioremap(phys, len); in sram_init() 72 phys, len, -1); in sram_init()
|
/arch/sh/mm/ |
D | cache-sh4.c | 31 static void __flush_cache_one(unsigned long addr, unsigned long phys, 89 static inline void flush_cache_one(unsigned long start, unsigned long phys) in flush_cache_one() argument 102 __flush_cache_one(start, phys, exec_offset); in flush_cache_one() 209 unsigned long address, pfn, phys; in sh4_flush_cache_page() local 220 phys = pfn << PAGE_SHIFT; in sh4_flush_cache_page() 254 (address & shm_align_mask), phys); in sh4_flush_cache_page() 316 static void __flush_cache_one(unsigned long addr, unsigned long phys, in __flush_cache_one() argument 355 p = phys; in __flush_cache_one()
|