/arch/m68k/mm/ |
D | sun3kmap.c | 30 static inline void do_page_mapin(unsigned long phys, unsigned long virt, in do_page_mapin() argument 40 sun3_put_pte(virt, pte); in do_page_mapin() 44 print_pte_vaddr(virt); in do_page_mapin() 49 static inline void do_pmeg_mapin(unsigned long phys, unsigned long virt, in do_pmeg_mapin() argument 53 if(sun3_get_segmap(virt & ~SUN3_PMEG_MASK) == SUN3_INVALID_PMEG) in do_pmeg_mapin() 54 mmu_emu_map_pmeg(sun3_get_context(), virt); in do_pmeg_mapin() 57 do_page_mapin(phys, virt, type); in do_pmeg_mapin() 59 virt += PAGE_SIZE; in do_pmeg_mapin() 68 unsigned long offset, virt, ret; in sun3_ioremap() local 89 virt = (unsigned long)area->addr; in sun3_ioremap() [all …]
|
/arch/unicore32/mm/ |
D | mm.h | 17 static inline pmd_t *pmd_off(pgd_t *pgd, unsigned long virt) in pmd_off() argument 19 return pmd_offset((pud_t *)pgd, virt); in pmd_off() 22 static inline pmd_t *pmd_off_k(unsigned long virt) in pmd_off_k() argument 24 return pmd_off(pgd_offset_k(virt), virt); in pmd_off_k()
|
D | ioremap.c | 44 int ioremap_page(unsigned long virt, unsigned long phys, in ioremap_page() argument 47 return ioremap_page_range(virt, virt + PAGE_SIZE, phys, in ioremap_page() 62 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument 64 unsigned long addr = virt, end = virt + (size & ~(SZ_4M - 1)); in unmap_area_sections() 94 flush_tlb_kernel_range(virt, end); in unmap_area_sections() 98 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 101 unsigned long addr = virt, end = virt + size; in remap_area_sections() 108 unmap_area_sections(virt, size); in remap_area_sections()
|
/arch/parisc/kernel/ |
D | kexec.c | 79 void *virt = (void *)__fix_to_virt(FIX_TEXT_KEXEC); in machine_kexec() local 88 desc.addr = (long long)virt; in machine_kexec() 90 reloc = (void *)virt; in machine_kexec() 93 memcpy(virt, dereference_function_descriptor(relocate_new_kernel), in machine_kexec() 96 *(unsigned long *)(virt + kexec_cmdline_offset) = arch->cmdline; in machine_kexec() 97 *(unsigned long *)(virt + kexec_initrd_start_offset) = arch->initrd_start; in machine_kexec() 98 *(unsigned long *)(virt + kexec_initrd_end_offset) = arch->initrd_end; in machine_kexec() 99 *(unsigned long *)(virt + kexec_free_mem_offset) = PAGE0->mem_free; in machine_kexec()
|
/arch/arm/mach-hisi/ |
D | platsmp.c | 108 void __iomem *virt; in hix5hd2_set_scu_boot_addr() local 110 virt = ioremap(start_addr, PAGE_SIZE); in hix5hd2_set_scu_boot_addr() 112 writel_relaxed(0xe51ff004, virt); /* ldr pc, [pc, #-4] */ in hix5hd2_set_scu_boot_addr() 113 writel_relaxed(jump_addr, virt + 4); /* pc jump phy address */ in hix5hd2_set_scu_boot_addr() 114 iounmap(virt); in hix5hd2_set_scu_boot_addr() 144 void __iomem *virt; in hip01_set_boot_addr() local 146 virt = phys_to_virt(start_addr); in hip01_set_boot_addr() 148 writel_relaxed(0xe51ff004, virt); in hip01_set_boot_addr() 149 writel_relaxed(jump_addr, virt + 4); in hip01_set_boot_addr()
|
/arch/x86/pci/ |
D | mmconfig_64.c | 23 if (cfg && cfg->virt) in pci_dev_base() 24 return cfg->virt + (PCI_MMCFG_BUS_OFFSET(bus) | (devfn << 12)); in pci_dev_base() 139 cfg->virt = mcfg_ioremap(cfg); in pci_mmcfg_arch_map() 140 if (!cfg->virt) { in pci_mmcfg_arch_map() 150 if (cfg && cfg->virt) { in pci_mmcfg_arch_unmap() 151 iounmap(cfg->virt + PCI_MMCFG_BUS_OFFSET(cfg->start_bus)); in pci_mmcfg_arch_unmap() 152 cfg->virt = NULL; in pci_mmcfg_arch_unmap()
|
/arch/powerpc/mm/nohash/ |
D | fsl_booke.c | 105 static void settlbcam(int index, unsigned long virt, phys_addr_t phys, in settlbcam() argument 119 TLBCAM[index].MAS2 = virt & PAGE_MASK; in settlbcam() 138 tlbcam_addrs[index].start = virt; in settlbcam() 139 tlbcam_addrs[index].limit = virt + size - 1; in settlbcam() 143 unsigned long calc_cam_sz(unsigned long ram, unsigned long virt, in calc_cam_sz() argument 147 unsigned int align = __ffs(virt | phys); in calc_cam_sz() 168 static unsigned long map_mem_in_cams_addr(phys_addr_t phys, unsigned long virt, in map_mem_in_cams_addr() argument 179 cam_sz = calc_cam_sz(ram, virt, phys); in map_mem_in_cams_addr() 181 settlbcam(i, virt, phys, cam_sz, in map_mem_in_cams_addr() 186 virt += cam_sz; in map_mem_in_cams_addr() [all …]
|
D | 44x.c | 54 static void __init ppc44x_pin_tlb(unsigned int virt, unsigned int phys) in ppc44x_pin_tlb() argument 69 "r" (virt | PPC44x_TLB_VALID | PPC44x_TLB_256M), in ppc44x_pin_tlb() 124 static void ppc47x_pin_tlb(unsigned int virt, unsigned int phys) in ppc47x_pin_tlb() argument 140 virt, phys, bolted); in ppc47x_pin_tlb() 156 "r" (virt | PPC47x_TLB0_VALID | PPC47x_TLB0_256M), in ppc47x_pin_tlb()
|
/arch/arm/mm/ |
D | ioremap.c | 109 int ioremap_page(unsigned long virt, unsigned long phys, in ioremap_page() argument 112 return ioremap_page_range(virt, virt + PAGE_SIZE, phys, in ioremap_page() 142 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument 144 unsigned long addr = virt, end = virt + (size & ~(SZ_1M - 1)); in unmap_area_sections() 185 flush_tlb_kernel_range(virt, end); in unmap_area_sections() 189 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 192 unsigned long addr = virt, end = virt + size; in remap_area_sections() 201 unmap_area_sections(virt, size); in remap_area_sections() 221 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument 224 unsigned long addr = virt, end = virt + size; in remap_area_supersections() [all …]
|
D | mm.h | 39 static inline pmd_t *pmd_off_k(unsigned long virt) in pmd_off_k() argument 41 return pmd_offset(pud_offset(pgd_offset_k(virt), virt), virt); in pmd_off_k()
|
/arch/xtensa/mm/ |
D | cache.c | 150 unsigned long virt; in flush_dcache_page() local 163 virt = TLBTEMP_BASE_1 + (phys & DCACHE_ALIAS_MASK); in flush_dcache_page() 164 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_page() 166 virt = TLBTEMP_BASE_1 + (temp & DCACHE_ALIAS_MASK); in flush_dcache_page() 169 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_page() 172 __invalidate_icache_page_alias(virt, phys); in flush_dcache_page() 205 unsigned long virt = TLBTEMP_BASE_1 + (address & DCACHE_ALIAS_MASK); in local_flush_cache_page() local 208 __flush_invalidate_dcache_page_alias(virt, phys); in local_flush_cache_page() 209 __invalidate_icache_page_alias(virt, phys); in local_flush_cache_page()
|
/arch/powerpc/boot/ |
D | oflib.c | 152 unsigned int of_claim(unsigned long virt, unsigned long size, in of_claim() argument 161 return of_call_prom("claim", 3, 1, virt, size, align); in of_claim() 164 align, size, virt); in of_claim() 168 align, size, virt); in of_claim() 171 0x12, size, virt, virt); in of_claim() 172 return virt; in of_claim()
|
/arch/um/kernel/ |
D | physmem.c | 41 void map_memory(unsigned long virt, unsigned long phys, unsigned long len, in map_memory() argument 48 err = os_map_memory((void *) virt, fd, offset, len, r, w, x); in map_memory() 55 "err = %d\n", virt, fd, offset, len, r, w, x, err); in map_memory() 188 return region->virt; in find_iomem() 211 region->virt = iomem_start; in setup_iomem() 212 region->phys = __pa(region->virt); in setup_iomem()
|
/arch/arm/kvm/ |
D | Makefile | 6 plus_virt := $(call as-instr,.arch_extension virt,+virt) 7 ifeq ($(plus_virt),+virt) 11 KVM := ../../../virt/kvm 13 ccflags-y += -I $(srctree)/$(src) -I $(srctree)/virt/kvm/arm/vgic
|
/arch/arm/crypto/ |
D | aes-ce-glue.c | 189 ce_aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_encrypt() 209 ce_aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_decrypt() 227 ce_aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_encrypt_walk() 257 ce_aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_decrypt_walk() 327 ce_aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_encrypt() 385 ce_aes_cbc_cts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_decrypt() 404 ce_aes_ctr_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ctr_encrypt() 413 u8 *tdst = walk.dst.virt.addr; in ctr_encrypt() 414 u8 *tsrc = walk.src.virt.addr; in ctr_encrypt() 496 ce_aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt() [all …]
|
/arch/x86/crypto/ |
D | glue_helper.c | 34 const u8 *src = walk.src.virt.addr; in glue_ecb_req_128bit() 35 u8 *dst = walk.dst.virt.addr; in glue_ecb_req_128bit() 78 const u128 *src = (u128 *)walk.src.virt.addr; in glue_cbc_encrypt_req_128bit() 79 u128 *dst = (u128 *)walk.dst.virt.addr; in glue_cbc_encrypt_req_128bit() 111 const u128 *src = walk.src.virt.addr; in glue_cbc_decrypt_req_128bit() 112 u128 *dst = walk.dst.virt.addr; in glue_cbc_decrypt_req_128bit() 172 const u128 *src = walk.src.virt.addr; in glue_ctr_req_128bit() 173 u128 *dst = walk.dst.virt.addr; in glue_ctr_req_128bit() 215 memcpy(&tmp, walk.src.virt.addr, nbytes); in glue_ctr_req_128bit() 219 memcpy(walk.dst.virt.addr, &tmp, nbytes); in glue_ctr_req_128bit() [all …]
|
D | cast5_avx_glue.c | 61 u8 *wsrc = walk.src.virt.addr; in ecb_crypt() 62 u8 *wdst = walk.dst.virt.addr; in ecb_crypt() 122 u64 *src = (u64 *)walk.src.virt.addr; in cbc_encrypt() 123 u64 *dst = (u64 *)walk.dst.virt.addr; in cbc_encrypt() 147 u64 *src = (u64 *)walk->src.virt.addr; in __cbc_decrypt() 148 u64 *dst = (u64 *)walk->dst.virt.addr; in __cbc_decrypt() 221 u8 *src = walk->src.virt.addr; in ctr_crypt_final() 222 u8 *dst = walk->dst.virt.addr; in ctr_crypt_final() 236 u64 *src = (u64 *)walk->src.virt.addr; in __ctr_crypt() 237 u64 *dst = (u64 *)walk->dst.virt.addr; in __ctr_crypt()
|
/arch/um/include/shared/ |
D | mem.h | 12 static inline unsigned long to_phys(void *virt) in to_phys() argument 14 return(((unsigned long) virt) - uml_physmem); in to_phys()
|
D | mem_user.h | 41 unsigned long virt; member 59 extern void map_memory(unsigned long virt, unsigned long phys,
|
/arch/arm/mach-axxia/ |
D | platsmp.c | 24 u32 *virt = (u32 *) phys_to_virt(release_phys); in write_release_addr() local 25 writel_relaxed(__pa_symbol(secondary_startup), virt); in write_release_addr() 28 __cpuc_flush_dcache_area(virt, sizeof(u32)); in write_release_addr()
|
/arch/x86/oprofile/ |
D | op_model_amd.c | 276 int virt = op_x86_phys_to_virt(i); in op_mux_switch_ctrl() local 277 if (!reset_value[virt]) in op_mux_switch_ctrl() 281 val |= op_x86_get_ctrl(model, &counter_config[virt]); in op_mux_switch_ctrl() 366 int virt = op_x86_phys_to_virt(i); in op_amd_setup_ctrs() local 367 if (!reset_value[virt]) in op_amd_setup_ctrs() 371 wrmsrl(msrs->counters[i].addr, -(u64)reset_value[virt]); in op_amd_setup_ctrs() 376 val |= op_x86_get_ctrl(model, &counter_config[virt]); in op_amd_setup_ctrs() 388 int virt = op_x86_phys_to_virt(i); in op_amd_check_ctrs() local 389 if (!reset_value[virt]) in op_amd_check_ctrs() 395 oprofile_add_sample(regs, virt); in op_amd_check_ctrs() [all …]
|
/arch/powerpc/crypto/ |
D | aes-spe-glue.c | 199 ppc_encrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ecb_encrypt() 227 ppc_decrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ecb_decrypt() 255 ppc_encrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, in ppc_cbc_encrypt() 283 ppc_decrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, in ppc_cbc_decrypt() 312 ppc_crypt_ctr(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ctr_crypt() 343 ppc_encrypt_xts(walk.dst.virt.addr, walk.src.virt.addr, in ppc_xts_encrypt() 374 ppc_decrypt_xts(walk.dst.virt.addr, walk.src.virt.addr, in ppc_xts_decrypt()
|
/arch/arm64/crypto/ |
D | aes-glue.c | 203 aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_encrypt() 223 aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_decrypt() 241 aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_encrypt_walk() 270 aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_decrypt_walk() 339 aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_encrypt() 396 aes_cbc_cts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_decrypt() 432 aes_essiv_cbc_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in essiv_cbc_encrypt() 454 aes_essiv_cbc_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in essiv_cbc_decrypt() 475 aes_ctr_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ctr_encrypt() 483 u8 *tdst = walk.dst.virt.addr; in ctr_encrypt() [all …]
|
D | aes-neonbs-glue.c | 116 fn(walk.dst.virt.addr, walk.src.virt.addr, ctx->rk, in __ecb_crypt() 172 neon_aes_cbc_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cbc_encrypt() 198 aesbs_cbc_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cbc_decrypt() 249 aesbs_ctr_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ctr_encrypt() 254 u8 *dst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; in ctr_encrypt() 255 u8 *src = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; in ctr_encrypt() 363 out = walk.dst.virt.addr; in __xts_crypt() 364 in = walk.src.virt.addr; in __xts_crypt() 405 out = walk.dst.virt.addr; in __xts_crypt() 406 in = walk.src.virt.addr; in __xts_crypt()
|
/arch/csky/include/asm/ |
D | highmem.h | 30 #define PKMAP_NR(virt) ((virt-PKMAP_BASE) >> PAGE_SHIFT) argument
|