/kernel/linux/linux-5.10/arch/arm/mm/ |
D | kasan_init.c | 207 phys_addr_t pa_start, pa_end; in kasan_init() local 242 for_each_mem_range(i, &pa_start, &pa_end) { in kasan_init() 243 void *start = __va(pa_start); in kasan_init() 247 if (pa_start >= arm_lowmem_limit) { in kasan_init() 248 pr_info("Skip highmem block at %pa-%pa\n", &pa_start, &pa_end); in kasan_init() 253 &pa_start, &pa_end, &arm_lowmem_limit); in kasan_init() 258 &pa_start, &pa_end, start, end); in kasan_init()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/usnic/ |
D | usnic_uiom.c | 252 dma_addr_t pa_start = 0; in usnic_uiom_map_sorted_intervals() local 272 pa_start = pa; in usnic_uiom_map_sorted_intervals() 279 (pa != pa_start)) { in usnic_uiom_map_sorted_intervals() 281 size = pa_end - pa_start + PAGE_SIZE; in usnic_uiom_map_sorted_intervals() 283 va_start, &pa_start, size, flags); in usnic_uiom_map_sorted_intervals() 284 err = iommu_map(pd->domain, va_start, pa_start, in usnic_uiom_map_sorted_intervals() 288 va_start, &pa_start, size, err); in usnic_uiom_map_sorted_intervals() 292 pa_start = pa; in usnic_uiom_map_sorted_intervals() 298 size = pa - pa_start + PAGE_SIZE; in usnic_uiom_map_sorted_intervals() 300 va_start, &pa_start, size, flags); in usnic_uiom_map_sorted_intervals() [all …]
|
/kernel/linux/linux-5.10/arch/arm/boot/compressed/ |
D | kaslr.c | 20 u32 pa_start; member 168 for (pa = regions->pa_start; pa < regions->pa_end; pa += SZ_2M, i++) { in count_suitable_regions() 359 regions.pa_start = round_down(image_base, SZ_128M); in kaslr_early_init() 360 regions.pa_end = lowmem_top - PAGE_OFFSET + regions.pa_start; in kaslr_early_init() 422 puthex32(regions.pa_start); in kaslr_early_init()
|
/kernel/linux/linux-5.10/arch/arm64/mm/ |
D | kasan_init.c | 215 phys_addr_t pa_start, pa_end; in kasan_init() local 249 for_each_mem_range(i, &pa_start, &pa_end) { in kasan_init() 250 void *start = (void *)__phys_to_virt(pa_start); in kasan_init()
|
D | mmu.c | 592 phys_addr_t pa_start = __pa_symbol(va_start); in map_kernel_segment() local 595 BUG_ON(!PAGE_ALIGNED(pa_start)); in map_kernel_segment() 598 __create_pgd_mapping(pgdp, pa_start, (unsigned long)va_start, size, prot, in map_kernel_segment() 605 vma->phys_addr = pa_start; in map_kernel_segment() 637 phys_addr_t pa_start = __pa_symbol(__entry_tramp_text_start); in map_entry_trampoline() local 644 __create_pgd_mapping(tramp_pg_dir, pa_start, TRAMP_VALIAS, in map_entry_trampoline() 651 pa_start + i * PAGE_SIZE, prot); in map_entry_trampoline()
|
/kernel/linux/linux-5.10/arch/powerpc/kvm/ |
D | book3s_mmu_hpte.c | 298 void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end) in kvmppc_mmu_pte_pflush() argument 304 trace_kvm_book3s_mmu_flush("p", vcpu, pa_start, pa_end); in kvmppc_mmu_pte_pflush() 312 if ((pte->pte.raddr >= pa_start) && in kvmppc_mmu_pte_pflush()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/nohash/ |
D | kaslr_booke.c | 24 unsigned long pa_start; member 327 regions.pa_start = memstart_addr; in kaslr_choose_location()
|
/kernel/linux/linux-5.10/arch/sparc/mm/ |
D | init_64.c | 979 u64 ret_end, pa_start, m_mask, m_match, m_end; in memblock_nid_range() local 996 pa_start = start + mblock->offset; in memblock_nid_range() 1003 if ((pa_start & m->mask) == m->match) { in memblock_nid_range() 1020 if ((pa_start & m->mask) == m->match) { in memblock_nid_range() 1045 m_end += pa_start & ~((1ul << fls64(m_mask)) - 1); in memblock_nid_range()
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
D | kvm_book3s.h | 147 extern void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end);
|
/kernel/linux/patches/linux-5.10/imx8mm_patch/patches/ |
D | 0001_linux_arch.patch | 15917 phys_addr_t pa_start = __pa_symbol(__entry_tramp_text_start); 15923 - __create_pgd_mapping(tramp_pg_dir, pa_start, TRAMP_VALIAS, 15926 + __create_pgd_mapping(tramp_pg_dir, pa_start, TRAMP_VALIAS, PAGE_SIZE, 15932 - pa_start + i * PAGE_SIZE, prot); 15934 + __set_fixmap(FIX_ENTRY_TRAMP_TEXT, pa_start, prot);
|