/arch/s390/kernel/ |
D | vdso.c | 95 unsigned long segment_table, page_table, page_frame; in vdso_alloc_per_cpu() local 106 page_table = get_zeroed_page(GFP_KERNEL | GFP_DMA); in vdso_alloc_per_cpu() 108 if (!segment_table || !page_table || !page_frame) in vdso_alloc_per_cpu() 119 clear_table((unsigned long *) page_table, _PAGE_INVALID, in vdso_alloc_per_cpu() 122 *(unsigned long *) segment_table = _SEGMENT_ENTRY + page_table; in vdso_alloc_per_cpu() 123 *(unsigned long *) page_table = _PAGE_PROTECT + page_frame; in vdso_alloc_per_cpu() 125 psal = (u32 *) (page_table + 256*sizeof(unsigned long)); in vdso_alloc_per_cpu() 143 free_page(page_table); in vdso_alloc_per_cpu() 150 unsigned long segment_table, page_table, page_frame; in vdso_free_per_cpu() local 159 page_table = *(unsigned long *) segment_table; in vdso_free_per_cpu() [all …]
|
/arch/unicore32/kernel/ |
D | hibernate.c | 54 pte_t *page_table = (pte_t *)get_safe_page(GFP_ATOMIC); in resume_one_page_table_init() local 55 if (!page_table) in resume_one_page_table_init() 58 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_KERNEL_TABLE)); in resume_one_page_table_init() 60 BUG_ON(page_table != pte_offset_kernel(pmd, 0)); in resume_one_page_table_init() 62 return page_table; in resume_one_page_table_init()
|
/arch/x86/power/ |
D | hibernate_32.c | 62 pte_t *page_table = (pte_t *)get_safe_page(GFP_ATOMIC); in resume_one_page_table_init() local 63 if (!page_table) in resume_one_page_table_init() 66 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in resume_one_page_table_init() 68 BUG_ON(page_table != pte_offset_kernel(pmd, 0)); in resume_one_page_table_init() 70 return page_table; in resume_one_page_table_init()
|
/arch/sparc/mm/ |
D | iommu.c | 100 iommu->page_table = (iopte_t *)tmp; in sbus_iommu_init() 103 memset(iommu->page_table, 0, IOMMU_NPTES*sizeof(iopte_t)); in sbus_iommu_init() 107 base = __pa((unsigned long)iommu->page_table) >> 4; in sbus_iommu_init() 127 impl, vers, iommu->page_table, in sbus_iommu_init() 190 iopte0 = &iommu->page_table[ioptex]; in iommu_get_one() 292 iopte_val(iommu->page_table[ioptex + i]) = 0; in iommu_release_one() 329 iopte_t *iopte = iommu->page_table; in iommu_map_dma_area() 394 iopte_t *iopte = iommu->page_table; in iommu_unmap_dma_area()
|
D | io-unit.c | 66 iounit->page_table = xpt; in iounit_iommu_init() 69 xptend = iounit->page_table + (16 * PAGE_SIZE) / sizeof(iopte_t); in iounit_iommu_init() 136 sbus_writel(iopte_val(iopte), &iounit->page_table[scan]); in iounit_get_area() 230 iopte = iounit->page_table + i; in iounit_map_dma_area()
|
/arch/sparc/kernel/ |
D | iommu.c | 138 iommu->page_table = (iopte_t *)page_address(page); in iommu_table_init() 141 iopte_make_dummy(iommu, &iommu->page_table[i]); in iommu_table_init() 167 return iommu->page_table + entry; in alloc_npages() 231 ((iopte - iommu->page_table) << IO_PAGE_SHIFT)); in dma_4u_alloc_coherent() 297 ((base - iommu->page_table) << IO_PAGE_SHIFT)); in dma_4u_map_page() 406 base = iommu->page_table + in dma_4u_unmap_page() 502 base = iommu->page_table + entry; in dma_4u_map_sg() 568 base = iommu->page_table + entry; in dma_4u_map_sg() 600 base = iommu->page_table + in fetch_sg_ctx() 640 base = iommu->page_table + entry; in dma_4u_unmap_sg() [all …]
|
D | ldc.c | 108 struct ldc_mtable_entry *page_table; member 1023 base = iommu->page_table + entry; in ldc_demap() 1073 ldc_iommu->page_table = table; in ldc_iommu_init() 1085 ldc_iommu->page_table = NULL; in ldc_iommu_init() 1106 free_pages((unsigned long) ldc_iommu->page_table, order); in ldc_iommu_release() 1107 ldc_iommu->page_table = NULL; in ldc_iommu_release() 1964 return iommu->page_table + entry; in alloc_npages() 2013 struct ldc_mtable_entry *page_table; member 2028 sp->page_table[sp->pte_idx].mte = sp->mte_base | pa; in fill_cookies() 2114 state.page_table = iommu->page_table; in ldc_map_sg() [all …]
|
D | psycho_common.c | 427 upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); in psycho_iommu_init()
|
D | pci_fire.c | 62 upa_writeq(__pa(iommu->page_table) | 0x7UL, iommu->iommu_tsbbase); in pci_fire_pbm_iommu_init()
|
D | sbus.c | 623 upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); in sbus_iommu_init()
|
D | pci_schizo.c | 1203 upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); in schizo_pbm_iommu_init()
|
/arch/tile/mm/ |
D | migrate.h | 31 extern int flush_and_install_context(HV_PhysAddr page_table, HV_PTE access,
|
D | init.c | 118 static void __init assign_pte(pmd_t *pmd, pte_t *page_table) in assign_pte() argument 120 phys_addr_t pa = __pa(page_table); in assign_pte() 126 if (page_table != (pte_t *)pmd_page_vaddr(*pmd)) in assign_pte()
|
D | fault.c | 186 unsigned long pgd_pfn = ctx.page_table >> PAGE_SHIFT; in get_current_pgd() 189 return (pgd_t *) __va(ctx.page_table); in get_current_pgd()
|
/arch/arm64/kernel/ |
D | hibernate-asm.S | 36 .macro break_before_make_ttbr_switch zero_page, page_table argument 41 msr ttbr1_el1, \page_table
|
/arch/sparc/include/asm/ |
D | io-unit.h | 46 iopte_t __iomem *page_table; member
|
D | iommu_32.h | 103 iopte_t *page_table; member
|
D | iommu_64.h | 59 iopte_t *page_table; member
|
/arch/x86/mm/ |
D | init_32.c | 97 pte_t *page_table = (pte_t *)alloc_low_page(); in one_page_table_init() local 99 paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT); in one_page_table_init() 100 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in one_page_table_init() 101 BUG_ON(page_table != pte_offset_kernel(pmd, 0)); in one_page_table_init()
|
/arch/tile/include/hv/ |
D | hypervisor.h | 775 int hv_install_context(HV_PhysAddr page_table, HV_PTE access, HV_ASID asid, 817 HV_PhysAddr page_table; member
|
/arch/tile/kernel/ |
D | hvglue_trace.c | 208 HV_WRAP4(int, hv_install_context, HV_PhysAddr, page_table, HV_PTE, access,
|