/arch/s390/kernel/ |
D | vdso.c | 98 unsigned long segment_table, page_table, page_frame; in vdso_alloc_per_cpu() local 108 page_table = get_zeroed_page(GFP_KERNEL | GFP_DMA); in vdso_alloc_per_cpu() 110 if (!segment_table || !page_table || !page_frame) in vdso_alloc_per_cpu() 115 clear_table((unsigned long *) page_table, _PAGE_INVALID, in vdso_alloc_per_cpu() 118 *(unsigned long *) segment_table = _SEGMENT_ENTRY + page_table; in vdso_alloc_per_cpu() 119 *(unsigned long *) page_table = _PAGE_PROTECT + page_frame; in vdso_alloc_per_cpu() 121 psal = (u32 *) (page_table + 256*sizeof(unsigned long)); in vdso_alloc_per_cpu() 139 free_page(page_table); in vdso_alloc_per_cpu() 146 unsigned long segment_table, page_table, page_frame; in vdso_free_per_cpu() local 155 page_table = *(unsigned long *) segment_table; in vdso_free_per_cpu() [all …]
|
/arch/unicore32/kernel/ |
D | hibernate.c | 54 pte_t *page_table = (pte_t *)get_safe_page(GFP_ATOMIC); in resume_one_page_table_init() local 55 if (!page_table) in resume_one_page_table_init() 58 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_KERNEL_TABLE)); in resume_one_page_table_init() 60 BUG_ON(page_table != pte_offset_kernel(pmd, 0)); in resume_one_page_table_init() 62 return page_table; in resume_one_page_table_init()
|
/arch/x86/power/ |
D | hibernate_32.c | 62 pte_t *page_table = (pte_t *)get_safe_page(GFP_ATOMIC); in resume_one_page_table_init() local 63 if (!page_table) in resume_one_page_table_init() 66 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in resume_one_page_table_init() 68 BUG_ON(page_table != pte_offset_kernel(pmd, 0)); in resume_one_page_table_init() 70 return page_table; in resume_one_page_table_init()
|
/arch/sparc/mm/ |
D | iommu.c | 100 iommu->page_table = (iopte_t *)tmp; in sbus_iommu_init() 103 memset(iommu->page_table, 0, IOMMU_NPTES*sizeof(iopte_t)); in sbus_iommu_init() 107 base = __pa((unsigned long)iommu->page_table) >> 4; in sbus_iommu_init() 127 impl, vers, iommu->page_table, in sbus_iommu_init() 190 iopte0 = &iommu->page_table[ioptex]; in iommu_get_one() 292 iopte_val(iommu->page_table[ioptex + i]) = 0; in iommu_release_one() 329 iopte_t *iopte = iommu->page_table; in iommu_map_dma_area() 394 iopte_t *iopte = iommu->page_table; in iommu_unmap_dma_area()
|
D | io-unit.c | 66 iounit->page_table = xpt; in iounit_iommu_init() 69 xptend = iounit->page_table + (16 * PAGE_SIZE) / sizeof(iopte_t); in iounit_iommu_init() 136 sbus_writel(iopte, &iounit->page_table[scan]); in iounit_get_area() 230 iopte = iounit->page_table + i; in iounit_map_dma_area()
|
/arch/sparc/kernel/ |
D | iommu.c | 226 iommu->page_table = (iopte_t *)page_address(page); in iommu_table_init() 229 iopte_make_dummy(iommu, &iommu->page_table[i]); in iommu_table_init() 253 return iommu->page_table + entry; in alloc_npages() 319 ((iopte - iommu->page_table) << IO_PAGE_SHIFT)); in dma_4u_alloc_coherent() 389 ((base - iommu->page_table) << IO_PAGE_SHIFT)); in dma_4u_map_page() 498 base = iommu->page_table + in dma_4u_unmap_page() 594 base = iommu->page_table + entry; in dma_4u_map_sg() 661 base = iommu->page_table + entry; in dma_4u_map_sg() 689 base = iommu->page_table + in fetch_sg_ctx() 730 base = iommu->page_table + entry; in dma_4u_unmap_sg() [all …]
|
D | ldc.c | 104 struct ldc_mtable_entry *page_table; member 1040 iommu->page_table = table; in ldc_iommu_init() 1052 iommu->page_table = NULL; in ldc_iommu_init() 1072 free_pages((unsigned long) iommu->page_table, order); in ldc_iommu_release() 1073 iommu->page_table = NULL; in ldc_iommu_release() 1962 return iommu->page_table + entry; in alloc_npages() 2011 struct ldc_mtable_entry *page_table; member 2026 sp->page_table[sp->pte_idx].mte = sp->mte_base | pa; in fill_cookies() 2113 state.page_table = iommu->page_table; in ldc_map_sg() 2117 state.pte_idx = (base - iommu->page_table); in ldc_map_sg() [all …]
|
D | psycho_common.c | 427 upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); in psycho_iommu_init()
|
D | pci_fire.c | 62 upa_writeq(__pa(iommu->page_table) | 0x7UL, iommu->iommu_tsbbase); in pci_fire_pbm_iommu_init()
|
D | sbus.c | 623 upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); in sbus_iommu_init()
|
D | pci_schizo.c | 1203 upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); in schizo_pbm_iommu_init()
|
/arch/tile/mm/ |
D | migrate.h | 31 extern int flush_and_install_context(HV_PhysAddr page_table, HV_PTE access,
|
D | init.c | 118 static void __init assign_pte(pmd_t *pmd, pte_t *page_table) in assign_pte() argument 120 phys_addr_t pa = __pa(page_table); in assign_pte() 126 if (page_table != (pte_t *)pmd_page_vaddr(*pmd)) in assign_pte()
|
D | fault.c | 187 unsigned long pgd_pfn = ctx.page_table >> PAGE_SHIFT; in get_current_pgd() 190 return (pgd_t *) __va(ctx.page_table); in get_current_pgd()
|
/arch/sparc/include/asm/ |
D | io-unit.h | 46 iopte_t __iomem *page_table; member
|
D | iommu_64.h | 30 iopte_t *page_table; member
|
D | iommu_32.h | 103 iopte_t *page_table; member
|
/arch/x86/mm/ |
D | init_32.c | 98 pte_t *page_table = (pte_t *)alloc_low_page(); in one_page_table_init() local 100 paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT); in one_page_table_init() 101 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in one_page_table_init() 102 BUG_ON(page_table != pte_offset_kernel(pmd, 0)); in one_page_table_init()
|
/arch/tile/include/hv/ |
D | hypervisor.h | 772 int hv_install_context(HV_PhysAddr page_table, HV_PTE access, HV_ASID asid, 814 HV_PhysAddr page_table; member
|
/arch/tile/kernel/ |
D | hvglue_trace.c | 206 HV_WRAP4(int, hv_install_context, HV_PhysAddr, page_table, HV_PTE, access,
|