Home
last modified time | relevance | path

Searched refs:ptes (Results 1 – 17 of 17) sorted by relevance

/arch/x86/xen/
Dgrant-table.c27 pte_t **ptes; member
45 set_pte_at(&init_mm, addr, gnttab_shared_vm_area.ptes[i], in arch_gnttab_map_shared()
67 set_pte_at(&init_mm, addr, gnttab_status_vm_area.ptes[i], in arch_gnttab_map_status()
77 pte_t **ptes; in arch_gnttab_unmap() local
82 ptes = gnttab_status_vm_area.ptes; in arch_gnttab_unmap()
84 ptes = gnttab_shared_vm_area.ptes; in arch_gnttab_unmap()
89 set_pte_at(&init_mm, addr, ptes[i], __pte(0)); in arch_gnttab_unmap()
98 area->ptes[area->idx++] = pte; in gnttab_apply()
104 area->ptes = kmalloc_array(nr_frames, sizeof(*area->ptes), GFP_KERNEL); in arch_gnttab_valloc()
105 if (area->ptes == NULL) in arch_gnttab_valloc()
[all …]
/arch/alpha/kernel/
Dpci_iommu.c78 arena->ptes = memblock_alloc(mem_size, align); in iommu_arena_new_node()
79 if (!arena->ptes) in iommu_arena_new_node()
108 unsigned long *ptes; in iommu_arena_find_pages() local
118 ptes = arena->ptes; in iommu_arena_find_pages()
130 if (ptes[p+i]) { in iommu_arena_find_pages()
164 unsigned long *ptes; in iommu_arena_alloc() local
170 ptes = arena->ptes; in iommu_arena_alloc()
183 ptes[p+i] = IOMMU_INVALID_PTE; in iommu_arena_alloc()
197 p = arena->ptes + ofs; in iommu_arena_free()
295 arena->ptes[i + dma_ofs] = mk_iommu_pte(paddr); in pci_map_single_1()
[all …]
Dcore_titan.c328 port->tba[0].csr = virt_to_phys(hose->sg_isa->ptes); in titan_init_one_pachip_port()
336 port->tba[2].csr = virt_to_phys(hose->sg_pci->ptes); in titan_init_one_pachip_port()
463 unsigned long *ptes; in titan_ioremap() local
518 ptes = hose->sg_pci->ptes; in titan_ioremap()
522 pfn = ptes[baddr >> PAGE_SHIFT]; in titan_ioremap()
711 pte = aper->arena->ptes[baddr >> PAGE_SHIFT]; in titan_agp_translate()
Dcore_marvel.c295 csrs->POx_TBASE[0].csr = virt_to_phys(hose->sg_isa->ptes); in io7_init_hose()
312 csrs->POx_TBASE[2].csr = virt_to_phys(hose->sg_pci->ptes); in io7_init_hose()
690 unsigned long *ptes; in marvel_ioremap() local
745 ptes = hose->sg_pci->ptes; in marvel_ioremap()
749 pfn = ptes[baddr >> PAGE_SHIFT]; in marvel_ioremap()
1003 pte = aper->arena->ptes[baddr >> PAGE_SHIFT]; in marvel_agp_translate()
Dpci_impl.h139 unsigned long *ptes; member
Dcore_cia.c464 arena->ptes[4] = pte0; in verify_tb_operation()
488 arena->ptes[5] = pte0; in verify_tb_operation()
524 arena->ptes[4] = 0; in verify_tb_operation()
525 arena->ptes[5] = 0; in verify_tb_operation()
737 *(vip)CIA_IOC_PCI_T0_BASE = virt_to_phys(hose->sg_isa->ptes) >> 2; in do_init_arch()
Dcore_tsunami.c337 pchip->tba[0].csr = virt_to_phys(hose->sg_isa->ptes); in tsunami_init_one_pchip()
341 pchip->tba[1].csr = virt_to_phys(hose->sg_pci->ptes); in tsunami_init_one_pchip()
Dcore_mcpcia.c378 *(vuip)MCPCIA_T0_BASE(mid) = virt_to_phys(hose->sg_isa->ptes) >> 8; in mcpcia_startup_hose()
382 *(vuip)MCPCIA_T1_BASE(mid) = virt_to_phys(hose->sg_pci->ptes) >> 8; in mcpcia_startup_hose()
Dcore_apecs.c361 *(vuip)APECS_IOC_TB2R = virt_to_phys(hose->sg_isa->ptes) >> 1; in apecs_init_arch()
Dcore_wildfire.c123 pci->pci_window[0].tbase.csr = virt_to_phys(hose->sg_isa->ptes); in wildfire_init_hose()
135 pci->pci_window[3].tbase.csr = virt_to_phys(hose->sg_pci->ptes); in wildfire_init_hose()
Dcore_lca.c286 *(vulp)LCA_IOC_T_BASE0 = virt_to_phys(hose->sg_isa->ptes); in lca_init_arch()
Dcore_t2.c361 *(vulp)T2_TBASE2 = virt_to_phys(hose->sg_isa->ptes) >> 1; in t2_sg_map_window2()
/arch/powerpc/include/asm/
Dplpar_wrappers.h175 unsigned long *ptes) in plpar_pte_read_4() argument
183 memcpy(ptes, retbuf, 8*sizeof(unsigned long)); in plpar_pte_read_4()
193 unsigned long *ptes) in plpar_pte_read_4_raw() argument
201 memcpy(ptes, retbuf, 8*sizeof(unsigned long)); in plpar_pte_read_4_raw()
399 unsigned long *ptes) in plpar_pte_read_4() argument
/arch/powerpc/mm/ptdump/
Dhashpagetable.c244 } ptes[4]; in pseries_find() local
262 lpar_rc = plpar_pte_read_4(0, hpte_group, (void *)ptes); in pseries_find()
267 if (HPTE_V_COMPARE(ptes[j].v, want_v) && in pseries_find()
268 (ptes[j].v & HPTE_V_VALID)) { in pseries_find()
270 *v = ptes[j].v; in pseries_find()
271 *r = ptes[j].r; in pseries_find()
/arch/powerpc/platforms/pseries/
Dlpar.c854 } ptes[4]; in manual_hpte_clear_all() local
863 lpar_rc = plpar_pte_read_4_raw(0, i, (void *)ptes); in manual_hpte_clear_all()
870 if ((ptes[j].pteh & HPTE_V_VRMA_MASK) == in manual_hpte_clear_all()
873 if (ptes[j].pteh & HPTE_V_VALID) in manual_hpte_clear_all()
875 &(ptes[j].pteh), &(ptes[j].ptel)); in manual_hpte_clear_all()
965 } ptes[4]; in __pSeries_lpar_hpte_find() local
969 lpar_rc = plpar_pte_read_4(0, hpte_group, (void *)ptes); in __pSeries_lpar_hpte_find()
977 if (HPTE_V_COMPARE(ptes[j].pteh, want_v) && in __pSeries_lpar_hpte_find()
978 (ptes[j].pteh & HPTE_V_VALID)) in __pSeries_lpar_hpte_find()
/arch/powerpc/mm/
Dhugetlbpage.c258 void *ptes[]; member
270 kmem_cache_free(PGT_CACHE(PTE_T_ORDER), batch->ptes[i]); in hugepd_free_rcu_callback()
293 (*batchp)->ptes[(*batchp)->index++] = hugepte; in hugepd_free()
/arch/x86/kvm/mmu/
Dpaging_tmpl.h84 pt_element_t ptes[PT_MAX_FULL_LEVELS]; member
213 pte = orig_pte = walker->ptes[level - 1]; in FNAME()
254 walker->ptes[level - 1] = pte; in FNAME()
425 walker->ptes[walker->level - 1] = pte; in FNAME()
581 return r || curr_pte != gw->ptes[level - 1]; in FNAME()