• Home
  • Raw
  • Download

Lines Matching refs:page_size

16 	return hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size,  in hl_is_dram_va()
145 int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, in hl_mmu_unmap_page() argument
163 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_unmap_page()
173 if ((page_size % mmu_prop->page_size) == 0) { in hl_mmu_unmap_page()
174 real_page_size = mmu_prop->page_size; in hl_mmu_unmap_page()
182 if (is_dram_addr && ((page_size % prop->dram_page_size) == 0)) { in hl_mmu_unmap_page()
187 page_size, mmu_prop->page_size >> 10); in hl_mmu_unmap_page()
193 npages = page_size / real_page_size; in hl_mmu_unmap_page()
234 u32 page_size, bool flush_pte) in hl_mmu_map_page() argument
252 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_map_page()
263 if ((page_size % mmu_prop->page_size) == 0) { in hl_mmu_map_page()
264 real_page_size = mmu_prop->page_size; in hl_mmu_map_page()
265 } else if (is_dram_addr && ((page_size % prop->dram_page_size) == 0) && in hl_mmu_map_page()
266 (prop->dram_page_size < mmu_prop->page_size)) { in hl_mmu_map_page()
279 page_size, mmu_prop->page_size >> 10); in hl_mmu_map_page()
291 (mmu_prop->page_size - 1)) || in hl_mmu_map_page()
293 (mmu_prop->page_size - 1)))) || in hl_mmu_map_page()
300 npages = page_size / real_page_size; in hl_mmu_map_page()
353 u32 page_size; in hl_mmu_map_contiguous() local
359 page_size = prop->dmmu.page_size; in hl_mmu_map_contiguous()
362 page_size = prop->pmmu.page_size; in hl_mmu_map_contiguous()
365 page_size = prop->pmmu_huge.page_size; in hl_mmu_map_contiguous()
369 for (off = 0 ; off < size ; off += page_size) { in hl_mmu_map_contiguous()
372 flush_pte = (off + page_size) >= size; in hl_mmu_map_contiguous()
373 rc = hl_mmu_map_page(ctx, curr_va, curr_pa, page_size, in hl_mmu_map_contiguous()
386 for (; off >= 0 ; off -= page_size) { in hl_mmu_map_contiguous()
388 flush_pte = (off - (s32) page_size) < 0; in hl_mmu_map_contiguous()
389 if (hl_mmu_unmap_page(ctx, curr_va, page_size, flush_pte)) in hl_mmu_map_contiguous()
411 u32 page_size; in hl_mmu_unmap_contiguous() local
417 page_size = prop->dmmu.page_size; in hl_mmu_unmap_contiguous()
420 page_size = prop->pmmu.page_size; in hl_mmu_unmap_contiguous()
423 page_size = prop->pmmu_huge.page_size; in hl_mmu_unmap_contiguous()
427 for (off = 0 ; off < size ; off += page_size) { in hl_mmu_unmap_contiguous()
429 flush_pte = (off + page_size) >= size; in hl_mmu_unmap_contiguous()
430 rc = hl_mmu_unmap_page(ctx, curr_va, page_size, flush_pte); in hl_mmu_unmap_contiguous()
568 is_dram_addr = hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size, in hl_mmu_get_tlb_info()