• Home
  • Raw
  • Download

Lines Matching refs:vm_area_struct

107 	struct vm_area_struct *vma;
122 void hugetlb_dup_vma_private(struct vm_area_struct *vma);
123 void clear_vma_resv_huge_pages(struct vm_area_struct *vma);
124 int move_hugetlb_page_tables(struct vm_area_struct *vma,
125 struct vm_area_struct *new_vma,
129 struct vm_area_struct *, struct vm_area_struct *);
130 void unmap_hugepage_range(struct vm_area_struct *,
134 struct vm_area_struct *vma,
141 vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
145 struct vm_area_struct *dst_vma,
152 struct vm_area_struct *vma,
166 pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
169 struct vm_area_struct *vma,
196 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
238 int huge_pmd_unshare(struct mm_struct *mm, struct vm_area_struct *vma,
240 void adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma,
243 extern void __hugetlb_zap_begin(struct vm_area_struct *vma,
245 extern void __hugetlb_zap_end(struct vm_area_struct *vma,
248 static inline void hugetlb_zap_begin(struct vm_area_struct *vma, in hugetlb_zap_begin()
255 static inline void hugetlb_zap_end(struct vm_area_struct *vma, in hugetlb_zap_end()
262 void hugetlb_vma_lock_read(struct vm_area_struct *vma);
263 void hugetlb_vma_unlock_read(struct vm_area_struct *vma);
264 void hugetlb_vma_lock_write(struct vm_area_struct *vma);
265 void hugetlb_vma_unlock_write(struct vm_area_struct *vma);
266 int hugetlb_vma_trylock_write(struct vm_area_struct *vma);
267 void hugetlb_vma_assert_locked(struct vm_area_struct *vma);
269 long hugetlb_change_protection(struct vm_area_struct *vma,
274 void hugetlb_unshare_all_pmds(struct vm_area_struct *vma);
275 void hugetlb_split(struct vm_area_struct *vma, unsigned long addr);
279 static inline void hugetlb_dup_vma_private(struct vm_area_struct *vma) in hugetlb_dup_vma_private()
283 static inline void clear_vma_resv_huge_pages(struct vm_area_struct *vma) in clear_vma_resv_huge_pages()
299 struct vm_area_struct *vma, in huge_pmd_unshare()
306 struct vm_area_struct *vma, in adjust_range_if_pmd_sharing_possible()
312 struct vm_area_struct *vma, in hugetlb_zap_begin()
318 struct vm_area_struct *vma, in hugetlb_zap_end()
325 struct vm_area_struct *dst_vma, in copy_hugetlb_page_range()
326 struct vm_area_struct *src_vma) in copy_hugetlb_page_range()
332 static inline int move_hugetlb_page_tables(struct vm_area_struct *vma, in move_hugetlb_page_tables()
333 struct vm_area_struct *new_vma, in move_hugetlb_page_tables()
361 static inline void hugetlb_vma_lock_read(struct vm_area_struct *vma) in hugetlb_vma_lock_read()
365 static inline void hugetlb_vma_unlock_read(struct vm_area_struct *vma) in hugetlb_vma_unlock_read()
369 static inline void hugetlb_vma_lock_write(struct vm_area_struct *vma) in hugetlb_vma_lock_write()
373 static inline void hugetlb_vma_unlock_write(struct vm_area_struct *vma) in hugetlb_vma_unlock_write()
377 static inline int hugetlb_vma_trylock_write(struct vm_area_struct *vma) in hugetlb_vma_trylock_write()
382 static inline void hugetlb_vma_assert_locked(struct vm_area_struct *vma) in hugetlb_vma_assert_locked()
401 struct vm_area_struct *dst_vma, in hugetlb_mfill_atomic_pte()
444 struct vm_area_struct *vma, unsigned long address, in hugetlb_change_protection()
452 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range()
460 struct vm_area_struct *vma, unsigned long address, in hugetlb_fault()
467 static inline void hugetlb_unshare_all_pmds(struct vm_area_struct *vma) { } in hugetlb_unshare_all_pmds()
469 static inline void hugetlb_split(struct vm_area_struct *vma, unsigned long addr) {} in hugetlb_split()
694 struct folio *alloc_hugetlb_folio(struct vm_area_struct *vma,
704 void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
752 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma()
762 extern unsigned long vma_kernel_pagesize(struct vm_area_struct *vma);
764 extern unsigned long vma_mmu_pagesize(struct vm_area_struct *vma);
1013 static inline pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma, in huge_ptep_modify_prot_start()
1024 static inline void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, in huge_ptep_modify_prot_commit()
1068 static inline struct folio *alloc_hugetlb_folio(struct vm_area_struct *vma,
1105 static inline struct hstate *hstate_vma(struct vm_area_struct *vma)
1130 static inline unsigned long vma_kernel_pagesize(struct vm_area_struct *vma)
1135 static inline unsigned long vma_mmu_pagesize(struct vm_area_struct *vma)
1224 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma,
1248 struct hstate *h, struct vm_area_struct *vma, unsigned long address)
1284 bool want_pmd_share(struct vm_area_struct *vma, unsigned long addr);
1294 static inline bool __vma_shareable_lock(struct vm_area_struct *vma) in __vma_shareable_lock()
1299 bool __vma_private_lock(struct vm_area_struct *vma);
1306 hugetlb_walk(struct vm_area_struct *vma, unsigned long addr, unsigned long sz) in hugetlb_walk()