Lines Matching refs:mapping
760 struct address_space *mapping = vma->vm_file->f_mapping; in vma_resv_map() local
761 struct inode *inode = mapping->host; in vma_resv_map()
1246 return (unsigned long)page[2].mapping == -1U; in PageHugeTemporary()
1251 page[2].mapping = (void *)-1U; in SetPageHugeTemporary()
1256 page[2].mapping = NULL; in ClearPageHugeTemporary()
1275 page->mapping = NULL; in __free_huge_page()
1343 struct page, mapping); in free_hpage_workfn()
1361 if (llist_add((struct llist_node *)&page->mapping, in free_huge_page()
3676 struct address_space *mapping; in unmap_ref_private() local
3686 mapping = vma->vm_file->f_mapping; in unmap_ref_private()
3693 i_mmap_lock_write(mapping); in unmap_ref_private()
3694 vma_interval_tree_foreach(iter_vma, &mapping->i_mmap, pgoff, pgoff) { in unmap_ref_private()
3718 i_mmap_unlock_write(mapping); in unmap_ref_private()
3855 struct address_space *mapping; in hugetlbfs_pagecache_page() local
3858 mapping = vma->vm_file->f_mapping; in hugetlbfs_pagecache_page()
3861 return find_lock_page(mapping, idx); in hugetlbfs_pagecache_page()
3871 struct address_space *mapping; in hugetlbfs_pagecache_present() local
3875 mapping = vma->vm_file->f_mapping; in hugetlbfs_pagecache_present()
3878 page = find_get_page(mapping, idx); in hugetlbfs_pagecache_present()
3884 int huge_add_to_page_cache(struct page *page, struct address_space *mapping, in huge_add_to_page_cache() argument
3887 struct inode *inode = mapping->host; in huge_add_to_page_cache()
3889 int err = add_to_page_cache(page, mapping, idx, GFP_KERNEL); in huge_add_to_page_cache()
3909 struct address_space *mapping, pgoff_t idx, in hugetlb_no_page() argument
3938 page = find_lock_page(mapping, idx); in hugetlb_no_page()
3940 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_no_page()
3967 hash = hugetlb_fault_mutex_hash(h, mapping, idx, haddr); in hugetlb_no_page()
4003 int err = huge_add_to_page_cache(page, mapping, idx); in hugetlb_no_page()
4047 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_no_page()
4094 u32 hugetlb_fault_mutex_hash(struct hstate *h, struct address_space *mapping, in hugetlb_fault_mutex_hash() argument
4100 key[0] = (unsigned long) mapping; in hugetlb_fault_mutex_hash()
4112 u32 hugetlb_fault_mutex_hash(struct hstate *h, struct address_space *mapping, in hugetlb_fault_mutex_hash() argument
4130 struct address_space *mapping; in hugetlb_fault() local
4149 mapping = vma->vm_file->f_mapping; in hugetlb_fault()
4157 hash = hugetlb_fault_mutex_hash(h, mapping, idx, haddr); in hugetlb_fault()
4162 ret = hugetlb_no_page(mm, vma, mapping, idx, address, ptep, flags); in hugetlb_fault()
4267 struct address_space *mapping; in hugetlb_mcopy_atomic_pte() local
4306 mapping = dst_vma->vm_file->f_mapping; in hugetlb_mcopy_atomic_pte()
4313 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_mcopy_atomic_pte()
4324 ret = huge_add_to_page_cache(page, mapping, idx); in hugetlb_mcopy_atomic_pte()
4341 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_mcopy_atomic_pte()
4882 struct address_space *mapping = vma->vm_file->f_mapping; in huge_pmd_share() local
4894 i_mmap_lock_write(mapping); in huge_pmd_share()
4895 vma_interval_tree_foreach(svma, &mapping->i_mmap, idx, idx) { in huge_pmd_share()
4924 i_mmap_unlock_write(mapping); in huge_pmd_share()