• Home
  • Raw
  • Download

Lines Matching refs:mapping

757 		struct address_space *mapping = vma->vm_file->f_mapping;  in vma_resv_map()  local
758 struct inode *inode = mapping->host; in vma_resv_map()
1266 page->mapping = NULL; in free_huge_page()
3451 struct address_space *mapping; in unmap_ref_private() local
3461 mapping = vma->vm_file->f_mapping; in unmap_ref_private()
3468 i_mmap_lock_write(mapping); in unmap_ref_private()
3469 vma_interval_tree_foreach(iter_vma, &mapping->i_mmap, pgoff, pgoff) { in unmap_ref_private()
3493 i_mmap_unlock_write(mapping); in unmap_ref_private()
3632 struct address_space *mapping; in hugetlbfs_pagecache_page() local
3635 mapping = vma->vm_file->f_mapping; in hugetlbfs_pagecache_page()
3638 return find_lock_page(mapping, idx); in hugetlbfs_pagecache_page()
3648 struct address_space *mapping; in hugetlbfs_pagecache_present() local
3652 mapping = vma->vm_file->f_mapping; in hugetlbfs_pagecache_present()
3655 page = find_get_page(mapping, idx); in hugetlbfs_pagecache_present()
3661 int huge_add_to_page_cache(struct page *page, struct address_space *mapping, in huge_add_to_page_cache() argument
3664 struct inode *inode = mapping->host; in huge_add_to_page_cache()
3666 int err = add_to_page_cache(page, mapping, idx, GFP_KERNEL); in huge_add_to_page_cache()
3685 struct address_space *mapping, pgoff_t idx, in hugetlb_no_page() argument
3713 page = find_lock_page(mapping, idx); in hugetlb_no_page()
3715 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_no_page()
3742 hash = hugetlb_fault_mutex_hash(h, mapping, idx, in hugetlb_no_page()
3764 int err = huge_add_to_page_cache(page, mapping, idx); in hugetlb_no_page()
3808 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_no_page()
3855 u32 hugetlb_fault_mutex_hash(struct hstate *h, struct address_space *mapping, in hugetlb_fault_mutex_hash() argument
3861 key[0] = (unsigned long) mapping; in hugetlb_fault_mutex_hash()
3873 u32 hugetlb_fault_mutex_hash(struct hstate *h, struct address_space *mapping, in hugetlb_fault_mutex_hash() argument
3891 struct address_space *mapping; in hugetlb_fault() local
3911 mapping = vma->vm_file->f_mapping; in hugetlb_fault()
3919 hash = hugetlb_fault_mutex_hash(h, mapping, idx, address); in hugetlb_fault()
3924 ret = hugetlb_no_page(mm, vma, mapping, idx, address, ptep, flags); in hugetlb_fault()
4029 struct address_space *mapping; in hugetlb_mcopy_atomic_pte() local
4068 mapping = dst_vma->vm_file->f_mapping; in hugetlb_mcopy_atomic_pte()
4075 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_mcopy_atomic_pte()
4086 ret = huge_add_to_page_cache(page, mapping, idx); in hugetlb_mcopy_atomic_pte()
4103 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_mcopy_atomic_pte()
4611 struct address_space *mapping = vma->vm_file->f_mapping; in huge_pmd_share() local
4623 i_mmap_lock_write(mapping); in huge_pmd_share()
4624 vma_interval_tree_foreach(svma, &mapping->i_mmap, idx, idx) { in huge_pmd_share()
4653 i_mmap_unlock_write(mapping); in huge_pmd_share()