• Home
  • Raw
  • Download

Lines Matching refs:mapping

964 		struct address_space *mapping = vma->vm_file->f_mapping;  in vma_resv_map()  local
965 struct inode *inode = mapping->host; in vma_resv_map()
1420 return (unsigned long)page[2].mapping == -1U; in PageHugeTemporary()
1425 page[2].mapping = (void *)-1U; in SetPageHugeTemporary()
1430 page[2].mapping = NULL; in ClearPageHugeTemporary()
1449 page->mapping = NULL; in __free_huge_page()
1519 struct page, mapping); in free_hpage_workfn()
1537 if (llist_add((struct llist_node *)&page->mapping, in free_huge_page()
1626 struct address_space *mapping = page_mapping(hpage); in hugetlb_page_mapping_lock_write() local
1628 if (!mapping) in hugetlb_page_mapping_lock_write()
1629 return mapping; in hugetlb_page_mapping_lock_write()
1631 if (i_mmap_trylock_write(mapping)) in hugetlb_page_mapping_lock_write()
1632 return mapping; in hugetlb_page_mapping_lock_write()
3824 struct address_space *mapping = vma->vm_file->f_mapping; in copy_hugetlb_page_range() local
3842 i_mmap_lock_read(mapping); in copy_hugetlb_page_range()
3921 i_mmap_unlock_read(mapping); in copy_hugetlb_page_range()
4098 struct address_space *mapping; in unmap_ref_private() local
4108 mapping = vma->vm_file->f_mapping; in unmap_ref_private()
4115 i_mmap_lock_write(mapping); in unmap_ref_private()
4116 vma_interval_tree_foreach(iter_vma, &mapping->i_mmap, pgoff, pgoff) { in unmap_ref_private()
4140 i_mmap_unlock_write(mapping); in unmap_ref_private()
4204 struct address_space *mapping = vma->vm_file->f_mapping; in hugetlb_cow() local
4220 hash = hugetlb_fault_mutex_hash(mapping, idx); in hugetlb_cow()
4222 i_mmap_unlock_read(mapping); in hugetlb_cow()
4226 i_mmap_lock_read(mapping); in hugetlb_cow()
4297 struct address_space *mapping; in hugetlbfs_pagecache_page() local
4300 mapping = vma->vm_file->f_mapping; in hugetlbfs_pagecache_page()
4303 return find_lock_page(mapping, idx); in hugetlbfs_pagecache_page()
4313 struct address_space *mapping; in hugetlbfs_pagecache_present() local
4317 mapping = vma->vm_file->f_mapping; in hugetlbfs_pagecache_present()
4320 page = find_get_page(mapping, idx); in hugetlbfs_pagecache_present()
4326 int huge_add_to_page_cache(struct page *page, struct address_space *mapping, in huge_add_to_page_cache() argument
4329 struct inode *inode = mapping->host; in huge_add_to_page_cache()
4331 int err = add_to_page_cache(page, mapping, idx, GFP_KERNEL); in huge_add_to_page_cache()
4350 struct address_space *mapping, in hugetlb_handle_userfault() argument
4356 u32 hash = hugetlb_fault_mutex_hash(mapping, idx); in hugetlb_handle_userfault()
4377 i_mmap_unlock_read(mapping); in hugetlb_handle_userfault()
4383 struct address_space *mapping, pgoff_t idx, in hugetlb_no_page() argument
4395 u32 hash = hugetlb_fault_mutex_hash(mapping, idx); in hugetlb_no_page()
4413 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_no_page()
4418 page = find_lock_page(mapping, idx); in hugetlb_no_page()
4422 ret = hugetlb_handle_userfault(vma, mapping, idx, in hugetlb_no_page()
4457 int err = huge_add_to_page_cache(page, mapping, idx); in hugetlb_no_page()
4488 ret = hugetlb_handle_userfault(vma, mapping, idx, in hugetlb_no_page()
4543 i_mmap_unlock_read(mapping); in hugetlb_no_page()
4556 u32 hugetlb_fault_mutex_hash(struct address_space *mapping, pgoff_t idx) in hugetlb_fault_mutex_hash() argument
4561 key[0] = (unsigned long) mapping; in hugetlb_fault_mutex_hash()
4573 u32 hugetlb_fault_mutex_hash(struct address_space *mapping, pgoff_t idx) in hugetlb_fault_mutex_hash() argument
4590 struct address_space *mapping; in hugetlb_fault() local
4621 mapping = vma->vm_file->f_mapping; in hugetlb_fault()
4622 i_mmap_lock_read(mapping); in hugetlb_fault()
4625 i_mmap_unlock_read(mapping); in hugetlb_fault()
4635 hash = hugetlb_fault_mutex_hash(mapping, idx); in hugetlb_fault()
4644 return hugetlb_no_page(mm, vma, mapping, idx, address, ptep, flags); in hugetlb_fault()
4724 i_mmap_unlock_read(mapping); in hugetlb_fault()
4751 struct address_space *mapping; in hugetlb_mcopy_atomic_pte() local
4762 mapping = dst_vma->vm_file->f_mapping; in hugetlb_mcopy_atomic_pte()
4767 page = find_lock_page(mapping, idx); in hugetlb_mcopy_atomic_pte()
4811 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_mcopy_atomic_pte()
4822 ret = huge_add_to_page_cache(page, mapping, idx); in hugetlb_mcopy_atomic_pte()
4839 size = i_size_read(mapping->host) >> huge_page_shift(h); in hugetlb_mcopy_atomic_pte()
5449 struct address_space *mapping = vma->vm_file->f_mapping; in huge_pmd_share() local
5458 i_mmap_assert_locked(mapping); in huge_pmd_share()
5459 vma_interval_tree_foreach(svma, &mapping->i_mmap, idx, idx) { in huge_pmd_share()