• Home
  • Raw
  • Download

Lines Matching refs:page

52 	struct page **compressed_pages;
120 struct page *page; in check_compressed_csum() local
130 page = cb->compressed_pages[i]; in check_compressed_csum()
133 kaddr = kmap_atomic(page, KM_USER0); in check_compressed_csum()
170 struct page *page; in end_compressed_bio_read() local
204 page = cb->compressed_pages[index]; in end_compressed_bio_read()
205 page->mapping = NULL; in end_compressed_bio_read()
206 page_cache_release(page); in end_compressed_bio_read()
244 struct page *pages[16]; in end_compressed_writeback()
282 struct page *page; in end_compressed_bio_write() local
315 page = cb->compressed_pages[index]; in end_compressed_bio_write()
316 page->mapping = NULL; in end_compressed_bio_write()
317 page_cache_release(page); in end_compressed_bio_write()
339 struct page **compressed_pages, in btrfs_submit_compressed_write()
348 struct page *page; in btrfs_submit_compressed_write() local
376 page = compressed_pages[page_index]; in btrfs_submit_compressed_write()
377 page->mapping = inode->i_mapping; in btrfs_submit_compressed_write()
379 ret = io_tree->ops->merge_bio_hook(page, 0, in btrfs_submit_compressed_write()
385 page->mapping = NULL; in btrfs_submit_compressed_write()
386 if (ret || bio_add_page(bio, page, PAGE_CACHE_SIZE, 0) < in btrfs_submit_compressed_write()
411 bio_add_page(bio, page, PAGE_CACHE_SIZE, 0); in btrfs_submit_compressed_write()
445 struct page *page; in add_ra_bio_pages() local
455 page = cb->orig_bio->bi_io_vec[cb->orig_bio->bi_vcnt - 1].bv_page; in add_ra_bio_pages()
456 last_offset = (page_offset(page) + PAGE_CACHE_SIZE); in add_ra_bio_pages()
473 page = radix_tree_lookup(&mapping->page_tree, page_index); in add_ra_bio_pages()
475 if (page) { in add_ra_bio_pages()
482 page = alloc_page(mapping_gfp_mask(mapping) | GFP_NOFS); in add_ra_bio_pages()
483 if (!page) in add_ra_bio_pages()
486 page->index = page_index; in add_ra_bio_pages()
491 if (add_to_page_cache(page, mapping, in add_ra_bio_pages()
492 page->index, GFP_NOFS)) { in add_ra_bio_pages()
493 page_cache_release(page); in add_ra_bio_pages()
498 page_cache_get(page); in add_ra_bio_pages()
499 if (!pagevec_add(&pvec, page)) in add_ra_bio_pages()
508 set_page_extent_mapped(page); in add_ra_bio_pages()
520 unlock_page(page); in add_ra_bio_pages()
521 page_cache_release(page); in add_ra_bio_pages()
526 if (page->index == end_index) { in add_ra_bio_pages()
533 userpage = kmap_atomic(page, KM_USER0); in add_ra_bio_pages()
535 flush_dcache_page(page); in add_ra_bio_pages()
540 ret = bio_add_page(cb->orig_bio, page, in add_ra_bio_pages()
545 page_cache_release(page); in add_ra_bio_pages()
548 unlock_page(page); in add_ra_bio_pages()
549 page_cache_release(page); in add_ra_bio_pages()
583 struct page *page; in btrfs_submit_compressed_read() local
624 cb->compressed_pages = kmalloc(sizeof(struct page *) * nr_pages, in btrfs_submit_compressed_read()
646 page = cb->compressed_pages[page_index]; in btrfs_submit_compressed_read()
647 page->mapping = inode->i_mapping; in btrfs_submit_compressed_read()
648 page->index = em_start >> PAGE_CACHE_SHIFT; in btrfs_submit_compressed_read()
651 ret = tree->ops->merge_bio_hook(page, 0, in btrfs_submit_compressed_read()
657 page->mapping = NULL; in btrfs_submit_compressed_read()
658 if (ret || bio_add_page(comp_bio, page, PAGE_CACHE_SIZE, 0) < in btrfs_submit_compressed_read()
691 bio_add_page(comp_bio, page, PAGE_CACHE_SIZE, 0); in btrfs_submit_compressed_read()