• Home
  • Raw
  • Download

Lines Matching full:ordered

1395  * This function locks the extent and properly waits for data=ordered extents
1421 struct btrfs_ordered_extent *ordered; in lock_and_cleanup_extent_if_need() local
1425 ordered = btrfs_lookup_ordered_range(inode, start_pos, in lock_and_cleanup_extent_if_need()
1427 if (ordered && in lock_and_cleanup_extent_if_need()
1428 ordered->file_offset + ordered->num_bytes > start_pos && in lock_and_cleanup_extent_if_need()
1429 ordered->file_offset <= last_pos) { in lock_and_cleanup_extent_if_need()
1436 btrfs_start_ordered_extent(ordered, 1); in lock_and_cleanup_extent_if_need()
1437 btrfs_put_ordered_extent(ordered); in lock_and_cleanup_extent_if_need()
1440 if (ordered) in lock_and_cleanup_extent_if_need()
1441 btrfs_put_ordered_extent(ordered); in lock_and_cleanup_extent_if_need()
1491 struct btrfs_ordered_extent *ordered; in check_can_nocow() local
1496 ordered = btrfs_lookup_ordered_range(inode, lockstart, in check_can_nocow()
1498 if (ordered) { in check_can_nocow()
1499 btrfs_put_ordered_extent(ordered); in check_can_nocow()
1537 * This function will flush ordered extents in the range to ensure proper
1543 * -EAGAIN if we can't get the needed lock or there are ordered extents
2072 * update the last_trans of the inode during ordered extent completion, in skip_inode_logging()
2088 * It needs to call filemap_fdatawait so that all ordered extent updates are
2114 * races between hole detection during logging and completion of ordered in btrfs_sync_file()
2115 * extents outside the range, to missing checksums due to ordered extents in btrfs_sync_file()
2167 * wait for all ordered extents to complete below. in btrfs_sync_file()
2180 * For a full fsync we wait for the ordered extents to complete while in btrfs_sync_file()
2182 * attach the ordered extents to the transaction so that a transaction in btrfs_sync_file()
2184 * the current transaction commits before the ordered extents complete in btrfs_sync_file()
2191 * Get our ordered extents as soon as possible to avoid doing in btrfs_sync_file()
2193 * checksums attached to the ordered extents. in btrfs_sync_file()
2215 * An ordered extent might have started before and completed in btrfs_sync_file()
2481 struct btrfs_ordered_extent *ordered; in btrfs_punch_hole_lock_range() local
2488 ordered = btrfs_lookup_first_ordered_extent(BTRFS_I(inode), in btrfs_punch_hole_lock_range()
2492 * We need to make sure we have no ordered extents in this range in btrfs_punch_hole_lock_range()
2496 if ((!ordered || in btrfs_punch_hole_lock_range()
2497 (ordered->file_offset + ordered->num_bytes <= lockstart || in btrfs_punch_hole_lock_range()
2498 ordered->file_offset > lockend)) && in btrfs_punch_hole_lock_range()
2501 if (ordered) in btrfs_punch_hole_lock_range()
2502 btrfs_put_ordered_extent(ordered); in btrfs_punch_hole_lock_range()
2505 if (ordered) in btrfs_punch_hole_lock_range()
2506 btrfs_put_ordered_extent(ordered); in btrfs_punch_hole_lock_range()
3359 * wait for ordered IO before we have any locks. We'll loop again in btrfs_fallocate()
3375 struct btrfs_ordered_extent *ordered; in btrfs_fallocate() local
3377 /* the extent lock is ordered inside the running in btrfs_fallocate()
3382 ordered = btrfs_lookup_first_ordered_extent(BTRFS_I(inode), in btrfs_fallocate()
3385 if (ordered && in btrfs_fallocate()
3386 ordered->file_offset + ordered->num_bytes > alloc_start && in btrfs_fallocate()
3387 ordered->file_offset < alloc_end) { in btrfs_fallocate()
3388 btrfs_put_ordered_extent(ordered); in btrfs_fallocate()
3401 if (ordered) in btrfs_fallocate()
3402 btrfs_put_ordered_extent(ordered); in btrfs_fallocate()
3650 * from there. We have to do this otherwise we'll miss the ordered in btrfs_fdatawrite_range()