Lines Matching refs:end
472 const u64 end = start + len - 1; in btrfs_find_new_delalloc_bytes() local
474 while (search_start < end) { in btrfs_find_new_delalloc_bytes()
475 const u64 search_len = end - search_start + 1; in btrfs_find_new_delalloc_bytes()
587 void btrfs_drop_extent_cache(struct btrfs_inode *inode, u64 start, u64 end, in btrfs_drop_extent_cache() argument
594 u64 len = end - start + 1; in btrfs_drop_extent_cache()
602 WARN_ON(end < start); in btrfs_drop_extent_cache()
603 if (end == (u64)-1) { in btrfs_drop_extent_cache()
749 struct btrfs_path *path, u64 start, u64 end, in __btrfs_drop_extents() argument
779 btrfs_drop_extent_cache(BTRFS_I(inode), start, end - 1, 0); in __btrfs_drop_extents()
827 if (key.type > BTRFS_EXTENT_DATA_KEY || key.offset >= end) in __btrfs_drop_extents()
880 if (start > key.offset && end < extent_end) { in __btrfs_drop_extents()
936 if (start <= key.offset && end < extent_end) { in __btrfs_drop_extents()
943 new_key.offset = end; in __btrfs_drop_extents()
946 extent_offset += end - key.offset; in __btrfs_drop_extents()
949 extent_end - end); in __btrfs_drop_extents()
952 inode_sub_bytes(inode, end - key.offset); in __btrfs_drop_extents()
961 if (start > key.offset && end >= extent_end) { in __btrfs_drop_extents()
973 if (end == extent_end) in __btrfs_drop_extents()
984 if (start <= key.offset && end >= extent_end) { in __btrfs_drop_extents()
1014 if (end == extent_end) in __btrfs_drop_extents()
1085 *drop_end = found ? min(end, last_end) : end; in __btrfs_drop_extents()
1091 u64 end, int drop_cache) in btrfs_drop_extents() argument
1099 ret = __btrfs_drop_extents(trans, root, inode, path, start, end, NULL, in btrfs_drop_extents()
1107 u64 *start, u64 *end) in extent_mergeable() argument
1130 if ((*start && *start != key.offset) || (*end && *end != extent_end)) in extent_mergeable()
1134 *end = extent_end; in extent_mergeable()
1146 struct btrfs_inode *inode, u64 start, u64 end) in btrfs_mark_extent_written() argument
1201 if (key.offset > start || extent_end < end) { in btrfs_mark_extent_written()
1212 if (start == key.offset && end < extent_end) { in btrfs_mark_extent_written()
1218 new_key.offset = end; in btrfs_mark_extent_written()
1225 extent_end - end); in btrfs_mark_extent_written()
1227 end - orig_offset); in btrfs_mark_extent_written()
1233 end - other_start); in btrfs_mark_extent_written()
1239 if (start > key.offset && end == extent_end) { in btrfs_mark_extent_written()
1240 other_start = end; in btrfs_mark_extent_written()
1268 while (start > key.offset || end < extent_end) { in btrfs_mark_extent_written()
1270 split = end; in btrfs_mark_extent_written()
1318 extent_end = end; in btrfs_mark_extent_written()
1323 other_start = end; in btrfs_mark_extent_written()
2031 static int start_ordered_ops(struct inode *inode, loff_t start, loff_t end) in start_ordered_ops() argument
2044 ret = btrfs_fdatawrite_range(inode, start, end); in start_ordered_ops()
2062 int btrfs_sync_file(struct file *file, loff_t start, loff_t end, int datasync) in btrfs_sync_file() argument
2082 ret = start_ordered_ops(inode, start, end); in btrfs_sync_file()
2107 end = LLONG_MAX; in btrfs_sync_file()
2128 ret = start_ordered_ops(inode, start, end); in btrfs_sync_file()
2141 ret = btrfs_wait_ordered_range(inode, start, (u64)end - (u64)start + 1); in btrfs_sync_file()
2191 ret = btrfs_log_dentry_safe(trans, dentry, start, end, &ctx); in btrfs_sync_file()
2250 int slot, u64 start, u64 end) in hole_mergeable() argument
2271 if (key.offset == end) in hole_mergeable()
2280 struct btrfs_path *path, u64 offset, u64 end) in fill_holes() argument
2310 if (hole_mergeable(inode, leaf, path->slots[0] - 1, offset, end)) { in fill_holes()
2317 end - offset; in fill_holes()
2325 if (hole_mergeable(inode, leaf, path->slots[0], offset, end)) { in fill_holes()
2332 num_bytes = btrfs_file_extent_num_bytes(leaf, fi) + end - in fill_holes()
2343 offset, 0, 0, end - offset, 0, end - offset, 0, 0, 0); in fill_holes()
2352 btrfs_drop_extent_cache(inode, offset, end - 1, 0); in fill_holes()
2356 hole_em->len = end - offset; in fill_holes()
2368 btrfs_drop_extent_cache(inode, offset, end - 1, 0); in fill_holes()
2519 const u64 start, const u64 end, in btrfs_punch_hole_range() argument
2532 u64 len = end - start; in btrfs_punch_hole_range()
2535 if (end <= start) in btrfs_punch_hole_range()
2570 while (cur_offset < end) { in btrfs_punch_hole_range()
2572 cur_offset, end + 1, &drop_end, in btrfs_punch_hole_range()
2678 if (drop_end <= end) in btrfs_punch_hole_range()
2679 drop_end = end + 1; in btrfs_punch_hole_range()
2925 const u64 end, in btrfs_fallocate_update_isize() argument
2933 if (mode & FALLOC_FL_KEEP_SIZE || end <= i_size_read(inode)) in btrfs_fallocate_update_isize()
2941 i_size_write(inode, end); in btrfs_fallocate_update_isize()
2942 btrfs_ordered_update_i_size(inode, end, NULL); in btrfs_fallocate_update_isize()
3490 int btrfs_fdatawrite_range(struct inode *inode, loff_t start, loff_t end) in btrfs_fdatawrite_range() argument
3508 ret = filemap_fdatawrite_range(inode->i_mapping, start, end); in btrfs_fdatawrite_range()
3511 ret = filemap_fdatawrite_range(inode->i_mapping, start, end); in btrfs_fdatawrite_range()