Lines Matching refs:start
310 range.start = defrag->last_offset; in __btrfs_run_defrag_inode()
322 defrag->last_offset = range.start; in __btrfs_run_defrag_inode()
467 const u64 start, in btrfs_find_new_delalloc_bytes() argument
471 u64 search_start = start; in btrfs_find_new_delalloc_bytes()
472 const u64 end = start + len - 1; in btrfs_find_new_delalloc_bytes()
489 if (em->start < search_start) in btrfs_find_new_delalloc_bytes()
490 em_len -= search_start - em->start; in btrfs_find_new_delalloc_bytes()
587 void btrfs_drop_extent_cache(struct btrfs_inode *inode, u64 start, u64 end, in btrfs_drop_extent_cache() argument
594 u64 len = end - start + 1; in btrfs_drop_extent_cache()
602 WARN_ON(end < start); in btrfs_drop_extent_cache()
619 em = lookup_extent_mapping(em_tree, start, len); in btrfs_drop_extent_cache()
627 if (testend && em->start + em->len >= start + len) { in btrfs_drop_extent_cache()
632 start = em->start + em->len; in btrfs_drop_extent_cache()
634 len = start + len - (em->start + em->len); in btrfs_drop_extent_cache()
646 if (em->start < start) { in btrfs_drop_extent_cache()
647 split->start = em->start; in btrfs_drop_extent_cache()
648 split->len = start - em->start; in btrfs_drop_extent_cache()
662 split->orig_start = split->start; in btrfs_drop_extent_cache()
678 if (testend && em->start + em->len > start + len) { in btrfs_drop_extent_cache()
679 u64 diff = start + len - em->start; in btrfs_drop_extent_cache()
681 split->start = start + len; in btrfs_drop_extent_cache()
682 split->len = em->start + em->len - (start + len); in btrfs_drop_extent_cache()
705 split->orig_start = split->start; in btrfs_drop_extent_cache()
749 struct btrfs_path *path, u64 start, u64 end, in __btrfs_drop_extents() argument
762 u64 search_start = start; in __btrfs_drop_extents()
767 u64 last_end = start; in __btrfs_drop_extents()
779 btrfs_drop_extent_cache(BTRFS_I(inode), start, end - 1, 0); in __btrfs_drop_extents()
781 if (start >= BTRFS_I(inode)->disk_i_size && !replace_extent) in __btrfs_drop_extents()
792 if (ret > 0 && path->slots[0] > 0 && search_start == start) { in __btrfs_drop_extents()
869 search_start = max(key.offset, start); in __btrfs_drop_extents()
880 if (start > key.offset && end < extent_end) { in __btrfs_drop_extents()
888 new_key.offset = start; in __btrfs_drop_extents()
902 start - key.offset); in __btrfs_drop_extents()
907 extent_offset += start - key.offset; in __btrfs_drop_extents()
910 extent_end - start); in __btrfs_drop_extents()
920 start - extent_offset); in __btrfs_drop_extents()
924 key.offset = start; in __btrfs_drop_extents()
936 if (start <= key.offset && end < extent_end) { in __btrfs_drop_extents()
961 if (start > key.offset && end >= extent_end) { in __btrfs_drop_extents()
969 start - key.offset); in __btrfs_drop_extents()
972 inode_sub_bytes(inode, extent_end - start); in __btrfs_drop_extents()
984 if (start <= key.offset && end >= extent_end) { in __btrfs_drop_extents()
1066 key.offset = start; in __btrfs_drop_extents()
1090 struct btrfs_root *root, struct inode *inode, u64 start, in btrfs_drop_extents() argument
1099 ret = __btrfs_drop_extents(trans, root, inode, path, start, end, NULL, in btrfs_drop_extents()
1107 u64 *start, u64 *end) in extent_mergeable() argument
1130 if ((*start && *start != key.offset) || (*end && *end != extent_end)) in extent_mergeable()
1133 *start = key.offset; in extent_mergeable()
1146 struct btrfs_inode *inode, u64 start, u64 end) in btrfs_mark_extent_written() argument
1174 split = start; in btrfs_mark_extent_written()
1201 if (key.offset > start || extent_end < end) { in btrfs_mark_extent_written()
1212 if (start == key.offset && end < extent_end) { in btrfs_mark_extent_written()
1214 other_end = start; in btrfs_mark_extent_written()
1239 if (start > key.offset && end == extent_end) { in btrfs_mark_extent_written()
1248 start - key.offset); in btrfs_mark_extent_written()
1252 new_key.offset = start; in btrfs_mark_extent_written()
1260 other_end - start); in btrfs_mark_extent_written()
1262 start - orig_offset); in btrfs_mark_extent_written()
1268 while (start > key.offset || end < extent_end) { in btrfs_mark_extent_written()
1269 if (key.offset == start) in btrfs_mark_extent_written()
1309 if (split == start) { in btrfs_mark_extent_written()
1310 key.offset = start; in btrfs_mark_extent_written()
1312 if (start != key.offset) { in btrfs_mark_extent_written()
1345 other_end = start; in btrfs_mark_extent_written()
2031 static int start_ordered_ops(struct inode *inode, loff_t start, loff_t end) in start_ordered_ops() argument
2044 ret = btrfs_fdatawrite_range(inode, start, end); in start_ordered_ops()
2062 int btrfs_sync_file(struct file *file, loff_t start, loff_t end, int datasync) in btrfs_sync_file() argument
2082 ret = start_ordered_ops(inode, start, end); in btrfs_sync_file()
2106 start = 0; in btrfs_sync_file()
2128 ret = start_ordered_ops(inode, start, end); in btrfs_sync_file()
2141 ret = btrfs_wait_ordered_range(inode, start, (u64)end - (u64)start + 1); in btrfs_sync_file()
2191 ret = btrfs_log_dentry_safe(trans, dentry, start, end, &ctx); in btrfs_sync_file()
2250 int slot, u64 start, u64 end) in hole_mergeable() argument
2273 if (key.offset + btrfs_file_extent_num_bytes(leaf, fi) == start) in hole_mergeable()
2355 hole_em->start = offset; in fill_holes()
2388 static int find_first_non_hole(struct inode *inode, u64 *start, u64 *len) in find_first_non_hole() argument
2395 round_down(*start, fs_info->sectorsize), in find_first_non_hole()
2403 *len = em->start + em->len > *start + *len ? in find_first_non_hole()
2404 0 : *start + *len - em->start - em->len; in find_first_non_hole()
2405 *start = em->start + em->len; in find_first_non_hole()
2519 const u64 start, const u64 end, in btrfs_punch_hole_range() argument
2532 u64 len = end - start; in btrfs_punch_hole_range()
2535 if (end <= start) in btrfs_punch_hole_range()
2569 cur_offset = start; in btrfs_punch_hole_range()
2887 u64 start; member
2897 static int add_falloc_range(struct list_head *head, u64 start, u64 len) in add_falloc_range() argument
2910 if (prev->start + prev->len == start) { in add_falloc_range()
2918 range->start = start; in add_falloc_range()
3011 if (em->start <= alloc_start && in btrfs_zero_range()
3013 const u64 em_end = em->start + em->len; in btrfs_zero_range()
3323 range->start, in btrfs_fallocate()
3328 data_reserved, range->start, in btrfs_fallocate()
3361 u64 start; in find_desired_extent() local
3372 start = max_t(loff_t, 0, *offset); in find_desired_extent()
3374 lockstart = round_down(start, fs_info->sectorsize); in find_desired_extent()
3385 while (start < inode->i_size) { in find_desired_extent()
3386 em = btrfs_get_extent_fiemap(BTRFS_I(inode), start, len); in find_desired_extent()
3402 start = em->start + em->len; in find_desired_extent()
3409 if (whence == SEEK_DATA && start >= inode->i_size) in find_desired_extent()
3412 *offset = min_t(loff_t, start, inode->i_size); in find_desired_extent()
3490 int btrfs_fdatawrite_range(struct inode *inode, loff_t start, loff_t end) in btrfs_fdatawrite_range() argument
3508 ret = filemap_fdatawrite_range(inode->i_mapping, start, end); in btrfs_fdatawrite_range()
3511 ret = filemap_fdatawrite_range(inode->i_mapping, start, end); in btrfs_fdatawrite_range()