Home
last modified time | relevance | path

Searched refs:start (Results 1 – 25 of 200) sorted by relevance

12345678

/fs/btrfs/
Dextent_io.c40 u64 start; member
85 (unsigned long long)state->start, in extent_io_exit()
96 "refs %d\n", (unsigned long long)eb->start, in extent_io_exit()
171 if (offset < entry->start) in tree_insert()
201 if (offset < entry->start) in __etree_search()
221 while (prev && offset < prev_entry->start) { in __etree_search()
254 if (offset < eb->start) in buffer_tree_insert()
256 else if (offset > eb->start) in buffer_tree_insert()
276 if (offset < eb->start) in buffer_search()
278 else if (offset > eb->start) in buffer_search()
[all …]
Dextent_io.h43 u64 start, u64 end, int *page_started,
45 int (*writepage_start_hook)(struct page *page, u64 start, u64 end);
46 int (*writepage_io_hook)(struct page *page, u64 start, u64 end);
51 int (*readpage_io_hook)(struct page *page, u64 start, u64 end);
53 u64 start, u64 end,
56 u64 start, u64 end,
58 int (*readpage_end_io_hook)(struct page *page, u64 start, u64 end,
60 int (*writepage_end_io_hook)(struct page *page, u64 start, u64 end,
62 int (*set_bit_hook)(struct inode *inode, u64 start, u64 end,
64 int (*clear_bit_hook)(struct inode *inode, u64 start, u64 end,
[all …]
Dextent_map.c101 if (offset < entry->start) in tree_insert()
137 if (offset < entry->start) in __tree_search()
157 while (prev && offset < prev_entry->start) { in __tree_search()
193 if (extent_map_end(prev) == next->start && in mergable_maps()
227 exist = lookup_extent_mapping(tree, em->start, em->len); in add_extent_mapping()
234 rb = tree_insert(&tree->map, em->start, &em->rb_node); in add_extent_mapping()
241 if (em->start != 0) { in add_extent_mapping()
246 em->start = merge->start; in add_extent_mapping()
271 static u64 range_end(u64 start, u64 len) in range_end() argument
273 if (start + len < start) in range_end()
[all …]
Dfile.c166 int btrfs_drop_extent_cache(struct inode *inode, u64 start, u64 end, in btrfs_drop_extent_cache() argument
173 u64 len = end - start + 1; in btrfs_drop_extent_cache()
179 WARN_ON(end < start); in btrfs_drop_extent_cache()
191 em = lookup_extent_mapping(em_tree, start, len); in btrfs_drop_extent_cache()
199 if (em->start <= start && in btrfs_drop_extent_cache()
200 (!testend || em->start + em->len >= start + len)) { in btrfs_drop_extent_cache()
204 if (start < em->start) { in btrfs_drop_extent_cache()
205 len = em->start - start; in btrfs_drop_extent_cache()
207 len = start + len - (em->start + em->len); in btrfs_drop_extent_cache()
208 start = em->start + em->len; in btrfs_drop_extent_cache()
[all …]
Dinode.c88 static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end);
91 u64 start, u64 end, int *page_started,
111 u64 start, size_t size, size_t compressed_size, in insert_inline_extent() argument
140 key.offset = start; in insert_inline_extent()
182 start >> PAGE_CACHE_SHIFT); in insert_inline_extent()
185 offset = start & (PAGE_CACHE_SIZE - 1); in insert_inline_extent()
209 struct inode *inode, u64 start, u64 end, in cow_file_range_inline() argument
215 u64 inline_len = actual_end - start; in cow_file_range_inline()
225 if (start > 0 || in cow_file_range_inline()
235 ret = btrfs_drop_extents(trans, root, inode, start, in cow_file_range_inline()
[all …]
Dordered-data.c168 u64 start, u64 len, u64 disk_len, int type) in btrfs_add_ordered_extent() argument
181 entry->start = start; in btrfs_add_ordered_extent()
383 u64 start = entry->file_offset; in btrfs_start_ordered_extent() local
384 u64 end = start + entry->len - 1; in btrfs_start_ordered_extent()
391 btrfs_fdatawrite_range(inode->i_mapping, start, end, WB_SYNC_ALL); in btrfs_start_ordered_extent()
401 int btrfs_wait_ordered_range(struct inode *inode, u64 start, u64 len) in btrfs_wait_ordered_range() argument
408 if (start + len < start) { in btrfs_wait_ordered_range()
411 orig_end = start + len - 1; in btrfs_wait_ordered_range()
420 btrfs_fdatawrite_range(inode->i_mapping, start, orig_end, WB_SYNC_NONE); in btrfs_wait_ordered_range()
426 btrfs_fdatawrite_range(inode->i_mapping, start, orig_end, WB_SYNC_ALL); in btrfs_wait_ordered_range()
[all …]
Dextent_map.h21 u64 start; member
39 if (em->start + em->len < em->start) in extent_map_end()
41 return em->start + em->len; in extent_map_end()
53 u64 start, u64 len);
Dcompression.c58 u64 start; member
193 cb->start, in end_compressed_bio_read()
239 static noinline int end_compressed_writeback(struct inode *inode, u64 start, in end_compressed_writeback() argument
242 unsigned long index = start >> PAGE_CACHE_SHIFT; in end_compressed_writeback()
243 unsigned long end_index = (start + ram_size - 1) >> PAGE_CACHE_SHIFT; in end_compressed_writeback()
301 cb->start, in end_compressed_bio_write()
302 cb->start + cb->len - 1, in end_compressed_bio_write()
306 end_compressed_writeback(inode, cb->start, cb->len); in end_compressed_bio_write()
336 int btrfs_submit_compressed_write(struct inode *inode, u64 start, in btrfs_submit_compressed_write() argument
353 WARN_ON(start & ((u64)PAGE_CACHE_SIZE - 1)); in btrfs_submit_compressed_write()
[all …]
/fs/adfs/
Dmap.c83 unsigned int start = dm->dm_startbit; in lookup_zone() local
88 frag = GET_FRAG_ID(map, start, idmask); in lookup_zone()
89 mapptr = start + idlen; in lookup_zone()
110 start = mapptr; in lookup_zone()
116 frag, start, mapptr); in lookup_zone()
121 int length = mapptr - start; in lookup_zone()
127 return start + *offset; in lookup_zone()
144 unsigned int start = 8, mapptr; in scan_free_map() local
151 frag = GET_FRAG_ID(map, start, idmask); in scan_free_map()
161 start += frag; in scan_free_map()
[all …]
/fs/hfs/
Dbitmap.c32 u32 mask, start, len, n; in hfs_find_set_zero_bits() local
70 start = (curr - bitmap) * 32 + i; in hfs_find_set_zero_bits()
71 if (start >= size) in hfs_find_set_zero_bits()
72 return start; in hfs_find_set_zero_bits()
74 len = min(size - start, len); in hfs_find_set_zero_bits()
108 *max = (curr - bitmap) * 32 + i - start; in hfs_find_set_zero_bits()
109 return start; in hfs_find_set_zero_bits()
193 int hfs_clear_vbm_bits(struct super_block *sb, u16 start, u16 count) in hfs_clear_vbm_bits() argument
203 dprint(DBG_BITMAP, "clear_bits: %u,%u\n", start, count); in hfs_clear_vbm_bits()
205 if ((start + count) > HFS_SB(sb)->fs_ablocks) in hfs_clear_vbm_bits()
[all …]
Dextent.c208 u16 count, start; in hfs_add_extent() local
215 start = be16_to_cpu(extent->block); in hfs_add_extent()
216 if (alloc_block != start + count) { in hfs_add_extent()
236 u16 count, start; in hfs_free_extents() local
252 start = be16_to_cpu(extent->block); in hfs_free_extents()
254 hfs_clear_vbm_bits(sb, start, count); in hfs_free_extents()
260 hfs_clear_vbm_bits(sb, start + count, block_nr); in hfs_free_extents()
275 u32 total_blocks, blocks, start; in hfs_free_fork() local
306 start = be16_to_cpu(fd.key->ext.FABN); in hfs_free_fork()
307 hfs_free_extents(sb, extent, total_blocks - start, total_blocks); in hfs_free_fork()
[all …]
/fs/partitions/
Dacorn.c58 __le32 start; member
98 le32_to_cpu(rr->part[part].start), in riscix_partition()
307 __le32 start; member
375 u32 start = le32_to_cpu(p->start); in adfspart_check_ICS() local
395 if (size > 1 && adfspart_check_ICSLinux(bdev, start)) { in adfspart_check_ICS()
396 start += 1; in adfspart_check_ICS()
402 put_partition(state, slot++, start, size); in adfspart_check_ICS()
415 __le32 start; member
470 u32 start = le32_to_cpu(p->start); in adfspart_check_POWERTEC() local
474 put_partition(state, slot++, start, size); in adfspart_check_POWERTEC()
[all …]
/fs/qnx4/
Dbitmap.c66 int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1; in qnx4_count_free_blocks() local
74 if ((bh = sb_bread(sb, start + offset)) == NULL) { in qnx4_count_free_blocks()
91 int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1; in qnx4_is_free() local
97 start += block / (QNX4_BLOCK_SIZE * 8); in qnx4_is_free()
99 (unsigned long) block, (unsigned long) start)); in qnx4_is_free()
101 bh = sb_bread(sb, start); in qnx4_is_free()
120 int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1; in qnx4_set_bitmap() local
125 start += block / (QNX4_BLOCK_SIZE * 8); in qnx4_set_bitmap()
127 (unsigned long) block, (unsigned long) start)); in qnx4_set_bitmap()
129 bh = sb_bread(sb, start); in qnx4_set_bitmap()
/fs/hfsplus/
Dextents.c244 u32 count, start; in hfsplus_add_extent() local
251 start = be32_to_cpu(extent->start_block); in hfsplus_add_extent()
252 if (alloc_block != start + count) { in hfsplus_add_extent()
273 u32 count, start; in hfsplus_free_extents() local
289 start = be32_to_cpu(extent->start_block); in hfsplus_free_extents()
291 hfsplus_block_free(sb, start, count); in hfsplus_free_extents()
297 hfsplus_block_free(sb, start + count, block_nr); in hfsplus_free_extents()
313 u32 total_blocks, blocks, start; in hfsplus_free_fork() local
336 start = be32_to_cpu(fd.key->ext.start_block); in hfsplus_free_fork()
338 total_blocks - start, in hfsplus_free_fork()
[all …]
Dbitmap.c23 u32 mask, start, len, n; in hfsplus_block_allocate() local
36 start = size; in hfsplus_block_allocate()
81 start = size; in hfsplus_block_allocate()
91 start = size; in hfsplus_block_allocate()
95 start = offset + (curr - pptr) * 32 + i; in hfsplus_block_allocate()
96 if (start >= size) { in hfsplus_block_allocate()
101 len = min(size - start, len); in hfsplus_block_allocate()
132 start = size; in hfsplus_block_allocate()
152 *max = offset + (curr - pptr) * 32 + i - start; in hfsplus_block_allocate()
155 dprint(DBG_BITMAP, "-> %u,%u\n", start, *max); in hfsplus_block_allocate()
[all …]
/fs/xfs/quota/
Dxfs_qm_stats.c55 char **start, in xfs_qm_read_xfsquota() argument
71 *start = buffer; in xfs_qm_read_xfsquota()
75 *start = buffer + offset; in xfs_qm_read_xfsquota()
86 char **start, in xfs_qm_read_stats() argument
106 *start = buffer; in xfs_qm_read_stats()
110 *start = buffer + offset; in xfs_qm_read_stats()
/fs/befs/
Dendian.h77 run.start = le16_to_cpu((__force __le16)n.start); in fsrun_to_cpu()
81 run.start = be16_to_cpu((__force __be16)n.start); in fsrun_to_cpu()
94 run.start = cpu_to_le16(n.start); in cpu_to_fsrun()
98 run.start = cpu_to_be16(n.start); in cpu_to_fsrun()
Ddebug.c105 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_inode()
118 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_inode()
122 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_inode()
136 tmp_run.allocation_group, tmp_run.start, in befs_dump_inode()
147 tmp_run.start, tmp_run.len); in befs_dump_inode()
157 tmp_run.allocation_group, tmp_run.start, in befs_dump_inode()
206 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_super_block()
215 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_super_block()
219 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_super_block()
239 befs_debug(sb, "[%u, %hu, %hu]", n.allocation_group, n.start, n.len);
/fs/xfs/
Dxfs_rtalloc.c739 xfs_rtblock_t start, /* start block to allocate */ in xfs_rtallocate_range() argument
749 end = start + len - 1; in xfs_rtallocate_range()
755 error = xfs_rtfind_back(mp, tp, start, 0, &preblock); in xfs_rtallocate_range()
781 if (preblock < start) { in xfs_rtallocate_range()
783 XFS_RTBLOCKLOG(start - preblock), in xfs_rtallocate_range()
804 error = xfs_rtmodify_range(mp, tp, start, len, 0); in xfs_rtallocate_range()
924 xfs_rtblock_t start, /* starting block number of extent */ in xfs_rtcheck_range() argument
945 block = XFS_BITTOBLOCK(mp, start); in xfs_rtcheck_range()
957 word = XFS_BITTOWORD(mp, start); in xfs_rtcheck_range()
959 bit = (int)(start & (XFS_NBWORD - 1)); in xfs_rtcheck_range()
[all …]
/fs/ext3/
Dresize.c28 ext3_fsblk_t start = le32_to_cpu(es->s_blocks_count); in verify_group_input() local
29 ext3_fsblk_t end = start + input->blocks_count; in verify_group_input()
35 ext3_fsblk_t metaend = start + overhead; in verify_group_input()
54 else if ((start - le32_to_cpu(es->s_first_data_block)) % in verify_group_input()
67 else if (outside(input->block_bitmap, start, end)) in verify_group_input()
71 else if (outside(input->inode_bitmap, start, end)) in verify_group_input()
75 else if (outside(input->inode_table, start, end) || in verify_group_input()
76 outside(itend - 1, start, end)) in verify_group_input()
92 else if (inside(input->block_bitmap, start, metaend)) in verify_group_input()
96 input->block_bitmap, start, metaend - 1); in verify_group_input()
[all …]
/fs/ext4/
Dresize.c28 ext4_fsblk_t start = ext4_blocks_count(es); in verify_group_input() local
29 ext4_fsblk_t end = start + input->blocks_count; in verify_group_input()
35 ext4_fsblk_t metaend = start + overhead; in verify_group_input()
50 ext4_get_group_no_and_offset(sb, start, NULL, &offset); in verify_group_input()
67 else if (outside(input->block_bitmap, start, end)) in verify_group_input()
71 else if (outside(input->inode_bitmap, start, end)) in verify_group_input()
75 else if (outside(input->inode_table, start, end) || in verify_group_input()
76 outside(itend - 1, start, end)) in verify_group_input()
94 else if (inside(input->block_bitmap, start, metaend)) in verify_group_input()
99 start, metaend - 1); in verify_group_input()
[all …]
/fs/cifs/
Dcifs_unicode.h45 wchar_t start; member
288 while (rp->start) { in UniToupper()
289 if (uc < rp->start) /* Before start of range */ in UniToupper()
292 return uc + rp->table[uc - rp->start]; in UniToupper()
330 while (rp->start) { in UniTolower()
331 if (uc < rp->start) /* Before start of range */ in UniTolower()
334 return uc + rp->table[uc - rp->start]; in UniTolower()
/fs/proc/
Dkcore.c265 unsigned long start; in read_kcore() local
312 start = kc_offset_to_vaddr(*fpos - elf_buflen); in read_kcore()
313 if ((tsz = (PAGE_SIZE - (start & ~PAGE_MASK))) > buflen) in read_kcore()
321 if (start >= m->addr && start < (m->addr+m->size)) in read_kcore()
329 } else if (is_vmalloc_addr((void *)start)) { in read_kcore()
332 unsigned long curstart = start; in read_kcore()
364 memcpy(elf_buf + (vmstart - start), in read_kcore()
374 if (kern_addr_valid(start)) { in read_kcore()
377 n = copy_to_user(buffer, (char *)start, tsz); in read_kcore()
398 start += tsz; in read_kcore()
/fs/afs/
Dwrite.c323 pgoff_t start, first, last; in afs_write_back_from_locked_page() local
337 start = primary_page->index; in afs_write_back_from_locked_page()
338 if (start >= wb->last) in afs_write_back_from_locked_page()
340 start++; in afs_write_back_from_locked_page()
342 _debug("more %lx [%lx]", start, count); in afs_write_back_from_locked_page()
343 n = wb->last - start + 1; in afs_write_back_from_locked_page()
347 start, n, pages); in afs_write_back_from_locked_page()
351 if (pages[0]->index != start) { in afs_write_back_from_locked_page()
383 start += loop; in afs_write_back_from_locked_page()
384 } while (start <= wb->last && count < 65536); in afs_write_back_from_locked_page()
[all …]
/fs/ext2/
Dballoc.c344 ext2_fsblk_t start = rsv->rsv_start; in ext2_rsv_window_add() local
355 if (start < this->rsv_start) in ext2_rsv_window_add()
357 else if (start > this->rsv_end) in ext2_rsv_window_add()
586 bitmap_search_next_usable_block(ext2_grpblk_t start, struct buffer_head *bh, in bitmap_search_next_usable_block() argument
591 next = ext2_find_next_zero_bit(bh->b_data, maxblocks, start); in bitmap_search_next_usable_block()
610 find_next_usable_block(int start, struct buffer_head *bh, int maxblocks) in find_next_usable_block() argument
615 if (start > 0) { in find_next_usable_block()
624 ext2_grpblk_t end_goal = (start + 63) & ~63; in find_next_usable_block()
627 here = ext2_find_next_zero_bit(bh->b_data, end_goal, start); in find_next_usable_block()
633 here = start; in find_next_usable_block()
[all …]

12345678