/fs/btrfs/ |
D | extent_io.c | 40 u64 start; member 85 (unsigned long long)state->start, in extent_io_exit() 96 "refs %d\n", (unsigned long long)eb->start, in extent_io_exit() 171 if (offset < entry->start) in tree_insert() 201 if (offset < entry->start) in __etree_search() 221 while (prev && offset < prev_entry->start) { in __etree_search() 254 if (offset < eb->start) in buffer_tree_insert() 256 else if (offset > eb->start) in buffer_tree_insert() 276 if (offset < eb->start) in buffer_search() 278 else if (offset > eb->start) in buffer_search() [all …]
|
D | extent_io.h | 43 u64 start, u64 end, int *page_started, 45 int (*writepage_start_hook)(struct page *page, u64 start, u64 end); 46 int (*writepage_io_hook)(struct page *page, u64 start, u64 end); 51 int (*readpage_io_hook)(struct page *page, u64 start, u64 end); 53 u64 start, u64 end, 56 u64 start, u64 end, 58 int (*readpage_end_io_hook)(struct page *page, u64 start, u64 end, 60 int (*writepage_end_io_hook)(struct page *page, u64 start, u64 end, 62 int (*set_bit_hook)(struct inode *inode, u64 start, u64 end, 64 int (*clear_bit_hook)(struct inode *inode, u64 start, u64 end, [all …]
|
D | extent_map.c | 101 if (offset < entry->start) in tree_insert() 137 if (offset < entry->start) in __tree_search() 157 while (prev && offset < prev_entry->start) { in __tree_search() 193 if (extent_map_end(prev) == next->start && in mergable_maps() 227 exist = lookup_extent_mapping(tree, em->start, em->len); in add_extent_mapping() 234 rb = tree_insert(&tree->map, em->start, &em->rb_node); in add_extent_mapping() 241 if (em->start != 0) { in add_extent_mapping() 246 em->start = merge->start; in add_extent_mapping() 271 static u64 range_end(u64 start, u64 len) in range_end() argument 273 if (start + len < start) in range_end() [all …]
|
D | file.c | 166 int btrfs_drop_extent_cache(struct inode *inode, u64 start, u64 end, in btrfs_drop_extent_cache() argument 173 u64 len = end - start + 1; in btrfs_drop_extent_cache() 179 WARN_ON(end < start); in btrfs_drop_extent_cache() 191 em = lookup_extent_mapping(em_tree, start, len); in btrfs_drop_extent_cache() 199 if (em->start <= start && in btrfs_drop_extent_cache() 200 (!testend || em->start + em->len >= start + len)) { in btrfs_drop_extent_cache() 204 if (start < em->start) { in btrfs_drop_extent_cache() 205 len = em->start - start; in btrfs_drop_extent_cache() 207 len = start + len - (em->start + em->len); in btrfs_drop_extent_cache() 208 start = em->start + em->len; in btrfs_drop_extent_cache() [all …]
|
D | inode.c | 88 static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end); 91 u64 start, u64 end, int *page_started, 111 u64 start, size_t size, size_t compressed_size, in insert_inline_extent() argument 140 key.offset = start; in insert_inline_extent() 182 start >> PAGE_CACHE_SHIFT); in insert_inline_extent() 185 offset = start & (PAGE_CACHE_SIZE - 1); in insert_inline_extent() 209 struct inode *inode, u64 start, u64 end, in cow_file_range_inline() argument 215 u64 inline_len = actual_end - start; in cow_file_range_inline() 225 if (start > 0 || in cow_file_range_inline() 235 ret = btrfs_drop_extents(trans, root, inode, start, in cow_file_range_inline() [all …]
|
D | ordered-data.c | 168 u64 start, u64 len, u64 disk_len, int type) in btrfs_add_ordered_extent() argument 181 entry->start = start; in btrfs_add_ordered_extent() 383 u64 start = entry->file_offset; in btrfs_start_ordered_extent() local 384 u64 end = start + entry->len - 1; in btrfs_start_ordered_extent() 391 btrfs_fdatawrite_range(inode->i_mapping, start, end, WB_SYNC_ALL); in btrfs_start_ordered_extent() 401 int btrfs_wait_ordered_range(struct inode *inode, u64 start, u64 len) in btrfs_wait_ordered_range() argument 408 if (start + len < start) { in btrfs_wait_ordered_range() 411 orig_end = start + len - 1; in btrfs_wait_ordered_range() 420 btrfs_fdatawrite_range(inode->i_mapping, start, orig_end, WB_SYNC_NONE); in btrfs_wait_ordered_range() 426 btrfs_fdatawrite_range(inode->i_mapping, start, orig_end, WB_SYNC_ALL); in btrfs_wait_ordered_range() [all …]
|
D | extent_map.h | 21 u64 start; member 39 if (em->start + em->len < em->start) in extent_map_end() 41 return em->start + em->len; in extent_map_end() 53 u64 start, u64 len);
|
D | compression.c | 58 u64 start; member 193 cb->start, in end_compressed_bio_read() 239 static noinline int end_compressed_writeback(struct inode *inode, u64 start, in end_compressed_writeback() argument 242 unsigned long index = start >> PAGE_CACHE_SHIFT; in end_compressed_writeback() 243 unsigned long end_index = (start + ram_size - 1) >> PAGE_CACHE_SHIFT; in end_compressed_writeback() 301 cb->start, in end_compressed_bio_write() 302 cb->start + cb->len - 1, in end_compressed_bio_write() 306 end_compressed_writeback(inode, cb->start, cb->len); in end_compressed_bio_write() 336 int btrfs_submit_compressed_write(struct inode *inode, u64 start, in btrfs_submit_compressed_write() argument 353 WARN_ON(start & ((u64)PAGE_CACHE_SIZE - 1)); in btrfs_submit_compressed_write() [all …]
|
/fs/adfs/ |
D | map.c | 83 unsigned int start = dm->dm_startbit; in lookup_zone() local 88 frag = GET_FRAG_ID(map, start, idmask); in lookup_zone() 89 mapptr = start + idlen; in lookup_zone() 110 start = mapptr; in lookup_zone() 116 frag, start, mapptr); in lookup_zone() 121 int length = mapptr - start; in lookup_zone() 127 return start + *offset; in lookup_zone() 144 unsigned int start = 8, mapptr; in scan_free_map() local 151 frag = GET_FRAG_ID(map, start, idmask); in scan_free_map() 161 start += frag; in scan_free_map() [all …]
|
/fs/hfs/ |
D | bitmap.c | 32 u32 mask, start, len, n; in hfs_find_set_zero_bits() local 70 start = (curr - bitmap) * 32 + i; in hfs_find_set_zero_bits() 71 if (start >= size) in hfs_find_set_zero_bits() 72 return start; in hfs_find_set_zero_bits() 74 len = min(size - start, len); in hfs_find_set_zero_bits() 108 *max = (curr - bitmap) * 32 + i - start; in hfs_find_set_zero_bits() 109 return start; in hfs_find_set_zero_bits() 193 int hfs_clear_vbm_bits(struct super_block *sb, u16 start, u16 count) in hfs_clear_vbm_bits() argument 203 dprint(DBG_BITMAP, "clear_bits: %u,%u\n", start, count); in hfs_clear_vbm_bits() 205 if ((start + count) > HFS_SB(sb)->fs_ablocks) in hfs_clear_vbm_bits() [all …]
|
D | extent.c | 208 u16 count, start; in hfs_add_extent() local 215 start = be16_to_cpu(extent->block); in hfs_add_extent() 216 if (alloc_block != start + count) { in hfs_add_extent() 236 u16 count, start; in hfs_free_extents() local 252 start = be16_to_cpu(extent->block); in hfs_free_extents() 254 hfs_clear_vbm_bits(sb, start, count); in hfs_free_extents() 260 hfs_clear_vbm_bits(sb, start + count, block_nr); in hfs_free_extents() 275 u32 total_blocks, blocks, start; in hfs_free_fork() local 306 start = be16_to_cpu(fd.key->ext.FABN); in hfs_free_fork() 307 hfs_free_extents(sb, extent, total_blocks - start, total_blocks); in hfs_free_fork() [all …]
|
/fs/partitions/ |
D | acorn.c | 58 __le32 start; member 98 le32_to_cpu(rr->part[part].start), in riscix_partition() 307 __le32 start; member 375 u32 start = le32_to_cpu(p->start); in adfspart_check_ICS() local 395 if (size > 1 && adfspart_check_ICSLinux(bdev, start)) { in adfspart_check_ICS() 396 start += 1; in adfspart_check_ICS() 402 put_partition(state, slot++, start, size); in adfspart_check_ICS() 415 __le32 start; member 470 u32 start = le32_to_cpu(p->start); in adfspart_check_POWERTEC() local 474 put_partition(state, slot++, start, size); in adfspart_check_POWERTEC() [all …]
|
/fs/qnx4/ |
D | bitmap.c | 66 int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1; in qnx4_count_free_blocks() local 74 if ((bh = sb_bread(sb, start + offset)) == NULL) { in qnx4_count_free_blocks() 91 int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1; in qnx4_is_free() local 97 start += block / (QNX4_BLOCK_SIZE * 8); in qnx4_is_free() 99 (unsigned long) block, (unsigned long) start)); in qnx4_is_free() 101 bh = sb_bread(sb, start); in qnx4_is_free() 120 int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1; in qnx4_set_bitmap() local 125 start += block / (QNX4_BLOCK_SIZE * 8); in qnx4_set_bitmap() 127 (unsigned long) block, (unsigned long) start)); in qnx4_set_bitmap() 129 bh = sb_bread(sb, start); in qnx4_set_bitmap()
|
/fs/hfsplus/ |
D | extents.c | 244 u32 count, start; in hfsplus_add_extent() local 251 start = be32_to_cpu(extent->start_block); in hfsplus_add_extent() 252 if (alloc_block != start + count) { in hfsplus_add_extent() 273 u32 count, start; in hfsplus_free_extents() local 289 start = be32_to_cpu(extent->start_block); in hfsplus_free_extents() 291 hfsplus_block_free(sb, start, count); in hfsplus_free_extents() 297 hfsplus_block_free(sb, start + count, block_nr); in hfsplus_free_extents() 313 u32 total_blocks, blocks, start; in hfsplus_free_fork() local 336 start = be32_to_cpu(fd.key->ext.start_block); in hfsplus_free_fork() 338 total_blocks - start, in hfsplus_free_fork() [all …]
|
D | bitmap.c | 23 u32 mask, start, len, n; in hfsplus_block_allocate() local 36 start = size; in hfsplus_block_allocate() 81 start = size; in hfsplus_block_allocate() 91 start = size; in hfsplus_block_allocate() 95 start = offset + (curr - pptr) * 32 + i; in hfsplus_block_allocate() 96 if (start >= size) { in hfsplus_block_allocate() 101 len = min(size - start, len); in hfsplus_block_allocate() 132 start = size; in hfsplus_block_allocate() 152 *max = offset + (curr - pptr) * 32 + i - start; in hfsplus_block_allocate() 155 dprint(DBG_BITMAP, "-> %u,%u\n", start, *max); in hfsplus_block_allocate() [all …]
|
/fs/xfs/quota/ |
D | xfs_qm_stats.c | 55 char **start, in xfs_qm_read_xfsquota() argument 71 *start = buffer; in xfs_qm_read_xfsquota() 75 *start = buffer + offset; in xfs_qm_read_xfsquota() 86 char **start, in xfs_qm_read_stats() argument 106 *start = buffer; in xfs_qm_read_stats() 110 *start = buffer + offset; in xfs_qm_read_stats()
|
/fs/befs/ |
D | endian.h | 77 run.start = le16_to_cpu((__force __le16)n.start); in fsrun_to_cpu() 81 run.start = be16_to_cpu((__force __be16)n.start); in fsrun_to_cpu() 94 run.start = cpu_to_le16(n.start); in cpu_to_fsrun() 98 run.start = cpu_to_be16(n.start); in cpu_to_fsrun()
|
D | debug.c | 105 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_inode() 118 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_inode() 122 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_inode() 136 tmp_run.allocation_group, tmp_run.start, in befs_dump_inode() 147 tmp_run.start, tmp_run.len); in befs_dump_inode() 157 tmp_run.allocation_group, tmp_run.start, in befs_dump_inode() 206 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_super_block() 215 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_super_block() 219 tmp_run.allocation_group, tmp_run.start, tmp_run.len); in befs_dump_super_block() 239 befs_debug(sb, "[%u, %hu, %hu]", n.allocation_group, n.start, n.len);
|
/fs/xfs/ |
D | xfs_rtalloc.c | 739 xfs_rtblock_t start, /* start block to allocate */ in xfs_rtallocate_range() argument 749 end = start + len - 1; in xfs_rtallocate_range() 755 error = xfs_rtfind_back(mp, tp, start, 0, &preblock); in xfs_rtallocate_range() 781 if (preblock < start) { in xfs_rtallocate_range() 783 XFS_RTBLOCKLOG(start - preblock), in xfs_rtallocate_range() 804 error = xfs_rtmodify_range(mp, tp, start, len, 0); in xfs_rtallocate_range() 924 xfs_rtblock_t start, /* starting block number of extent */ in xfs_rtcheck_range() argument 945 block = XFS_BITTOBLOCK(mp, start); in xfs_rtcheck_range() 957 word = XFS_BITTOWORD(mp, start); in xfs_rtcheck_range() 959 bit = (int)(start & (XFS_NBWORD - 1)); in xfs_rtcheck_range() [all …]
|
/fs/ext3/ |
D | resize.c | 28 ext3_fsblk_t start = le32_to_cpu(es->s_blocks_count); in verify_group_input() local 29 ext3_fsblk_t end = start + input->blocks_count; in verify_group_input() 35 ext3_fsblk_t metaend = start + overhead; in verify_group_input() 54 else if ((start - le32_to_cpu(es->s_first_data_block)) % in verify_group_input() 67 else if (outside(input->block_bitmap, start, end)) in verify_group_input() 71 else if (outside(input->inode_bitmap, start, end)) in verify_group_input() 75 else if (outside(input->inode_table, start, end) || in verify_group_input() 76 outside(itend - 1, start, end)) in verify_group_input() 92 else if (inside(input->block_bitmap, start, metaend)) in verify_group_input() 96 input->block_bitmap, start, metaend - 1); in verify_group_input() [all …]
|
/fs/ext4/ |
D | resize.c | 28 ext4_fsblk_t start = ext4_blocks_count(es); in verify_group_input() local 29 ext4_fsblk_t end = start + input->blocks_count; in verify_group_input() 35 ext4_fsblk_t metaend = start + overhead; in verify_group_input() 50 ext4_get_group_no_and_offset(sb, start, NULL, &offset); in verify_group_input() 67 else if (outside(input->block_bitmap, start, end)) in verify_group_input() 71 else if (outside(input->inode_bitmap, start, end)) in verify_group_input() 75 else if (outside(input->inode_table, start, end) || in verify_group_input() 76 outside(itend - 1, start, end)) in verify_group_input() 94 else if (inside(input->block_bitmap, start, metaend)) in verify_group_input() 99 start, metaend - 1); in verify_group_input() [all …]
|
/fs/cifs/ |
D | cifs_unicode.h | 45 wchar_t start; member 288 while (rp->start) { in UniToupper() 289 if (uc < rp->start) /* Before start of range */ in UniToupper() 292 return uc + rp->table[uc - rp->start]; in UniToupper() 330 while (rp->start) { in UniTolower() 331 if (uc < rp->start) /* Before start of range */ in UniTolower() 334 return uc + rp->table[uc - rp->start]; in UniTolower()
|
/fs/proc/ |
D | kcore.c | 265 unsigned long start; in read_kcore() local 312 start = kc_offset_to_vaddr(*fpos - elf_buflen); in read_kcore() 313 if ((tsz = (PAGE_SIZE - (start & ~PAGE_MASK))) > buflen) in read_kcore() 321 if (start >= m->addr && start < (m->addr+m->size)) in read_kcore() 329 } else if (is_vmalloc_addr((void *)start)) { in read_kcore() 332 unsigned long curstart = start; in read_kcore() 364 memcpy(elf_buf + (vmstart - start), in read_kcore() 374 if (kern_addr_valid(start)) { in read_kcore() 377 n = copy_to_user(buffer, (char *)start, tsz); in read_kcore() 398 start += tsz; in read_kcore()
|
/fs/afs/ |
D | write.c | 323 pgoff_t start, first, last; in afs_write_back_from_locked_page() local 337 start = primary_page->index; in afs_write_back_from_locked_page() 338 if (start >= wb->last) in afs_write_back_from_locked_page() 340 start++; in afs_write_back_from_locked_page() 342 _debug("more %lx [%lx]", start, count); in afs_write_back_from_locked_page() 343 n = wb->last - start + 1; in afs_write_back_from_locked_page() 347 start, n, pages); in afs_write_back_from_locked_page() 351 if (pages[0]->index != start) { in afs_write_back_from_locked_page() 383 start += loop; in afs_write_back_from_locked_page() 384 } while (start <= wb->last && count < 65536); in afs_write_back_from_locked_page() [all …]
|
/fs/ext2/ |
D | balloc.c | 344 ext2_fsblk_t start = rsv->rsv_start; in ext2_rsv_window_add() local 355 if (start < this->rsv_start) in ext2_rsv_window_add() 357 else if (start > this->rsv_end) in ext2_rsv_window_add() 586 bitmap_search_next_usable_block(ext2_grpblk_t start, struct buffer_head *bh, in bitmap_search_next_usable_block() argument 591 next = ext2_find_next_zero_bit(bh->b_data, maxblocks, start); in bitmap_search_next_usable_block() 610 find_next_usable_block(int start, struct buffer_head *bh, int maxblocks) in find_next_usable_block() argument 615 if (start > 0) { in find_next_usable_block() 624 ext2_grpblk_t end_goal = (start + 63) & ~63; in find_next_usable_block() 627 here = ext2_find_next_zero_bit(bh->b_data, end_goal, start); in find_next_usable_block() 633 here = start; in find_next_usable_block() [all …]
|