/fs/ntfs3/ |
D | run.c | 23 CLST lcn; /* Logical cluster number. */ member 124 if (n->lcn != SPARSE_LCN) in run_consolidate() 125 n->lcn += dl; in run_consolidate() 133 if ((n->lcn == SPARSE_LCN) != (r->lcn == SPARSE_LCN)) { in run_consolidate() 144 if (n->lcn != SPARSE_LCN && n->lcn != r->lcn + r->len) in run_consolidate() 192 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, in run_lookup_entry() argument 215 *lcn = r->lcn == SPARSE_LCN ? SPARSE_LCN : (r->lcn + gap); in run_lookup_entry() 241 if (r->lcn != SPARSE_LCN) in run_truncate_head() 242 r->lcn += dlen; in run_truncate_head() 316 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, in run_add_entry() argument [all …]
|
D | fsntfs.c | 352 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, in ntfs_look_for_free_space() argument 380 lcn = wnd_zone_bit(wnd); in ntfs_look_for_free_space() 383 wnd_zone_set(wnd, lcn + alen, zlen - alen); in ntfs_look_for_free_space() 385 err = wnd_set_used(wnd, lcn, alen); in ntfs_look_for_free_space() 389 alcn = lcn; in ntfs_look_for_free_space() 396 if (!lcn) in ntfs_look_for_free_space() 397 lcn = sbi->used.next_free_lcn; in ntfs_look_for_free_space() 399 if (lcn >= wnd->nbits) in ntfs_look_for_free_space() 400 lcn = 0; in ntfs_look_for_free_space() 402 alen = wnd_find(wnd, len, lcn, BITMAP_FIND_MARK_AS_USED, &alcn); in ntfs_look_for_free_space() [all …]
|
D | attrib.c | 97 CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0; in run_deallocate_ex() local 103 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex() 119 if (lcn != SPARSE_LCN) { in run_deallocate_ex() 122 mark_as_free_ex(sbi, lcn, clen, trim); in run_deallocate_ex() 132 if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) || in run_deallocate_ex() 150 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, in attr_allocate_clusters() argument 159 err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen, in attr_allocate_clusters() 175 *new_lcn = lcn; in attr_allocate_clusters() 181 if (!run_add_entry(run, vcn, lcn, flen, opt & ALLOCATE_MFT)) { in attr_allocate_clusters() 183 mark_as_free_ex(sbi, lcn, len, false); in attr_allocate_clusters() [all …]
|
D | file.c | 120 CLST vcn, lcn, clen; in ntfs_extend_initialized_size() local 126 err = attr_data_get_block(ni, vcn, 1, &lcn, &clen, NULL, in ntfs_extend_initialized_size() 131 if (lcn == SPARSE_LCN) { in ntfs_extend_initialized_size() 288 CLST lcn, len; in ntfs_file_mmap() local 294 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap() 607 CLST lcn, clen; in ntfs_fallocate() local 619 &lcn, &clen, &new, in ntfs_fallocate() 629 &lcn, &clen, &new, in ntfs_fallocate() 886 CLST lcn, clen; in ntfs_compress_write() local 892 err = attr_data_get_block(ni, frame << NTFS_LZNT_CUNIT, 1, &lcn, in ntfs_compress_write() [all …]
|
D | bitmap.c | 510 CLST lcn, clen; in wnd_rescan() local 546 &lcn, &clen, NULL)) { in wnd_rescan() 551 lbo = ((u64)lcn << cluster_bits) + off; in wnd_rescan() 683 CLST lcn, clen; in wnd_map() local 692 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen, in wnd_map() 697 lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask); in wnd_map() 1416 void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len) in wnd_zone_set() argument 1424 wnd_remove_free_ext(wnd, lcn, len); in wnd_zone_set() 1426 wnd->zone_bit = lcn; in wnd_zone_set() 1427 wnd->zone_end = lcn + len; in wnd_zone_set() [all …]
|
D | super.c | 1157 CLST vcn, lcn, len; in ntfs_fill_super() local 1377 for (i = 0; run_get_entry(&ni->file.run, i, &vcn, &lcn, &len); i++) { in ntfs_fill_super() 1378 if (lcn == SPARSE_LCN) in ntfs_fill_super() 1386 if (wnd_set_used_safe(&sbi->used.bitmap, lcn, len, &tt) || tt) { in ntfs_fill_super() 1627 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len) in ntfs_unmap_meta() argument 1631 sector_t devblock = (u64)lcn * sbi->blocks_per_cluster; in ntfs_unmap_meta() 1656 int ntfs_discard(struct ntfs_sb_info *sbi, CLST lcn, CLST len) in ntfs_discard() argument 1662 if (sbi->used.next_free_lcn == lcn + len) in ntfs_discard() 1663 sbi->used.next_free_lcn = lcn; in ntfs_discard() 1671 lbo = (u64)lcn << sbi->cluster_bits; in ntfs_discard()
|
D | ntfs_fs.h | 425 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, 436 int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn, 596 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, 654 void mark_as_free_ex(struct ntfs_sb_info *sbi, CLST lcn, CLST len, bool trim); 792 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, 797 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, 802 CLST *lcn, CLST *len); 824 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len);
|
D | frecord.c | 1958 CLST lcn, clen; in ni_fiemap() local 2020 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap() 2024 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && in ni_fiemap() 2044 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap() 2057 if (lcn == SPARSE_LCN) { in ni_fiemap() 2081 lbo = (u64)lcn << cluster_bits; in ni_fiemap() 2251 CLST vcn, cend, lcn, clen, end; in ni_decompress_file() local 2299 err = attr_data_get_block(ni, vcn, cend - vcn, &lcn, in ni_decompress_file()
|
D | fslog.c | 228 __le64 lcn; member 3740 CLST rno, lcn, lcn0, len0, clen; in log_replay() local 4540 u64 lcn0 = le64_to_cpu(r->lcn); in log_replay() 4839 lcn = le64_to_cpu(dp->page_lcns[i]); in log_replay() 4842 lcn0 != lcn) && in log_replay() 4843 !run_add_entry(oa->run1, vcn, lcn, 1, false)) { in log_replay() 4932 if (!run_lookup_entry(oa->run1, vcn, &lcn, NULL, NULL) || in log_replay() 4933 lcn == SPARSE_LCN) { in log_replay() 5056 &lcn, &clen, NULL); in log_replay() 5063 if (is_mapped && lcn != SPARSE_LCN && clen >= t16) in log_replay()
|
D | index.c | 379 CLST lcn, clen, vcn, vcn_next; in scan_nres_bitmap() local 393 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 408 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 415 block = (sector_t)lcn * sbi->blocks_per_cluster; in scan_nres_bitmap() 452 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && vcn == vcn_next; in scan_nres_bitmap()
|
D | inode.c | 566 CLST vcn, lcn, len; in ntfs_get_block_vbo() local 593 err = attr_data_get_block(ni, vcn, 1, &lcn, &len, create ? &new : NULL, in ntfs_get_block_vbo() 603 if (lcn == SPARSE_LCN) { in ntfs_get_block_vbo() 615 lbo = ((u64)lcn << cluster_bits) + off; in ntfs_get_block_vbo()
|
/fs/ntfs/ |
D | runlist.c | 148 if ((dst->lcn == LCN_RL_NOT_MAPPED) && (src->lcn == LCN_RL_NOT_MAPPED)) in ntfs_are_rl_mergeable() 154 if ((dst->lcn >= 0) && (src->lcn >= 0) && in ntfs_are_rl_mergeable() 155 ((dst->lcn + dst->length) == src->lcn)) in ntfs_are_rl_mergeable() 158 if ((dst->lcn == LCN_HOLE) && (src->lcn == LCN_HOLE)) in ntfs_are_rl_mergeable() 241 if (dst[marker].lcn == LCN_ENOENT) in ntfs_rl_append() 327 if (dst[marker].lcn == LCN_HOLE || dst[marker].lcn == LCN_RL_NOT_MAPPED) in ntfs_rl_insert() 339 dst[loc].lcn = LCN_RL_NOT_MAPPED; in ntfs_rl_insert() 427 if (dsize - tail > 0 && dst[marker].lcn == LCN_ENOENT) in ntfs_rl_replace() 557 drl[0].lcn = LCN_RL_NOT_MAPPED; in ntfs_runlists_merge() 566 while (srl[si].length && srl[si].lcn < LCN_HOLE) in ntfs_runlists_merge() [all …]
|
D | lcnalloc.c | 48 if (rl->lcn < 0) in ntfs_cluster_free_from_rl_nolock() 50 err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length); in ntfs_cluster_free_from_rl_nolock() 137 LCN zone_start, zone_end, bmp_pos, bmp_initial_pos, last_read_pos, lcn; in ntfs_cluster_alloc() local 284 lcn = bmp_pos & 7; in ntfs_cluster_alloc() 288 (unsigned long long)lcn, in ntfs_cluster_alloc() 290 while (lcn < buf_size && lcn + bmp_pos < zone_end) { in ntfs_cluster_alloc() 291 byte = buf + (lcn >> 3); in ntfs_cluster_alloc() 296 (unsigned long long)lcn, in ntfs_cluster_alloc() 299 (unsigned int)(lcn >> 3), in ntfs_cluster_alloc() 303 lcn = (lcn + 8) & ~(LCN)7; in ntfs_cluster_alloc() [all …]
|
D | aops.c | 168 LCN lcn; in ntfs_read_block() local 255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block() 257 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_block() 259 if (lcn >= 0) { in ntfs_read_block() 261 bh->b_blocknr = ((lcn << vol->cluster_size_bits) in ntfs_read_block() 273 if (lcn == LCN_HOLE) in ntfs_read_block() 276 if (!is_retry && lcn == LCN_RL_NOT_MAPPED) { in ntfs_read_block() 294 if (err == -ENOENT || lcn == LCN_ENOENT) { in ntfs_read_block() 534 LCN lcn; in ntfs_write_block() local 713 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_write_block() [all …]
|
D | debug.c | 135 LCN lcn = (rl + i)->lcn; in ntfs_debug_dump_runlist() local 137 if (lcn < (LCN)0) { in ntfs_debug_dump_runlist() 138 int index = -lcn - 1; in ntfs_debug_dump_runlist() 150 (long long)(rl + i)->lcn, in ntfs_debug_dump_runlist()
|
D | file.c | 567 LCN lcn; in ntfs_prepare_pages_for_non_resident_write() local 621 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write() 702 BUG_ON(lcn < 0); in ntfs_prepare_pages_for_non_resident_write() 831 lcn = ntfs_rl_vcn_to_lcn(rl, bh_cpos); in ntfs_prepare_pages_for_non_resident_write() 832 if (likely(lcn >= 0)) { in ntfs_prepare_pages_for_non_resident_write() 840 lcn_block = lcn << (vol->cluster_size_bits - in ntfs_prepare_pages_for_non_resident_write() 861 lcn = LCN_RL_NOT_MAPPED; in ntfs_prepare_pages_for_non_resident_write() 866 if (unlikely(lcn != LCN_HOLE && lcn != LCN_ENOENT)) { in ntfs_prepare_pages_for_non_resident_write() 867 if (likely(!is_retry && lcn == LCN_RL_NOT_MAPPED)) { in ntfs_prepare_pages_for_non_resident_write() 894 lcn = LCN_ENOENT; in ntfs_prepare_pages_for_non_resident_write() [all …]
|
D | attrib.c | 330 LCN lcn; in ntfs_attr_vcn_to_lcn_nolock() local 350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock() 351 if (likely(lcn >= LCN_HOLE)) { in ntfs_attr_vcn_to_lcn_nolock() 352 ntfs_debug("Done, lcn 0x%llx.", (long long)lcn); in ntfs_attr_vcn_to_lcn_nolock() 353 return lcn; in ntfs_attr_vcn_to_lcn_nolock() 355 if (lcn != LCN_RL_NOT_MAPPED) { in ntfs_attr_vcn_to_lcn_nolock() 356 if (lcn != LCN_ENOENT) in ntfs_attr_vcn_to_lcn_nolock() 357 lcn = LCN_EIO; in ntfs_attr_vcn_to_lcn_nolock() 381 lcn = LCN_ENOENT; in ntfs_attr_vcn_to_lcn_nolock() 383 lcn = LCN_ENOMEM; in ntfs_attr_vcn_to_lcn_nolock() [all …]
|
D | mft.c | 526 LCN lcn; in ntfs_sync_mft_mirror() local 548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror() 550 if (likely(lcn >= 0)) { in ntfs_sync_mft_mirror() 552 bh->b_blocknr = ((lcn << in ntfs_sync_mft_mirror() 563 (long long)lcn); in ntfs_sync_mft_mirror() 719 LCN lcn; in write_mft_record_nolock() local 736 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in write_mft_record_nolock() 738 if (likely(lcn >= 0)) { in write_mft_record_nolock() 740 bh->b_blocknr = ((lcn << in write_mft_record_nolock() 750 ni->mft_no, (long long)lcn); in write_mft_record_nolock() [all …]
|
D | logfile.c | 761 LCN lcn; in ntfs_empty_logfile() local 769 lcn = rl->lcn; in ntfs_empty_logfile() 770 if (unlikely(lcn == LCN_RL_NOT_MAPPED)) { in ntfs_empty_logfile() 775 if (unlikely(!rl->length || lcn < LCN_HOLE)) in ntfs_empty_logfile() 778 if (lcn == LCN_HOLE) in ntfs_empty_logfile() 780 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_empty_logfile() 784 end_block = (lcn + len) << vol->cluster_size_bits >> in ntfs_empty_logfile()
|
D | compress.c | 479 LCN lcn; in ntfs_read_compressed_block() local 609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block() 611 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_compressed_block() 614 (unsigned long long)lcn); in ntfs_read_compressed_block() 615 if (lcn < 0) { in ntfs_read_compressed_block() 620 if (lcn == LCN_HOLE) in ntfs_read_compressed_block() 622 if (is_retry || lcn != LCN_RL_NOT_MAPPED) in ntfs_read_compressed_block() 634 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_read_compressed_block()
|
D | runlist.h | 30 LCN lcn; /* lcn = Starting logical cluster number. */ member
|
D | super.c | 1175 rl2[0].lcn = vol->mftmirr_lcn; in check_mft_mirror() 1179 rl2[1].lcn = LCN_ENOENT; in check_mft_mirror() 1191 if (rl2[i].vcn != rl[i].vcn || rl2[i].lcn != rl[i].lcn || in check_mft_mirror()
|
/fs/erofs/ |
D | zmap.c | 15 unsigned long lcn; member 26 unsigned long lcn) in z_erofs_load_full_lcluster() argument 32 lcn * sizeof(struct z_erofs_lcluster_index); in z_erofs_load_full_lcluster() 42 m->lcn = lcn; in z_erofs_load_full_lcluster() 226 unsigned long lcn, bool lookahead) in z_erofs_load_compact_lcluster() argument 237 if (lcn >= totalidx) in z_erofs_load_compact_lcluster() 240 m->lcn = lcn; in z_erofs_load_compact_lcluster() 253 if (lcn < compacted_4b_initial) { in z_erofs_load_compact_lcluster() 258 lcn -= compacted_4b_initial; in z_erofs_load_compact_lcluster() 260 if (lcn < compacted_2b) { in z_erofs_load_compact_lcluster() [all …]
|