/fs/ntfs3/ |
D | run.c | 23 CLST lcn; /* Logical cluster number. */ member 124 if (n->lcn != SPARSE_LCN) in run_consolidate() 125 n->lcn += dl; in run_consolidate() 133 if ((n->lcn == SPARSE_LCN) != (r->lcn == SPARSE_LCN)) { in run_consolidate() 144 if (n->lcn != SPARSE_LCN && n->lcn != r->lcn + r->len) in run_consolidate() 192 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, in run_lookup_entry() argument 215 *lcn = r->lcn == SPARSE_LCN ? SPARSE_LCN : (r->lcn + gap); in run_lookup_entry() 241 if (r->lcn != SPARSE_LCN) in run_truncate_head() 242 r->lcn += dlen; in run_truncate_head() 316 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, in run_add_entry() argument [all …]
|
D | fsntfs.c | 356 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, in ntfs_look_for_free_space() argument 384 lcn = wnd_zone_bit(wnd); in ntfs_look_for_free_space() 387 wnd_zone_set(wnd, lcn + alen, zlen - alen); in ntfs_look_for_free_space() 389 err = wnd_set_used(wnd, lcn, alen); in ntfs_look_for_free_space() 393 alcn = lcn; in ntfs_look_for_free_space() 400 if (!lcn) in ntfs_look_for_free_space() 401 lcn = sbi->used.next_free_lcn; in ntfs_look_for_free_space() 403 if (lcn >= wnd->nbits) in ntfs_look_for_free_space() 404 lcn = 0; in ntfs_look_for_free_space() 406 alen = wnd_find(wnd, len, lcn, BITMAP_FIND_MARK_AS_USED, &alcn); in ntfs_look_for_free_space() [all …]
|
D | attrib.c | 124 CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0; in run_deallocate_ex() local 130 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex() 146 if (lcn != SPARSE_LCN) { in run_deallocate_ex() 147 mark_as_free_ex(sbi, lcn, clen, trim); in run_deallocate_ex() 156 if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) || in run_deallocate_ex() 174 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, in attr_allocate_clusters() argument 184 err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen, in attr_allocate_clusters() 198 *new_lcn = lcn; in attr_allocate_clusters() 201 if (!run_add_entry(run, vcn, lcn, flen, opt == ALLOCATE_MFT)) { in attr_allocate_clusters() 204 wnd_set_free(wnd, lcn, flen); in attr_allocate_clusters() [all …]
|
D | bitmap.c | 515 CLST lcn, clen; in wnd_rescan() local 551 &lcn, &clen, NULL)) { in wnd_rescan() 556 lbo = ((u64)lcn << cluster_bits) + off; in wnd_rescan() 690 CLST lcn, clen; in wnd_map() local 699 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen, in wnd_map() 704 lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask); in wnd_map() 1398 void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len) in wnd_zone_set() argument 1407 wnd_remove_free_ext(wnd, lcn, len); in wnd_zone_set() 1409 wnd->zone_bit = lcn; in wnd_zone_set() 1410 wnd->zone_end = lcn + len; in wnd_zone_set() [all …]
|
D | file.c | 119 CLST vcn, lcn, clen; in ntfs_extend_initialized_size() local 125 err = attr_data_get_block(ni, vcn, 0, &lcn, &clen, in ntfs_extend_initialized_size() 130 if (lcn == SPARSE_LCN) { in ntfs_extend_initialized_size() 383 CLST lcn, len; in ntfs_file_mmap() local 389 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap() 678 CLST lcn, clen; in ntfs_fallocate() local 689 &lcn, &clen, &new); in ntfs_fallocate() 922 CLST lcn, clen; in ntfs_compress_write() local 928 err = attr_data_get_block(ni, frame << NTFS_LZNT_CUNIT, 0, &lcn, in ntfs_compress_write() 933 if (lcn == SPARSE_LCN) { in ntfs_compress_write()
|
D | super.c | 894 CLST vcn, lcn, len; in ntfs_fill_super() local 1075 for (i = 0; run_get_entry(&ni->file.run, i, &vcn, &lcn, &len); i++) { in ntfs_fill_super() 1076 if (lcn == SPARSE_LCN) in ntfs_fill_super() 1286 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len) in ntfs_unmap_meta() argument 1290 sector_t devblock = (u64)lcn * sbi->blocks_per_cluster; in ntfs_unmap_meta() 1315 int ntfs_discard(struct ntfs_sb_info *sbi, CLST lcn, CLST len) in ntfs_discard() argument 1321 if (sbi->used.next_free_lcn == lcn + len) in ntfs_discard() 1322 sbi->used.next_free_lcn = lcn; in ntfs_discard() 1330 lbo = (u64)lcn << sbi->cluster_bits; in ntfs_discard()
|
D | ntfs_fs.h | 417 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, 428 int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn, 592 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, 647 void mark_as_free_ex(struct ntfs_sb_info *sbi, CLST lcn, CLST len, bool trim); 778 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, 784 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, 788 CLST *lcn, CLST *len); 809 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len);
|
D | frecord.c | 1867 CLST lcn, clen; in ni_fiemap() local 1917 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap() 1921 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && in ni_fiemap() 1941 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap() 1954 if (lcn == SPARSE_LCN) { in ni_fiemap() 1978 lbo = (u64)lcn << cluster_bits; in ni_fiemap() 2134 CLST vcn, cend, lcn, clen, end; in ni_decompress_file() local 2182 err = attr_data_get_block(ni, vcn, cend - vcn, &lcn, in ni_decompress_file()
|
D | fslog.c | 228 __le64 lcn; member 3739 CLST rno, lcn, lcn0, len0, clen; in log_replay() local 4527 u64 lcn0 = le64_to_cpu(r->lcn); in log_replay() 4826 lcn = le64_to_cpu(dp->page_lcns[i]); in log_replay() 4829 lcn0 != lcn) && in log_replay() 4830 !run_add_entry(oa->run1, vcn, lcn, 1, false)) { in log_replay() 4920 if (!run_lookup_entry(oa->run1, vcn, &lcn, NULL, NULL) || in log_replay() 4921 lcn == SPARSE_LCN) { in log_replay() 5044 &lcn, &clen, NULL); in log_replay() 5051 if (is_mapped && lcn != SPARSE_LCN && clen >= t16) in log_replay()
|
D | index.c | 379 CLST lcn, clen, vcn, vcn_next; in scan_nres_bitmap() local 393 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 408 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 415 block = (sector_t)lcn * sbi->blocks_per_cluster; in scan_nres_bitmap() 451 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && vcn == vcn_next; in scan_nres_bitmap()
|
D | inode.c | 557 CLST vcn, lcn, len; in ntfs_get_block_vbo() local 590 err = attr_data_get_block(ni, vcn, 1, &lcn, &len, create ? &new : NULL); in ntfs_get_block_vbo() 599 if (lcn == SPARSE_LCN) { in ntfs_get_block_vbo() 614 lbo = ((u64)lcn << cluster_bits) + off; in ntfs_get_block_vbo()
|
/fs/ntfs/ |
D | runlist.c | 148 if ((dst->lcn == LCN_RL_NOT_MAPPED) && (src->lcn == LCN_RL_NOT_MAPPED)) in ntfs_are_rl_mergeable() 154 if ((dst->lcn >= 0) && (src->lcn >= 0) && in ntfs_are_rl_mergeable() 155 ((dst->lcn + dst->length) == src->lcn)) in ntfs_are_rl_mergeable() 158 if ((dst->lcn == LCN_HOLE) && (src->lcn == LCN_HOLE)) in ntfs_are_rl_mergeable() 241 if (dst[marker].lcn == LCN_ENOENT) in ntfs_rl_append() 327 if (dst[marker].lcn == LCN_HOLE || dst[marker].lcn == LCN_RL_NOT_MAPPED) in ntfs_rl_insert() 339 dst[loc].lcn = LCN_RL_NOT_MAPPED; in ntfs_rl_insert() 427 if (dsize - tail > 0 && dst[marker].lcn == LCN_ENOENT) in ntfs_rl_replace() 557 drl[0].lcn = LCN_RL_NOT_MAPPED; in ntfs_runlists_merge() 566 while (srl[si].length && srl[si].lcn < LCN_HOLE) in ntfs_runlists_merge() [all …]
|
D | lcnalloc.c | 48 if (rl->lcn < 0) in ntfs_cluster_free_from_rl_nolock() 50 err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length); in ntfs_cluster_free_from_rl_nolock() 137 LCN zone_start, zone_end, bmp_pos, bmp_initial_pos, last_read_pos, lcn; in ntfs_cluster_alloc() local 284 lcn = bmp_pos & 7; in ntfs_cluster_alloc() 288 (unsigned long long)lcn, in ntfs_cluster_alloc() 290 while (lcn < buf_size && lcn + bmp_pos < zone_end) { in ntfs_cluster_alloc() 291 byte = buf + (lcn >> 3); in ntfs_cluster_alloc() 296 (unsigned long long)lcn, in ntfs_cluster_alloc() 299 (unsigned int)(lcn >> 3), in ntfs_cluster_alloc() 303 lcn = (lcn + 8) & ~(LCN)7; in ntfs_cluster_alloc() [all …]
|
D | aops.c | 168 LCN lcn; in ntfs_read_block() local 255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block() 257 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_block() 259 if (lcn >= 0) { in ntfs_read_block() 261 bh->b_blocknr = ((lcn << vol->cluster_size_bits) in ntfs_read_block() 273 if (lcn == LCN_HOLE) in ntfs_read_block() 276 if (!is_retry && lcn == LCN_RL_NOT_MAPPED) { in ntfs_read_block() 294 if (err == -ENOENT || lcn == LCN_ENOENT) { in ntfs_read_block() 533 LCN lcn; in ntfs_write_block() local 711 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_write_block() [all …]
|
D | debug.c | 135 LCN lcn = (rl + i)->lcn; in ntfs_debug_dump_runlist() local 137 if (lcn < (LCN)0) { in ntfs_debug_dump_runlist() 138 int index = -lcn - 1; in ntfs_debug_dump_runlist() 150 (long long)(rl + i)->lcn, in ntfs_debug_dump_runlist()
|
D | file.c | 571 LCN lcn; in ntfs_prepare_pages_for_non_resident_write() local 625 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write() 706 BUG_ON(lcn < 0); in ntfs_prepare_pages_for_non_resident_write() 835 lcn = ntfs_rl_vcn_to_lcn(rl, bh_cpos); in ntfs_prepare_pages_for_non_resident_write() 836 if (likely(lcn >= 0)) { in ntfs_prepare_pages_for_non_resident_write() 844 lcn_block = lcn << (vol->cluster_size_bits - in ntfs_prepare_pages_for_non_resident_write() 865 lcn = LCN_RL_NOT_MAPPED; in ntfs_prepare_pages_for_non_resident_write() 870 if (unlikely(lcn != LCN_HOLE && lcn != LCN_ENOENT)) { in ntfs_prepare_pages_for_non_resident_write() 871 if (likely(!is_retry && lcn == LCN_RL_NOT_MAPPED)) { in ntfs_prepare_pages_for_non_resident_write() 898 lcn = LCN_ENOENT; in ntfs_prepare_pages_for_non_resident_write() [all …]
|
D | attrib.c | 330 LCN lcn; in ntfs_attr_vcn_to_lcn_nolock() local 350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock() 351 if (likely(lcn >= LCN_HOLE)) { in ntfs_attr_vcn_to_lcn_nolock() 352 ntfs_debug("Done, lcn 0x%llx.", (long long)lcn); in ntfs_attr_vcn_to_lcn_nolock() 353 return lcn; in ntfs_attr_vcn_to_lcn_nolock() 355 if (lcn != LCN_RL_NOT_MAPPED) { in ntfs_attr_vcn_to_lcn_nolock() 356 if (lcn != LCN_ENOENT) in ntfs_attr_vcn_to_lcn_nolock() 357 lcn = LCN_EIO; in ntfs_attr_vcn_to_lcn_nolock() 381 lcn = LCN_ENOENT; in ntfs_attr_vcn_to_lcn_nolock() 383 lcn = LCN_ENOMEM; in ntfs_attr_vcn_to_lcn_nolock() [all …]
|
D | mft.c | 526 LCN lcn; in ntfs_sync_mft_mirror() local 548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror() 550 if (likely(lcn >= 0)) { in ntfs_sync_mft_mirror() 552 bh->b_blocknr = ((lcn << in ntfs_sync_mft_mirror() 563 (long long)lcn); in ntfs_sync_mft_mirror() 719 LCN lcn; in write_mft_record_nolock() local 736 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in write_mft_record_nolock() 738 if (likely(lcn >= 0)) { in write_mft_record_nolock() 740 bh->b_blocknr = ((lcn << in write_mft_record_nolock() 750 ni->mft_no, (long long)lcn); in write_mft_record_nolock() [all …]
|
D | logfile.c | 761 LCN lcn; in ntfs_empty_logfile() local 769 lcn = rl->lcn; in ntfs_empty_logfile() 770 if (unlikely(lcn == LCN_RL_NOT_MAPPED)) { in ntfs_empty_logfile() 775 if (unlikely(!rl->length || lcn < LCN_HOLE)) in ntfs_empty_logfile() 778 if (lcn == LCN_HOLE) in ntfs_empty_logfile() 780 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_empty_logfile() 784 end_block = (lcn + len) << vol->cluster_size_bits >> in ntfs_empty_logfile()
|
D | compress.c | 479 LCN lcn; in ntfs_read_compressed_block() local 609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block() 611 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_compressed_block() 614 (unsigned long long)lcn); in ntfs_read_compressed_block() 615 if (lcn < 0) { in ntfs_read_compressed_block() 620 if (lcn == LCN_HOLE) in ntfs_read_compressed_block() 622 if (is_retry || lcn != LCN_RL_NOT_MAPPED) in ntfs_read_compressed_block() 634 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_read_compressed_block()
|
D | runlist.h | 30 LCN lcn; /* lcn = Starting logical cluster number. */ member
|
D | super.c | 1167 rl2[0].lcn = vol->mftmirr_lcn; in check_mft_mirror() 1171 rl2[1].lcn = LCN_ENOENT; in check_mft_mirror() 1183 if (rl2[i].vcn != rl[i].vcn || rl2[i].lcn != rl[i].lcn || in check_mft_mirror()
|
/fs/erofs/ |
D | zmap.c | 114 unsigned long lcn; member 155 unsigned long lcn) in legacy_load_cluster_from_disk() argument 163 lcn * sizeof(struct z_erofs_vle_decompressed_index); in legacy_load_cluster_from_disk() 172 m->lcn = lcn; in legacy_load_cluster_from_disk() 354 unsigned long lcn, bool lookahead) in compacted_load_cluster_from_disk() argument 367 if (lcn >= totalidx) in compacted_load_cluster_from_disk() 370 m->lcn = lcn; in compacted_load_cluster_from_disk() 383 if (lcn < compacted_4b_initial) { in compacted_load_cluster_from_disk() 388 lcn -= compacted_4b_initial; in compacted_load_cluster_from_disk() 390 if (lcn < compacted_2b) { in compacted_load_cluster_from_disk() [all …]
|