Home
last modified time | relevance | path

Searched refs:lcn (Results 1 – 23 of 23) sorted by relevance

/fs/ntfs3/
Drun.c23 CLST lcn; /* Logical cluster number. */ member
124 if (n->lcn != SPARSE_LCN) in run_consolidate()
125 n->lcn += dl; in run_consolidate()
133 if ((n->lcn == SPARSE_LCN) != (r->lcn == SPARSE_LCN)) { in run_consolidate()
144 if (n->lcn != SPARSE_LCN && n->lcn != r->lcn + r->len) in run_consolidate()
192 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, in run_lookup_entry() argument
215 *lcn = r->lcn == SPARSE_LCN ? SPARSE_LCN : (r->lcn + gap); in run_lookup_entry()
241 if (r->lcn != SPARSE_LCN) in run_truncate_head()
242 r->lcn += dlen; in run_truncate_head()
316 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, in run_add_entry() argument
[all …]
Dfsntfs.c356 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, in ntfs_look_for_free_space() argument
384 lcn = wnd_zone_bit(wnd); in ntfs_look_for_free_space()
387 wnd_zone_set(wnd, lcn + alen, zlen - alen); in ntfs_look_for_free_space()
389 err = wnd_set_used(wnd, lcn, alen); in ntfs_look_for_free_space()
393 alcn = lcn; in ntfs_look_for_free_space()
400 if (!lcn) in ntfs_look_for_free_space()
401 lcn = sbi->used.next_free_lcn; in ntfs_look_for_free_space()
403 if (lcn >= wnd->nbits) in ntfs_look_for_free_space()
404 lcn = 0; in ntfs_look_for_free_space()
406 alen = wnd_find(wnd, len, lcn, BITMAP_FIND_MARK_AS_USED, &alcn); in ntfs_look_for_free_space()
[all …]
Dattrib.c124 CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0; in run_deallocate_ex() local
130 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex()
146 if (lcn != SPARSE_LCN) { in run_deallocate_ex()
147 mark_as_free_ex(sbi, lcn, clen, trim); in run_deallocate_ex()
156 if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) || in run_deallocate_ex()
174 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, in attr_allocate_clusters() argument
184 err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen, in attr_allocate_clusters()
198 *new_lcn = lcn; in attr_allocate_clusters()
201 if (!run_add_entry(run, vcn, lcn, flen, opt == ALLOCATE_MFT)) { in attr_allocate_clusters()
204 wnd_set_free(wnd, lcn, flen); in attr_allocate_clusters()
[all …]
Dbitmap.c515 CLST lcn, clen; in wnd_rescan() local
551 &lcn, &clen, NULL)) { in wnd_rescan()
556 lbo = ((u64)lcn << cluster_bits) + off; in wnd_rescan()
690 CLST lcn, clen; in wnd_map() local
699 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen, in wnd_map()
704 lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask); in wnd_map()
1398 void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len) in wnd_zone_set() argument
1407 wnd_remove_free_ext(wnd, lcn, len); in wnd_zone_set()
1409 wnd->zone_bit = lcn; in wnd_zone_set()
1410 wnd->zone_end = lcn + len; in wnd_zone_set()
[all …]
Dfile.c119 CLST vcn, lcn, clen; in ntfs_extend_initialized_size() local
125 err = attr_data_get_block(ni, vcn, 0, &lcn, &clen, in ntfs_extend_initialized_size()
130 if (lcn == SPARSE_LCN) { in ntfs_extend_initialized_size()
383 CLST lcn, len; in ntfs_file_mmap() local
389 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap()
678 CLST lcn, clen; in ntfs_fallocate() local
689 &lcn, &clen, &new); in ntfs_fallocate()
922 CLST lcn, clen; in ntfs_compress_write() local
928 err = attr_data_get_block(ni, frame << NTFS_LZNT_CUNIT, 0, &lcn, in ntfs_compress_write()
933 if (lcn == SPARSE_LCN) { in ntfs_compress_write()
Dsuper.c894 CLST vcn, lcn, len; in ntfs_fill_super() local
1075 for (i = 0; run_get_entry(&ni->file.run, i, &vcn, &lcn, &len); i++) { in ntfs_fill_super()
1076 if (lcn == SPARSE_LCN) in ntfs_fill_super()
1286 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len) in ntfs_unmap_meta() argument
1290 sector_t devblock = (u64)lcn * sbi->blocks_per_cluster; in ntfs_unmap_meta()
1315 int ntfs_discard(struct ntfs_sb_info *sbi, CLST lcn, CLST len) in ntfs_discard() argument
1321 if (sbi->used.next_free_lcn == lcn + len) in ntfs_discard()
1322 sbi->used.next_free_lcn = lcn; in ntfs_discard()
1330 lbo = (u64)lcn << sbi->cluster_bits; in ntfs_discard()
Dntfs_fs.h417 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc,
428 int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn,
592 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len,
647 void mark_as_free_ex(struct ntfs_sb_info *sbi, CLST lcn, CLST len, bool trim);
778 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn,
784 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len,
788 CLST *lcn, CLST *len);
809 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len);
Dfrecord.c1867 CLST lcn, clen; in ni_fiemap() local
1917 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap()
1921 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && in ni_fiemap()
1941 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap()
1954 if (lcn == SPARSE_LCN) { in ni_fiemap()
1978 lbo = (u64)lcn << cluster_bits; in ni_fiemap()
2134 CLST vcn, cend, lcn, clen, end; in ni_decompress_file() local
2182 err = attr_data_get_block(ni, vcn, cend - vcn, &lcn, in ni_decompress_file()
Dfslog.c228 __le64 lcn; member
3739 CLST rno, lcn, lcn0, len0, clen; in log_replay() local
4527 u64 lcn0 = le64_to_cpu(r->lcn); in log_replay()
4826 lcn = le64_to_cpu(dp->page_lcns[i]); in log_replay()
4829 lcn0 != lcn) && in log_replay()
4830 !run_add_entry(oa->run1, vcn, lcn, 1, false)) { in log_replay()
4920 if (!run_lookup_entry(oa->run1, vcn, &lcn, NULL, NULL) || in log_replay()
4921 lcn == SPARSE_LCN) { in log_replay()
5044 &lcn, &clen, NULL); in log_replay()
5051 if (is_mapped && lcn != SPARSE_LCN && clen >= t16) in log_replay()
Dindex.c379 CLST lcn, clen, vcn, vcn_next; in scan_nres_bitmap() local
393 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap()
408 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap()
415 block = (sector_t)lcn * sbi->blocks_per_cluster; in scan_nres_bitmap()
451 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && vcn == vcn_next; in scan_nres_bitmap()
Dinode.c557 CLST vcn, lcn, len; in ntfs_get_block_vbo() local
590 err = attr_data_get_block(ni, vcn, 1, &lcn, &len, create ? &new : NULL); in ntfs_get_block_vbo()
599 if (lcn == SPARSE_LCN) { in ntfs_get_block_vbo()
614 lbo = ((u64)lcn << cluster_bits) + off; in ntfs_get_block_vbo()
/fs/ntfs/
Drunlist.c148 if ((dst->lcn == LCN_RL_NOT_MAPPED) && (src->lcn == LCN_RL_NOT_MAPPED)) in ntfs_are_rl_mergeable()
154 if ((dst->lcn >= 0) && (src->lcn >= 0) && in ntfs_are_rl_mergeable()
155 ((dst->lcn + dst->length) == src->lcn)) in ntfs_are_rl_mergeable()
158 if ((dst->lcn == LCN_HOLE) && (src->lcn == LCN_HOLE)) in ntfs_are_rl_mergeable()
241 if (dst[marker].lcn == LCN_ENOENT) in ntfs_rl_append()
327 if (dst[marker].lcn == LCN_HOLE || dst[marker].lcn == LCN_RL_NOT_MAPPED) in ntfs_rl_insert()
339 dst[loc].lcn = LCN_RL_NOT_MAPPED; in ntfs_rl_insert()
427 if (dsize - tail > 0 && dst[marker].lcn == LCN_ENOENT) in ntfs_rl_replace()
557 drl[0].lcn = LCN_RL_NOT_MAPPED; in ntfs_runlists_merge()
566 while (srl[si].length && srl[si].lcn < LCN_HOLE) in ntfs_runlists_merge()
[all …]
Dlcnalloc.c48 if (rl->lcn < 0) in ntfs_cluster_free_from_rl_nolock()
50 err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length); in ntfs_cluster_free_from_rl_nolock()
137 LCN zone_start, zone_end, bmp_pos, bmp_initial_pos, last_read_pos, lcn; in ntfs_cluster_alloc() local
284 lcn = bmp_pos & 7; in ntfs_cluster_alloc()
288 (unsigned long long)lcn, in ntfs_cluster_alloc()
290 while (lcn < buf_size && lcn + bmp_pos < zone_end) { in ntfs_cluster_alloc()
291 byte = buf + (lcn >> 3); in ntfs_cluster_alloc()
296 (unsigned long long)lcn, in ntfs_cluster_alloc()
299 (unsigned int)(lcn >> 3), in ntfs_cluster_alloc()
303 lcn = (lcn + 8) & ~(LCN)7; in ntfs_cluster_alloc()
[all …]
Daops.c168 LCN lcn; in ntfs_read_block() local
255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block()
257 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_block()
259 if (lcn >= 0) { in ntfs_read_block()
261 bh->b_blocknr = ((lcn << vol->cluster_size_bits) in ntfs_read_block()
273 if (lcn == LCN_HOLE) in ntfs_read_block()
276 if (!is_retry && lcn == LCN_RL_NOT_MAPPED) { in ntfs_read_block()
294 if (err == -ENOENT || lcn == LCN_ENOENT) { in ntfs_read_block()
533 LCN lcn; in ntfs_write_block() local
711 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_write_block()
[all …]
Ddebug.c135 LCN lcn = (rl + i)->lcn; in ntfs_debug_dump_runlist() local
137 if (lcn < (LCN)0) { in ntfs_debug_dump_runlist()
138 int index = -lcn - 1; in ntfs_debug_dump_runlist()
150 (long long)(rl + i)->lcn, in ntfs_debug_dump_runlist()
Dfile.c571 LCN lcn; in ntfs_prepare_pages_for_non_resident_write() local
625 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write()
706 BUG_ON(lcn < 0); in ntfs_prepare_pages_for_non_resident_write()
835 lcn = ntfs_rl_vcn_to_lcn(rl, bh_cpos); in ntfs_prepare_pages_for_non_resident_write()
836 if (likely(lcn >= 0)) { in ntfs_prepare_pages_for_non_resident_write()
844 lcn_block = lcn << (vol->cluster_size_bits - in ntfs_prepare_pages_for_non_resident_write()
865 lcn = LCN_RL_NOT_MAPPED; in ntfs_prepare_pages_for_non_resident_write()
870 if (unlikely(lcn != LCN_HOLE && lcn != LCN_ENOENT)) { in ntfs_prepare_pages_for_non_resident_write()
871 if (likely(!is_retry && lcn == LCN_RL_NOT_MAPPED)) { in ntfs_prepare_pages_for_non_resident_write()
898 lcn = LCN_ENOENT; in ntfs_prepare_pages_for_non_resident_write()
[all …]
Dattrib.c330 LCN lcn; in ntfs_attr_vcn_to_lcn_nolock() local
350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock()
351 if (likely(lcn >= LCN_HOLE)) { in ntfs_attr_vcn_to_lcn_nolock()
352 ntfs_debug("Done, lcn 0x%llx.", (long long)lcn); in ntfs_attr_vcn_to_lcn_nolock()
353 return lcn; in ntfs_attr_vcn_to_lcn_nolock()
355 if (lcn != LCN_RL_NOT_MAPPED) { in ntfs_attr_vcn_to_lcn_nolock()
356 if (lcn != LCN_ENOENT) in ntfs_attr_vcn_to_lcn_nolock()
357 lcn = LCN_EIO; in ntfs_attr_vcn_to_lcn_nolock()
381 lcn = LCN_ENOENT; in ntfs_attr_vcn_to_lcn_nolock()
383 lcn = LCN_ENOMEM; in ntfs_attr_vcn_to_lcn_nolock()
[all …]
Dmft.c526 LCN lcn; in ntfs_sync_mft_mirror() local
548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror()
550 if (likely(lcn >= 0)) { in ntfs_sync_mft_mirror()
552 bh->b_blocknr = ((lcn << in ntfs_sync_mft_mirror()
563 (long long)lcn); in ntfs_sync_mft_mirror()
719 LCN lcn; in write_mft_record_nolock() local
736 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in write_mft_record_nolock()
738 if (likely(lcn >= 0)) { in write_mft_record_nolock()
740 bh->b_blocknr = ((lcn << in write_mft_record_nolock()
750 ni->mft_no, (long long)lcn); in write_mft_record_nolock()
[all …]
Dlogfile.c761 LCN lcn; in ntfs_empty_logfile() local
769 lcn = rl->lcn; in ntfs_empty_logfile()
770 if (unlikely(lcn == LCN_RL_NOT_MAPPED)) { in ntfs_empty_logfile()
775 if (unlikely(!rl->length || lcn < LCN_HOLE)) in ntfs_empty_logfile()
778 if (lcn == LCN_HOLE) in ntfs_empty_logfile()
780 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_empty_logfile()
784 end_block = (lcn + len) << vol->cluster_size_bits >> in ntfs_empty_logfile()
Dcompress.c479 LCN lcn; in ntfs_read_compressed_block() local
609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block()
611 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_compressed_block()
614 (unsigned long long)lcn); in ntfs_read_compressed_block()
615 if (lcn < 0) { in ntfs_read_compressed_block()
620 if (lcn == LCN_HOLE) in ntfs_read_compressed_block()
622 if (is_retry || lcn != LCN_RL_NOT_MAPPED) in ntfs_read_compressed_block()
634 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_read_compressed_block()
Drunlist.h30 LCN lcn; /* lcn = Starting logical cluster number. */ member
Dsuper.c1167 rl2[0].lcn = vol->mftmirr_lcn; in check_mft_mirror()
1171 rl2[1].lcn = LCN_ENOENT; in check_mft_mirror()
1183 if (rl2[i].vcn != rl[i].vcn || rl2[i].lcn != rl[i].lcn || in check_mft_mirror()
/fs/erofs/
Dzmap.c114 unsigned long lcn; member
155 unsigned long lcn) in legacy_load_cluster_from_disk() argument
163 lcn * sizeof(struct z_erofs_vle_decompressed_index); in legacy_load_cluster_from_disk()
172 m->lcn = lcn; in legacy_load_cluster_from_disk()
354 unsigned long lcn, bool lookahead) in compacted_load_cluster_from_disk() argument
367 if (lcn >= totalidx) in compacted_load_cluster_from_disk()
370 m->lcn = lcn; in compacted_load_cluster_from_disk()
383 if (lcn < compacted_4b_initial) { in compacted_load_cluster_from_disk()
388 lcn -= compacted_4b_initial; in compacted_load_cluster_from_disk()
390 if (lcn < compacted_2b) { in compacted_load_cluster_from_disk()
[all …]