Home
last modified time | relevance | path

Searched refs:lcn (Results 1 – 23 of 23) sorted by relevance

/fs/ntfs3/
Drun.c23 CLST lcn; /* Logical cluster number. */ member
124 if (n->lcn != SPARSE_LCN) in run_consolidate()
125 n->lcn += dl; in run_consolidate()
133 if ((n->lcn == SPARSE_LCN) != (r->lcn == SPARSE_LCN)) { in run_consolidate()
144 if (n->lcn != SPARSE_LCN && n->lcn != r->lcn + r->len) in run_consolidate()
192 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, in run_lookup_entry() argument
215 *lcn = r->lcn == SPARSE_LCN ? SPARSE_LCN : (r->lcn + gap); in run_lookup_entry()
241 if (r->lcn != SPARSE_LCN) in run_truncate_head()
242 r->lcn += dlen; in run_truncate_head()
316 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, in run_add_entry() argument
[all …]
Dfsntfs.c352 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, in ntfs_look_for_free_space() argument
380 lcn = wnd_zone_bit(wnd); in ntfs_look_for_free_space()
383 wnd_zone_set(wnd, lcn + alen, zlen - alen); in ntfs_look_for_free_space()
385 err = wnd_set_used(wnd, lcn, alen); in ntfs_look_for_free_space()
389 alcn = lcn; in ntfs_look_for_free_space()
396 if (!lcn) in ntfs_look_for_free_space()
397 lcn = sbi->used.next_free_lcn; in ntfs_look_for_free_space()
399 if (lcn >= wnd->nbits) in ntfs_look_for_free_space()
400 lcn = 0; in ntfs_look_for_free_space()
402 alen = wnd_find(wnd, len, lcn, BITMAP_FIND_MARK_AS_USED, &alcn); in ntfs_look_for_free_space()
[all …]
Dattrib.c97 CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0; in run_deallocate_ex() local
103 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex()
119 if (lcn != SPARSE_LCN) { in run_deallocate_ex()
122 mark_as_free_ex(sbi, lcn, clen, trim); in run_deallocate_ex()
132 if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) || in run_deallocate_ex()
150 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, in attr_allocate_clusters() argument
159 err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen, in attr_allocate_clusters()
175 *new_lcn = lcn; in attr_allocate_clusters()
181 if (!run_add_entry(run, vcn, lcn, flen, opt & ALLOCATE_MFT)) { in attr_allocate_clusters()
183 mark_as_free_ex(sbi, lcn, len, false); in attr_allocate_clusters()
[all …]
Dfile.c120 CLST vcn, lcn, clen; in ntfs_extend_initialized_size() local
126 err = attr_data_get_block(ni, vcn, 1, &lcn, &clen, NULL, in ntfs_extend_initialized_size()
131 if (lcn == SPARSE_LCN) { in ntfs_extend_initialized_size()
288 CLST lcn, len; in ntfs_file_mmap() local
294 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap()
607 CLST lcn, clen; in ntfs_fallocate() local
619 &lcn, &clen, &new, in ntfs_fallocate()
629 &lcn, &clen, &new, in ntfs_fallocate()
886 CLST lcn, clen; in ntfs_compress_write() local
892 err = attr_data_get_block(ni, frame << NTFS_LZNT_CUNIT, 1, &lcn, in ntfs_compress_write()
[all …]
Dbitmap.c510 CLST lcn, clen; in wnd_rescan() local
546 &lcn, &clen, NULL)) { in wnd_rescan()
551 lbo = ((u64)lcn << cluster_bits) + off; in wnd_rescan()
683 CLST lcn, clen; in wnd_map() local
692 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen, in wnd_map()
697 lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask); in wnd_map()
1416 void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len) in wnd_zone_set() argument
1424 wnd_remove_free_ext(wnd, lcn, len); in wnd_zone_set()
1426 wnd->zone_bit = lcn; in wnd_zone_set()
1427 wnd->zone_end = lcn + len; in wnd_zone_set()
[all …]
Dsuper.c1157 CLST vcn, lcn, len; in ntfs_fill_super() local
1377 for (i = 0; run_get_entry(&ni->file.run, i, &vcn, &lcn, &len); i++) { in ntfs_fill_super()
1378 if (lcn == SPARSE_LCN) in ntfs_fill_super()
1386 if (wnd_set_used_safe(&sbi->used.bitmap, lcn, len, &tt) || tt) { in ntfs_fill_super()
1627 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len) in ntfs_unmap_meta() argument
1631 sector_t devblock = (u64)lcn * sbi->blocks_per_cluster; in ntfs_unmap_meta()
1656 int ntfs_discard(struct ntfs_sb_info *sbi, CLST lcn, CLST len) in ntfs_discard() argument
1662 if (sbi->used.next_free_lcn == lcn + len) in ntfs_discard()
1663 sbi->used.next_free_lcn = lcn; in ntfs_discard()
1671 lbo = (u64)lcn << sbi->cluster_bits; in ntfs_discard()
Dntfs_fs.h425 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc,
436 int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn,
596 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len,
654 void mark_as_free_ex(struct ntfs_sb_info *sbi, CLST lcn, CLST len, bool trim);
792 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn,
797 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len,
802 CLST *lcn, CLST *len);
824 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len);
Dfrecord.c1958 CLST lcn, clen; in ni_fiemap() local
2020 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap()
2024 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && in ni_fiemap()
2044 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in ni_fiemap()
2057 if (lcn == SPARSE_LCN) { in ni_fiemap()
2081 lbo = (u64)lcn << cluster_bits; in ni_fiemap()
2251 CLST vcn, cend, lcn, clen, end; in ni_decompress_file() local
2299 err = attr_data_get_block(ni, vcn, cend - vcn, &lcn, in ni_decompress_file()
Dfslog.c228 __le64 lcn; member
3740 CLST rno, lcn, lcn0, len0, clen; in log_replay() local
4540 u64 lcn0 = le64_to_cpu(r->lcn); in log_replay()
4839 lcn = le64_to_cpu(dp->page_lcns[i]); in log_replay()
4842 lcn0 != lcn) && in log_replay()
4843 !run_add_entry(oa->run1, vcn, lcn, 1, false)) { in log_replay()
4932 if (!run_lookup_entry(oa->run1, vcn, &lcn, NULL, NULL) || in log_replay()
4933 lcn == SPARSE_LCN) { in log_replay()
5056 &lcn, &clen, NULL); in log_replay()
5063 if (is_mapped && lcn != SPARSE_LCN && clen >= t16) in log_replay()
Dindex.c379 CLST lcn, clen, vcn, vcn_next; in scan_nres_bitmap() local
393 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap()
408 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap()
415 block = (sector_t)lcn * sbi->blocks_per_cluster; in scan_nres_bitmap()
452 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && vcn == vcn_next; in scan_nres_bitmap()
Dinode.c566 CLST vcn, lcn, len; in ntfs_get_block_vbo() local
593 err = attr_data_get_block(ni, vcn, 1, &lcn, &len, create ? &new : NULL, in ntfs_get_block_vbo()
603 if (lcn == SPARSE_LCN) { in ntfs_get_block_vbo()
615 lbo = ((u64)lcn << cluster_bits) + off; in ntfs_get_block_vbo()
/fs/ntfs/
Drunlist.c148 if ((dst->lcn == LCN_RL_NOT_MAPPED) && (src->lcn == LCN_RL_NOT_MAPPED)) in ntfs_are_rl_mergeable()
154 if ((dst->lcn >= 0) && (src->lcn >= 0) && in ntfs_are_rl_mergeable()
155 ((dst->lcn + dst->length) == src->lcn)) in ntfs_are_rl_mergeable()
158 if ((dst->lcn == LCN_HOLE) && (src->lcn == LCN_HOLE)) in ntfs_are_rl_mergeable()
241 if (dst[marker].lcn == LCN_ENOENT) in ntfs_rl_append()
327 if (dst[marker].lcn == LCN_HOLE || dst[marker].lcn == LCN_RL_NOT_MAPPED) in ntfs_rl_insert()
339 dst[loc].lcn = LCN_RL_NOT_MAPPED; in ntfs_rl_insert()
427 if (dsize - tail > 0 && dst[marker].lcn == LCN_ENOENT) in ntfs_rl_replace()
557 drl[0].lcn = LCN_RL_NOT_MAPPED; in ntfs_runlists_merge()
566 while (srl[si].length && srl[si].lcn < LCN_HOLE) in ntfs_runlists_merge()
[all …]
Dlcnalloc.c48 if (rl->lcn < 0) in ntfs_cluster_free_from_rl_nolock()
50 err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length); in ntfs_cluster_free_from_rl_nolock()
137 LCN zone_start, zone_end, bmp_pos, bmp_initial_pos, last_read_pos, lcn; in ntfs_cluster_alloc() local
284 lcn = bmp_pos & 7; in ntfs_cluster_alloc()
288 (unsigned long long)lcn, in ntfs_cluster_alloc()
290 while (lcn < buf_size && lcn + bmp_pos < zone_end) { in ntfs_cluster_alloc()
291 byte = buf + (lcn >> 3); in ntfs_cluster_alloc()
296 (unsigned long long)lcn, in ntfs_cluster_alloc()
299 (unsigned int)(lcn >> 3), in ntfs_cluster_alloc()
303 lcn = (lcn + 8) & ~(LCN)7; in ntfs_cluster_alloc()
[all …]
Daops.c168 LCN lcn; in ntfs_read_block() local
255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block()
257 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_block()
259 if (lcn >= 0) { in ntfs_read_block()
261 bh->b_blocknr = ((lcn << vol->cluster_size_bits) in ntfs_read_block()
273 if (lcn == LCN_HOLE) in ntfs_read_block()
276 if (!is_retry && lcn == LCN_RL_NOT_MAPPED) { in ntfs_read_block()
294 if (err == -ENOENT || lcn == LCN_ENOENT) { in ntfs_read_block()
534 LCN lcn; in ntfs_write_block() local
713 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_write_block()
[all …]
Ddebug.c135 LCN lcn = (rl + i)->lcn; in ntfs_debug_dump_runlist() local
137 if (lcn < (LCN)0) { in ntfs_debug_dump_runlist()
138 int index = -lcn - 1; in ntfs_debug_dump_runlist()
150 (long long)(rl + i)->lcn, in ntfs_debug_dump_runlist()
Dfile.c567 LCN lcn; in ntfs_prepare_pages_for_non_resident_write() local
621 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write()
702 BUG_ON(lcn < 0); in ntfs_prepare_pages_for_non_resident_write()
831 lcn = ntfs_rl_vcn_to_lcn(rl, bh_cpos); in ntfs_prepare_pages_for_non_resident_write()
832 if (likely(lcn >= 0)) { in ntfs_prepare_pages_for_non_resident_write()
840 lcn_block = lcn << (vol->cluster_size_bits - in ntfs_prepare_pages_for_non_resident_write()
861 lcn = LCN_RL_NOT_MAPPED; in ntfs_prepare_pages_for_non_resident_write()
866 if (unlikely(lcn != LCN_HOLE && lcn != LCN_ENOENT)) { in ntfs_prepare_pages_for_non_resident_write()
867 if (likely(!is_retry && lcn == LCN_RL_NOT_MAPPED)) { in ntfs_prepare_pages_for_non_resident_write()
894 lcn = LCN_ENOENT; in ntfs_prepare_pages_for_non_resident_write()
[all …]
Dattrib.c330 LCN lcn; in ntfs_attr_vcn_to_lcn_nolock() local
350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock()
351 if (likely(lcn >= LCN_HOLE)) { in ntfs_attr_vcn_to_lcn_nolock()
352 ntfs_debug("Done, lcn 0x%llx.", (long long)lcn); in ntfs_attr_vcn_to_lcn_nolock()
353 return lcn; in ntfs_attr_vcn_to_lcn_nolock()
355 if (lcn != LCN_RL_NOT_MAPPED) { in ntfs_attr_vcn_to_lcn_nolock()
356 if (lcn != LCN_ENOENT) in ntfs_attr_vcn_to_lcn_nolock()
357 lcn = LCN_EIO; in ntfs_attr_vcn_to_lcn_nolock()
381 lcn = LCN_ENOENT; in ntfs_attr_vcn_to_lcn_nolock()
383 lcn = LCN_ENOMEM; in ntfs_attr_vcn_to_lcn_nolock()
[all …]
Dmft.c526 LCN lcn; in ntfs_sync_mft_mirror() local
548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror()
550 if (likely(lcn >= 0)) { in ntfs_sync_mft_mirror()
552 bh->b_blocknr = ((lcn << in ntfs_sync_mft_mirror()
563 (long long)lcn); in ntfs_sync_mft_mirror()
719 LCN lcn; in write_mft_record_nolock() local
736 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in write_mft_record_nolock()
738 if (likely(lcn >= 0)) { in write_mft_record_nolock()
740 bh->b_blocknr = ((lcn << in write_mft_record_nolock()
750 ni->mft_no, (long long)lcn); in write_mft_record_nolock()
[all …]
Dlogfile.c761 LCN lcn; in ntfs_empty_logfile() local
769 lcn = rl->lcn; in ntfs_empty_logfile()
770 if (unlikely(lcn == LCN_RL_NOT_MAPPED)) { in ntfs_empty_logfile()
775 if (unlikely(!rl->length || lcn < LCN_HOLE)) in ntfs_empty_logfile()
778 if (lcn == LCN_HOLE) in ntfs_empty_logfile()
780 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_empty_logfile()
784 end_block = (lcn + len) << vol->cluster_size_bits >> in ntfs_empty_logfile()
Dcompress.c479 LCN lcn; in ntfs_read_compressed_block() local
609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block()
611 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_compressed_block()
614 (unsigned long long)lcn); in ntfs_read_compressed_block()
615 if (lcn < 0) { in ntfs_read_compressed_block()
620 if (lcn == LCN_HOLE) in ntfs_read_compressed_block()
622 if (is_retry || lcn != LCN_RL_NOT_MAPPED) in ntfs_read_compressed_block()
634 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_read_compressed_block()
Drunlist.h30 LCN lcn; /* lcn = Starting logical cluster number. */ member
Dsuper.c1175 rl2[0].lcn = vol->mftmirr_lcn; in check_mft_mirror()
1179 rl2[1].lcn = LCN_ENOENT; in check_mft_mirror()
1191 if (rl2[i].vcn != rl[i].vcn || rl2[i].lcn != rl[i].lcn || in check_mft_mirror()
/fs/erofs/
Dzmap.c15 unsigned long lcn; member
26 unsigned long lcn) in z_erofs_load_full_lcluster() argument
32 lcn * sizeof(struct z_erofs_lcluster_index); in z_erofs_load_full_lcluster()
42 m->lcn = lcn; in z_erofs_load_full_lcluster()
226 unsigned long lcn, bool lookahead) in z_erofs_load_compact_lcluster() argument
237 if (lcn >= totalidx) in z_erofs_load_compact_lcluster()
240 m->lcn = lcn; in z_erofs_load_compact_lcluster()
253 if (lcn < compacted_4b_initial) { in z_erofs_load_compact_lcluster()
258 lcn -= compacted_4b_initial; in z_erofs_load_compact_lcluster()
260 if (lcn < compacted_2b) { in z_erofs_load_compact_lcluster()
[all …]