/fs/erofs/ |
D | namei.c | 50 int head, back; in find_target_dirent() local 56 back = ndirents - 1; in find_target_dirent() 59 while (head <= back) { in find_target_dirent() 60 const int mid = head + (back - head) / 2; in find_target_dirent() 81 back = mid - 1; in find_target_dirent() 93 int head = 0, back = erofs_iblks(dir) - 1; in erofs_find_target_block() local 97 while (head <= back) { in erofs_find_target_block() 98 const int mid = head + (back - head) / 2; in erofs_find_target_block() 134 back = mid - 1; in erofs_find_target_block()
|
/fs/xfs/libxfs/ |
D | xfs_da_btree.c | 148 to->back = be32_to_cpu(from3->hdr.info.hdr.back); in xfs_da3_node_hdr_from_disk() 156 to->back = be32_to_cpu(from->hdr.info.back); in xfs_da3_node_hdr_from_disk() 176 to3->hdr.info.hdr.back = cpu_to_be32(from->back); in xfs_da3_node_hdr_to_disk() 183 to->hdr.info.back = cpu_to_be32(from->back); in xfs_da3_node_hdr_to_disk() 619 node->hdr.info.back = cpu_to_be32(oldblk->blkno); in xfs_da3_split() 625 if (node->hdr.info.back) { in xfs_da3_split() 626 if (be32_to_cpu(node->hdr.info.back) != addblk->blkno) { in xfs_da3_split() 1172 ASSERT(!blkinfo->back); in xfs_da_blkinfo_onlychild_validate() 1203 ASSERT(oldroothdr.back == 0); in xfs_da3_root_join() 1322 forward = nodehdr.forw < nodehdr.back; in xfs_da3_node_toosmall() [all …]
|
D | xfs_dir2_leaf.c | 44 to->back = be32_to_cpu(from3->hdr.info.hdr.back); in xfs_dir2_leaf_hdr_from_disk() 54 to->back = be32_to_cpu(from->hdr.info.back); in xfs_dir2_leaf_hdr_from_disk() 78 to3->hdr.info.hdr.back = cpu_to_be32(from->back); in xfs_dir2_leaf_hdr_to_disk() 87 to->hdr.info.back = cpu_to_be32(from->back); in xfs_dir2_leaf_hdr_to_disk()
|
D | xfs_attr_leaf.c | 166 to->back = be32_to_cpu(hdr3->info.hdr.back); in xfs_attr3_leaf_hdr_from_disk() 180 to->back = be32_to_cpu(from->hdr.info.back); in xfs_attr3_leaf_hdr_from_disk() 208 hdr3->info.hdr.back = cpu_to_be32(from->back); in xfs_attr3_leaf_hdr_to_disk() 223 to->hdr.info.back = cpu_to_be32(from->back); in xfs_attr3_leaf_hdr_to_disk() 2068 forward = ichdr.forw < ichdr.back; in xfs_attr3_leaf_toosmall() 2074 blkno = ichdr.back; in xfs_attr3_leaf_toosmall() 2349 tmphdr.back = savehdr.back; in xfs_attr3_leaf_unbalance()
|
D | xfs_attr_leaf.h | 23 uint32_t back; member
|
D | xfs_da_btree.h | 148 uint32_t back; member
|
D | xfs_dir2_priv.h | 17 uint32_t back; member
|
D | xfs_da_format.h | 23 __be32 back; /* following block in list */ member
|
D | xfs_dir2_node.c | 1549 forward = leafhdr.forw < leafhdr.back; in xfs_dir2_leafn_toosmall() 1553 blkno = forward ? leafhdr.forw : leafhdr.back; in xfs_dir2_leafn_toosmall()
|
/fs/f2fs/ |
D | extent_cache.c | 127 static bool __is_extent_mergeable(struct extent_info *back, in __is_extent_mergeable() argument 132 if (back->c_len && back->len != back->c_len) in __is_extent_mergeable() 137 return (back->fofs + back->len == front->fofs && in __is_extent_mergeable() 138 back->blk + back->len == front->blk); in __is_extent_mergeable() 140 return (back->fofs + back->len == front->fofs && in __is_extent_mergeable() 141 abs(back->age - front->age) <= SAME_AGE_REGION && in __is_extent_mergeable() 142 abs(back->last_blocks - front->last_blocks) <= in __is_extent_mergeable() 149 struct extent_info *back, enum extent_type type) in __is_back_mergeable() argument 151 return __is_extent_mergeable(back, cur, type); in __is_back_mergeable()
|
D | Kconfig | 100 multiple back-end compression algorithms are supported.
|
D | f2fs.h | 877 static inline bool __is_discard_mergeable(struct discard_info *back, in __is_discard_mergeable() argument 880 return (back->lstart + back->len == front->lstart) && in __is_discard_mergeable() 881 (back->len + front->len <= max_len); in __is_discard_mergeable() 885 struct discard_info *back, unsigned int max_len) in __is_discard_back_mergeable() argument 887 return __is_discard_mergeable(back, cur, max_len); in __is_discard_back_mergeable()
|
/fs/xfs/scrub/ |
D | dabtree.c | 282 xfs_dablk_t back; in xchk_da_btree_block_check_siblings() local 286 back = be32_to_cpu(hdr->back); in xchk_da_btree_block_check_siblings() 290 if (forw != 0 || back != 0) in xchk_da_btree_block_check_siblings() 299 error = xchk_da_btree_block_check_sibling(ds, level, 0, back); in xchk_da_btree_block_check_siblings()
|
/fs/jfs/ |
D | jfs_imap.c | 849 int back, fwd; in diFree() local 1024 fwd = back = -1; in diFree() 1059 if ((back = le32_to_cpu(iagp->extfreeback)) >= 0) { in diFree() 1060 if ((rc = diIAGRead(imap, back, &bmp))) in diFree() 1086 else if (inofreefwd == back) in diFree() 1100 else if (inofreeback == back) in diFree() 1142 if (back >= 0) in diFree() 2154 int agno, iagno, fwd, back, freei = 0, sword, rc; in diNewExt() local 2199 if ((back = le32_to_cpu(iagp->extfreeback)) >= 0) { in diNewExt() 2200 if ((rc = diIAGRead(imap, back, &bmp))) in diNewExt() [all …]
|
/fs/ufs/ |
D | balloc.c | 871 int i, start, end, forw, back; in ufs_clusteracct() local 904 back = start - i; in ufs_clusteracct() 910 i = back + forw + 1; in ufs_clusteracct() 914 if (back > 0) in ufs_clusteracct() 915 fs32_sub(sb, (__fs32*)ubh_get_addr(UCPI_UBH(ucpi), ucpi->c_clustersumoff + (back << 2)), cnt); in ufs_clusteracct()
|
D | Kconfig | 36 experimental, so you should back up your UFS partitions beforehand.
|
/fs/nilfs2/ |
D | Kconfig | 17 periods until they are changed back to checkpoints. Each
|
/fs/coda/ |
D | Kconfig | 12 persistent client caches and write back caching.
|
/fs/verity/ |
D | Kconfig | 32 data each time it is paged back in, which provides better
|
/fs/pstore/ |
D | Kconfig | 159 buffer in RAM where it can be read back at some later point. 181 where it can be read back at some later point.
|
/fs/jffs2/ |
D | Kconfig | 50 This causes JFFS2 to read back every page written through the
|