/fs/ext4/ |
D | extents.c | 181 int depth = path->p_depth; in ext4_ext_find_goal() local 201 ex = path[depth].p_ext; in ext4_ext_find_goal() 214 if (path[depth].p_bh) in ext4_ext_find_goal() 215 return path[depth].p_bh->b_blocknr; in ext4_ext_find_goal() 354 ext4_ext_max_entries(struct inode *inode, int depth) in ext4_ext_max_entries() argument 358 if (depth == ext_depth(inode)) { in ext4_ext_max_entries() 359 if (depth == 0) in ext4_ext_max_entries() 364 if (depth == 0) in ext4_ext_max_entries() 395 int depth) in ext4_valid_extent_entries() argument 403 if (depth == 0) { in ext4_valid_extent_entries() [all …]
|
D | indirect.c | 142 static Indirect *ext4_get_branch(struct inode *inode, int depth, in ext4_get_branch() argument 156 while (--depth) { in ext4_get_branch() 524 int depth; in ext4_ind_map_blocks() local 531 depth = ext4_block_to_path(inode, map->m_lblk, offsets, in ext4_ind_map_blocks() 534 if (depth == 0) in ext4_ind_map_blocks() 537 partial = ext4_get_branch(inode, depth, offsets, chain, &err); in ext4_ind_map_blocks() 541 first_block = le32_to_cpu(chain[depth - 1].key); in ext4_ind_map_blocks() 547 blk = le32_to_cpu(*(chain[depth-1].p + count)); in ext4_ind_map_blocks() 583 indirect_blks = (chain + depth) - partial - 1; in ext4_ind_map_blocks() 616 map->m_pblk = le32_to_cpu(chain[depth-1].key); in ext4_ind_map_blocks() [all …]
|
/fs/minix/ |
D | itree_common.c | 30 int depth, in get_branch() argument 44 while (--depth) { in get_branch() 153 int depth = block_to_path(inode, block, offsets); in get_block() local 155 if (depth == 0) in get_block() 159 partial = get_branch(inode, depth, offsets, chain, &err); in get_block() 164 map_bh(bh, inode->i_sb, block_to_cpu(chain[depth-1].key)); in get_block() 166 partial = chain+depth-1; /* the whole chain */ in get_block() 189 left = (chain + depth) - partial; in get_block() 217 int depth, in find_shared() argument 226 for (k = depth; k > 1 && !offsets[k-1]; k--) in find_shared() [all …]
|
/fs/quota/ |
D | quota_tree.c | 25 static int get_index(struct qtree_mem_dqinfo *info, struct kqid qid, int depth) in get_index() argument 30 depth = info->dqi_qtree_depth - depth - 1; in get_index() 31 while (depth--) in get_index() 292 uint *treeblk, int depth) in do_insert_tree() argument 317 newblk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); in do_insert_tree() 320 if (depth == info->dqi_qtree_depth - 1) { in do_insert_tree() 326 dquot->dq_id, depth)])); in do_insert_tree() 333 ret = do_insert_tree(info, dquot, &newblk, depth+1); in do_insert_tree() 336 ref[get_index(info, dquot->dq_id, depth)] = in do_insert_tree() 461 uint *blk, int depth) in remove_tree() argument [all …]
|
/fs/configfs/ |
D | symlink.c | 41 int depth = 0; in item_depth() local 42 do { depth++; } while ((p = p->ci_parent) && !configfs_is_root(p)); in item_depth() 43 return depth; in item_depth() 239 int depth, size; in configfs_get_target_path() local 241 depth = item_depth(item); in configfs_get_target_path() 242 size = item_path_length(target) + depth * 3 - 1; in configfs_get_target_path() 246 pr_debug("%s: depth = %d, size = %d\n", __func__, depth, size); in configfs_get_target_path() 248 for (s = path; depth--; s += 3) in configfs_get_target_path()
|
D | inode.c | 160 int depth = sd->s_depth; in configfs_set_inode_lock_class() local 162 if (depth > 0) { in configfs_set_inode_lock_class() 163 if (depth <= ARRAY_SIZE(default_group_class)) { in configfs_set_inode_lock_class() 165 &default_group_class[depth - 1]); in configfs_set_inode_lock_class()
|
/fs/sysv/ |
D | itree.c | 88 int depth, in get_branch() argument 101 while (--depth) { in get_branch() 210 int depth = block_to_path(inode, iblock, offsets); in get_block() local 212 if (depth == 0) in get_block() 217 partial = get_branch(inode, depth, offsets, chain, &err); in get_block() 224 chain[depth-1].key)); in get_block() 226 partial = chain+depth-1; /* the whole chain */ in get_block() 249 left = (chain + depth) - partial; in get_block() 277 int depth, in find_shared() argument 286 for (k = depth; k > 1 && !offsets[k-1]; k--) in find_shared() [all …]
|
/fs/ext2/ |
D | inode.c | 232 int depth, in ext2_get_branch() argument 246 while (--depth) { in ext2_get_branch() 630 int depth; in ext2_get_blocks() local 635 depth = ext2_block_to_path(inode,iblock,offsets,&blocks_to_boundary); in ext2_get_blocks() 637 if (depth == 0) in ext2_get_blocks() 640 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks() 643 first_block = le32_to_cpu(chain[depth - 1].key); in ext2_get_blocks() 650 if (!verify_chain(chain, chain + depth - 1)) { in ext2_get_blocks() 661 blk = le32_to_cpu(*(chain[depth-1].p + count)); in ext2_get_blocks() 693 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks() [all …]
|
/fs/reiserfs/ |
D | ioctl.c | 170 int depth; in reiserfs_unpack() local 186 depth = reiserfs_write_lock_once(inode->i_sb); in reiserfs_unpack() 224 reiserfs_write_unlock_once(inode->i_sb, depth); in reiserfs_unpack()
|
D | xattr_acl.c | 439 int depth; in reiserfs_acl_chmod() local 465 depth = reiserfs_write_lock_once(inode->i_sb); in reiserfs_acl_chmod() 474 reiserfs_write_unlock_once(inode->i_sb, depth); in reiserfs_acl_chmod()
|
D | inode.c | 33 int depth; in reiserfs_evict_inode() local 43 depth = reiserfs_write_lock_once(inode->i_sb); in reiserfs_evict_inode() 80 reiserfs_write_unlock_once(inode->i_sb, depth); in reiserfs_evict_inode() 3108 int depth; in reiserfs_setattr() local 3120 depth = reiserfs_write_lock_once(inode->i_sb); in reiserfs_setattr() 3184 reiserfs_write_unlock_once(inode->i_sb, depth); in reiserfs_setattr() 3186 depth = reiserfs_write_lock_once(inode->i_sb); in reiserfs_setattr() 3210 reiserfs_write_unlock_once(inode->i_sb, depth); in reiserfs_setattr() 3224 depth = reiserfs_write_lock_once(inode->i_sb); in reiserfs_setattr() 3232 reiserfs_write_unlock_once(inode->i_sb, depth); in reiserfs_setattr()
|
/fs/sdcardfs/ |
D | derived_perm.c | 279 static void __fixup_perms_recursive(struct dentry *dentry, struct limit_search *limit, int depth) in __fixup_perms_recursive() argument 289 WARN(depth > 3, "%s: Max expected depth exceeded!\n", __func__); in __fixup_perms_recursive() 290 spin_lock_nested(&dentry->d_lock, depth); in __fixup_perms_recursive() 299 spin_lock_nested(&child->d_lock, depth + 1); in __fixup_perms_recursive() 312 __fixup_perms_recursive(child, limit, depth + 1); in __fixup_perms_recursive()
|
/fs/hfsplus/ |
D | btree.c | 64 tree->depth = be16_to_cpu(head->depth); in hfs_btree_open() 191 head->depth = cpu_to_be16(tree->depth); in hfs_btree_write()
|
/fs/ext3/ |
D | inode.c | 410 static Indirect *ext3_get_branch(struct inode *inode, int depth, int *offsets, in ext3_get_branch() argument 422 while (--depth) { in ext3_get_branch() 869 int depth; in ext3_get_blocks_handle() local 877 depth = ext3_block_to_path(inode,iblock,offsets,&blocks_to_boundary); in ext3_get_blocks_handle() 879 if (depth == 0) in ext3_get_blocks_handle() 882 partial = ext3_get_branch(inode, depth, offsets, chain, &err); in ext3_get_blocks_handle() 886 first_block = le32_to_cpu(chain[depth - 1].key); in ext3_get_blocks_handle() 893 if (!verify_chain(chain, chain + depth - 1)) { in ext3_get_blocks_handle() 905 blk = le32_to_cpu(*(chain[depth-1].p + count)); in ext3_get_blocks_handle() 942 partial = ext3_get_branch(inode, depth, offsets, chain, &err); in ext3_get_blocks_handle() [all …]
|
/fs/hfs/ |
D | btree.c | 92 tree->depth = be16_to_cpu(head->depth); in hfs_btree_open() 180 head->depth = cpu_to_be16(tree->depth); in hfs_btree_write()
|
D | btree.h | 33 unsigned int depth; member 139 __be16 depth; /* (V) The number of levels in this B-tree */ member
|
/fs/f2fs/ |
D | node.c | 543 int ofs, int depth) in truncate_nodes() argument 566 if (depth < 3) { in truncate_nodes() 586 ret = truncate_nodes(&rdn, child_nofs, 0, depth - 1); in truncate_nodes() 615 struct f2fs_inode *ri, int *offset, int depth) in truncate_partial_nodes() argument 623 int idx = depth - 2; in truncate_partial_nodes() 630 for (i = 0; i < depth - 1; i++) { in truncate_partial_nodes() 634 depth = i + 1; in truncate_partial_nodes() 642 for (i = offset[depth - 1]; i < NIDS_PER_BLOCK; i++) { in truncate_partial_nodes() 653 if (offset[depth - 1] == 0) { in truncate_partial_nodes() 661 offset[depth - 1] = 0; in truncate_partial_nodes() [all …]
|
/fs/ocfs2/ |
D | ocfs2_trace.h | 423 TP_PROTO(unsigned long long owner, int depth), 424 TP_ARGS(owner, depth), 427 __field(int, depth) 431 __entry->depth = depth; 433 TP_printk("%llu %d", __entry->owner, __entry->depth) 438 int depth), 439 TP_ARGS(subtree_root, blkno, depth), 443 __field(int, depth) 448 __entry->depth = depth; 451 __entry->blkno, __entry->depth) [all …]
|
D | alloc.c | 582 int i, start = 0, depth = 0; in ocfs2_reinit_path() local 602 depth = le16_to_cpu(path_root_el(path)->l_tree_depth); in ocfs2_reinit_path() 606 path->p_tree_depth = depth; in ocfs2_reinit_path() 1516 int depth = le16_to_cpu(el->l_tree_depth); in ocfs2_grow_tree() local 1536 depth); in ocfs2_grow_tree() 1546 depth++; in ocfs2_grow_tree() 1547 if (depth == 1) { in ocfs2_grow_tree() 1575 *final_depth = depth; in ocfs2_grow_tree() 4859 int ret = 0, depth; in ocfs2_split_and_insert() local 4877 depth = le16_to_cpu(rightmost_el->l_tree_depth); in ocfs2_split_and_insert() [all …]
|
/fs/gfs2/ |
D | glops.c | 310 u16 height, depth; in gfs2_dinode_in() local 351 depth = be16_to_cpu(str->di_depth); in gfs2_dinode_in() 352 if (unlikely(depth > GFS2_DIR_MAX_DEPTH)) in gfs2_dinode_in() 354 ip->i_depth = (u8)depth; in gfs2_dinode_in()
|
D | dir.c | 827 static struct gfs2_leaf *new_leaf(struct inode *inode, struct buffer_head **pbh, u16 depth) in new_leaf() argument 849 leaf->lf_depth = cpu_to_be16(depth); in new_leaf() 1311 filldir_t filldir, int *copied, unsigned *depth, in gfs2_dir_read_leaf() argument 1333 *depth = be16_to_cpu(lf->lf_depth); in gfs2_dir_read_leaf() 1465 unsigned depth = 0; in dir_e_read() local 1481 &copied, &depth, in dir_e_read() 1486 len = 1 << (dip->i_depth - depth); in dir_e_read()
|
/fs/afs/ |
D | vnode.c | 21 int depth, char lr) 30 bad = dump_tree_aux(node->rb_left, node, depth + 2, '/'); 35 depth, depth, "", lr, 43 bad |= dump_tree_aux(node->rb_right, node, depth + 2, '\\');
|
/fs/ufs/ |
D | inode.c | 91 int depth = ufs_block_to_path(inode, frag >> uspi->s_fpbshift, offsets); in ufs_frag_map() local 98 UFSD(": frag = %llu depth = %d\n", (unsigned long long)frag, depth); in ufs_frag_map() 103 if (depth == 0) in ufs_frag_map() 116 while (--depth) { in ufs_frag_map() 136 while (--depth) { in ufs_frag_map()
|
/fs/ |
D | exec.c | 1371 unsigned int depth = bprm->recursion_depth; in search_binary_handler() local 1377 if (depth > 5) in search_binary_handler() 1404 bprm->recursion_depth = depth + 1; in search_binary_handler() 1406 bprm->recursion_depth = depth; in search_binary_handler() 1408 if (depth == 0) { in search_binary_handler()
|
/fs/qnx6/ |
D | inode.c | 123 int depth = ei->di_filelevels; in qnx6_block_map() local 126 bitdelta = ptrbits * depth; in qnx6_block_map() 137 for (i = 0; i < depth; i++) { in qnx6_block_map()
|