/fs/jfs/ |
D | jfs_txnmgr.c | 204 TxAnchor.freelock = TxLock[lid].next; in txLockAlloc() 218 TxLock[lid].next = TxAnchor.freelock; in txLockFree() 289 TxBlock[k].next = k + 1; in txInit() 293 TxBlock[k].next = 0; in txInit() 317 TxLock[k].next = k + 1; in txInit() 318 TxLock[k].next = 0; in txInit() 411 if ((tblk->next == 0) && !(flag & COMMIT_FORCE)) { in txBegin() 419 TxAnchor.freetid = tblk->next; in txBegin() 431 tblk->next = tblk->last = tblk->xflag = tblk->flag = tblk->lsn = 0; in txBegin() 536 assert(tblk->next == 0); in txEnd() [all …]
|
D | jfs_dtree.c | 496 static void free_index(tid_t tid, struct inode *ip, u32 index, u32 next) in free_index() argument 509 dirtab_slot->addr2 = cpu_to_le32(next); in free_index() 1410 nextbn = le64_to_cpu(sp->header.next); in dtSplitPage() 1411 rp->header.next = cpu_to_le64(nextbn); in dtSplitPage() 1413 sp->header.next = cpu_to_le64(rbn); in dtSplitPage() 1458 f->next = fsi; in dtSplitPage() 1459 f->next = -1; in dtSplitPage() 1575 f->next = fsi; in dtSplitPage() 1576 f->next = -1; in dtSplitPage() 1787 f->next = last; in dtExtendPage() [all …]
|
D | jfs_txnmgr.h | 48 lid_t next; /* index of first tlock of tid */ member 96 lid_t next; /* 2: index next lockword on tid locklist member 171 lid_t next; /* 2: next linelock */ member 187 lid_t next; /* 2: */ member 218 lid_t next; /* 2: */ member 246 lid_t next; /* 2: */ member
|
/fs/cachefiles/ |
D | namei.c | 488 struct dentry *dir, *next = NULL; in cachefiles_walk_to_object() local 530 next = lookup_one_len(name, dir, nlen); in cachefiles_walk_to_object() 532 if (IS_ERR(next)) in cachefiles_walk_to_object() 535 _debug("next -> %p %s", next, d_backing_inode(next) ? "positive" : "negative"); in cachefiles_walk_to_object() 538 object->new = !d_backing_inode(next); in cachefiles_walk_to_object() 543 if (d_is_negative(next)) in cachefiles_walk_to_object() 549 if (d_is_negative(next)) { in cachefiles_walk_to_object() 555 ret = security_path_mkdir(&path, next, 0); in cachefiles_walk_to_object() 559 ret = vfs_mkdir(d_inode(dir), next, 0); in cachefiles_walk_to_object() 564 ASSERT(d_backing_inode(next)); in cachefiles_walk_to_object() [all …]
|
/fs/ |
D | char_dev.c | 32 struct char_device_struct *next; member 54 for (cd = chrdevs[offset]; cd; cd = cd->next) in chrdev_show() 112 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) in __register_chrdev_region() 139 cd->next = *cp; in __register_chrdev_region() 156 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) in __unregister_chrdev_region() 163 *cp = cd->next; in __unregister_chrdev_region() 182 dev_t n, next; in register_chrdev_region() local 184 for (n = from; n < to; n = next) { in register_chrdev_region() 185 next = MKDEV(MAJOR(n)+1, 0); in register_chrdev_region() 186 if (next > to) in register_chrdev_region() [all …]
|
D | pnode.c | 19 return list_entry(p->mnt_share.next, struct mount, mnt_share); in next_peer() 24 return list_entry(p->mnt_slave_list.next, struct mount, mnt_slave); in first_slave() 34 return list_entry(p->mnt_slave.next, struct mount, mnt_slave); in next_slave() 162 struct mount *next = next_peer(m); in propagation_next() local 163 return (next == origin) ? NULL : next; in propagation_next() 164 } else if (m->mnt_slave.next != &master->mnt_slave_list) in propagation_next() 189 struct mount *next; in next_group() local 192 next = next_peer(m); in next_group() 194 if (next == origin) in next_group() 196 } else if (m->mnt_slave.next != &next->mnt_slave) in next_group() [all …]
|
D | filesystems.c | 49 for (p = &file_systems; *p; p = &(*p)->next) in find_filesystem() 75 if (fs->next) in register_filesystem() 109 *tmp = fs->next; in unregister_filesystem() 110 fs->next = NULL; in unregister_filesystem() 115 tmp = &(*tmp)->next; in unregister_filesystem() 138 for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) { in fs_index() 155 for (tmp = file_systems; tmp; tmp = tmp->next, index--) in fs_name() 175 for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index++) in fs_maxindex() 216 tmp = tmp->next; in get_filesystem_list() 233 tmp = tmp->next; in filesystems_proc_show()
|
/fs/jffs2/ |
D | nodelist.c | 43 new->next = (*prev)->next; in jffs2_add_fd_to_list() 53 prev = &((*prev)->next); in jffs2_add_fd_to_list() 55 new->next = *prev; in jffs2_add_fd_to_list() 73 struct jffs2_node_frag *next = frag_next(frag); in jffs2_truncate_fragtree() local 77 frag = next; in jffs2_truncate_fragtree() 395 struct jffs2_node_frag *next = frag_next(newfrag); in jffs2_add_full_dnode_to_inode() local 397 if (next) { in jffs2_add_full_dnode_to_inode() 399 if (next->node) in jffs2_add_full_dnode_to_inode() 400 mark_ref_normal(next->node->raw); in jffs2_add_full_dnode_to_inode() 427 ret = ret->next; in jffs2_get_ino_cache() [all …]
|
D | xattr.c | 570 ref->next = ic->xref; in create_xattr_ref() 586 ref->next = c->xref_dead_list; in delete_xattr_ref() 607 _ref = ref->next; in jffs2_xattr_delete_inode() 622 _ref = ref->next; in jffs2_xattr_free_inode() 648 for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) { in check_xattr_ref_inode() 652 *pref = ref->next; in check_xattr_ref_inode() 658 for (cmp=ref->next, pcmp=&ref->next; cmp; pcmp=&cmp->next, cmp=cmp->next) { in check_xattr_ref_inode() 664 *pcmp = cmp->next; in check_xattr_ref_inode() 673 *pcmp = cmp->next; in check_xattr_ref_inode() 676 *pref = ref->next; in check_xattr_ref_inode() [all …]
|
D | build.c | 40 if (ic->next) in next_inode() 41 return ic->next; in next_inode() 61 for(fd = ic->scan_dents; fd; fd = fd->next) { in jffs2_build_inode_pass1() 151 dead_fds = fd->next; in jffs2_build_filesystem() 181 ic->scan_dents = fd->next; in jffs2_build_filesystem() 229 ic->scan_dents = fd->next; in jffs2_build_filesystem() 250 struct jffs2_raw_node_ref *next = raw->next_in_ino; in jffs2_build_remove_unlinked_inode() local 253 raw = next; in jffs2_build_remove_unlinked_inode() 264 ic->scan_dents = fd->next; in jffs2_build_remove_unlinked_inode() 292 fd->next = *dead_fds; in jffs2_build_remove_unlinked_inode()
|
/fs/autofs4/ |
D | expire.c | 78 struct list_head *next; in get_next_positive_subdir() local 85 next = prev->d_child.next; in get_next_positive_subdir() 88 next = prev->d_subdirs.next; in get_next_positive_subdir() 92 if (next == &root->d_subdirs) { in get_next_positive_subdir() 99 q = list_entry(next, struct dentry, d_child); in get_next_positive_subdir() 105 next = q->d_child.next; in get_next_positive_subdir() 125 struct list_head *next; in get_next_positive_dentry() local 136 next = p->d_subdirs.next; in get_next_positive_dentry() 137 if (next == &p->d_subdirs) { in get_next_positive_dentry() 155 next = p->d_child.next; in get_next_positive_dentry() [all …]
|
/fs/omfs/ |
D | file.c | 37 u64 next, last; in omfs_shrink_inode() local 45 next = inode->i_ino; in omfs_shrink_inode() 52 bh = omfs_bread(inode->i_sb, next); in omfs_shrink_inode() 61 if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) in omfs_shrink_inode() 69 last = next; in omfs_shrink_inode() 70 next = be64_to_cpu(oe->e_next); in omfs_shrink_inode() 89 if (next == ~0) in omfs_shrink_inode() 92 bh = omfs_bread(inode->i_sb, next); in omfs_shrink_inode() 221 sector_t next, offset; in omfs_get_block() local 239 next = inode->i_ino; in omfs_get_block() [all …]
|
/fs/btrfs/ |
D | extent_map.c | 185 static int mergable_maps(struct extent_map *prev, struct extent_map *next) in mergable_maps() argument 198 test_bit(EXTENT_FLAG_LOGGING, &next->flags)) in mergable_maps() 206 if (!list_empty(&prev->list) || !list_empty(&next->list)) in mergable_maps() 209 if (extent_map_end(prev) == next->start && in mergable_maps() 210 prev->flags == next->flags && in mergable_maps() 211 prev->bdev == next->bdev && in mergable_maps() 212 ((next->block_start == EXTENT_MAP_HOLE && in mergable_maps() 214 (next->block_start == EXTENT_MAP_INLINE && in mergable_maps() 216 (next->block_start == EXTENT_MAP_DELALLOC && in mergable_maps() 218 (next->block_start < EXTENT_MAP_LAST_BYTE - 1 && in mergable_maps() [all …]
|
D | delayed-inode.c | 209 p = delayed_root->node_list.next; in btrfs_first_delayed_node() 223 struct btrfs_delayed_node *next = NULL; in btrfs_next_delayed_node() local 231 p = delayed_root->node_list.next; in btrfs_next_delayed_node() 235 p = node->n_list.next; in btrfs_next_delayed_node() 237 next = list_entry(p, struct btrfs_delayed_node, n_list); in btrfs_next_delayed_node() 238 atomic_inc(&next->refs); in btrfs_next_delayed_node() 242 return next; in btrfs_next_delayed_node() 293 p = delayed_root->prepare_list.next; in btrfs_first_prepared_delayed_node() 337 struct btrfs_delayed_item **next) in __btrfs_lookup_delayed_item() argument 370 if (next) { in __btrfs_lookup_delayed_item() [all …]
|
D | delayed-ref.c | 206 struct btrfs_delayed_ref_node *next; in merge_ref() local 209 next = list_first_entry(&head->ref_list, struct btrfs_delayed_ref_node, in merge_ref() 211 while (!done && &next->list != &head->ref_list) { in merge_ref() 215 next2 = list_next_entry(next, list); in merge_ref() 217 if (next == ref) in merge_ref() 218 goto next; in merge_ref() 220 if (seq && next->seq >= seq) in merge_ref() 221 goto next; in merge_ref() 223 if (next->type != ref->type) in merge_ref() 224 goto next; in merge_ref() [all …]
|
D | relocation.c | 225 node = list_entry(cache->detached.next, in backref_cache_cleanup() 231 node = list_entry(cache->leaves.next, in backref_cache_cleanup() 356 edge = list_entry(node->upper.next, in walk_up_backref() 383 edge = list_entry(edge->list[LOWER].next, in walk_down_backref() 437 edge = list_entry(node->upper.next, struct backref_edge, in remove_backref_node() 498 node = list_entry(cache->detached.next, in update_backref_cache() 504 node = list_entry(cache->changed.next, in update_backref_cache() 754 edge = list_entry(cur->upper.next, struct backref_edge, 795 goto next; 815 goto next; [all …]
|
D | print-tree.c | 342 struct extent_buffer *next = read_tree_block(root, in btrfs_print_tree() local 345 if (IS_ERR(next)) { in btrfs_print_tree() 347 } else if (!extent_buffer_uptodate(next)) { in btrfs_print_tree() 348 free_extent_buffer(next); in btrfs_print_tree() 352 if (btrfs_is_leaf(next) && in btrfs_print_tree() 355 if (btrfs_header_level(next) != in btrfs_print_tree() 358 btrfs_print_tree(root, next); in btrfs_print_tree() 359 free_extent_buffer(next); in btrfs_print_tree()
|
/fs/nilfs2/ |
D | segbuf.h | 50 sector_t next; member 96 #define NILFS_NEXT_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.next) 99 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next) 100 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head)) 107 (list_entry((head)->next, struct buffer_head, b_assoc_buffers)) 109 (list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \ 111 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head)
|
/fs/xfs/ |
D | xfs_mru_cache.c | 243 struct xfs_mru_cache_elem *elem, *next; in _xfs_mru_cache_clear_reap_list() local 247 list_for_each_entry_safe(elem, next, &mru->reap_list, list_node) { in _xfs_mru_cache_clear_reap_list() 260 list_for_each_entry_safe(elem, next, &tmp, list_node) { in _xfs_mru_cache_clear_reap_list() 281 unsigned long now, next; in _xfs_mru_cache_reap() local 288 next = _xfs_mru_cache_migrate(mru, jiffies); in _xfs_mru_cache_reap() 291 mru->queued = next; in _xfs_mru_cache_reap() 294 if (next <= now) in _xfs_mru_cache_reap() 295 next = 0; in _xfs_mru_cache_reap() 297 next -= now; in _xfs_mru_cache_reap() 298 queue_delayed_work(xfs_mru_reap_wq, &mru->work, next); in _xfs_mru_cache_reap()
|
/fs/f2fs/ |
D | shrinker.c | 50 p = f2fs_list.next; in f2fs_shrink_count() 56 p = p->next; in f2fs_shrink_count() 71 p = p->next; in f2fs_shrink_count() 91 p = f2fs_list.next; in f2fs_shrink_scan() 100 p = p->next; in f2fs_shrink_scan() 119 p = p->next; in f2fs_shrink_scan()
|
/fs/proc/ |
D | generic.c | 144 const char *cp = name, *next; in __xlate_proc_name() local 153 next = strchr(cp, '/'); in __xlate_proc_name() 154 if (!next) in __xlate_proc_name() 157 len = next - cp; in __xlate_proc_name() 288 struct proc_dir_entry *next; in proc_readdir_de() local 298 next = pde_subdir_next(de); in proc_readdir_de() 300 de = next; in proc_readdir_de() 585 struct proc_dir_entry *root = NULL, *de, *next; in remove_proc_subtree() local 605 next = pde_subdir_first(de); in remove_proc_subtree() 606 if (next) { in remove_proc_subtree() [all …]
|
/fs/hfs/ |
D | brec.c | 244 node->this, new_node->this, node->next); in hfs_bnode_split() 245 new_node->next = node->next; in hfs_bnode_split() 251 if (node->next) in hfs_bnode_split() 252 next_node = hfs_bnode_find(tree, node->next); in hfs_bnode_split() 315 node_desc.next = cpu_to_be32(new_node->next); in hfs_bnode_split() 324 node->next = new_node->this; in hfs_bnode_split() 326 node_desc.next = cpu_to_be32(node->next); in hfs_bnode_split() 480 new_node->next = 0; in hfs_btree_inc_height() 484 node_desc.next = cpu_to_be32(new_node->next); in hfs_btree_inc_height()
|
/fs/hfsplus/ |
D | brec.c | 248 node->this, new_node->this, node->next); in hfs_bnode_split() 249 new_node->next = node->next; in hfs_bnode_split() 255 if (node->next) in hfs_bnode_split() 256 next_node = hfs_bnode_find(tree, node->next); in hfs_bnode_split() 319 node_desc.next = cpu_to_be32(new_node->next); in hfs_bnode_split() 328 node->next = new_node->this; in hfs_bnode_split() 330 node_desc.next = cpu_to_be32(node->next); in hfs_bnode_split() 484 new_node->next = 0; in hfs_btree_inc_height() 488 node_desc.next = cpu_to_be32(new_node->next); in hfs_btree_inc_height()
|
/fs/fat/ |
D | fatent.c | 120 int next; in fat12_ent_get() local 124 next = (*ent12_p[0] >> 4) | (*ent12_p[1] << 4); in fat12_ent_get() 126 next = (*ent12_p[1] << 8) | *ent12_p[0]; in fat12_ent_get() 129 next &= 0x0fff; in fat12_ent_get() 130 if (next >= BAD_FAT12) in fat12_ent_get() 131 next = FAT_ENT_EOF; in fat12_ent_get() 132 return next; in fat12_ent_get() 137 int next = le16_to_cpu(*fatent->u.ent16_p); in fat16_ent_get() local 139 if (next >= BAD_FAT16) in fat16_ent_get() 140 next = FAT_ENT_EOF; in fat16_ent_get() [all …]
|
/fs/overlayfs/ |
D | copy_up.c | 394 struct dentry *next; in ovl_copy_up() local 403 next = dget(dentry); in ovl_copy_up() 406 parent = dget_parent(next); in ovl_copy_up() 412 dput(next); in ovl_copy_up() 413 next = parent; in ovl_copy_up() 416 ovl_path_lower(next, &lowerpath); in ovl_copy_up() 419 err = ovl_copy_up_one(parent, next, &lowerpath, &stat); in ovl_copy_up() 422 dput(next); in ovl_copy_up()
|