/fs/hfs/ |
D | bnode.c | 17 void hfs_bnode_read(struct hfs_bnode *node, void *buf, in hfs_bnode_read() argument 22 off += node->page_offset; in hfs_bnode_read() 23 page = node->page[0]; in hfs_bnode_read() 29 u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off) in hfs_bnode_read_u16() argument 33 hfs_bnode_read(node, &data, off, 2); in hfs_bnode_read_u16() 37 u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off) in hfs_bnode_read_u8() argument 41 hfs_bnode_read(node, &data, off, 1); in hfs_bnode_read_u8() 45 void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off) in hfs_bnode_read_key() argument 50 tree = node->tree; in hfs_bnode_read_key() 51 if (node->type == HFS_NODE_LEAF || in hfs_bnode_read_key() [all …]
|
D | brec.c | 18 u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off) in hfs_brec_lenoff() argument 23 dataoff = node->tree->node_size - (rec + 2) * 2; in hfs_brec_lenoff() 24 hfs_bnode_read(node, retval, dataoff, 4); in hfs_brec_lenoff() 30 u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec) in hfs_brec_keylen() argument 34 if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) in hfs_brec_keylen() 37 if ((node->type == HFS_NODE_INDEX) && in hfs_brec_keylen() 38 !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) { in hfs_brec_keylen() 39 if (node->tree->attributes & HFS_TREE_BIGKEYS) in hfs_brec_keylen() 40 retval = node->tree->max_key_len + 2; in hfs_brec_keylen() 42 retval = node->tree->max_key_len + 1; in hfs_brec_keylen() [all …]
|
D | btree.c | 138 struct hfs_bnode *node; in hfs_btree_close() local 145 while ((node = tree->node_hash[i])) { in hfs_btree_close() 146 tree->node_hash[i] = node->next_hash; in hfs_btree_close() 147 if (atomic_read(&node->refcnt)) in hfs_btree_close() 149 node->tree->cnid, node->this, in hfs_btree_close() 150 atomic_read(&node->refcnt)); in hfs_btree_close() 151 hfs_bnode_free(node); in hfs_btree_close() 162 struct hfs_bnode *node; in hfs_btree_write() local 165 node = hfs_bnode_find(tree, 0); in hfs_btree_write() 166 if (IS_ERR(node)) in hfs_btree_write() [all …]
|
/fs/hfsplus/ |
D | bnode.c | 21 void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len) in hfs_bnode_read() argument 26 off += node->page_offset; in hfs_bnode_read() 27 pagep = node->page + (off >> PAGE_CACHE_SHIFT); in hfs_bnode_read() 42 u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off) in hfs_bnode_read_u16() argument 46 hfs_bnode_read(node, &data, off, 2); in hfs_bnode_read_u16() 50 u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off) in hfs_bnode_read_u8() argument 54 hfs_bnode_read(node, &data, off, 1); in hfs_bnode_read_u8() 58 void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off) in hfs_bnode_read_key() argument 63 tree = node->tree; in hfs_bnode_read_key() 64 if (node->type == HFS_NODE_LEAF || in hfs_bnode_read_key() [all …]
|
D | brec.c | 19 u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off) in hfs_brec_lenoff() argument 24 dataoff = node->tree->node_size - (rec + 2) * 2; in hfs_brec_lenoff() 25 hfs_bnode_read(node, retval, dataoff, 4); in hfs_brec_lenoff() 31 u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec) in hfs_brec_keylen() argument 35 if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) in hfs_brec_keylen() 38 if ((node->type == HFS_NODE_INDEX) && in hfs_brec_keylen() 39 !(node->tree->attributes & HFS_TREE_VARIDXKEYS) && in hfs_brec_keylen() 40 (node->tree->cnid != HFSPLUS_ATTR_CNID)) { in hfs_brec_keylen() 41 retval = node->tree->max_key_len + 2; in hfs_brec_keylen() 43 recoff = hfs_bnode_read_u16(node, in hfs_brec_keylen() [all …]
|
D | btree.c | 147 struct hfs_bnode *node; in hfs_btree_close() local 154 while ((node = tree->node_hash[i])) { in hfs_btree_close() 155 tree->node_hash[i] = node->next_hash; in hfs_btree_close() 156 if (atomic_read(&node->refcnt)) in hfs_btree_close() 159 node->tree->cnid, node->this, in hfs_btree_close() 160 atomic_read(&node->refcnt)); in hfs_btree_close() 161 hfs_bnode_free(node); in hfs_btree_close() 172 struct hfs_bnode *node; in hfs_btree_write() local 175 node = hfs_bnode_find(tree, 0); in hfs_btree_write() 176 if (IS_ERR(node)) in hfs_btree_write() [all …]
|
/fs/ocfs2/cluster/ |
D | quorum.c | 168 static void o2quo_set_hold(struct o2quo_state *qs, u8 node) in o2quo_set_hold() argument 172 if (!test_and_set_bit(node, qs->qs_hold_bm)) { in o2quo_set_hold() 175 "node %u\n", node); in o2quo_set_hold() 176 mlog(0, "node %u, %d total\n", node, qs->qs_holds); in o2quo_set_hold() 180 static void o2quo_clear_hold(struct o2quo_state *qs, u8 node) in o2quo_clear_hold() argument 184 if (test_and_clear_bit(node, qs->qs_hold_bm)) { in o2quo_clear_hold() 185 mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1); in o2quo_clear_hold() 193 node, qs->qs_holds); in o2quo_clear_hold() 201 void o2quo_hb_up(u8 node) in o2quo_hb_up() argument 209 "node %u\n", node); in o2quo_hb_up() [all …]
|
D | nodemanager.c | 46 struct o2nm_node *node = NULL; in o2nm_get_node_by_num() local 52 node = o2nm_single_cluster->cl_nodes[node_num]; in o2nm_get_node_by_num() 53 if (node) in o2nm_get_node_by_num() 54 config_item_get(&node->nd_item); in o2nm_get_node_by_num() 57 return node; in o2nm_get_node_by_num() 85 struct o2nm_node *node, *ret = NULL; in o2nm_node_ip_tree_lookup() local 91 node = rb_entry(parent, struct o2nm_node, nd_ip_node); in o2nm_node_ip_tree_lookup() 93 cmp = memcmp(&ip_needle, &node->nd_ipv4_address, in o2nm_node_ip_tree_lookup() 100 ret = node; in o2nm_node_ip_tree_lookup() 115 struct o2nm_node *node = NULL; in o2nm_get_node_by_ip() local [all …]
|
/fs/btrfs/ |
D | delayed-inode.c | 92 struct btrfs_delayed_node *node; in btrfs_get_delayed_node() local 94 node = ACCESS_ONCE(btrfs_inode->delayed_node); in btrfs_get_delayed_node() 95 if (node) { in btrfs_get_delayed_node() 96 atomic_inc(&node->refs); in btrfs_get_delayed_node() 97 return node; in btrfs_get_delayed_node() 101 node = radix_tree_lookup(&root->delayed_nodes_tree, ino); in btrfs_get_delayed_node() 102 if (node) { in btrfs_get_delayed_node() 104 atomic_inc(&node->refs); /* can be accessed */ in btrfs_get_delayed_node() 105 BUG_ON(btrfs_inode->delayed_node != node); in btrfs_get_delayed_node() 107 return node; in btrfs_get_delayed_node() [all …]
|
D | relocation.c | 92 struct backref_node *node[2]; member 195 struct backref_node *node); 197 struct backref_node *node); 218 struct backref_node *node; in backref_cache_cleanup() local 222 node = list_entry(cache->detached.next, in backref_cache_cleanup() 224 remove_backref_node(cache, node); in backref_cache_cleanup() 228 node = list_entry(cache->leaves.next, in backref_cache_cleanup() 230 remove_backref_node(cache, node); in backref_cache_cleanup() 246 struct backref_node *node; in alloc_backref_node() local 248 node = kzalloc(sizeof(*node), GFP_NOFS); in alloc_backref_node() [all …]
|
D | delayed-ref.h | 76 struct btrfs_delayed_ref_node node; member 104 struct btrfs_delayed_ref_node node; member 111 struct btrfs_delayed_ref_node node; member 260 static int btrfs_delayed_ref_is_head(struct btrfs_delayed_ref_node *node) in btrfs_delayed_ref_is_head() argument 262 return node->is_head; in btrfs_delayed_ref_is_head() 269 btrfs_delayed_node_to_tree_ref(struct btrfs_delayed_ref_node *node) in btrfs_delayed_node_to_tree_ref() argument 271 WARN_ON(btrfs_delayed_ref_is_head(node)); in btrfs_delayed_node_to_tree_ref() 272 return container_of(node, struct btrfs_delayed_tree_ref, node); in btrfs_delayed_node_to_tree_ref() 276 btrfs_delayed_node_to_data_ref(struct btrfs_delayed_ref_node *node) in btrfs_delayed_node_to_data_ref() argument 278 WARN_ON(btrfs_delayed_ref_is_head(node)); in btrfs_delayed_node_to_data_ref() [all …]
|
D | delayed-ref.c | 65 if (ref1->node.type == BTRFS_EXTENT_DATA_REF_KEY) { in comp_data_refs() 136 struct rb_node *node) in tree_insert() argument 144 ins = rb_entry(node, struct btrfs_delayed_ref_node, rb_node); in tree_insert() 159 rb_link_node(node, parent_node, p); in tree_insert() 160 rb_insert_color(node, root); in tree_insert() 230 atomic_inc(&head->node.refs); in btrfs_delayed_ref_lock() 235 if (!head->node.in_tree) { in btrfs_delayed_ref_lock() 237 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock() 240 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock() 260 struct rb_node *node; in merge_ref() local [all …]
|
D | ordered-data.c | 41 struct rb_node *node) in tree_insert() argument 59 rb_link_node(node, parent, p); in tree_insert() 60 rb_insert_color(node, root); in tree_insert() 188 struct rb_node *node; in __btrfs_add_ordered_extent() local 224 node = tree_insert(&tree->tree, file_offset, in __btrfs_add_ordered_extent() 226 if (node) in __btrfs_add_ordered_extent() 301 struct rb_node *node; in btrfs_dec_test_first_ordered_pending() local 311 node = tree_search(tree, *file_offset); in btrfs_dec_test_first_ordered_pending() 312 if (!node) { in btrfs_dec_test_first_ordered_pending() 317 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_dec_test_first_ordered_pending() [all …]
|
/fs/nilfs2/ |
D | btree.c | 84 static int nilfs_btree_node_get_flags(const struct nilfs_btree_node *node) in nilfs_btree_node_get_flags() argument 86 return node->bn_flags; in nilfs_btree_node_get_flags() 90 nilfs_btree_node_set_flags(struct nilfs_btree_node *node, int flags) in nilfs_btree_node_set_flags() argument 92 node->bn_flags = flags; in nilfs_btree_node_set_flags() 95 static int nilfs_btree_node_root(const struct nilfs_btree_node *node) in nilfs_btree_node_root() argument 97 return nilfs_btree_node_get_flags(node) & NILFS_BTREE_NODE_ROOT; in nilfs_btree_node_root() 100 static int nilfs_btree_node_get_level(const struct nilfs_btree_node *node) in nilfs_btree_node_get_level() argument 102 return node->bn_level; in nilfs_btree_node_get_level() 106 nilfs_btree_node_set_level(struct nilfs_btree_node *node, int level) in nilfs_btree_node_set_level() argument 108 node->bn_level = level; in nilfs_btree_node_set_level() [all …]
|
/fs/befs/ |
D | btree.c | 100 befs_btree_node * node, befs_off_t node_off); 102 static int befs_leafnode(befs_btree_node * node); 104 static fs16 *befs_bt_keylen_index(befs_btree_node * node); 106 static fs64 *befs_bt_valarray(befs_btree_node * node); 108 static char *befs_bt_keydata(befs_btree_node * node); 110 static int befs_find_key(struct super_block *sb, befs_btree_node * node, 113 static char *befs_bt_get_key(struct super_block *sb, befs_btree_node * node, 194 befs_btree_node * node, befs_off_t node_off) in befs_bt_read_node() argument 200 if (node->bh) in befs_bt_read_node() 201 brelse(node->bh); in befs_bt_read_node() [all …]
|
/fs/ext4/ |
D | block_validity.c | 24 struct rb_node node; member 62 struct rb_node **n = &sbi->system_blks.rb_node, *node; in add_system_zone() local 67 entry = rb_entry(parent, struct ext4_system_zone, node); in add_system_zone() 79 node); in add_system_zone() 91 new_node = &new_entry->node; in add_system_zone() 98 node = rb_prev(new_node); in add_system_zone() 99 if (node) { in add_system_zone() 100 entry = rb_entry(node, struct ext4_system_zone, node); in add_system_zone() 104 rb_erase(node, &sbi->system_blks); in add_system_zone() 110 node = rb_next(new_node); in add_system_zone() [all …]
|
D | extents_status.c | 181 struct rb_node *node; in ext4_es_print_tree() local 185 node = rb_first(&tree->root); in ext4_es_print_tree() 186 while (node) { in ext4_es_print_tree() 188 es = rb_entry(node, struct extent_status, rb_node); in ext4_es_print_tree() 192 node = rb_next(node); in ext4_es_print_tree() 213 struct rb_node *node = root->rb_node; in __es_tree_search() local 216 while (node) { in __es_tree_search() 217 es = rb_entry(node, struct extent_status, rb_node); in __es_tree_search() 219 node = node->rb_left; in __es_tree_search() 221 node = node->rb_right; in __es_tree_search() [all …]
|
/fs/jffs2/ |
D | nodelist.c | 93 if (frag->node && (frag->ofs & (PAGE_CACHE_SIZE - 1)) == 0) { in jffs2_truncate_fragtree() 96 frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE; in jffs2_truncate_fragtree() 104 if (this->node) { in jffs2_obsolete_node_frag() 105 this->node->frags--; in jffs2_obsolete_node_frag() 106 if (!this->node->frags) { in jffs2_obsolete_node_frag() 109 ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size); in jffs2_obsolete_node_frag() 110 jffs2_mark_node_obsolete(c, this->node->raw); in jffs2_obsolete_node_frag() 111 jffs2_free_full_dnode(this->node); in jffs2_obsolete_node_frag() 114 …ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags); in jffs2_obsolete_node_frag() 115 mark_ref_normal(this->node->raw); in jffs2_obsolete_node_frag() [all …]
|
D | debug.c | 81 struct jffs2_full_dnode *fn = frag->node; in __jffs2_dbg_fragtree_paranoia_check_nolock() 99 && frag_prev(frag)->size < PAGE_CACHE_SIZE && frag_prev(frag)->node) { in __jffs2_dbg_fragtree_paranoia_check_nolock() 106 && frag_next(frag)->size < PAGE_CACHE_SIZE && frag_next(frag)->node) { in __jffs2_dbg_fragtree_paranoia_check_nolock() 711 if (this->node) in __jffs2_dbg_dump_fragtree_nolock() 713 this->ofs, this->ofs+this->size, ref_offset(this->node->raw), in __jffs2_dbg_dump_fragtree_nolock() 714 ref_flags(this->node->raw), this, frag_left(this), frag_right(this), in __jffs2_dbg_dump_fragtree_nolock() 775 union jffs2_node_union node; in __jffs2_dbg_dump_node() local 783 ret = jffs2_flash_read(c, ofs, len, &retlen, (unsigned char *)&node); in __jffs2_dbg_dump_node() 790 printk(JFFS2_DBG "magic:\t%#04x\n", je16_to_cpu(node.u.magic)); in __jffs2_dbg_dump_node() 791 printk(JFFS2_DBG "nodetype:\t%#04x\n", je16_to_cpu(node.u.nodetype)); in __jffs2_dbg_dump_node() [all …]
|
D | scan.c | 364 raw->next_in_ino = xd->node->next_in_ino; in jffs2_scan_xattr_node() 365 xd->node->next_in_ino = raw; in jffs2_scan_xattr_node() 444 struct jffs2_unknown_node *node; in jffs2_scan_eraseblock() local 630 if (jeb->offset + c->sector_size < ofs + sizeof(*node)) { in jffs2_scan_eraseblock() 634 sizeof(*node)); in jffs2_scan_eraseblock() 640 if (buf_ofs + buf_len < ofs + sizeof(*node)) { in jffs2_scan_eraseblock() 651 node = (struct jffs2_unknown_node *)&buf[ofs-buf_ofs]; in jffs2_scan_eraseblock() 713 if (ofs == jeb->offset && je16_to_cpu(node->magic) == KSAMTIB_CIGAM_2SFFJ) { in jffs2_scan_eraseblock() 721 if (je16_to_cpu(node->magic) == JFFS2_DIRTY_BITMASK) { in jffs2_scan_eraseblock() 728 if (je16_to_cpu(node->magic) == JFFS2_OLD_MAGIC_BITMASK) { in jffs2_scan_eraseblock() [all …]
|
D | gc.c | 515 if (frag->node && frag->node->raw == raw) { in jffs2_garbage_collect_live() 516 fn = frag->node; in jffs2_garbage_collect_live() 520 if (nrfrags == frag->node->frags) in jffs2_garbage_collect_live() 529 frag->node->raw = f->inocache->nodes; in jffs2_garbage_collect_live() 575 union jffs2_node_union *node; in jffs2_garbage_collect_pristine() local 604 node = kmalloc(rawlen, GFP_KERNEL); in jffs2_garbage_collect_pristine() 605 if (!node) in jffs2_garbage_collect_pristine() 608 ret = jffs2_flash_read(c, ref_offset(raw), rawlen, &retlen, (char *)node); in jffs2_garbage_collect_pristine() 614 crc = crc32(0, node, sizeof(struct jffs2_unknown_node)-4); in jffs2_garbage_collect_pristine() 615 if (je32_to_cpu(node->u.hdr_crc) != crc) { in jffs2_garbage_collect_pristine() [all …]
|
/fs/ocfs2/dlm/ |
D | dlmdomain.c | 502 unsigned int node; in dlm_begin_exit_domain_handler() local 508 node = exit_msg->node_idx; in dlm_begin_exit_domain_handler() 509 mlog(0, "%s: Node %u sent a begin exit domain message\n", dlm->name, node); in dlm_begin_exit_domain_handler() 512 set_bit(node, dlm->exit_domain_map); in dlm_begin_exit_domain_handler() 545 int node = -1, num = 0; in __dlm_print_nodes() local 550 while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES, in __dlm_print_nodes() 551 node + 1)) < O2NM_MAX_NODES) { in __dlm_print_nodes() 552 printk("%d ", node); in __dlm_print_nodes() 562 unsigned int node; in dlm_exit_domain_handler() local 570 node = exit_msg->node_idx; in dlm_exit_domain_handler() [all …]
|
/fs/nfs/ |
D | pnfs_dev.c | 81 hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[hash], node) in _lookup_deviceid() 143 hlist_del_init_rcu(&d->node); in nfs4_delete_deviceid() 159 INIT_HLIST_NODE(&d->node); in nfs4_init_deviceid_node() 194 hlist_add_head_rcu(&new->node, &nfs4_deviceid_cache[hash]); in nfs4_insert_deviceid_node() 223 nfs4_mark_deviceid_unavailable(struct nfs4_deviceid_node *node) in nfs4_mark_deviceid_unavailable() argument 225 node->timestamp_unavailable = jiffies; in nfs4_mark_deviceid_unavailable() 226 set_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags); in nfs4_mark_deviceid_unavailable() 231 nfs4_test_deviceid_unavailable(struct nfs4_deviceid_node *node) in nfs4_test_deviceid_unavailable() argument 233 if (test_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags)) { in nfs4_test_deviceid_unavailable() 238 if (time_in_range(node->timestamp_unavailable, start, end)) in nfs4_test_deviceid_unavailable() [all …]
|
/fs/xfs/ |
D | xfs_da_btree.c | 351 struct xfs_da_intnode *node; in xfs_da3_node_create() local 366 node = bp->b_addr; in xfs_da3_node_create() 380 xfs_da3_node_hdr_to_disk(node, &ichdr); in xfs_da3_node_create() 382 XFS_DA_LOGRANGE(node, &node->hdr, xfs_da3_node_hdr_size(node))); in xfs_da3_node_create() 399 struct xfs_da_intnode *node; in xfs_da3_split() local 511 node = oldblk->bp->b_addr; in xfs_da3_split() 512 if (node->hdr.info.forw) { in xfs_da3_split() 513 if (be32_to_cpu(node->hdr.info.forw) == addblk->blkno) { in xfs_da3_split() 519 node = bp->b_addr; in xfs_da3_split() 520 node->hdr.info.back = cpu_to_be32(oldblk->blkno); in xfs_da3_split() [all …]
|
/fs/ocfs2/ |
D | reservations.c | 89 struct rb_node *node; in ocfs2_dump_resv() local 96 node = rb_first(&resmap->m_reservations); in ocfs2_dump_resv() 97 while (node) { in ocfs2_dump_resv() 98 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv() 105 node = rb_next(node); in ocfs2_dump_resv() 148 struct rb_node *node; in ocfs2_check_resmap() local 151 node = rb_first(&resmap->m_reservations); in ocfs2_check_resmap() 152 while (node) { in ocfs2_check_resmap() 153 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_check_resmap() 183 node = rb_next(node); in ocfs2_check_resmap() [all …]
|