/fs/xfs/ |
D | xfs_btree.c | 52 #define xfs_btree_magic(cur) \ argument 53 xfs_magics[!!((cur)->bc_flags & XFS_BTREE_CRC_BLOCKS)][cur->bc_btnum] 58 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_btree_check_lblock() argument 66 mp = cur->bc_mp; in xfs_btree_check_lblock() 76 be32_to_cpu(block->bb_magic) == xfs_btree_magic(cur) && in xfs_btree_check_lblock() 79 cur->bc_ops->get_maxrecs(cur, level) && in xfs_btree_check_lblock() 102 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_btree_check_sblock() argument 113 mp = cur->bc_mp; in xfs_btree_check_sblock() 114 agbp = cur->bc_private.a.agbp; in xfs_btree_check_sblock() 126 be32_to_cpu(block->bb_magic) == xfs_btree_magic(cur) && in xfs_btree_check_sblock() [all …]
|
D | xfs_alloc_btree.c | 41 struct xfs_btree_cur *cur) in xfs_allocbt_dup_cursor() argument 43 return xfs_allocbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_allocbt_dup_cursor() 44 cur->bc_private.a.agbp, cur->bc_private.a.agno, in xfs_allocbt_dup_cursor() 45 cur->bc_btnum); in xfs_allocbt_dup_cursor() 50 struct xfs_btree_cur *cur, in xfs_allocbt_set_root() argument 54 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_allocbt_set_root() 57 int btnum = cur->bc_btnum; in xfs_allocbt_set_root() 58 struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); in xfs_allocbt_set_root() 67 xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS); in xfs_allocbt_set_root() 72 struct xfs_btree_cur *cur, in xfs_allocbt_alloc_block() argument [all …]
|
D | xfs_ialloc_btree.c | 42 struct xfs_btree_cur *cur, in xfs_inobt_get_minrecs() argument 45 return cur->bc_mp->m_inobt_mnr[level != 0]; in xfs_inobt_get_minrecs() 50 struct xfs_btree_cur *cur) in xfs_inobt_dup_cursor() argument 52 return xfs_inobt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_inobt_dup_cursor() 53 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_inobt_dup_cursor() 58 struct xfs_btree_cur *cur, in xfs_inobt_set_root() argument 62 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_inobt_set_root() 67 xfs_ialloc_log_agi(cur->bc_tp, agbp, XFS_AGI_ROOT | XFS_AGI_LEVEL); in xfs_inobt_set_root() 72 struct xfs_btree_cur *cur, in xfs_inobt_alloc_block() argument 82 XFS_BTREE_TRACE_CURSOR(cur, XBT_ENTRY); in xfs_inobt_alloc_block() [all …]
|
D | xfs_bmap_btree.c | 485 struct xfs_btree_cur *cur) in xfs_bmbt_dup_cursor() argument 489 new = xfs_bmbt_init_cursor(cur->bc_mp, cur->bc_tp, in xfs_bmbt_dup_cursor() 490 cur->bc_private.b.ip, cur->bc_private.b.whichfork); in xfs_bmbt_dup_cursor() 496 new->bc_private.b.firstblock = cur->bc_private.b.firstblock; in xfs_bmbt_dup_cursor() 497 new->bc_private.b.flist = cur->bc_private.b.flist; in xfs_bmbt_dup_cursor() 498 new->bc_private.b.flags = cur->bc_private.b.flags; in xfs_bmbt_dup_cursor() 520 struct xfs_btree_cur *cur, in xfs_bmbt_alloc_block() argument 530 args.tp = cur->bc_tp; in xfs_bmbt_alloc_block() 531 args.mp = cur->bc_mp; in xfs_bmbt_alloc_block() 532 args.fsbno = cur->bc_private.b.firstblock; in xfs_bmbt_alloc_block() [all …]
|
D | xfs_trans_ail.c | 159 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_init() argument 161 cur->item = NULL; in xfs_trans_ail_cursor_init() 162 list_add_tail(&cur->list, &ailp->xa_cursors); in xfs_trans_ail_cursor_init() 172 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_next() argument 174 struct xfs_log_item *lip = cur->item; in xfs_trans_ail_cursor_next() 179 cur->item = xfs_ail_next(ailp, lip); in xfs_trans_ail_cursor_next() 190 struct xfs_ail_cursor *cur) in xfs_trans_ail_cursor_done() argument 192 cur->item = NULL; in xfs_trans_ail_cursor_done() 193 list_del_init(&cur->list); in xfs_trans_ail_cursor_done() 209 struct xfs_ail_cursor *cur; in xfs_trans_ail_cursor_clear() local [all …]
|
D | xfs_ialloc.c | 62 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_inobt_lookup() argument 67 cur->bc_rec.i.ir_startino = ino; in xfs_inobt_lookup() 68 cur->bc_rec.i.ir_freecount = 0; in xfs_inobt_lookup() 69 cur->bc_rec.i.ir_free = 0; in xfs_inobt_lookup() 70 return xfs_btree_lookup(cur, dir, stat); in xfs_inobt_lookup() 79 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_inobt_update() argument 87 return xfs_btree_update(cur, &rec); in xfs_inobt_update() 95 struct xfs_btree_cur *cur, /* btree cursor */ in xfs_inobt_get_rec() argument 102 error = xfs_btree_get_rec(cur, &rec, stat); in xfs_inobt_get_rec() 117 struct xfs_btree_cur *cur, in xfs_check_agi_freecount() argument [all …]
|
D | xfs_bmap.c | 126 struct xfs_btree_cur *cur, in xfs_bmbt_lookup_eq() argument 132 cur->bc_rec.b.br_startoff = off; in xfs_bmbt_lookup_eq() 133 cur->bc_rec.b.br_startblock = bno; in xfs_bmbt_lookup_eq() 134 cur->bc_rec.b.br_blockcount = len; in xfs_bmbt_lookup_eq() 135 return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat); in xfs_bmbt_lookup_eq() 140 struct xfs_btree_cur *cur, in xfs_bmbt_lookup_ge() argument 146 cur->bc_rec.b.br_startoff = off; in xfs_bmbt_lookup_ge() 147 cur->bc_rec.b.br_startblock = bno; in xfs_bmbt_lookup_ge() 148 cur->bc_rec.b.br_blockcount = len; in xfs_bmbt_lookup_ge() 149 return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat); in xfs_bmbt_lookup_ge() [all …]
|
D | xfs_btree.h | 142 #define XFS_BTREE_STATS_INC(cur, stat) \ argument 144 switch (cur->bc_btnum) { \ 155 #define XFS_BTREE_STATS_ADD(cur, stat, val) \ argument 157 switch (cur->bc_btnum) { \ 179 void (*set_root)(struct xfs_btree_cur *cur, 183 int (*alloc_block)(struct xfs_btree_cur *cur, 187 int (*free_block)(struct xfs_btree_cur *cur, struct xfs_buf *bp); 190 void (*update_lastrec)(struct xfs_btree_cur *cur, 196 int (*get_minrecs)(struct xfs_btree_cur *cur, int level); 197 int (*get_maxrecs)(struct xfs_btree_cur *cur, int level); [all …]
|
D | xfs_itable.c | 201 xfs_btree_cur_t *cur; /* btree cursor for ialloc btree */ in xfs_bulkstat() local 280 cur = xfs_inobt_init_cursor(mp, NULL, agbp, agno); in xfs_bulkstat() 294 error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_LE, in xfs_bulkstat() 299 !(error = xfs_inobt_get_rec(cur, &r, &i)) && in xfs_bulkstat() 338 error = xfs_btree_increment(cur, 0, &tmp); in xfs_bulkstat() 343 error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &tmp); in xfs_bulkstat() 362 error = xfs_inobt_lookup(cur, agino, in xfs_bulkstat() 375 error = xfs_inobt_get_rec(cur, &r, &i); in xfs_bulkstat() 412 error = xfs_btree_increment(cur, 0, &tmp); in xfs_bulkstat() 420 xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); in xfs_bulkstat() [all …]
|
D | xfs_trans_priv.h | 83 struct xfs_ail_cursor *cur, 115 struct xfs_ail_cursor *cur, 118 struct xfs_ail_cursor *cur, 121 struct xfs_ail_cursor *cur); 123 struct xfs_ail_cursor *cur);
|
D | xfs_discard.c | 46 struct xfs_btree_cur *cur; in xfs_trim_extents() local 58 cur = xfs_allocbt_init_cursor(mp, NULL, agbp, agno, XFS_BTNUM_CNT); in xfs_trim_extents() 70 error = xfs_alloc_lookup_ge(cur, 0, in xfs_trim_extents() 85 error = xfs_alloc_get_rec(cur, &fbno, &flen, &i); in xfs_trim_extents() 133 error = xfs_btree_decrement(cur, 0, &i); in xfs_trim_extents() 139 xfs_btree_del_cursor(cur, error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR); in xfs_trim_extents()
|
D | xfs_alloc.h | 215 struct xfs_btree_cur *cur, /* btree cursor */ 222 struct xfs_btree_cur *cur, /* btree cursor */ 229 struct xfs_btree_cur *cur, /* btree cursor */
|
D | xfs_ialloc.h | 144 int xfs_inobt_lookup(struct xfs_btree_cur *cur, xfs_agino_t ino, 150 int xfs_inobt_get_rec(struct xfs_btree_cur *cur,
|
/fs/btrfs/ |
D | raid56.c | 188 struct btrfs_stripe_hash *cur; in btrfs_alloc_stripe_hash_table() local 218 cur = h + i; in btrfs_alloc_stripe_hash_table() 219 INIT_LIST_HEAD(&cur->hash_list); in btrfs_alloc_stripe_hash_table() 220 spin_lock_init(&cur->lock); in btrfs_alloc_stripe_hash_table() 221 init_waitqueue_head(&cur->wait); in btrfs_alloc_stripe_hash_table() 560 struct btrfs_raid_bio *cur) in rbio_can_merge() argument 563 test_bit(RBIO_RMW_LOCKED_BIT, &cur->flags)) in rbio_can_merge() 574 test_bit(RBIO_CACHE_BIT, &cur->flags)) in rbio_can_merge() 578 cur->raid_map[0]) in rbio_can_merge() 583 cur->read_rebuild) { in rbio_can_merge() [all …]
|
D | extent_io.c | 2718 u64 cur = start; in __extent_read_full_page() local 2767 while (cur <= end) { in __extent_read_full_page() 2770 if (cur >= last_byte) { in __extent_read_full_page() 2779 set_extent_uptodate(tree, cur, cur + iosize - 1, in __extent_read_full_page() 2781 unlock_extent_cached(tree, cur, cur + iosize - 1, in __extent_read_full_page() 2785 em = get_extent(inode, page, pg_offset, cur, in __extent_read_full_page() 2786 end - cur + 1, 0); in __extent_read_full_page() 2789 unlock_extent(tree, cur, end); in __extent_read_full_page() 2792 extent_offset = cur - em->start; in __extent_read_full_page() 2793 BUG_ON(extent_map_end(em) <= cur); in __extent_read_full_page() [all …]
|
D | async-thread.c | 261 struct list_head *cur = NULL; in get_next_work() local 264 cur = prio_head->next; in get_next_work() 271 cur = head->next; in get_next_work() 273 if (cur) in get_next_work() 282 cur = prio_head->next; in get_next_work() 284 cur = head->next; in get_next_work() 287 if (!cur) in get_next_work() 291 work = list_entry(cur, struct btrfs_work, list); in get_next_work() 410 struct list_head *cur; in btrfs_stop_workers() local 417 cur = workers->worker_list.next; in btrfs_stop_workers() [all …]
|
D | send.c | 767 u32 cur = 0; in iterate_inode_ref() local 803 while (cur < total) { in iterate_inode_ref() 807 iref = (struct btrfs_inode_ref *)(ptr + cur); in iterate_inode_ref() 813 extref = (struct btrfs_inode_extref *)(ptr + cur); in iterate_inode_ref() 852 cur += elem_size + name_len; in iterate_inode_ref() 893 u32 cur; in iterate_dir_item() local 911 cur = 0; in iterate_dir_item() 916 while (cur < total) { in iterate_dir_item() 953 cur += len; in iterate_dir_item() 1861 struct name_cache_entry *cur; in name_cache_search() local [all …]
|
D | ulist.c | 149 struct ulist_node *cur = NULL; in ulist_rbtree_insert() local 153 cur = rb_entry(parent, struct ulist_node, rb_node); in ulist_rbtree_insert() 155 if (cur->val < ins->val) in ulist_rbtree_insert() 157 else if (cur->val > ins->val) in ulist_rbtree_insert()
|
D | ctree.c | 486 struct tree_mod_elem *cur; in __tree_mod_log_insert() local 493 cur = container_of(*new, struct tree_mod_elem, node); in __tree_mod_log_insert() 495 if (cur->index < tm->index) in __tree_mod_log_insert() 497 else if (cur->index > tm->index) in __tree_mod_log_insert() 499 else if (cur->seq < tm->seq) in __tree_mod_log_insert() 501 else if (cur->seq > tm->seq) in __tree_mod_log_insert() 716 struct tree_mod_elem *cur = NULL; in __tree_mod_log_search() local 724 cur = container_of(node, struct tree_mod_elem, node); in __tree_mod_log_search() 725 if (cur->index < index) { in __tree_mod_log_search() 727 } else if (cur->index > index) { in __tree_mod_log_search() [all …]
|
/fs/cifs/ |
D | smb2file.c | 121 struct smb2_lock_element *buf, *cur; in smb2_unlock_range() local 143 cur = buf; in smb2_unlock_range() 163 cur->Length = cpu_to_le64(li->length); in smb2_unlock_range() 164 cur->Offset = cpu_to_le64(li->offset); in smb2_unlock_range() 165 cur->Flags = cpu_to_le32(SMB2_LOCKFLAG_UNLOCK); in smb2_unlock_range() 191 cur = buf; in smb2_unlock_range() 194 cur++; in smb2_unlock_range() 220 struct smb2_lock_element *cur = buf; in smb2_push_mand_fdlocks() local 224 cur->Length = cpu_to_le64(li->length); in smb2_push_mand_fdlocks() 225 cur->Offset = cpu_to_le64(li->offset); in smb2_push_mand_fdlocks() [all …]
|
/fs/efs/ |
D | inode.c | 205 int cur, last, first = 1; in efs_map_block() local 230 cur = (last + dirext) % in->numextents; in efs_map_block() 231 if ((result = efs_extent_check(&in->extents[cur], block, sb))) { in efs_map_block() 232 in->lastextent = cur; in efs_map_block() 248 cur = (last + indext) % indexts; in efs_map_block() 258 for(dirext = 0; cur < ibase && dirext < direxts; dirext++) { in efs_map_block() 265 …printk(KERN_ERR "EFS: couldn't find direct extent for indirect extent %d (block %u)\n", cur, block… in efs_map_block() 272 (cur - ibase) / in efs_map_block() 274 ioffset = (cur - ibase) % in efs_map_block() 297 printk(KERN_ERR "EFS: extent %d has bad magic number in block %d\n", cur, iblock); in efs_map_block() [all …]
|
/fs/ |
D | pnode.c | 408 static struct mount *next_descendent(struct mount *root, struct mount *cur) in next_descendent() argument 410 if (!IS_MNT_NEW(cur) && !list_empty(&cur->mnt_slave_list)) in next_descendent() 411 return first_slave(cur); in next_descendent() 413 struct mount *master = cur->mnt_master; in next_descendent() 415 if (!master || cur->mnt_slave.next != &master->mnt_slave_list) { in next_descendent() 416 struct mount *next = next_slave(cur); in next_descendent() 420 cur = master; in next_descendent() 421 } while (cur != root); in next_descendent()
|
/fs/reiserfs/ |
D | ibalance.c | 123 struct buffer_head *cur = cur_bi->bi_bh; in internal_insert_childs() local 134 blkh = B_BLK_HEAD(cur); in internal_insert_childs() 138 RFALSE(B_FREE_SPACE(cur) < count * (KEY_SIZE + DC_SIZE), in internal_insert_childs() 140 B_FREE_SPACE(cur), count * (KEY_SIZE + DC_SIZE)); in internal_insert_childs() 143 dc = B_N_CHILD(cur, to + 1); in internal_insert_childs() 156 ih = B_N_PDELIM_KEY(cur, ((to == -1) ? 0 : to)); in internal_insert_childs() 172 do_balance_mark_internal_dirty(cur_bi->tb, cur, 0); in internal_insert_childs() 175 check_internal(cur); in internal_insert_childs() 199 struct buffer_head *cur = cur_bi->bi_bh; in internal_delete_pointers_items() local 205 RFALSE(cur == NULL, "buffer is 0"); in internal_delete_pointers_items() [all …]
|
D | journal.c | 1823 struct reiserfs_journal_cnode *cur; in remove_journal_hash() local 1830 cur = *head; in remove_journal_hash() 1831 while (cur) { in remove_journal_hash() 1832 if (cur->blocknr == block && cur->sb == sb in remove_journal_hash() 1833 && (jl == NULL || jl == cur->jlist) in remove_journal_hash() 1834 && (!test_bit(BLOCK_FREED, &cur->state) || remove_freed)) { in remove_journal_hash() 1835 if (cur->hnext) { in remove_journal_hash() 1836 cur->hnext->hprev = cur->hprev; in remove_journal_hash() 1838 if (cur->hprev) { in remove_journal_hash() 1839 cur->hprev->hnext = cur->hnext; in remove_journal_hash() [all …]
|
/fs/ext4/ |
D | mballoc.c | 567 struct list_head *cur; in __mb_check_buddy() local 639 list_for_each(cur, &grp->bb_prealloc_list) { in __mb_check_buddy() 642 pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); in __mb_check_buddy() 1269 static void mb_clear_bits(void *bm, int cur, int len) in mb_clear_bits() argument 1273 len = cur + len; in mb_clear_bits() 1274 while (cur < len) { in mb_clear_bits() 1275 if ((cur & 31) == 0 && (len - cur) >= 32) { in mb_clear_bits() 1277 addr = bm + (cur >> 3); in mb_clear_bits() 1279 cur += 32; in mb_clear_bits() 1282 mb_clear_bit(cur, bm); in mb_clear_bits() [all …]
|