/fs/btrfs/ |
D | check-integrity.c | 185 struct btrfsic_state *state; member 246 static int btrfsic_process_metablock(struct btrfsic_state *state, 254 struct btrfsic_state *state, 265 static int btrfsic_handle_extent_data(struct btrfsic_state *state, 269 static int btrfsic_map_block(struct btrfsic_state *state, u64 bytenr, u32 len, 273 static int btrfsic_read_block(struct btrfsic_state *state, 276 struct btrfsic_state *state, 280 static int btrfsic_is_block_ref_by_superblock(const struct btrfsic_state *state, 283 static int btrfsic_check_all_ref_blocks(struct btrfsic_state *state, 286 static void btrfsic_print_add_link(const struct btrfsic_state *state, [all …]
|
D | extent_io.c | 37 static inline bool extent_state_in_tree(const struct extent_state *state) in extent_state_in_tree() argument 39 return !RB_EMPTY_NODE(&state->rb_node); in extent_state_in_tree() 95 struct extent_state *state; in btrfs_extent_state_leak_debug_check() local 98 state = list_entry(states.next, struct extent_state, leak_list); in btrfs_extent_state_leak_debug_check() 100 state->start, state->end, state->state, in btrfs_extent_state_leak_debug_check() 101 extent_state_in_tree(state), in btrfs_extent_state_leak_debug_check() 102 refcount_read(&state->refs)); in btrfs_extent_state_leak_debug_check() 103 list_del(&state->leak_list); in btrfs_extent_state_leak_debug_check() 104 kmem_cache_free(extent_state_cache, state); in btrfs_extent_state_leak_debug_check() 150 static int add_extent_changeset(struct extent_state *state, u32 bits, in add_extent_changeset() argument [all …]
|
/fs/nfs/ |
D | nfs4state.c | 672 struct nfs4_state *state; in nfs4_alloc_open_state() local 674 state = kzalloc(sizeof(*state), GFP_NOFS); in nfs4_alloc_open_state() 675 if (!state) in nfs4_alloc_open_state() 677 refcount_set(&state->count, 1); in nfs4_alloc_open_state() 678 INIT_LIST_HEAD(&state->lock_states); in nfs4_alloc_open_state() 679 spin_lock_init(&state->state_lock); in nfs4_alloc_open_state() 680 seqlock_init(&state->seqlock); in nfs4_alloc_open_state() 681 init_waitqueue_head(&state->waitq); in nfs4_alloc_open_state() 682 return state; in nfs4_alloc_open_state() 686 nfs4_state_set_mode_locked(struct nfs4_state *state, fmode_t fmode) in nfs4_state_set_mode_locked() argument [all …]
|
D | nfs4proc.c | 475 struct nfs4_state *state = exception->state; in nfs4_do_handle_exception() local 485 if (stateid == NULL && state != NULL) in nfs4_do_handle_exception() 486 stateid = nfs4_recoverable_stateid(&state->stateid); in nfs4_do_handle_exception() 520 if (state == NULL) in nfs4_do_handle_exception() 522 ret = nfs4_schedule_stateid_recovery(server, state); in nfs4_do_handle_exception() 659 struct nfs4_state *state, long *timeout) in nfs4_async_handle_error() argument 662 .state = state, in nfs4_async_handle_error() 1472 if (p->state != NULL) in nfs4_opendata_free() 1473 nfs4_put_open_state(p->state); in nfs4_opendata_free() 1493 static bool nfs4_mode_match_open_stateid(struct nfs4_state *state, in nfs4_mode_match_open_stateid() argument [all …]
|
D | nfs4trace.h | 593 #define show_nfs4_clp_state(state) \ argument 594 __print_flags(state, "|", \ 624 __field(unsigned long, state) 629 __entry->state = clp->cl_state; 635 show_nfs4_clp_state(__entry->state) 650 __field(unsigned long, state) 657 __entry->state = clp->cl_state; 665 show_nfs4_clp_state(__entry->state), -__entry->error, 812 const struct nfs4_state *state = ctx->state; 819 if (!IS_ERR_OR_NULL(state)) { [all …]
|
/fs/nfsd/ |
D | nfs4acl.c | 455 init_state(struct posix_acl_state *state, int cnt) in init_state() argument 459 memset(state, 0, sizeof(struct posix_acl_state)); in init_state() 460 state->empty = 1; in init_state() 468 state->users = kzalloc(alloc, GFP_KERNEL); in init_state() 469 if (!state->users) in init_state() 471 state->groups = kzalloc(alloc, GFP_KERNEL); in init_state() 472 if (!state->groups) { in init_state() 473 kfree(state->users); in init_state() 480 free_state(struct posix_acl_state *state) { in free_state() argument 481 kfree(state->users); in free_state() [all …]
|
/fs/xfs/libxfs/ |
D | xfs_attr.c | 61 struct xfs_da_state **state); 62 STATIC int xfs_attr_fillstate(xfs_da_state_t *state); 63 STATIC int xfs_attr_refillstate(xfs_da_state_t *state); 67 struct xfs_da_state *state); 1075 struct xfs_da_state *state; in xfs_attr_node_hasname() local 1078 state = xfs_da_state_alloc(args); in xfs_attr_node_hasname() 1080 *statep = state; in xfs_attr_node_hasname() 1085 error = xfs_da3_node_lookup_int(state, &retval); in xfs_attr_node_hasname() 1090 xfs_da_state_free(state); in xfs_attr_node_hasname() 1163 struct xfs_da_state *state = dac->da_state; in xfs_attr_node_addname() local [all …]
|
D | xfs_da_btree.c | 39 STATIC int xfs_da3_root_split(xfs_da_state_t *state, 42 STATIC int xfs_da3_node_split(xfs_da_state_t *state, 48 STATIC void xfs_da3_node_rebalance(xfs_da_state_t *state, 51 STATIC void xfs_da3_node_add(xfs_da_state_t *state, 58 STATIC int xfs_da3_root_join(xfs_da_state_t *state, 60 STATIC int xfs_da3_node_toosmall(xfs_da_state_t *state, int *retval); 61 STATIC void xfs_da3_node_remove(xfs_da_state_t *state, 63 STATIC void xfs_da3_node_unbalance(xfs_da_state_t *state, 70 STATIC int xfs_da3_blk_unlink(xfs_da_state_t *state, 85 struct xfs_da_state *state; in xfs_da_state_alloc() local [all …]
|
D | xfs_dir2_node.c | 29 static void xfs_dir2_leafn_rebalance(xfs_da_state_t *state, 629 xfs_da_state_t *state) /* state to fill in */ in xfs_dir2_leafn_lookup_for_addname() argument 664 if (state->extravalid) { in xfs_dir2_leafn_lookup_for_addname() 666 curbp = state->extrablk.bp; in xfs_dir2_leafn_lookup_for_addname() 667 curfdb = state->extrablk.blkno; in xfs_dir2_leafn_lookup_for_addname() 752 state->extravalid = 1; in xfs_dir2_leafn_lookup_for_addname() 753 state->extrablk.bp = curbp; in xfs_dir2_leafn_lookup_for_addname() 754 state->extrablk.index = fi; in xfs_dir2_leafn_lookup_for_addname() 755 state->extrablk.blkno = curfdb; in xfs_dir2_leafn_lookup_for_addname() 762 state->extrablk.magic = XFS_DIR2_FREE_MAGIC; in xfs_dir2_leafn_lookup_for_addname() [all …]
|
D | xfs_attr_leaf.c | 54 STATIC void xfs_attr3_leaf_rebalance(xfs_da_state_t *state, 57 STATIC int xfs_attr3_leaf_figure_balance(xfs_da_state_t *state, 1300 struct xfs_da_state *state, in xfs_attr3_leaf_split() argument 1307 trace_xfs_attr_leaf_split(state->args); in xfs_attr3_leaf_split() 1313 error = xfs_da_grow_inode(state->args, &blkno); in xfs_attr3_leaf_split() 1316 error = xfs_attr3_leaf_create(state->args, blkno, &newblk->bp); in xfs_attr3_leaf_split() 1326 xfs_attr3_leaf_rebalance(state, oldblk, newblk); in xfs_attr3_leaf_split() 1327 error = xfs_da3_blk_link(state, oldblk, newblk); in xfs_attr3_leaf_split() 1338 if (state->inleaf) { in xfs_attr3_leaf_split() 1339 trace_xfs_attr_leaf_add_old(state->args); in xfs_attr3_leaf_split() [all …]
|
D | xfs_bmap.c | 1452 int state = xfs_bmap_fork_to_state(whichfork); in xfs_bmap_add_extent_delay_real() local 1487 state |= BMAP_LEFT_FILLING; in xfs_bmap_add_extent_delay_real() 1489 state |= BMAP_RIGHT_FILLING; in xfs_bmap_add_extent_delay_real() 1496 state |= BMAP_LEFT_VALID; in xfs_bmap_add_extent_delay_real() 1498 state |= BMAP_LEFT_DELAY; in xfs_bmap_add_extent_delay_real() 1501 if ((state & BMAP_LEFT_VALID) && !(state & BMAP_LEFT_DELAY) && in xfs_bmap_add_extent_delay_real() 1506 state |= BMAP_LEFT_CONTIG; in xfs_bmap_add_extent_delay_real() 1514 state |= BMAP_RIGHT_VALID; in xfs_bmap_add_extent_delay_real() 1516 state |= BMAP_RIGHT_DELAY; in xfs_bmap_add_extent_delay_real() 1519 if ((state & BMAP_RIGHT_VALID) && !(state & BMAP_RIGHT_DELAY) && in xfs_bmap_add_extent_delay_real() [all …]
|
/fs/fscache/ |
D | operation.c | 42 op->state = FSCACHE_OP_ST_INITIALISED; in fscache_operation_init() 72 ASSERTIFCMP(op->state != FSCACHE_OP_ST_IN_PROGRESS, in fscache_enqueue_operation() 73 op->state, ==, FSCACHE_OP_ST_CANCELLED); in fscache_enqueue_operation() 102 ASSERTCMP(op->state, ==, FSCACHE_OP_ST_PENDING); in fscache_run_op() 104 op->state = FSCACHE_OP_ST_IN_PROGRESS; in fscache_run_op() 131 op->debug_id, object->debug_id, object->state->name); in fscache_report_unexpected_submission() 132 kdebug("objstate=%s [%s]", object->state->name, ostate->name); in fscache_report_unexpected_submission() 168 ASSERTCMP(op->state, ==, FSCACHE_OP_ST_INITIALISED); in fscache_submit_exclusive_op() 176 ostate = object->state; in fscache_submit_exclusive_op() 179 op->state = FSCACHE_OP_ST_PENDING; in fscache_submit_exclusive_op() [all …]
|
D | object.c | 51 #define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); }) argument 70 #define TRANSIT_TO(state, emask) \ argument 71 { .events = (emask), .transit_to = STATE(state) } 169 const struct fscache_state *state, *new_state; in fscache_object_sm_dispatcher() local 177 object->debug_id, object->state->name, object->events); in fscache_object_sm_dispatcher() 182 state = object->state; in fscache_object_sm_dispatcher() 193 state = t->transit_to; in fscache_object_sm_dispatcher() 194 ASSERT(state->work != NULL); in fscache_object_sm_dispatcher() 205 if (!state->work) { in fscache_object_sm_dispatcher() 207 for (t = state->transitions; t->events; t++) { in fscache_object_sm_dispatcher() [all …]
|
/fs/dlm/ |
D | midcomms.c | 178 int state; member 223 static inline const char *dlm_state_str(int state) in dlm_state_str() argument 225 switch (state) { in dlm_state_str() 247 return dlm_state_str(node->state); in dlm_midcomms_state() 318 node->state = DLM_CLOSED; in midcomms_node_reset() 451 node->nodeid, dlm_state_str(node->state)); in dlm_pas_fin_ack_rcv() 453 switch (node->state) { in dlm_pas_fin_ack_rcv() 465 __func__, node->state); in dlm_pas_fin_ack_rcv() 483 node->nodeid, dlm_state_str(node->state)); in dlm_midcomms_receive_buffer() 485 switch (node->state) { in dlm_midcomms_receive_buffer() [all …]
|
/fs/quota/ |
D | quota.c | 120 struct qc_state state; in quota_getinfo() local 127 ret = sb->s_qcop->get_state(sb, &state); in quota_getinfo() 130 tstate = state.s_state + type; in quota_getinfo() 334 static int quota_state_to_flags(struct qc_state *state) in quota_state_to_flags() argument 338 if (state->s_state[USRQUOTA].flags & QCI_ACCT_ENABLED) in quota_state_to_flags() 340 if (state->s_state[USRQUOTA].flags & QCI_LIMITS_ENFORCED) in quota_state_to_flags() 342 if (state->s_state[GRPQUOTA].flags & QCI_ACCT_ENABLED) in quota_state_to_flags() 344 if (state->s_state[GRPQUOTA].flags & QCI_LIMITS_ENFORCED) in quota_state_to_flags() 346 if (state->s_state[PRJQUOTA].flags & QCI_ACCT_ENABLED) in quota_state_to_flags() 348 if (state->s_state[PRJQUOTA].flags & QCI_LIMITS_ENFORCED) in quota_state_to_flags() [all …]
|
/fs/ext4/ |
D | fast_commit.c | 1440 struct ext4_fc_replay_state *state; in ext4_fc_record_modified_inode() local 1443 state = &EXT4_SB(sb)->s_fc_replay_state; in ext4_fc_record_modified_inode() 1444 for (i = 0; i < state->fc_modified_inodes_used; i++) in ext4_fc_record_modified_inode() 1445 if (state->fc_modified_inodes[i] == ino) in ext4_fc_record_modified_inode() 1447 if (state->fc_modified_inodes_used == state->fc_modified_inodes_size) { in ext4_fc_record_modified_inode() 1450 fc_modified_inodes = krealloc(state->fc_modified_inodes, in ext4_fc_record_modified_inode() 1451 sizeof(int) * (state->fc_modified_inodes_size + in ext4_fc_record_modified_inode() 1456 state->fc_modified_inodes = fc_modified_inodes; in ext4_fc_record_modified_inode() 1457 state->fc_modified_inodes_size += in ext4_fc_record_modified_inode() 1460 state->fc_modified_inodes[state->fc_modified_inodes_used++] = ino; in ext4_fc_record_modified_inode() [all …]
|
/fs/jfs/ |
D | jfs_mount.c | 232 if (chkSuper(sb) || (sbi->state != FM_CLEAN)) in jfs_mount_rw() 353 sbi->state = le32_to_cpu(j_sb->s_state); in chkSuper() 400 int updateSuper(struct super_block *sb, uint state) in updateSuper() argument 408 if (state == FM_DIRTY) { in updateSuper() 409 sbi->p_state = state; in updateSuper() 411 } else if (state == FM_MOUNT) { in updateSuper() 412 sbi->p_state = sbi->state; in updateSuper() 413 state = FM_DIRTY; in updateSuper() 414 } else if (state == FM_CLEAN) { in updateSuper() 415 state = sbi->p_state; in updateSuper() [all …]
|
/fs/xfs/ |
D | xfs_quotaops.c | 57 struct qc_state *state) in xfs_fs_get_quota_state() argument 62 memset(state, 0, sizeof(*state)); in xfs_fs_get_quota_state() 65 state->s_incoredqs = q->qi_dquots; in xfs_fs_get_quota_state() 67 state->s_state[USRQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 69 state->s_state[USRQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 71 state->s_state[GRPQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 73 state->s_state[GRPQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 75 state->s_state[PRJQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 77 state->s_state[PRJQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 79 xfs_qm_fill_state(&state->s_state[USRQUOTA], mp, q->qi_uquotaip, in xfs_fs_get_quota_state() [all …]
|
/fs/afs/ |
D | cell.c | 240 enum afs_cell_state state; in afs_lookup_cell() local 301 wait_var_event(&cell->state, in afs_lookup_cell() 303 state = smp_load_acquire(&cell->state); /* vs error */ in afs_lookup_cell() 304 state == AFS_CELL_ACTIVE || state == AFS_CELL_REMOVED; in afs_lookup_cell() 308 if (state == AFS_CELL_REMOVED) { in afs_lookup_cell() 745 _debug("state %u", cell->state); in afs_manage_cell() 746 switch (cell->state) { in afs_manage_cell() 755 smp_store_release(&cell->state, AFS_CELL_REMOVED); in afs_manage_cell() 758 if (cell->state == AFS_CELL_REMOVED) { in afs_manage_cell() 759 wake_up_var(&cell->state); in afs_manage_cell() [all …]
|
D | rxrpc.c | 430 call->state = AFS_CALL_COMPLETE; in afs_make_call() 472 enum afs_call_state state; in afs_deliver_to_call() local 479 while (state = READ_ONCE(call->state), in afs_deliver_to_call() 480 state == AFS_CALL_CL_AWAIT_REPLY || in afs_deliver_to_call() 481 state == AFS_CALL_SV_AWAIT_OP_ID || in afs_deliver_to_call() 482 state == AFS_CALL_SV_AWAIT_REQUEST || in afs_deliver_to_call() 483 state == AFS_CALL_SV_AWAIT_ACK in afs_deliver_to_call() 485 if (state == AFS_CALL_SV_AWAIT_ACK) { in afs_deliver_to_call() 505 state = READ_ONCE(call->state); in afs_deliver_to_call() 511 if (state == AFS_CALL_CL_PROC_REPLY) { in afs_deliver_to_call() [all …]
|
D | flock.c | 25 static inline void afs_set_lock_state(struct afs_vnode *vnode, enum afs_lock_state state) in afs_set_lock_state() argument 27 _debug("STATE %u -> %u", vnode->lock_state, state); in afs_set_lock_state() 28 vnode->lock_state = state; in afs_set_lock_state() 100 p->fl_u.afs.state = AFS_LOCK_GRANTED; in afs_grant_locks() 127 p->fl_u.afs.state = error; in afs_next_locker() 143 next->fl_u.afs.state = AFS_LOCK_YOUR_TRY; in afs_next_locker() 168 p->fl_u.afs.state = -ENOENT; in afs_kill_lockers_enoent() 471 fl->fl_u.afs.state = AFS_LOCK_PENDING; in afs_do_setlk() 514 fl->fl_u.afs.state = AFS_LOCK_GRANTED; in afs_do_setlk() 521 fl->fl_u.afs.state = AFS_LOCK_GRANTED; in afs_do_setlk() [all …]
|
/fs/ksmbd/ |
D | oplock.c | 101 lease->state = lctx->req_state; in alloc_lease() 266 pr_err("lease state(0x%x)\n", lease->state); in opinfo_write_to_read() 272 lease->state = lease->new_state; in opinfo_write_to_read() 286 lease->state = lease->new_state; in opinfo_read_handle_to_read() 305 pr_err("lease state(0x%x)\n", lease->state); in opinfo_write_to_none() 310 lease->state = lease->new_state; in opinfo_write_to_none() 327 pr_err("lease state(0x%x)\n", lease->state); in opinfo_read_to_none() 332 lease->state = lease->new_state; in opinfo_read_to_none() 346 if (!(lease->state & SMB2_LEASE_READ_CACHING_LE)) { in lease_read_to_write() 347 ksmbd_debug(OPLOCK, "bad lease state(0x%x)\n", lease->state); in lease_read_to_write() [all …]
|
/fs/ocfs2/dlm/ |
D | dlmthread.c | 51 if (res->state & flags) { in __dlm_wait_on_lockres_flags() 87 if (!list_empty(&res->dirty) || res->state & DLM_LOCK_RES_DIRTY) in __dlm_lockres_unused() 90 if (res->state & (DLM_LOCK_RES_RECOVERING| in __dlm_lockres_unused() 187 res->state &= ~DLM_LOCK_RES_DROPPING_REF; in __dlm_do_purge_lockres() 205 if (res->state & DLM_LOCK_RES_DROPPING_REF) { in dlm_purge_lockres() 212 res->state |= DLM_LOCK_RES_DROPPING_REF; in dlm_purge_lockres() 269 res->state &= ~DLM_LOCK_RES_DROPPING_REF; in dlm_purge_lockres() 314 (lockres->state & DLM_LOCK_RES_MIGRATING) || in dlm_run_purge_list() 320 !unused, lockres->state, in dlm_run_purge_list() 354 BUG_ON((res->state & (DLM_LOCK_RES_MIGRATING| in dlm_shuffle_lists() [all …]
|
D | dlmdebug.c | 90 buf, res->owner, res->state); in __dlm_print_one_lock_resource() 501 res->owner, res->state, res->last_used, in dump_lockres() 664 char *state; in debug_state_print() local 672 state = "NEW"; break; in debug_state_print() 674 state = "JOINED"; break; in debug_state_print() 676 state = "SHUTDOWN"; break; in debug_state_print() 678 state = "LEAVING"; break; in debug_state_print() 680 state = "UNKNOWN"; break; in debug_state_print() 692 task_pid_nr(dlm->dlm_thread_task), dlm->node_num, state); in debug_state_print() 770 if (dlm->reco.state == DLM_RECO_STATE_ACTIVE) in debug_state_print() [all …]
|
/fs/xfs/scrub/ |
D | dabtree.c | 54 ds->state->path.blk[level].blkno), in xchk_da_process_error() 76 ds->state->path.blk[level].blkno), in xchk_da_set_corrupt() 85 struct xfs_da_state_blk *blk = &ds->state->path.blk[level]; in xchk_da_btree_node_entry() 222 struct xfs_da_state_path *path = &ds->state->path; in xchk_da_btree_block_check_sibling() 223 struct xfs_da_state_path *altpath = &ds->state->altpath; in xchk_da_btree_block_check_sibling() 228 memcpy(altpath, path, sizeof(ds->state->altpath)); in xchk_da_btree_block_check_sibling() 235 error = xfs_da3_path_shift(ds->state, altpath, direction, in xchk_da_btree_block_check_sibling() 244 error = xfs_da3_path_shift(ds->state, altpath, direction, false, in xchk_da_btree_block_check_sibling() 305 memset(&ds->state->altpath, 0, sizeof(ds->state->altpath)); in xchk_da_btree_block_check_siblings() 327 blk = &ds->state->path.blk[level]; in xchk_da_btree_block() [all …]
|