Home
last modified time | relevance | path

Searched refs:state (Results 1 – 25 of 151) sorted by relevance

1234567

/fs/btrfs/
Dcheck-integrity.c200 struct btrfsic_state *state; member
302 static int btrfsic_process_superblock(struct btrfsic_state *state,
304 static int btrfsic_process_metablock(struct btrfsic_state *state,
312 struct btrfsic_state *state,
323 static int btrfsic_handle_extent_data(struct btrfsic_state *state,
327 static int btrfsic_map_block(struct btrfsic_state *state, u64 bytenr, u32 len,
331 static int btrfsic_read_block(struct btrfsic_state *state,
333 static void btrfsic_dump_database(struct btrfsic_state *state);
334 static int btrfsic_test_for_metadata(struct btrfsic_state *state,
343 struct btrfsic_state *state,
[all …]
Dextent_io.c28 static inline bool extent_state_in_tree(const struct extent_state *state) in extent_state_in_tree() argument
30 return !RB_EMPTY_NODE(&state->rb_node); in extent_state_in_tree()
62 struct extent_state *state; in btrfs_leak_debug_check() local
66 state = list_entry(states.next, struct extent_state, leak_list); in btrfs_leak_debug_check()
68 state->start, state->end, state->state, in btrfs_leak_debug_check()
69 extent_state_in_tree(state), in btrfs_leak_debug_check()
70 atomic_read(&state->refs)); in btrfs_leak_debug_check()
71 list_del(&state->leak_list); in btrfs_leak_debug_check()
72 kmem_cache_free(extent_state_cache, state); in btrfs_leak_debug_check()
134 static void add_extent_changeset(struct extent_state *state, unsigned bits, in add_extent_changeset() argument
[all …]
Dtransaction.c91 while (!RB_EMPTY_ROOT(&tree->state)) { in clear_btree_io_tree()
93 struct extent_state *state; in clear_btree_io_tree() local
95 node = rb_first(&tree->state); in clear_btree_io_tree()
96 state = rb_entry(node, struct extent_state, rb_node); in clear_btree_io_tree()
97 rb_erase(&state->rb_node, &tree->state); in clear_btree_io_tree()
98 RB_CLEAR_NODE(&state->rb_node); in clear_btree_io_tree()
103 ASSERT(!waitqueue_active(&state->wq)); in clear_btree_io_tree()
104 free_extent_state(state); in clear_btree_io_tree()
188 if (btrfs_blocked_trans_types[cur_trans->state] & type) { in join_transaction()
236 cur_trans->state = TRANS_STATE_RUNNING; in join_transaction()
[all …]
/fs/nfsd/
Dnfs4acl.c455 init_state(struct posix_acl_state *state, int cnt) in init_state() argument
459 memset(state, 0, sizeof(struct posix_acl_state)); in init_state()
460 state->empty = 1; in init_state()
468 state->users = kzalloc(alloc, GFP_KERNEL); in init_state()
469 if (!state->users) in init_state()
471 state->groups = kzalloc(alloc, GFP_KERNEL); in init_state()
472 if (!state->groups) { in init_state()
473 kfree(state->users); in init_state()
480 free_state(struct posix_acl_state *state) { in free_state() argument
481 kfree(state->users); in free_state()
[all …]
/fs/nfs/
Dnfs4state.c655 struct nfs4_state *state; in nfs4_alloc_open_state() local
657 state = kzalloc(sizeof(*state), GFP_NOFS); in nfs4_alloc_open_state()
658 if (!state) in nfs4_alloc_open_state()
660 atomic_set(&state->count, 1); in nfs4_alloc_open_state()
661 INIT_LIST_HEAD(&state->lock_states); in nfs4_alloc_open_state()
662 spin_lock_init(&state->state_lock); in nfs4_alloc_open_state()
663 seqlock_init(&state->seqlock); in nfs4_alloc_open_state()
664 return state; in nfs4_alloc_open_state()
668 nfs4_state_set_mode_locked(struct nfs4_state *state, fmode_t fmode) in nfs4_state_set_mode_locked() argument
670 if (state->state == fmode) in nfs4_state_set_mode_locked()
[all …]
Dnfs4proc.c85 struct nfs4_state *state, struct nfs4_label *ilabel,
347 struct nfs4_state *state = exception->state; in nfs4_do_handle_exception() local
364 if (state == NULL) in nfs4_do_handle_exception()
366 ret = nfs4_schedule_stateid_recovery(server, state); in nfs4_do_handle_exception()
371 if (state != NULL) { in nfs4_do_handle_exception()
372 ret = nfs4_schedule_stateid_recovery(server, state); in nfs4_do_handle_exception()
494 struct nfs4_state *state, long *timeout) in nfs4_async_handle_error() argument
497 .state = state, in nfs4_async_handle_error()
990 struct nfs4_state *state; member
1172 if (p->state != NULL) in nfs4_opendata_free()
[all …]
/fs/xfs/libxfs/
Dxfs_da_btree.c57 STATIC int xfs_da3_root_split(xfs_da_state_t *state,
60 STATIC int xfs_da3_node_split(xfs_da_state_t *state,
66 STATIC void xfs_da3_node_rebalance(xfs_da_state_t *state,
69 STATIC void xfs_da3_node_add(xfs_da_state_t *state,
76 STATIC int xfs_da3_root_join(xfs_da_state_t *state,
78 STATIC int xfs_da3_node_toosmall(xfs_da_state_t *state, int *retval);
79 STATIC void xfs_da3_node_remove(xfs_da_state_t *state,
81 STATIC void xfs_da3_node_unbalance(xfs_da_state_t *state,
88 STATIC int xfs_da3_blk_unlink(xfs_da_state_t *state,
109 xfs_da_state_kill_altpath(xfs_da_state_t *state) in xfs_da_state_kill_altpath() argument
[all …]
Dxfs_dir2_node.c43 static void xfs_dir2_leafn_rebalance(xfs_da_state_t *state,
516 xfs_da_state_t *state) /* state to fill in */ in xfs_dir2_leafn_lookup_for_addname() argument
553 if (state->extravalid) { in xfs_dir2_leafn_lookup_for_addname()
555 curbp = state->extrablk.bp; in xfs_dir2_leafn_lookup_for_addname()
556 curfdb = state->extrablk.blkno; in xfs_dir2_leafn_lookup_for_addname()
641 state->extravalid = 1; in xfs_dir2_leafn_lookup_for_addname()
642 state->extrablk.bp = curbp; in xfs_dir2_leafn_lookup_for_addname()
643 state->extrablk.index = fi; in xfs_dir2_leafn_lookup_for_addname()
644 state->extrablk.blkno = curfdb; in xfs_dir2_leafn_lookup_for_addname()
651 state->extrablk.magic = XFS_DIR2_FREE_MAGIC; in xfs_dir2_leafn_lookup_for_addname()
[all …]
Dxfs_attr.c72 STATIC int xfs_attr_fillstate(xfs_da_state_t *state);
73 STATIC int xfs_attr_refillstate(xfs_da_state_t *state);
877 xfs_da_state_t *state; in xfs_attr_node_addname() local
891 state = xfs_da_state_alloc(); in xfs_attr_node_addname()
892 state->args = args; in xfs_attr_node_addname()
893 state->mp = mp; in xfs_attr_node_addname()
899 error = xfs_da3_node_lookup_int(state, &retval); in xfs_attr_node_addname()
902 blk = &state->path.blk[ state->path.active-1 ]; in xfs_attr_node_addname()
930 retval = xfs_attr3_leaf_add(blk->bp, state->args); in xfs_attr_node_addname()
932 if (state->path.active == 1) { in xfs_attr_node_addname()
[all …]
Dxfs_attr_leaf.c68 STATIC void xfs_attr3_leaf_rebalance(xfs_da_state_t *state,
71 STATIC int xfs_attr3_leaf_figure_balance(xfs_da_state_t *state,
1083 struct xfs_da_state *state, in xfs_attr3_leaf_split() argument
1090 trace_xfs_attr_leaf_split(state->args); in xfs_attr3_leaf_split()
1096 error = xfs_da_grow_inode(state->args, &blkno); in xfs_attr3_leaf_split()
1099 error = xfs_attr3_leaf_create(state->args, blkno, &newblk->bp); in xfs_attr3_leaf_split()
1109 xfs_attr3_leaf_rebalance(state, oldblk, newblk); in xfs_attr3_leaf_split()
1110 error = xfs_da3_blk_link(state, oldblk, newblk); in xfs_attr3_leaf_split()
1121 if (state->inleaf) { in xfs_attr3_leaf_split()
1122 trace_xfs_attr_leaf_add_old(state->args); in xfs_attr3_leaf_split()
[all …]
Dxfs_bmap.c166 xfs_exntst_t state) in xfs_bmbt_update() argument
170 xfs_bmbt_disk_set_allf(&rec.bmbt, off, bno, len, state); in xfs_bmbt_update()
503 int state = 0; in xfs_bmap_trace_exlist() local
506 state |= BMAP_ATTRFORK; in xfs_bmap_trace_exlist()
1722 int state = 0;/* state bits, accessed thru macros */ in xfs_bmap_add_extent_delay_real() local
1762 state |= BMAP_LEFT_FILLING; in xfs_bmap_add_extent_delay_real()
1764 state |= BMAP_RIGHT_FILLING; in xfs_bmap_add_extent_delay_real()
1771 state |= BMAP_LEFT_VALID; in xfs_bmap_add_extent_delay_real()
1775 state |= BMAP_LEFT_DELAY; in xfs_bmap_add_extent_delay_real()
1778 if ((state & BMAP_LEFT_VALID) && !(state & BMAP_LEFT_DELAY) && in xfs_bmap_add_extent_delay_real()
[all …]
Dxfs_da_btree.h167 int xfs_da3_split(xfs_da_state_t *state);
172 int xfs_da3_join(xfs_da_state_t *state);
173 void xfs_da3_fixhashpath(struct xfs_da_state *state,
179 int xfs_da3_node_lookup_int(xfs_da_state_t *state, int *result);
180 int xfs_da3_path_shift(xfs_da_state_t *state, xfs_da_state_path_t *path,
185 int xfs_da3_blk_link(xfs_da_state_t *state, xfs_da_state_blk_t *old_blk,
216 void xfs_da_state_free(xfs_da_state_t *state);
/fs/quota/
Dquota.c122 struct qc_state state; in quota_getinfo() local
131 ret = sb->s_qcop->get_state(sb, &state); in quota_getinfo()
134 tstate = state.s_state + type; in quota_getinfo()
292 static int quota_state_to_flags(struct qc_state *state) in quota_state_to_flags() argument
296 if (state->s_state[USRQUOTA].flags & QCI_ACCT_ENABLED) in quota_state_to_flags()
298 if (state->s_state[USRQUOTA].flags & QCI_LIMITS_ENFORCED) in quota_state_to_flags()
300 if (state->s_state[GRPQUOTA].flags & QCI_ACCT_ENABLED) in quota_state_to_flags()
302 if (state->s_state[GRPQUOTA].flags & QCI_LIMITS_ENFORCED) in quota_state_to_flags()
304 if (state->s_state[PRJQUOTA].flags & QCI_ACCT_ENABLED) in quota_state_to_flags()
306 if (state->s_state[PRJQUOTA].flags & QCI_LIMITS_ENFORCED) in quota_state_to_flags()
[all …]
/fs/fscache/
Doperation.c42 op->state = FSCACHE_OP_ST_INITIALISED; in fscache_operation_init()
69 ASSERTIFCMP(op->state != FSCACHE_OP_ST_IN_PROGRESS, in fscache_enqueue_operation()
70 op->state, ==, FSCACHE_OP_ST_CANCELLED); in fscache_enqueue_operation()
97 ASSERTCMP(op->state, ==, FSCACHE_OP_ST_PENDING); in fscache_run_op()
99 op->state = FSCACHE_OP_ST_IN_PROGRESS; in fscache_run_op()
124 op->debug_id, object->debug_id, object->state->name); in fscache_report_unexpected_submission()
125 kdebug("objstate=%s [%s]", object->state->name, ostate->name); in fscache_report_unexpected_submission()
159 ASSERTCMP(op->state, ==, FSCACHE_OP_ST_INITIALISED); in fscache_submit_exclusive_op()
167 ostate = object->state; in fscache_submit_exclusive_op()
170 op->state = FSCACHE_OP_ST_PENDING; in fscache_submit_exclusive_op()
[all …]
Dobject.c55 #define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); }) argument
74 #define TRANSIT_TO(state, emask) \ argument
75 { .events = (emask), .transit_to = STATE(state) }
171 const struct fscache_state *state, *new_state; in fscache_object_sm_dispatcher() local
178 object->debug_id, object->state->name, object->events); in fscache_object_sm_dispatcher()
183 state = object->state; in fscache_object_sm_dispatcher()
193 state = t->transit_to; in fscache_object_sm_dispatcher()
194 ASSERT(state->work != NULL); in fscache_object_sm_dispatcher()
204 if (!state->work) { in fscache_object_sm_dispatcher()
206 for (t = state->transitions; t->events; t++) { in fscache_object_sm_dispatcher()
[all …]
/fs/jfs/
Djfs_mount.c244 if (chkSuper(sb) || (sbi->state != FM_CLEAN)) in jfs_mount_rw()
365 sbi->state = le32_to_cpu(j_sb->s_state); in chkSuper()
412 int updateSuper(struct super_block *sb, uint state) in updateSuper() argument
420 if (state == FM_DIRTY) { in updateSuper()
421 sbi->p_state = state; in updateSuper()
423 } else if (state == FM_MOUNT) { in updateSuper()
424 sbi->p_state = sbi->state; in updateSuper()
425 state = FM_DIRTY; in updateSuper()
426 } else if (state == FM_CLEAN) { in updateSuper()
427 state = sbi->p_state; in updateSuper()
[all …]
/fs/xfs/
Dxfs_quotaops.c70 struct qc_state *state) in xfs_fs_get_quota_state() argument
75 memset(state, 0, sizeof(*state)); in xfs_fs_get_quota_state()
78 state->s_incoredqs = q->qi_dquots; in xfs_fs_get_quota_state()
80 state->s_state[USRQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state()
82 state->s_state[USRQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state()
84 state->s_state[GRPQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state()
86 state->s_state[GRPQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state()
88 state->s_state[PRJQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state()
90 state->s_state[PRJQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state()
92 xfs_qm_fill_state(&state->s_state[USRQUOTA], mp, q->qi_uquotaip, in xfs_fs_get_quota_state()
[all …]
/fs/afs/
Drxrpc.c316 call->state = AFS_CALL_AWAIT_REPLY; in afs_send_pages()
397 call->state = AFS_CALL_AWAIT_REPLY; in afs_make_call()
465 while ((call->state == AFS_CALL_AWAIT_REPLY || in afs_deliver_to_call()
466 call->state == AFS_CALL_AWAIT_OP_ID || in afs_deliver_to_call()
467 call->state == AFS_CALL_AWAIT_REQUEST || in afs_deliver_to_call()
468 call->state == AFS_CALL_AWAIT_ACK) && in afs_deliver_to_call()
478 call->state == AFS_CALL_AWAIT_REPLY) in afs_deliver_to_call()
479 call->state = AFS_CALL_COMPLETE; in afs_deliver_to_call()
489 if (call->state != AFS_CALL_AWAIT_REPLY) in afs_deliver_to_call()
495 call->state = AFS_CALL_ERROR; in afs_deliver_to_call()
[all …]
Dvlocation.c178 vl->state = AFS_VL_NEW; in afs_vlocation_alloc()
412 vl->state = AFS_VL_CREATING; in afs_vlocation_lookup()
420 vl->state = AFS_VL_VALID; in afs_vlocation_lookup()
447 while (vl->state != AFS_VL_VALID) { in afs_vlocation_lookup()
448 afs_vlocation_state_t state = vl->state; in afs_vlocation_lookup() local
450 _debug("invalid [state %d]", state); in afs_vlocation_lookup()
452 if (state == AFS_VL_NEW || state == AFS_VL_NO_VOLUME) { in afs_vlocation_lookup()
453 vl->state = AFS_VL_CREATING; in afs_vlocation_lookup()
464 vl->state == AFS_VL_NEW || in afs_vlocation_lookup()
465 vl->state == AFS_VL_VALID || in afs_vlocation_lookup()
[all …]
/fs/ceph/
Dmdsmap.c25 if (1 == m->m_max_mds && m->m_info[0].state > 0) in ceph_mdsmap_get_random_mds()
30 if (m->m_info[i].state > 0) in ceph_mdsmap_get_random_mds()
39 while (m->m_info[i].state <= 0) in ceph_mdsmap_get_random_mds()
88 s32 mds, inc, state; in ceph_mdsmap_decode() local
110 state = ceph_decode_32(p); in ceph_mdsmap_decode()
129 ceph_mds_state_name(state)); in ceph_mdsmap_decode()
131 if (mds < 0 || mds >= m->m_max_mds || state <= 0) in ceph_mdsmap_decode()
136 info->state = state; in ceph_mdsmap_decode()
/fs/ocfs2/dlm/
Ddlmthread.c70 if (res->state & flags) { in __dlm_wait_on_lockres_flags()
106 if (!list_empty(&res->dirty) || res->state & DLM_LOCK_RES_DIRTY) in __dlm_lockres_unused()
109 if (res->state & DLM_LOCK_RES_RECOVERING) in __dlm_lockres_unused()
177 res->state |= DLM_LOCK_RES_DROPPING_REF; in dlm_purge_lockres()
227 res->state &= ~DLM_LOCK_RES_DROPPING_REF; in dlm_purge_lockres()
272 (lockres->state & DLM_LOCK_RES_MIGRATING) || in dlm_run_purge_list()
278 !unused, lockres->state, in dlm_run_purge_list()
312 BUG_ON((res->state & (DLM_LOCK_RES_MIGRATING| in dlm_shuffle_lists()
474 if (res->state & (DLM_LOCK_RES_MIGRATING | in __dlm_dirty_lockres()
482 res->state |= DLM_LOCK_RES_DIRTY; in __dlm_dirty_lockres()
[all …]
Ddlmdebug.c107 buf, res->owner, res->state); in __dlm_print_one_lock_resource()
538 res->owner, res->state, res->last_used, in dump_lockres()
701 char *state; in debug_state_print() local
709 state = "NEW"; break; in debug_state_print()
711 state = "JOINED"; break; in debug_state_print()
713 state = "SHUTDOWN"; break; in debug_state_print()
715 state = "LEAVING"; break; in debug_state_print()
717 state = "UNKNOWN"; break; in debug_state_print()
729 task_pid_nr(dlm->dlm_thread_task), dlm->node_num, state); in debug_state_print()
807 if (dlm->reco.state == DLM_RECO_STATE_ACTIVE) in debug_state_print()
[all …]
/fs/gfs2/
Dglock.h185 extern void gfs2_holder_init(struct gfs2_glock *gl, unsigned int state,
187 extern void gfs2_holder_reinit(unsigned int state, unsigned flags,
198 unsigned int state, int flags,
218 unsigned int state, int flags, in gfs2_glock_nq_init() argument
223 gfs2_holder_init(gl, state, flags, gh); in gfs2_glock_nq_init()
232 extern void gfs2_glock_cb(struct gfs2_glock *gl, unsigned int state);
/fs/ntfs/
Dinode.h51 unsigned long state; /* NTFS specific flags describing this inode. member
185 return test_bit(NI_##flag, &(ni)->state); \
189 set_bit(NI_##flag, &(ni)->state); \
193 clear_bit(NI_##flag, &(ni)->state); \
202 return test_and_set_bit(NI_##flag, &(ni)->state); \
206 return test_and_clear_bit(NI_##flag, &(ni)->state); \
/fs/squashfs/
Dxz_wrapper.c38 struct xz_dec *state; member
106 stream->state = xz_dec_init(XZ_PREALLOC, comp_opts->dict_size); in squashfs_xz_init()
107 if (stream->state == NULL) { in squashfs_xz_init()
126 xz_dec_end(stream->state); in squashfs_xz_free()
141 xz_dec_reset(stream->state); in squashfs_xz_uncompress()
174 xz_err = xz_dec_run(stream->state, &stream->buf); in squashfs_xz_uncompress()

1234567