Home
last modified time | relevance | path

Searched refs:head (Results 1 – 25 of 198) sorted by relevance

12345678

/fs/proc/
Dproc_sysctl.c28 static bool is_empty_dir(struct ctl_table_header *head) in is_empty_dir() argument
30 return head->ctl_table[0].child == sysctl_mount_point; in is_empty_dir()
76 static int insert_links(struct ctl_table_header *head);
105 struct ctl_table_header *head; in find_entry() local
116 head = ctl_node->header; in find_entry()
117 entry = &head->ctl_table[ctl_node - head->node]; in find_entry()
126 *phead = head; in find_entry()
133 static int insert_entry(struct ctl_table_header *head, struct ctl_table *entry) in insert_entry() argument
135 struct rb_node *node = &head->node[entry - head->ctl_table].node; in insert_entry()
136 struct rb_node **p = &head->parent->root.rb_node; in insert_entry()
[all …]
Dkcore.c100 static void free_kclist_ents(struct list_head *head) in free_kclist_ents() argument
104 list_for_each_entry_safe(pos, tmp, head, list) { in free_kclist_ents()
145 LIST_HEAD(head); in kcore_update_ram()
155 list_add(&ent->list, &head); in kcore_update_ram()
156 __kcore_update_ram(&head); in kcore_update_ram()
165 get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head) in get_sparsemem_vmemmap_info() argument
177 list_for_each_entry(tmp, head, list) { in get_sparsemem_vmemmap_info()
191 list_add_tail(&vmm->list, head); in get_sparsemem_vmemmap_info()
198 get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head) in get_sparsemem_vmemmap_info() argument
208 struct list_head *head = (struct list_head *)arg; in kclist_add_private() local
[all …]
/fs/
Dmbcache2.c56 struct hlist_bl_head *head; in mb2_cache_entry_create() local
67 head = &cache->c_hash[hash_32(key, cache->c_bucket_bits)]; in mb2_cache_entry_create()
68 entry->e_hash_list_head = head; in mb2_cache_entry_create()
69 hlist_bl_lock(head); in mb2_cache_entry_create()
70 hlist_bl_for_each_entry(dup, dup_node, head, e_hash_list) { in mb2_cache_entry_create()
72 hlist_bl_unlock(head); in mb2_cache_entry_create()
77 hlist_bl_add_head(&entry->e_hash_list, head); in mb2_cache_entry_create()
78 hlist_bl_unlock(head); in mb2_cache_entry_create()
103 struct hlist_bl_head *head; in __entry_find() local
106 head = entry->e_hash_list_head; in __entry_find()
[all …]
Dbuffer.c90 struct buffer_head *head, *bh; in buffer_check_dirty_writeback() local
102 head = page_buffers(page); in buffer_check_dirty_writeback()
103 bh = head; in buffer_check_dirty_writeback()
112 } while (bh != head); in buffer_check_dirty_writeback()
209 struct buffer_head *head; in __find_get_block_slow() local
221 head = page_buffers(page); in __find_get_block_slow()
222 bh = head; in __find_get_block_slow()
232 } while (bh != head); in __find_get_block_slow()
684 struct buffer_head *head = page_buffers(page); in __set_page_dirty_buffers() local
685 struct buffer_head *bh = head; in __set_page_dirty_buffers()
[all …]
Dseq_file.c808 struct list_head *seq_list_start(struct list_head *head, loff_t pos) in seq_list_start() argument
812 list_for_each(lh, head) in seq_list_start()
820 struct list_head *seq_list_start_head(struct list_head *head, loff_t pos) in seq_list_start_head() argument
823 return head; in seq_list_start_head()
825 return seq_list_start(head, pos - 1); in seq_list_start_head()
829 struct list_head *seq_list_next(void *v, struct list_head *head, loff_t *ppos) in seq_list_next() argument
835 return lh == head ? NULL : lh; in seq_list_next()
846 struct hlist_node *seq_hlist_start(struct hlist_head *head, loff_t pos) in seq_hlist_start() argument
850 hlist_for_each(node, head) in seq_hlist_start()
865 struct hlist_node *seq_hlist_start_head(struct hlist_head *head, loff_t pos) in seq_hlist_start_head() argument
[all …]
Daio.c56 unsigned head; /* Written to by userland or under ring_lock member
525 ring->head = ring->tail = 0; in aio_setup_ring()
597 static void free_ioctx_rcufn(struct rcu_head *head) in free_ioctx_rcufn() argument
599 struct kioctx *ctx = container_of(head, struct kioctx, free_rcu); in free_ioctx_rcufn()
959 static void refill_reqs_available(struct kioctx *ctx, unsigned head, in refill_reqs_available() argument
965 head %= ctx->nr_events; in refill_reqs_available()
966 if (head <= tail) in refill_reqs_available()
967 events_in_ring = tail - head; in refill_reqs_available()
969 events_in_ring = ctx->nr_events - (head - tail); in refill_reqs_available()
993 unsigned head; in user_refill_reqs_available() local
[all …]
Dinode.c251 static void i_callback(struct rcu_head *head) in i_callback() argument
253 struct inode *inode = container_of(head, struct inode, i_rcu); in i_callback()
576 static void dispose_list(struct list_head *head) in dispose_list() argument
578 while (!list_empty(head)) { in dispose_list()
581 inode = list_first_entry(head, struct inode, i_lru); in dispose_list()
780 struct hlist_head *head, in find_inode() argument
787 hlist_for_each_entry(inode, head, i_hash) { in find_inode()
809 struct hlist_head *head, unsigned long ino) in find_inode_fast() argument
814 hlist_for_each_entry(inode, head, i_hash) { in find_inode_fast()
1020 struct hlist_head *head = inode_hashtable + hash(sb, hashval); in iget5_locked() local
[all …]
/fs/gfs2/
Drecovery.c57 struct list_head *head = &jd->jd_revoke_list; in gfs2_revoke_add() local
61 list_for_each_entry(rr, head, rr_list) { in gfs2_revoke_add()
79 list_add(&rr->rr_list, head); in gfs2_revoke_add()
110 struct list_head *head = &jd->jd_revoke_list; in gfs2_revoke_clean() local
113 while (!list_empty(head)) { in gfs2_revoke_clean()
114 rr = list_entry(head->next, struct gfs2_revoke_replay, rr_list); in gfs2_revoke_clean()
151 struct gfs2_log_header_host *head) in get_log_header() argument
173 *head = lh; in get_log_header()
192 struct gfs2_log_header_host *head) in find_good_lh() argument
198 error = get_log_header(jd, *blk, head); in find_good_lh()
[all …]
Dlops.c499 struct list_head *head; in buf_lo_after_commit() local
505 head = &tr->tr_buf; in buf_lo_after_commit()
506 while (!list_empty(head)) { in buf_lo_after_commit()
507 bd = list_entry(head->next, struct gfs2_bufdata, bd_list); in buf_lo_after_commit()
514 struct gfs2_log_header_host *head, int pass) in buf_lo_before_scan() argument
616 struct list_head *head = &sdp->sd_log_le_revoke; in revoke_lo_before_commit() local
629 list_for_each_entry(bd, head, bd_list) { in revoke_lo_before_commit()
654 struct list_head *head = &sdp->sd_log_le_revoke; in revoke_lo_after_commit() local
658 while (!list_empty(head)) { in revoke_lo_after_commit()
659 bd = list_entry(head->next, struct gfs2_bufdata, bd_list); in revoke_lo_after_commit()
[all …]
/fs/btrfs/
Ddelayed-ref.c158 struct btrfs_delayed_ref_head *head) in btrfs_delayed_ref_lock() argument
164 if (mutex_trylock(&head->mutex)) in btrfs_delayed_ref_lock()
167 atomic_inc(&head->node.refs); in btrfs_delayed_ref_lock()
170 mutex_lock(&head->mutex); in btrfs_delayed_ref_lock()
172 if (!head->node.in_tree) { in btrfs_delayed_ref_lock()
173 mutex_unlock(&head->mutex); in btrfs_delayed_ref_lock()
174 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock()
177 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock()
183 struct btrfs_delayed_ref_head *head, in drop_delayed_ref() argument
187 head = btrfs_delayed_node_to_head(ref); in drop_delayed_ref()
[all …]
/fs/hfsplus/
Dbtree.c135 struct hfs_btree_header_rec *head; in hfs_btree_open() local
165 head = (struct hfs_btree_header_rec *)(kmap(page) + in hfs_btree_open()
167 tree->root = be32_to_cpu(head->root); in hfs_btree_open()
168 tree->leaf_count = be32_to_cpu(head->leaf_count); in hfs_btree_open()
169 tree->leaf_head = be32_to_cpu(head->leaf_head); in hfs_btree_open()
170 tree->leaf_tail = be32_to_cpu(head->leaf_tail); in hfs_btree_open()
171 tree->node_count = be32_to_cpu(head->node_count); in hfs_btree_open()
172 tree->free_nodes = be32_to_cpu(head->free_nodes); in hfs_btree_open()
173 tree->attributes = be32_to_cpu(head->attributes); in hfs_btree_open()
174 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open()
[all …]
Dxattr.c63 struct hfs_btree_header_rec *head; in hfsplus_init_header_node() local
84 head = (struct hfs_btree_header_rec *)(buf + offset); in hfsplus_init_header_node()
85 head->node_size = cpu_to_be16(node_size); in hfsplus_init_header_node()
88 head->node_count = cpu_to_be32(tmp); in hfsplus_init_header_node()
89 head->free_nodes = cpu_to_be32(be32_to_cpu(head->node_count) - 1); in hfsplus_init_header_node()
90 head->clump_size = cpu_to_be32(clump_size); in hfsplus_init_header_node()
91 head->attributes |= cpu_to_be32(HFS_TREE_BIGKEYS | HFS_TREE_VARIDXKEYS); in hfsplus_init_header_node()
92 head->max_key_len = cpu_to_be16(HFSPLUS_ATTR_KEYLEN - sizeof(u16)); in hfsplus_init_header_node()
99 if (be32_to_cpu(head->node_count) > hdr_node_map_rec_bits) { in hfsplus_init_header_node()
103 desc->next = cpu_to_be32(be32_to_cpu(head->leaf_tail) + 1); in hfsplus_init_header_node()
[all …]
/fs/nilfs2/
Dsegbuf.h98 #define NILFS_LIST_SEGBUF(head) \ argument
99 list_entry((head), struct nilfs_segment_buffer, sb_list)
102 #define NILFS_LAST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->prev) argument
103 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next) argument
104 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head)) argument
110 #define NILFS_SEGBUF_FIRST_BH(head) \ argument
111 (list_entry((head)->next, struct buffer_head, b_assoc_buffers))
115 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head) argument
Dpage.c157 struct buffer_head *bh, *head; in nilfs_page_buffers_clean() local
159 bh = head = page_buffers(page); in nilfs_page_buffers_clean()
164 } while (bh != head); in nilfs_page_buffers_clean()
187 struct buffer_head *bh, *head; in nilfs_page_bug() local
190 bh = head = page_buffers(page); in nilfs_page_bug()
197 } while (bh != head); in nilfs_page_bug()
421 struct buffer_head *bh, *head; in nilfs_clear_dirty_page() local
427 bh = head = page_buffers(page); in nilfs_clear_dirty_page()
437 } while (bh = bh->b_this_page, bh != head); in nilfs_clear_dirty_page()
447 struct buffer_head *bh, *head; in nilfs_page_count_clean_buffers() local
[all …]
Drecovery.c318 struct list_head *head) in nilfs_scan_dsync_log() argument
371 list_add_tail(&rb->list, head); in nilfs_scan_dsync_log()
387 static void dispose_recovery_list(struct list_head *head) in dispose_recovery_list() argument
389 while (!list_empty(head)) { in dispose_recovery_list()
392 rb = list_first_entry(head, struct nilfs_recovery_block, list); in dispose_recovery_list()
403 static int nilfs_segment_list_add(struct list_head *head, __u64 segnum) in nilfs_segment_list_add() argument
412 list_add_tail(&ent->list, head); in nilfs_segment_list_add()
416 void nilfs_dispose_segment_list(struct list_head *head) in nilfs_dispose_segment_list() argument
418 while (!list_empty(head)) { in nilfs_dispose_segment_list()
421 ent = list_first_entry(head, struct nilfs_segment_entry, list); in nilfs_dispose_segment_list()
[all …]
/fs/9p/
Dvfs_dir.c55 int head; member
123 if (rdir->tail == rdir->head) { in v9fs_dir_readdir()
134 rdir->head = 0; in v9fs_dir_readdir()
137 while (rdir->head < rdir->tail) { in v9fs_dir_readdir()
138 err = p9stat_read(fid->clnt, rdir->buf + rdir->head, in v9fs_dir_readdir()
139 rdir->tail - rdir->head, &st); in v9fs_dir_readdir()
151 rdir->head += err; in v9fs_dir_readdir()
181 if (rdir->tail == rdir->head) { in v9fs_dir_readdir_dotl()
187 rdir->head = 0; in v9fs_dir_readdir_dotl()
191 while (rdir->head < rdir->tail) { in v9fs_dir_readdir_dotl()
[all …]
/fs/hfs/
Dbtree.c21 struct hfs_btree_header_rec *head; in hfs_btree_open() local
82 head = (struct hfs_btree_header_rec *)(kmap(page) + sizeof(struct hfs_bnode_desc)); in hfs_btree_open()
83 tree->root = be32_to_cpu(head->root); in hfs_btree_open()
84 tree->leaf_count = be32_to_cpu(head->leaf_count); in hfs_btree_open()
85 tree->leaf_head = be32_to_cpu(head->leaf_head); in hfs_btree_open()
86 tree->leaf_tail = be32_to_cpu(head->leaf_tail); in hfs_btree_open()
87 tree->node_count = be32_to_cpu(head->node_count); in hfs_btree_open()
88 tree->free_nodes = be32_to_cpu(head->free_nodes); in hfs_btree_open()
89 tree->attributes = be32_to_cpu(head->attributes); in hfs_btree_open()
90 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open()
[all …]
/fs/nfs/
Dwrite.c183 nfs_page_group_search_locked(struct nfs_page *head, unsigned int page_offset) in nfs_page_group_search_locked() argument
187 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_search_locked()
188 WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &head->wb_head->wb_flags)); in nfs_page_group_search_locked()
190 req = head; in nfs_page_group_search_locked()
197 } while (req != head); in nfs_page_group_search_locked()
324 nfs_unroll_locks_and_wait(struct inode *inode, struct nfs_page *head, in nfs_unroll_locks_and_wait() argument
332 for (tmp = head ; tmp != req; tmp = tmp->wb_this_page) in nfs_unroll_locks_and_wait()
340 nfs_page_group_unlock(head); in nfs_unroll_locks_and_wait()
344 nfs_release_request(head); in nfs_unroll_locks_and_wait()
431 struct nfs_page *head, *subreq; in nfs_lock_and_join_requests() local
[all …]
Dpagelist.c170 struct nfs_page *head = req->wb_head; in nfs_page_group_lock() local
172 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_lock()
174 if (!test_and_set_bit(PG_HEADLOCK, &head->wb_flags)) in nfs_page_group_lock()
178 return wait_on_bit_lock(&head->wb_flags, PG_HEADLOCK, in nfs_page_group_lock()
193 struct nfs_page *head = req->wb_head; in nfs_page_group_lock_wait() local
195 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_lock_wait()
197 wait_on_bit(&head->wb_flags, PG_HEADLOCK, in nfs_page_group_lock_wait()
208 struct nfs_page *head = req->wb_head; in nfs_page_group_unlock() local
210 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_unlock()
213 clear_bit(PG_HEADLOCK, &head->wb_flags); in nfs_page_group_unlock()
[all …]
/fs/befs/
Dbtree.c82 befs_host_btree_nodehead head; /* head of node converted to cpu byteorder */ member
219 node->head.left = fs64_to_cpu(sb, node->od_node->left); in befs_bt_read_node()
220 node->head.right = fs64_to_cpu(sb, node->od_node->right); in befs_bt_read_node()
221 node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow); in befs_bt_read_node()
222 node->head.all_key_count = in befs_bt_read_node()
224 node->head.all_key_length = in befs_bt_read_node()
287 node_off = this_node->head.overflow; in befs_btree_find()
356 last = node->head.all_key_count - 1; in befs_find_key()
467 while (key_sum + this_node->head.all_key_count <= key_no) { in befs_btree_read()
470 if (this_node->head.right == befs_bt_inval) { in befs_btree_read()
[all …]
/fs/ext4/
Dpage-io.c73 struct buffer_head *bh, *head; in ext4_finish_bio() local
95 bh = head = page_buffers(page); in ext4_finish_bio()
101 bit_spin_lock(BH_Uptodate_Lock, &head->b_state); in ext4_finish_bio()
112 } while ((bh = bh->b_this_page) != head); in ext4_finish_bio()
113 bit_spin_unlock(BH_Uptodate_Lock, &head->b_state); in ext4_finish_bio()
188 static void dump_completed_IO(struct inode *inode, struct list_head *head) in dump_completed_IO() argument
194 if (list_empty(head)) in dump_completed_IO()
198 list_for_each_entry(io, head, list) { in dump_completed_IO()
231 struct list_head *head) in ext4_do_flush_completed_IO() argument
240 dump_completed_IO(inode, head); in ext4_do_flush_completed_IO()
[all …]
/fs/notify/
Dmark.c114 u32 fsnotify_recalc_mask(struct hlist_head *head) in fsnotify_recalc_mask() argument
119 hlist_for_each_entry(mark, head, obj_list) in fsnotify_recalc_mask()
212 void fsnotify_destroy_marks(struct hlist_head *head, spinlock_t *lock) in fsnotify_destroy_marks() argument
225 if (hlist_empty(head)) { in fsnotify_destroy_marks()
229 mark = hlist_entry(head->first, struct fsnotify_mark, obj_list); in fsnotify_destroy_marks()
297 int fsnotify_add_mark_list(struct hlist_head *head, struct fsnotify_mark *mark, in fsnotify_add_mark_list() argument
304 if (hlist_empty(head)) { in fsnotify_add_mark_list()
305 hlist_add_head_rcu(&mark->obj_list, head); in fsnotify_add_mark_list()
310 hlist_for_each_entry(lmark, head, obj_list) { in fsnotify_add_mark_list()
410 struct fsnotify_mark *fsnotify_find_mark(struct hlist_head *head, in fsnotify_find_mark() argument
[all …]
Dfsnotify.h18 extern u32 fsnotify_recalc_mask(struct hlist_head *head);
27 extern int fsnotify_add_mark_list(struct hlist_head *head,
44 extern struct fsnotify_mark *fsnotify_find_mark(struct hlist_head *head,
47 extern void fsnotify_destroy_marks(struct hlist_head *head, spinlock_t *lock);
/fs/xfs/
Dxfs_log.c57 atomic64_t *head);
141 atomic64_t *head, in xlog_grant_sub_space() argument
144 int64_t head_val = atomic64_read(head); in xlog_grant_sub_space()
160 head_val = atomic64_cmpxchg(head, old, new); in xlog_grant_sub_space()
167 atomic64_t *head, in xlog_grant_add_space() argument
170 int64_t head_val = atomic64_read(head); in xlog_grant_add_space()
189 head_val = atomic64_cmpxchg(head, old, new); in xlog_grant_add_space()
195 struct xlog_grant_head *head) in xlog_grant_head_init() argument
197 xlog_assign_grant_head(&head->grant, 1, 0); in xlog_grant_head_init()
198 INIT_LIST_HEAD(&head->waiters); in xlog_grant_head_init()
[all …]
Dxfs_aops.c45 struct buffer_head *bh, *head; in xfs_count_page_state() local
49 bh = head = page_buffers(page); in xfs_count_page_state()
55 } while ((bh = bh->b_this_page) != head); in xfs_count_page_state()
476 xfs_ioend_t *head = ioend; in xfs_submit_ioend() local
490 ioend = head; in xfs_submit_ioend()
651 struct buffer_head *head; in xfs_check_page_type() local
660 bh = head = page_buffers(page); in xfs_check_page_type()
676 } while ((bh = bh->b_this_page) != head); in xfs_check_page_type()
696 struct buffer_head *bh, *head; in xfs_convert_page() local
766 bh = head = page_buffers(page); in xfs_convert_page()
[all …]

12345678