/fs/proc/ |
D | proc_sysctl.c | 28 static bool is_empty_dir(struct ctl_table_header *head) in is_empty_dir() argument 30 return head->ctl_table[0].child == sysctl_mount_point; in is_empty_dir() 76 static int insert_links(struct ctl_table_header *head); 105 struct ctl_table_header *head; in find_entry() local 116 head = ctl_node->header; in find_entry() 117 entry = &head->ctl_table[ctl_node - head->node]; in find_entry() 126 *phead = head; in find_entry() 133 static int insert_entry(struct ctl_table_header *head, struct ctl_table *entry) in insert_entry() argument 135 struct rb_node *node = &head->node[entry - head->ctl_table].node; in insert_entry() 136 struct rb_node **p = &head->parent->root.rb_node; in insert_entry() [all …]
|
D | kcore.c | 100 static void free_kclist_ents(struct list_head *head) in free_kclist_ents() argument 104 list_for_each_entry_safe(pos, tmp, head, list) { in free_kclist_ents() 145 LIST_HEAD(head); in kcore_update_ram() 155 list_add(&ent->list, &head); in kcore_update_ram() 156 __kcore_update_ram(&head); in kcore_update_ram() 165 get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head) in get_sparsemem_vmemmap_info() argument 177 list_for_each_entry(tmp, head, list) { in get_sparsemem_vmemmap_info() 191 list_add_tail(&vmm->list, head); in get_sparsemem_vmemmap_info() 198 get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head) in get_sparsemem_vmemmap_info() argument 208 struct list_head *head = (struct list_head *)arg; in kclist_add_private() local [all …]
|
/fs/ |
D | mbcache2.c | 56 struct hlist_bl_head *head; in mb2_cache_entry_create() local 67 head = &cache->c_hash[hash_32(key, cache->c_bucket_bits)]; in mb2_cache_entry_create() 68 entry->e_hash_list_head = head; in mb2_cache_entry_create() 69 hlist_bl_lock(head); in mb2_cache_entry_create() 70 hlist_bl_for_each_entry(dup, dup_node, head, e_hash_list) { in mb2_cache_entry_create() 72 hlist_bl_unlock(head); in mb2_cache_entry_create() 77 hlist_bl_add_head(&entry->e_hash_list, head); in mb2_cache_entry_create() 78 hlist_bl_unlock(head); in mb2_cache_entry_create() 103 struct hlist_bl_head *head; in __entry_find() local 106 head = entry->e_hash_list_head; in __entry_find() [all …]
|
D | buffer.c | 90 struct buffer_head *head, *bh; in buffer_check_dirty_writeback() local 102 head = page_buffers(page); in buffer_check_dirty_writeback() 103 bh = head; in buffer_check_dirty_writeback() 112 } while (bh != head); in buffer_check_dirty_writeback() 209 struct buffer_head *head; in __find_get_block_slow() local 221 head = page_buffers(page); in __find_get_block_slow() 222 bh = head; in __find_get_block_slow() 232 } while (bh != head); in __find_get_block_slow() 684 struct buffer_head *head = page_buffers(page); in __set_page_dirty_buffers() local 685 struct buffer_head *bh = head; in __set_page_dirty_buffers() [all …]
|
D | seq_file.c | 808 struct list_head *seq_list_start(struct list_head *head, loff_t pos) in seq_list_start() argument 812 list_for_each(lh, head) in seq_list_start() 820 struct list_head *seq_list_start_head(struct list_head *head, loff_t pos) in seq_list_start_head() argument 823 return head; in seq_list_start_head() 825 return seq_list_start(head, pos - 1); in seq_list_start_head() 829 struct list_head *seq_list_next(void *v, struct list_head *head, loff_t *ppos) in seq_list_next() argument 835 return lh == head ? NULL : lh; in seq_list_next() 846 struct hlist_node *seq_hlist_start(struct hlist_head *head, loff_t pos) in seq_hlist_start() argument 850 hlist_for_each(node, head) in seq_hlist_start() 865 struct hlist_node *seq_hlist_start_head(struct hlist_head *head, loff_t pos) in seq_hlist_start_head() argument [all …]
|
D | aio.c | 56 unsigned head; /* Written to by userland or under ring_lock member 525 ring->head = ring->tail = 0; in aio_setup_ring() 597 static void free_ioctx_rcufn(struct rcu_head *head) in free_ioctx_rcufn() argument 599 struct kioctx *ctx = container_of(head, struct kioctx, free_rcu); in free_ioctx_rcufn() 959 static void refill_reqs_available(struct kioctx *ctx, unsigned head, in refill_reqs_available() argument 965 head %= ctx->nr_events; in refill_reqs_available() 966 if (head <= tail) in refill_reqs_available() 967 events_in_ring = tail - head; in refill_reqs_available() 969 events_in_ring = ctx->nr_events - (head - tail); in refill_reqs_available() 993 unsigned head; in user_refill_reqs_available() local [all …]
|
D | inode.c | 251 static void i_callback(struct rcu_head *head) in i_callback() argument 253 struct inode *inode = container_of(head, struct inode, i_rcu); in i_callback() 576 static void dispose_list(struct list_head *head) in dispose_list() argument 578 while (!list_empty(head)) { in dispose_list() 581 inode = list_first_entry(head, struct inode, i_lru); in dispose_list() 780 struct hlist_head *head, in find_inode() argument 787 hlist_for_each_entry(inode, head, i_hash) { in find_inode() 809 struct hlist_head *head, unsigned long ino) in find_inode_fast() argument 814 hlist_for_each_entry(inode, head, i_hash) { in find_inode_fast() 1020 struct hlist_head *head = inode_hashtable + hash(sb, hashval); in iget5_locked() local [all …]
|
/fs/gfs2/ |
D | recovery.c | 57 struct list_head *head = &jd->jd_revoke_list; in gfs2_revoke_add() local 61 list_for_each_entry(rr, head, rr_list) { in gfs2_revoke_add() 79 list_add(&rr->rr_list, head); in gfs2_revoke_add() 110 struct list_head *head = &jd->jd_revoke_list; in gfs2_revoke_clean() local 113 while (!list_empty(head)) { in gfs2_revoke_clean() 114 rr = list_entry(head->next, struct gfs2_revoke_replay, rr_list); in gfs2_revoke_clean() 151 struct gfs2_log_header_host *head) in get_log_header() argument 173 *head = lh; in get_log_header() 192 struct gfs2_log_header_host *head) in find_good_lh() argument 198 error = get_log_header(jd, *blk, head); in find_good_lh() [all …]
|
D | lops.c | 499 struct list_head *head; in buf_lo_after_commit() local 505 head = &tr->tr_buf; in buf_lo_after_commit() 506 while (!list_empty(head)) { in buf_lo_after_commit() 507 bd = list_entry(head->next, struct gfs2_bufdata, bd_list); in buf_lo_after_commit() 514 struct gfs2_log_header_host *head, int pass) in buf_lo_before_scan() argument 616 struct list_head *head = &sdp->sd_log_le_revoke; in revoke_lo_before_commit() local 629 list_for_each_entry(bd, head, bd_list) { in revoke_lo_before_commit() 654 struct list_head *head = &sdp->sd_log_le_revoke; in revoke_lo_after_commit() local 658 while (!list_empty(head)) { in revoke_lo_after_commit() 659 bd = list_entry(head->next, struct gfs2_bufdata, bd_list); in revoke_lo_after_commit() [all …]
|
/fs/btrfs/ |
D | delayed-ref.c | 158 struct btrfs_delayed_ref_head *head) in btrfs_delayed_ref_lock() argument 164 if (mutex_trylock(&head->mutex)) in btrfs_delayed_ref_lock() 167 atomic_inc(&head->node.refs); in btrfs_delayed_ref_lock() 170 mutex_lock(&head->mutex); in btrfs_delayed_ref_lock() 172 if (!head->node.in_tree) { in btrfs_delayed_ref_lock() 173 mutex_unlock(&head->mutex); in btrfs_delayed_ref_lock() 174 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock() 177 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock() 183 struct btrfs_delayed_ref_head *head, in drop_delayed_ref() argument 187 head = btrfs_delayed_node_to_head(ref); in drop_delayed_ref() [all …]
|
/fs/hfsplus/ |
D | btree.c | 135 struct hfs_btree_header_rec *head; in hfs_btree_open() local 165 head = (struct hfs_btree_header_rec *)(kmap(page) + in hfs_btree_open() 167 tree->root = be32_to_cpu(head->root); in hfs_btree_open() 168 tree->leaf_count = be32_to_cpu(head->leaf_count); in hfs_btree_open() 169 tree->leaf_head = be32_to_cpu(head->leaf_head); in hfs_btree_open() 170 tree->leaf_tail = be32_to_cpu(head->leaf_tail); in hfs_btree_open() 171 tree->node_count = be32_to_cpu(head->node_count); in hfs_btree_open() 172 tree->free_nodes = be32_to_cpu(head->free_nodes); in hfs_btree_open() 173 tree->attributes = be32_to_cpu(head->attributes); in hfs_btree_open() 174 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open() [all …]
|
D | xattr.c | 63 struct hfs_btree_header_rec *head; in hfsplus_init_header_node() local 84 head = (struct hfs_btree_header_rec *)(buf + offset); in hfsplus_init_header_node() 85 head->node_size = cpu_to_be16(node_size); in hfsplus_init_header_node() 88 head->node_count = cpu_to_be32(tmp); in hfsplus_init_header_node() 89 head->free_nodes = cpu_to_be32(be32_to_cpu(head->node_count) - 1); in hfsplus_init_header_node() 90 head->clump_size = cpu_to_be32(clump_size); in hfsplus_init_header_node() 91 head->attributes |= cpu_to_be32(HFS_TREE_BIGKEYS | HFS_TREE_VARIDXKEYS); in hfsplus_init_header_node() 92 head->max_key_len = cpu_to_be16(HFSPLUS_ATTR_KEYLEN - sizeof(u16)); in hfsplus_init_header_node() 99 if (be32_to_cpu(head->node_count) > hdr_node_map_rec_bits) { in hfsplus_init_header_node() 103 desc->next = cpu_to_be32(be32_to_cpu(head->leaf_tail) + 1); in hfsplus_init_header_node() [all …]
|
/fs/nilfs2/ |
D | segbuf.h | 98 #define NILFS_LIST_SEGBUF(head) \ argument 99 list_entry((head), struct nilfs_segment_buffer, sb_list) 102 #define NILFS_LAST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->prev) argument 103 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next) argument 104 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head)) argument 110 #define NILFS_SEGBUF_FIRST_BH(head) \ argument 111 (list_entry((head)->next, struct buffer_head, b_assoc_buffers)) 115 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head) argument
|
D | page.c | 157 struct buffer_head *bh, *head; in nilfs_page_buffers_clean() local 159 bh = head = page_buffers(page); in nilfs_page_buffers_clean() 164 } while (bh != head); in nilfs_page_buffers_clean() 187 struct buffer_head *bh, *head; in nilfs_page_bug() local 190 bh = head = page_buffers(page); in nilfs_page_bug() 197 } while (bh != head); in nilfs_page_bug() 421 struct buffer_head *bh, *head; in nilfs_clear_dirty_page() local 427 bh = head = page_buffers(page); in nilfs_clear_dirty_page() 437 } while (bh = bh->b_this_page, bh != head); in nilfs_clear_dirty_page() 447 struct buffer_head *bh, *head; in nilfs_page_count_clean_buffers() local [all …]
|
D | recovery.c | 318 struct list_head *head) in nilfs_scan_dsync_log() argument 371 list_add_tail(&rb->list, head); in nilfs_scan_dsync_log() 387 static void dispose_recovery_list(struct list_head *head) in dispose_recovery_list() argument 389 while (!list_empty(head)) { in dispose_recovery_list() 392 rb = list_first_entry(head, struct nilfs_recovery_block, list); in dispose_recovery_list() 403 static int nilfs_segment_list_add(struct list_head *head, __u64 segnum) in nilfs_segment_list_add() argument 412 list_add_tail(&ent->list, head); in nilfs_segment_list_add() 416 void nilfs_dispose_segment_list(struct list_head *head) in nilfs_dispose_segment_list() argument 418 while (!list_empty(head)) { in nilfs_dispose_segment_list() 421 ent = list_first_entry(head, struct nilfs_segment_entry, list); in nilfs_dispose_segment_list() [all …]
|
/fs/9p/ |
D | vfs_dir.c | 55 int head; member 123 if (rdir->tail == rdir->head) { in v9fs_dir_readdir() 134 rdir->head = 0; in v9fs_dir_readdir() 137 while (rdir->head < rdir->tail) { in v9fs_dir_readdir() 138 err = p9stat_read(fid->clnt, rdir->buf + rdir->head, in v9fs_dir_readdir() 139 rdir->tail - rdir->head, &st); in v9fs_dir_readdir() 151 rdir->head += err; in v9fs_dir_readdir() 181 if (rdir->tail == rdir->head) { in v9fs_dir_readdir_dotl() 187 rdir->head = 0; in v9fs_dir_readdir_dotl() 191 while (rdir->head < rdir->tail) { in v9fs_dir_readdir_dotl() [all …]
|
/fs/hfs/ |
D | btree.c | 21 struct hfs_btree_header_rec *head; in hfs_btree_open() local 82 head = (struct hfs_btree_header_rec *)(kmap(page) + sizeof(struct hfs_bnode_desc)); in hfs_btree_open() 83 tree->root = be32_to_cpu(head->root); in hfs_btree_open() 84 tree->leaf_count = be32_to_cpu(head->leaf_count); in hfs_btree_open() 85 tree->leaf_head = be32_to_cpu(head->leaf_head); in hfs_btree_open() 86 tree->leaf_tail = be32_to_cpu(head->leaf_tail); in hfs_btree_open() 87 tree->node_count = be32_to_cpu(head->node_count); in hfs_btree_open() 88 tree->free_nodes = be32_to_cpu(head->free_nodes); in hfs_btree_open() 89 tree->attributes = be32_to_cpu(head->attributes); in hfs_btree_open() 90 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open() [all …]
|
/fs/nfs/ |
D | write.c | 183 nfs_page_group_search_locked(struct nfs_page *head, unsigned int page_offset) in nfs_page_group_search_locked() argument 187 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_search_locked() 188 WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &head->wb_head->wb_flags)); in nfs_page_group_search_locked() 190 req = head; in nfs_page_group_search_locked() 197 } while (req != head); in nfs_page_group_search_locked() 324 nfs_unroll_locks_and_wait(struct inode *inode, struct nfs_page *head, in nfs_unroll_locks_and_wait() argument 332 for (tmp = head ; tmp != req; tmp = tmp->wb_this_page) in nfs_unroll_locks_and_wait() 340 nfs_page_group_unlock(head); in nfs_unroll_locks_and_wait() 344 nfs_release_request(head); in nfs_unroll_locks_and_wait() 431 struct nfs_page *head, *subreq; in nfs_lock_and_join_requests() local [all …]
|
D | pagelist.c | 170 struct nfs_page *head = req->wb_head; in nfs_page_group_lock() local 172 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_lock() 174 if (!test_and_set_bit(PG_HEADLOCK, &head->wb_flags)) in nfs_page_group_lock() 178 return wait_on_bit_lock(&head->wb_flags, PG_HEADLOCK, in nfs_page_group_lock() 193 struct nfs_page *head = req->wb_head; in nfs_page_group_lock_wait() local 195 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_lock_wait() 197 wait_on_bit(&head->wb_flags, PG_HEADLOCK, in nfs_page_group_lock_wait() 208 struct nfs_page *head = req->wb_head; in nfs_page_group_unlock() local 210 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_unlock() 213 clear_bit(PG_HEADLOCK, &head->wb_flags); in nfs_page_group_unlock() [all …]
|
/fs/befs/ |
D | btree.c | 82 befs_host_btree_nodehead head; /* head of node converted to cpu byteorder */ member 219 node->head.left = fs64_to_cpu(sb, node->od_node->left); in befs_bt_read_node() 220 node->head.right = fs64_to_cpu(sb, node->od_node->right); in befs_bt_read_node() 221 node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow); in befs_bt_read_node() 222 node->head.all_key_count = in befs_bt_read_node() 224 node->head.all_key_length = in befs_bt_read_node() 287 node_off = this_node->head.overflow; in befs_btree_find() 356 last = node->head.all_key_count - 1; in befs_find_key() 467 while (key_sum + this_node->head.all_key_count <= key_no) { in befs_btree_read() 470 if (this_node->head.right == befs_bt_inval) { in befs_btree_read() [all …]
|
/fs/ext4/ |
D | page-io.c | 73 struct buffer_head *bh, *head; in ext4_finish_bio() local 95 bh = head = page_buffers(page); in ext4_finish_bio() 101 bit_spin_lock(BH_Uptodate_Lock, &head->b_state); in ext4_finish_bio() 112 } while ((bh = bh->b_this_page) != head); in ext4_finish_bio() 113 bit_spin_unlock(BH_Uptodate_Lock, &head->b_state); in ext4_finish_bio() 188 static void dump_completed_IO(struct inode *inode, struct list_head *head) in dump_completed_IO() argument 194 if (list_empty(head)) in dump_completed_IO() 198 list_for_each_entry(io, head, list) { in dump_completed_IO() 231 struct list_head *head) in ext4_do_flush_completed_IO() argument 240 dump_completed_IO(inode, head); in ext4_do_flush_completed_IO() [all …]
|
/fs/notify/ |
D | mark.c | 114 u32 fsnotify_recalc_mask(struct hlist_head *head) in fsnotify_recalc_mask() argument 119 hlist_for_each_entry(mark, head, obj_list) in fsnotify_recalc_mask() 212 void fsnotify_destroy_marks(struct hlist_head *head, spinlock_t *lock) in fsnotify_destroy_marks() argument 225 if (hlist_empty(head)) { in fsnotify_destroy_marks() 229 mark = hlist_entry(head->first, struct fsnotify_mark, obj_list); in fsnotify_destroy_marks() 297 int fsnotify_add_mark_list(struct hlist_head *head, struct fsnotify_mark *mark, in fsnotify_add_mark_list() argument 304 if (hlist_empty(head)) { in fsnotify_add_mark_list() 305 hlist_add_head_rcu(&mark->obj_list, head); in fsnotify_add_mark_list() 310 hlist_for_each_entry(lmark, head, obj_list) { in fsnotify_add_mark_list() 410 struct fsnotify_mark *fsnotify_find_mark(struct hlist_head *head, in fsnotify_find_mark() argument [all …]
|
D | fsnotify.h | 18 extern u32 fsnotify_recalc_mask(struct hlist_head *head); 27 extern int fsnotify_add_mark_list(struct hlist_head *head, 44 extern struct fsnotify_mark *fsnotify_find_mark(struct hlist_head *head, 47 extern void fsnotify_destroy_marks(struct hlist_head *head, spinlock_t *lock);
|
/fs/xfs/ |
D | xfs_log.c | 57 atomic64_t *head); 141 atomic64_t *head, in xlog_grant_sub_space() argument 144 int64_t head_val = atomic64_read(head); in xlog_grant_sub_space() 160 head_val = atomic64_cmpxchg(head, old, new); in xlog_grant_sub_space() 167 atomic64_t *head, in xlog_grant_add_space() argument 170 int64_t head_val = atomic64_read(head); in xlog_grant_add_space() 189 head_val = atomic64_cmpxchg(head, old, new); in xlog_grant_add_space() 195 struct xlog_grant_head *head) in xlog_grant_head_init() argument 197 xlog_assign_grant_head(&head->grant, 1, 0); in xlog_grant_head_init() 198 INIT_LIST_HEAD(&head->waiters); in xlog_grant_head_init() [all …]
|
D | xfs_aops.c | 45 struct buffer_head *bh, *head; in xfs_count_page_state() local 49 bh = head = page_buffers(page); in xfs_count_page_state() 55 } while ((bh = bh->b_this_page) != head); in xfs_count_page_state() 476 xfs_ioend_t *head = ioend; in xfs_submit_ioend() local 490 ioend = head; in xfs_submit_ioend() 651 struct buffer_head *head; in xfs_check_page_type() local 660 bh = head = page_buffers(page); in xfs_check_page_type() 676 } while ((bh = bh->b_this_page) != head); in xfs_check_page_type() 696 struct buffer_head *bh, *head; in xfs_convert_page() local 766 bh = head = page_buffers(page); in xfs_convert_page() [all …]
|