Home
last modified time | relevance | path

Searched refs:entry (Results 1 – 25 of 132) sorted by relevance

123456

/fs/
Dmbcache.c77 struct mb_cache_entry *entry, *dup; in mb_cache_entry_create() local
88 entry = kmem_cache_alloc(mb_entry_cache, mask); in mb_cache_entry_create()
89 if (!entry) in mb_cache_entry_create()
92 INIT_LIST_HEAD(&entry->e_list); in mb_cache_entry_create()
100 atomic_set(&entry->e_refcnt, 2); in mb_cache_entry_create()
101 entry->e_key = key; in mb_cache_entry_create()
102 entry->e_value = value; in mb_cache_entry_create()
103 entry->e_flags = 0; in mb_cache_entry_create()
105 set_bit(MBE_REUSABLE_B, &entry->e_flags); in mb_cache_entry_create()
111 kmem_cache_free(mb_entry_cache, entry); in mb_cache_entry_create()
[all …]
Ddax.c83 static unsigned long dax_to_pfn(void *entry) in dax_to_pfn() argument
85 return xa_to_value(entry) >> DAX_SHIFT; in dax_to_pfn()
93 static bool dax_is_locked(void *entry) in dax_is_locked() argument
95 return xa_to_value(entry) & DAX_LOCKED; in dax_is_locked()
98 static unsigned int dax_entry_order(void *entry) in dax_entry_order() argument
100 if (xa_to_value(entry) & DAX_PMD) in dax_entry_order()
105 static unsigned long dax_is_pmd_entry(void *entry) in dax_is_pmd_entry() argument
107 return xa_to_value(entry) & DAX_PMD; in dax_is_pmd_entry()
110 static bool dax_is_pte_entry(void *entry) in dax_is_pte_entry() argument
112 return !(xa_to_value(entry) & DAX_PMD); in dax_is_pte_entry()
[all …]
/fs/squashfs/
Dcache.c56 struct squashfs_cache_entry *entry; in squashfs_cache_get() local
62 if (cache->entry[i].block == block) { in squashfs_cache_get()
90 if (cache->entry[i].refcount == 0) in squashfs_cache_get()
96 entry = &cache->entry[i]; in squashfs_cache_get()
103 entry->block = block; in squashfs_cache_get()
104 entry->refcount = 1; in squashfs_cache_get()
105 entry->pending = 1; in squashfs_cache_get()
106 entry->num_waiters = 0; in squashfs_cache_get()
107 entry->error = 0; in squashfs_cache_get()
110 entry->length = squashfs_read_data(sb, block, length, in squashfs_cache_get()
[all …]
/fs/btrfs/
Dordered-data.c23 static u64 entry_end(struct btrfs_ordered_extent *entry) in entry_end() argument
25 if (entry->file_offset + entry->num_bytes < entry->file_offset) in entry_end()
27 return entry->file_offset + entry->num_bytes; in entry_end()
38 struct btrfs_ordered_extent *entry; in tree_insert() local
42 entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node); in tree_insert()
44 if (file_offset < entry->file_offset) in tree_insert()
46 else if (file_offset >= entry_end(entry)) in tree_insert()
67 struct btrfs_ordered_extent *entry; in __tree_search() local
71 entry = rb_entry(n, struct btrfs_ordered_extent, rb_node); in __tree_search()
73 prev_entry = entry; in __tree_search()
[all …]
Dfree-space-cache.c554 struct btrfs_free_space_entry *entry; in io_ctl_add_entry() local
559 entry = io_ctl->cur; in io_ctl_add_entry()
560 put_unaligned_le64(offset, &entry->offset); in io_ctl_add_entry()
561 put_unaligned_le64(bytes, &entry->bytes); in io_ctl_add_entry()
562 entry->type = (bitmap) ? BTRFS_FREE_SPACE_BITMAP : in io_ctl_add_entry()
622 struct btrfs_free_space *entry, u8 *type) in io_ctl_read_entry() argument
634 entry->offset = get_unaligned_le64(&e->offset); in io_ctl_read_entry()
635 entry->bytes = get_unaligned_le64(&e->bytes); in io_ctl_read_entry()
649 struct btrfs_free_space *entry) in io_ctl_read_bitmap() argument
657 copy_page(entry->bitmap, io_ctl->cur); in io_ctl_read_bitmap()
[all …]
/fs/nfs/
Dnfs42xattr.c128 nfs4_xattr_entry_lru_add(struct nfs4_xattr_entry *entry) in nfs4_xattr_entry_lru_add() argument
132 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? in nfs4_xattr_entry_lru_add()
135 return list_lru_add(lru, &entry->lru); in nfs4_xattr_entry_lru_add()
139 nfs4_xattr_entry_lru_del(struct nfs4_xattr_entry *entry) in nfs4_xattr_entry_lru_del() argument
143 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? in nfs4_xattr_entry_lru_del()
146 return list_lru_del(lru, &entry->lru); in nfs4_xattr_entry_lru_del()
178 struct nfs4_xattr_entry *entry; in nfs4_xattr_alloc_entry() local
205 entry = (struct nfs4_xattr_entry *)buf; in nfs4_xattr_alloc_entry()
233 entry->flags = flags; in nfs4_xattr_alloc_entry()
234 entry->xattr_value = valp; in nfs4_xattr_alloc_entry()
[all …]
/fs/nilfs2/
Ddat.c91 struct nilfs_dat_entry *entry; in nilfs_dat_commit_alloc() local
95 entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, in nilfs_dat_commit_alloc()
97 entry->de_start = cpu_to_le64(NILFS_CNO_MIN); in nilfs_dat_commit_alloc()
98 entry->de_end = cpu_to_le64(NILFS_CNO_MAX); in nilfs_dat_commit_alloc()
99 entry->de_blocknr = cpu_to_le64(0); in nilfs_dat_commit_alloc()
115 struct nilfs_dat_entry *entry; in nilfs_dat_commit_free() local
119 entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, in nilfs_dat_commit_free()
121 entry->de_start = cpu_to_le64(NILFS_CNO_MIN); in nilfs_dat_commit_free()
122 entry->de_end = cpu_to_le64(NILFS_CNO_MIN); in nilfs_dat_commit_free()
123 entry->de_blocknr = cpu_to_le64(0); in nilfs_dat_commit_free()
[all …]
/fs/ocfs2/
Dfilecheck.c100 struct ocfs2_filecheck_sysfs_entry *entry = container_of(kobj, in ocfs2_filecheck_release() local
103 complete(&entry->fs_kobj_unregister); in ocfs2_filecheck_release()
147 ocfs2_filecheck_sysfs_free(struct ocfs2_filecheck_sysfs_entry *entry) in ocfs2_filecheck_sysfs_free() argument
151 spin_lock(&entry->fs_fcheck->fc_lock); in ocfs2_filecheck_sysfs_free()
152 while (!list_empty(&entry->fs_fcheck->fc_head)) { in ocfs2_filecheck_sysfs_free()
153 p = list_first_entry(&entry->fs_fcheck->fc_head, in ocfs2_filecheck_sysfs_free()
159 spin_unlock(&entry->fs_fcheck->fc_lock); in ocfs2_filecheck_sysfs_free()
161 kfree(entry->fs_fcheck); in ocfs2_filecheck_sysfs_free()
162 entry->fs_fcheck = NULL; in ocfs2_filecheck_sysfs_free()
169 struct ocfs2_filecheck_sysfs_entry *entry = &osb->osb_fc_ent; in ocfs2_filecheck_create_sysfs() local
[all …]
/fs/ext4/
Dblock_validity.c57 struct ext4_system_zone *entry, *n; in release_system_zone() local
59 rbtree_postorder_for_each_entry_safe(entry, n, in release_system_zone()
61 kmem_cache_free(ext4_system_zone_cachep, entry); in release_system_zone()
73 struct ext4_system_zone *new_entry, *entry; in add_system_zone() local
79 entry = rb_entry(parent, struct ext4_system_zone, node); in add_system_zone()
80 if (start_blk < entry->start_blk) in add_system_zone()
82 else if (start_blk >= (entry->start_blk + entry->count)) in add_system_zone()
103 entry = rb_entry(node, struct ext4_system_zone, node); in add_system_zone()
104 if (can_merge(entry, new_entry)) { in add_system_zone()
105 new_entry->start_blk = entry->start_blk; in add_system_zone()
[all …]
Dxattr.c189 ext4_xattr_check_entries(struct ext4_xattr_entry *entry, void *end, in ext4_xattr_check_entries() argument
192 struct ext4_xattr_entry *e = entry; in ext4_xattr_check_entries()
205 while (!IS_LAST_ENTRY(entry)) { in ext4_xattr_check_entries()
206 u32 size = le32_to_cpu(entry->e_value_size); in ext4_xattr_check_entries()
211 if (size != 0 && entry->e_value_inum == 0) { in ext4_xattr_check_entries()
212 u16 offs = le16_to_cpu(entry->e_value_offs); in ext4_xattr_check_entries()
229 entry = EXT4_XATTR_NEXT(entry); in ext4_xattr_check_entries()
290 struct ext4_xattr_entry *entry, *next; in xattr_find_entry() local
297 for (entry = *pentry; !IS_LAST_ENTRY(entry); entry = next) { in xattr_find_entry()
298 next = EXT4_XATTR_NEXT(entry); in xattr_find_entry()
[all …]
/fs/proc/
Dproc_sysctl.c113 struct ctl_table *entry; in find_entry() local
124 entry = &head->ctl_table[ctl_node - head->node]; in find_entry()
125 procname = entry->procname; in find_entry()
134 return entry; in find_entry()
140 static int insert_entry(struct ctl_table_header *head, struct ctl_table *entry) in insert_entry() argument
142 struct rb_node *node = &head->node[entry - head->ctl_table].node; in insert_entry()
145 const char *name = entry->procname; in insert_entry()
169 pr_cont("/%s\n", entry->procname); in insert_entry()
179 static void erase_entry(struct ctl_table_header *head, struct ctl_table *entry) in erase_entry() argument
181 struct rb_node *node = &head->node[entry - head->ctl_table].node; in erase_entry()
[all …]
Dnamespaces.c121 const struct proc_ns_operations **entry, **last; in proc_ns_dir_readdir() local
130 entry = ns_entries + (ctx->pos - 2); in proc_ns_dir_readdir()
132 while (entry <= last) { in proc_ns_dir_readdir()
133 const struct proc_ns_operations *ops = *entry; in proc_ns_dir_readdir()
138 entry++; in proc_ns_dir_readdir()
155 const struct proc_ns_operations **entry, **last; in proc_ns_dir_lookup() local
163 for (entry = ns_entries; entry < last; entry++) { in proc_ns_dir_lookup()
164 if (strlen((*entry)->name) != len) in proc_ns_dir_lookup()
166 if (!memcmp(dentry->d_name.name, (*entry)->name, len)) in proc_ns_dir_lookup()
169 if (entry == last) in proc_ns_dir_lookup()
[all …]
/fs/ksmbd/mgmt/
Duser_session.c44 struct ksmbd_session_rpc *entry) in __session_rpc_close() argument
48 resp = ksmbd_rpc_close(sess, entry->id); in __session_rpc_close()
50 pr_err("Unable to close RPC pipe %d\n", entry->id); in __session_rpc_close()
53 ksmbd_rpc_id_free(entry->id); in __session_rpc_close()
54 kfree(entry); in __session_rpc_close()
59 struct ksmbd_session_rpc *entry; in ksmbd_session_rpc_clear_list() local
62 xa_for_each(&sess->rpc_handle_list, index, entry) { in ksmbd_session_rpc_clear_list()
64 __session_rpc_close(sess, entry); in ksmbd_session_rpc_clear_list()
93 struct ksmbd_session_rpc *entry; in ksmbd_session_rpc_open() local
101 entry = kzalloc(sizeof(struct ksmbd_session_rpc), GFP_KERNEL); in ksmbd_session_rpc_open()
[all …]
/fs/fuse/
Ddir.c37 static inline void __fuse_dentry_settime(struct dentry *entry, u64 time) in __fuse_dentry_settime() argument
39 entry->d_fsdata = (void *) time; in __fuse_dentry_settime()
42 static inline u64 fuse_dentry_time(const struct dentry *entry) in fuse_dentry_time() argument
44 return (u64)entry->d_fsdata; in fuse_dentry_time()
54 static inline u64 fuse_dentry_time(const struct dentry *entry) in fuse_dentry_time() argument
56 return ((struct fuse_dentry *) entry->d_fsdata)->time; in fuse_dentry_time()
97 void fuse_change_entry_timeout(struct dentry *entry, struct fuse_entry_out *o) in fuse_change_entry_timeout() argument
99 fuse_dentry_settime(entry, in fuse_change_entry_timeout()
146 void fuse_invalidate_entry_cache(struct dentry *entry) in fuse_invalidate_entry_cache() argument
148 fuse_dentry_settime(entry, 0); in fuse_invalidate_entry_cache()
[all …]
/fs/efivarfs/
Dsuper.c111 struct efivar_entry *entry; in efivarfs_callback() local
120 entry = kzalloc(sizeof(*entry), GFP_KERNEL); in efivarfs_callback()
121 if (!entry) in efivarfs_callback()
124 memcpy(entry->var.VariableName, name16, name_size); in efivarfs_callback()
125 memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t)); in efivarfs_callback()
127 len = ucs2_utf8size(entry->var.VariableName); in efivarfs_callback()
134 ucs2_as_utf8(name, entry->var.VariableName, len); in efivarfs_callback()
136 if (efivar_variable_is_removable(entry->var.VendorGuid, name, len)) in efivarfs_callback()
141 efi_guid_to_str(&entry->var.VendorGuid, name + len + 1); in efivarfs_callback()
159 efivar_entry_size(entry, &size); in efivarfs_callback()
[all …]
/fs/f2fs/
Dacl.c51 struct f2fs_acl_entry *entry = (struct f2fs_acl_entry *)(hdr + 1); in f2fs_acl_from_disk() local
72 if ((char *)entry > end) in f2fs_acl_from_disk()
75 acl->a_entries[i].e_tag = le16_to_cpu(entry->e_tag); in f2fs_acl_from_disk()
76 acl->a_entries[i].e_perm = le16_to_cpu(entry->e_perm); in f2fs_acl_from_disk()
83 entry = (struct f2fs_acl_entry *)((char *)entry + in f2fs_acl_from_disk()
90 le32_to_cpu(entry->e_id)); in f2fs_acl_from_disk()
91 entry = (struct f2fs_acl_entry *)((char *)entry + in f2fs_acl_from_disk()
97 le32_to_cpu(entry->e_id)); in f2fs_acl_from_disk()
98 entry = (struct f2fs_acl_entry *)((char *)entry + in f2fs_acl_from_disk()
105 if ((char *)entry != end) in f2fs_acl_from_disk()
[all …]
Dxattr.h62 #define ENTRY_SIZE(entry) (XATTR_ALIGN(sizeof(struct f2fs_xattr_entry) + \ argument
63 (entry)->e_name_len + le16_to_cpu((entry)->e_value_size)))
65 #define XATTR_NEXT_ENTRY(entry) ((struct f2fs_xattr_entry *)((char *)(entry) +\ argument
66 ENTRY_SIZE(entry)))
68 #define IS_XATTR_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0) argument
70 #define list_for_each_xattr(entry, addr) \ argument
71 for (entry = XATTR_FIRST_ENTRY(addr);\
72 !IS_XATTR_LAST_ENTRY(entry);\
73 entry = XATTR_NEXT_ENTRY(entry))
Drecovery.c68 struct fsync_inode_entry *entry; in get_fsync_inode() local
70 list_for_each_entry(entry, head, list) in get_fsync_inode()
71 if (entry->inode->i_ino == ino) in get_fsync_inode()
72 return entry; in get_fsync_inode()
81 struct fsync_inode_entry *entry; in add_fsync_inode() local
98 entry = f2fs_kmem_cache_alloc(fsync_entry_slab, in add_fsync_inode()
100 entry->inode = inode; in add_fsync_inode()
101 list_add_tail(&entry->list, head); in add_fsync_inode()
103 return entry; in add_fsync_inode()
109 static void del_fsync_inode(struct fsync_inode_entry *entry, int drop) in del_fsync_inode() argument
[all …]
Dxattr.c232 struct f2fs_xattr_entry *entry; in __find_xattr() local
234 list_for_each_xattr(entry, base_addr) { in __find_xattr()
235 if ((void *)(entry) + sizeof(__u32) > last_base_addr || in __find_xattr()
236 (void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) { in __find_xattr()
238 *last_addr = entry; in __find_xattr()
242 if (entry->e_name_index != index) in __find_xattr()
244 if (entry->e_name_len != len) in __find_xattr()
246 if (!memcmp(entry->e_name, name, len)) in __find_xattr()
249 return entry; in __find_xattr()
256 struct f2fs_xattr_entry *entry; in __find_inline_xattr() local
[all …]
/fs/ext2/
Dxattr.c72 #define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0) argument
149 ext2_xattr_entry_valid(struct ext2_xattr_entry *entry, in ext2_xattr_entry_valid() argument
155 next = EXT2_XATTR_NEXT(entry); in ext2_xattr_entry_valid()
159 if (entry->e_value_block != 0) in ext2_xattr_entry_valid()
162 size = le32_to_cpu(entry->e_value_size); in ext2_xattr_entry_valid()
164 le16_to_cpu(entry->e_value_offs) + size > end_offs) in ext2_xattr_entry_valid()
172 struct ext2_xattr_entry *entry) in ext2_xattr_cmp_entry() argument
176 cmp = name_index - entry->e_name_index; in ext2_xattr_cmp_entry()
178 cmp = name_len - entry->e_name_len; in ext2_xattr_cmp_entry()
180 cmp = memcmp(name, entry->e_name, name_len); in ext2_xattr_cmp_entry()
[all …]
/fs/fat/
Dfatent.c23 static void fat12_ent_blocknr(struct super_block *sb, int entry, in fat12_ent_blocknr() argument
27 int bytes = entry + (entry >> 1); in fat12_ent_blocknr()
28 WARN_ON(!fat_valid_entry(sbi, entry)); in fat12_ent_blocknr()
33 static void fat_ent_blocknr(struct super_block *sb, int entry, in fat_ent_blocknr() argument
37 int bytes = (entry << sbi->fatent_shift); in fat_ent_blocknr()
38 WARN_ON(!fat_valid_entry(sbi, entry)); in fat_ent_blocknr()
126 if (fatent->entry & 1) in fat12_ent_get()
164 if (fatent->entry & 1) { in fat12_ent_put()
199 u8 *nextp = ent12_p[1] + 1 + (fatent->entry & 1); in fat12_ent_next()
201 fatent->entry++; in fat12_ent_next()
[all …]
/fs/nfs_common/
Dnfsacl.c55 struct posix_acl_entry *entry = in xdr_nfsace_encode() local
58 *p++ = htonl(entry->e_tag | nfsacl_desc->typeflag); in xdr_nfsace_encode()
59 switch(entry->e_tag) { in xdr_nfsace_encode()
67 *p++ = htonl(from_kuid(&init_user_ns, entry->e_uid)); in xdr_nfsace_encode()
70 *p++ = htonl(from_kgid(&init_user_ns, entry->e_gid)); in xdr_nfsace_encode()
76 *p++ = htonl(entry->e_perm & S_IRWXO); in xdr_nfsace_encode()
222 struct posix_acl_entry *entry; in xdr_nfsace_decode() local
234 entry = &nfsacl_desc->acl->a_entries[nfsacl_desc->count++]; in xdr_nfsace_decode()
235 entry->e_tag = ntohl(*p++) & ~NFS_ACL_DEFAULT; in xdr_nfsace_decode()
237 entry->e_perm = ntohl(*p++); in xdr_nfsace_decode()
[all …]
/fs/omfs/
Dfile.c35 struct omfs_extent_entry *entry; in omfs_shrink_inode() local
71 entry = &oe->e_entry; in omfs_shrink_inode()
76 start = be64_to_cpu(entry->e_cluster); in omfs_shrink_inode()
77 count = be64_to_cpu(entry->e_blocks); in omfs_shrink_inode()
80 entry++; in omfs_shrink_inode()
120 struct omfs_extent_entry *entry = &oe->e_entry; in omfs_grow_extent() local
142 terminator = entry + extent_count - 1; in omfs_grow_extent()
144 entry = terminator-1; in omfs_grow_extent()
145 new_block = be64_to_cpu(entry->e_cluster) + in omfs_grow_extent()
146 be64_to_cpu(entry->e_blocks); in omfs_grow_extent()
[all …]
/fs/xfs/libxfs/
Dxfs_attr_leaf.c980 struct xfs_attr_leaf_entry *entry; in xfs_attr_shortform_allfit() local
989 entry = xfs_attr3_leaf_entryp(leaf); in xfs_attr_shortform_allfit()
992 for (i = 0; i < leafhdr.count; entry++, i++) { in xfs_attr_shortform_allfit()
993 if (entry->flags & XFS_ATTR_INCOMPLETE) in xfs_attr_shortform_allfit()
995 if (!(entry->flags & XFS_ATTR_LOCAL)) in xfs_attr_shortform_allfit()
1099 struct xfs_attr_leaf_entry *entry; in xfs_attr3_leaf_to_shortform() local
1117 entry = xfs_attr3_leaf_entryp(leaf); in xfs_attr3_leaf_to_shortform()
1149 for (i = 0; i < ichdr.count; entry++, i++) { in xfs_attr3_leaf_to_shortform()
1150 if (entry->flags & XFS_ATTR_INCOMPLETE) in xfs_attr3_leaf_to_shortform()
1152 if (!entry->nameidx) in xfs_attr3_leaf_to_shortform()
[all …]
/fs/afs/
Dvlclient.c20 struct afs_vldb_entry *entry; in afs_deliver_vl_get_entry_by_name_u() local
33 entry = call->ret_vldb; in afs_deliver_vl_get_entry_by_name_u()
40 entry->name[i] = (u8)ntohl(uvldb->name[i]); in afs_deliver_vl_get_entry_by_name_u()
41 entry->name[i] = 0; in afs_deliver_vl_get_entry_by_name_u()
42 entry->name_len = strlen(entry->name); in afs_deliver_vl_get_entry_by_name_u()
59 int n = entry->nr_servers; in afs_deliver_vl_get_entry_by_name_u()
66 entry->fs_mask[n] |= AFS_VOL_VTM_RW; in afs_deliver_vl_get_entry_by_name_u()
68 entry->fs_mask[n] |= AFS_VOL_VTM_BAK; in afs_deliver_vl_get_entry_by_name_u()
71 entry->fs_mask[n] |= AFS_VOL_VTM_RO; in afs_deliver_vl_get_entry_by_name_u()
72 if (!entry->fs_mask[n]) in afs_deliver_vl_get_entry_by_name_u()
[all …]

123456