/fs/ |
D | mbcache.c | 77 struct mb_cache_entry *entry, *dup; in mb_cache_entry_create() local 88 entry = kmem_cache_alloc(mb_entry_cache, mask); in mb_cache_entry_create() 89 if (!entry) in mb_cache_entry_create() 92 INIT_LIST_HEAD(&entry->e_list); in mb_cache_entry_create() 100 atomic_set(&entry->e_refcnt, 2); in mb_cache_entry_create() 101 entry->e_key = key; in mb_cache_entry_create() 102 entry->e_value = value; in mb_cache_entry_create() 103 entry->e_flags = 0; in mb_cache_entry_create() 105 set_bit(MBE_REUSABLE_B, &entry->e_flags); in mb_cache_entry_create() 111 kmem_cache_free(mb_entry_cache, entry); in mb_cache_entry_create() [all …]
|
D | dax.c | 83 static unsigned long dax_to_pfn(void *entry) in dax_to_pfn() argument 85 return xa_to_value(entry) >> DAX_SHIFT; in dax_to_pfn() 93 static bool dax_is_locked(void *entry) in dax_is_locked() argument 95 return xa_to_value(entry) & DAX_LOCKED; in dax_is_locked() 98 static unsigned int dax_entry_order(void *entry) in dax_entry_order() argument 100 if (xa_to_value(entry) & DAX_PMD) in dax_entry_order() 105 static unsigned long dax_is_pmd_entry(void *entry) in dax_is_pmd_entry() argument 107 return xa_to_value(entry) & DAX_PMD; in dax_is_pmd_entry() 110 static bool dax_is_pte_entry(void *entry) in dax_is_pte_entry() argument 112 return !(xa_to_value(entry) & DAX_PMD); in dax_is_pte_entry() [all …]
|
/fs/squashfs/ |
D | cache.c | 56 struct squashfs_cache_entry *entry; in squashfs_cache_get() local 62 if (cache->entry[i].block == block) { in squashfs_cache_get() 90 if (cache->entry[i].refcount == 0) in squashfs_cache_get() 96 entry = &cache->entry[i]; in squashfs_cache_get() 103 entry->block = block; in squashfs_cache_get() 104 entry->refcount = 1; in squashfs_cache_get() 105 entry->pending = 1; in squashfs_cache_get() 106 entry->num_waiters = 0; in squashfs_cache_get() 107 entry->error = 0; in squashfs_cache_get() 110 entry->length = squashfs_read_data(sb, block, length, in squashfs_cache_get() [all …]
|
/fs/btrfs/ |
D | ordered-data.c | 23 static u64 entry_end(struct btrfs_ordered_extent *entry) in entry_end() argument 25 if (entry->file_offset + entry->num_bytes < entry->file_offset) in entry_end() 27 return entry->file_offset + entry->num_bytes; in entry_end() 38 struct btrfs_ordered_extent *entry; in tree_insert() local 42 entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node); in tree_insert() 44 if (file_offset < entry->file_offset) in tree_insert() 46 else if (file_offset >= entry_end(entry)) in tree_insert() 67 struct btrfs_ordered_extent *entry; in __tree_search() local 71 entry = rb_entry(n, struct btrfs_ordered_extent, rb_node); in __tree_search() 73 prev_entry = entry; in __tree_search() [all …]
|
D | free-space-cache.c | 554 struct btrfs_free_space_entry *entry; in io_ctl_add_entry() local 559 entry = io_ctl->cur; in io_ctl_add_entry() 560 put_unaligned_le64(offset, &entry->offset); in io_ctl_add_entry() 561 put_unaligned_le64(bytes, &entry->bytes); in io_ctl_add_entry() 562 entry->type = (bitmap) ? BTRFS_FREE_SPACE_BITMAP : in io_ctl_add_entry() 622 struct btrfs_free_space *entry, u8 *type) in io_ctl_read_entry() argument 634 entry->offset = get_unaligned_le64(&e->offset); in io_ctl_read_entry() 635 entry->bytes = get_unaligned_le64(&e->bytes); in io_ctl_read_entry() 649 struct btrfs_free_space *entry) in io_ctl_read_bitmap() argument 657 copy_page(entry->bitmap, io_ctl->cur); in io_ctl_read_bitmap() [all …]
|
/fs/nfs/ |
D | nfs42xattr.c | 128 nfs4_xattr_entry_lru_add(struct nfs4_xattr_entry *entry) in nfs4_xattr_entry_lru_add() argument 132 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? in nfs4_xattr_entry_lru_add() 135 return list_lru_add(lru, &entry->lru); in nfs4_xattr_entry_lru_add() 139 nfs4_xattr_entry_lru_del(struct nfs4_xattr_entry *entry) in nfs4_xattr_entry_lru_del() argument 143 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? in nfs4_xattr_entry_lru_del() 146 return list_lru_del(lru, &entry->lru); in nfs4_xattr_entry_lru_del() 178 struct nfs4_xattr_entry *entry; in nfs4_xattr_alloc_entry() local 205 entry = (struct nfs4_xattr_entry *)buf; in nfs4_xattr_alloc_entry() 233 entry->flags = flags; in nfs4_xattr_alloc_entry() 234 entry->xattr_value = valp; in nfs4_xattr_alloc_entry() [all …]
|
/fs/nilfs2/ |
D | dat.c | 91 struct nilfs_dat_entry *entry; in nilfs_dat_commit_alloc() local 95 entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, in nilfs_dat_commit_alloc() 97 entry->de_start = cpu_to_le64(NILFS_CNO_MIN); in nilfs_dat_commit_alloc() 98 entry->de_end = cpu_to_le64(NILFS_CNO_MAX); in nilfs_dat_commit_alloc() 99 entry->de_blocknr = cpu_to_le64(0); in nilfs_dat_commit_alloc() 115 struct nilfs_dat_entry *entry; in nilfs_dat_commit_free() local 119 entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, in nilfs_dat_commit_free() 121 entry->de_start = cpu_to_le64(NILFS_CNO_MIN); in nilfs_dat_commit_free() 122 entry->de_end = cpu_to_le64(NILFS_CNO_MIN); in nilfs_dat_commit_free() 123 entry->de_blocknr = cpu_to_le64(0); in nilfs_dat_commit_free() [all …]
|
/fs/ocfs2/ |
D | filecheck.c | 100 struct ocfs2_filecheck_sysfs_entry *entry = container_of(kobj, in ocfs2_filecheck_release() local 103 complete(&entry->fs_kobj_unregister); in ocfs2_filecheck_release() 147 ocfs2_filecheck_sysfs_free(struct ocfs2_filecheck_sysfs_entry *entry) in ocfs2_filecheck_sysfs_free() argument 151 spin_lock(&entry->fs_fcheck->fc_lock); in ocfs2_filecheck_sysfs_free() 152 while (!list_empty(&entry->fs_fcheck->fc_head)) { in ocfs2_filecheck_sysfs_free() 153 p = list_first_entry(&entry->fs_fcheck->fc_head, in ocfs2_filecheck_sysfs_free() 159 spin_unlock(&entry->fs_fcheck->fc_lock); in ocfs2_filecheck_sysfs_free() 161 kfree(entry->fs_fcheck); in ocfs2_filecheck_sysfs_free() 162 entry->fs_fcheck = NULL; in ocfs2_filecheck_sysfs_free() 169 struct ocfs2_filecheck_sysfs_entry *entry = &osb->osb_fc_ent; in ocfs2_filecheck_create_sysfs() local [all …]
|
/fs/ext4/ |
D | block_validity.c | 57 struct ext4_system_zone *entry, *n; in release_system_zone() local 59 rbtree_postorder_for_each_entry_safe(entry, n, in release_system_zone() 61 kmem_cache_free(ext4_system_zone_cachep, entry); in release_system_zone() 73 struct ext4_system_zone *new_entry, *entry; in add_system_zone() local 79 entry = rb_entry(parent, struct ext4_system_zone, node); in add_system_zone() 80 if (start_blk < entry->start_blk) in add_system_zone() 82 else if (start_blk >= (entry->start_blk + entry->count)) in add_system_zone() 103 entry = rb_entry(node, struct ext4_system_zone, node); in add_system_zone() 104 if (can_merge(entry, new_entry)) { in add_system_zone() 105 new_entry->start_blk = entry->start_blk; in add_system_zone() [all …]
|
D | xattr.c | 189 ext4_xattr_check_entries(struct ext4_xattr_entry *entry, void *end, in ext4_xattr_check_entries() argument 192 struct ext4_xattr_entry *e = entry; in ext4_xattr_check_entries() 205 while (!IS_LAST_ENTRY(entry)) { in ext4_xattr_check_entries() 206 u32 size = le32_to_cpu(entry->e_value_size); in ext4_xattr_check_entries() 211 if (size != 0 && entry->e_value_inum == 0) { in ext4_xattr_check_entries() 212 u16 offs = le16_to_cpu(entry->e_value_offs); in ext4_xattr_check_entries() 229 entry = EXT4_XATTR_NEXT(entry); in ext4_xattr_check_entries() 290 struct ext4_xattr_entry *entry, *next; in xattr_find_entry() local 297 for (entry = *pentry; !IS_LAST_ENTRY(entry); entry = next) { in xattr_find_entry() 298 next = EXT4_XATTR_NEXT(entry); in xattr_find_entry() [all …]
|
/fs/proc/ |
D | proc_sysctl.c | 113 struct ctl_table *entry; in find_entry() local 124 entry = &head->ctl_table[ctl_node - head->node]; in find_entry() 125 procname = entry->procname; in find_entry() 134 return entry; in find_entry() 140 static int insert_entry(struct ctl_table_header *head, struct ctl_table *entry) in insert_entry() argument 142 struct rb_node *node = &head->node[entry - head->ctl_table].node; in insert_entry() 145 const char *name = entry->procname; in insert_entry() 169 pr_cont("/%s\n", entry->procname); in insert_entry() 179 static void erase_entry(struct ctl_table_header *head, struct ctl_table *entry) in erase_entry() argument 181 struct rb_node *node = &head->node[entry - head->ctl_table].node; in erase_entry() [all …]
|
D | namespaces.c | 121 const struct proc_ns_operations **entry, **last; in proc_ns_dir_readdir() local 130 entry = ns_entries + (ctx->pos - 2); in proc_ns_dir_readdir() 132 while (entry <= last) { in proc_ns_dir_readdir() 133 const struct proc_ns_operations *ops = *entry; in proc_ns_dir_readdir() 138 entry++; in proc_ns_dir_readdir() 155 const struct proc_ns_operations **entry, **last; in proc_ns_dir_lookup() local 163 for (entry = ns_entries; entry < last; entry++) { in proc_ns_dir_lookup() 164 if (strlen((*entry)->name) != len) in proc_ns_dir_lookup() 166 if (!memcmp(dentry->d_name.name, (*entry)->name, len)) in proc_ns_dir_lookup() 169 if (entry == last) in proc_ns_dir_lookup() [all …]
|
/fs/ksmbd/mgmt/ |
D | user_session.c | 44 struct ksmbd_session_rpc *entry) in __session_rpc_close() argument 48 resp = ksmbd_rpc_close(sess, entry->id); in __session_rpc_close() 50 pr_err("Unable to close RPC pipe %d\n", entry->id); in __session_rpc_close() 53 ksmbd_rpc_id_free(entry->id); in __session_rpc_close() 54 kfree(entry); in __session_rpc_close() 59 struct ksmbd_session_rpc *entry; in ksmbd_session_rpc_clear_list() local 62 xa_for_each(&sess->rpc_handle_list, index, entry) { in ksmbd_session_rpc_clear_list() 64 __session_rpc_close(sess, entry); in ksmbd_session_rpc_clear_list() 93 struct ksmbd_session_rpc *entry; in ksmbd_session_rpc_open() local 101 entry = kzalloc(sizeof(struct ksmbd_session_rpc), GFP_KERNEL); in ksmbd_session_rpc_open() [all …]
|
/fs/fuse/ |
D | dir.c | 37 static inline void __fuse_dentry_settime(struct dentry *entry, u64 time) in __fuse_dentry_settime() argument 39 entry->d_fsdata = (void *) time; in __fuse_dentry_settime() 42 static inline u64 fuse_dentry_time(const struct dentry *entry) in fuse_dentry_time() argument 44 return (u64)entry->d_fsdata; in fuse_dentry_time() 54 static inline u64 fuse_dentry_time(const struct dentry *entry) in fuse_dentry_time() argument 56 return ((struct fuse_dentry *) entry->d_fsdata)->time; in fuse_dentry_time() 97 void fuse_change_entry_timeout(struct dentry *entry, struct fuse_entry_out *o) in fuse_change_entry_timeout() argument 99 fuse_dentry_settime(entry, in fuse_change_entry_timeout() 146 void fuse_invalidate_entry_cache(struct dentry *entry) in fuse_invalidate_entry_cache() argument 148 fuse_dentry_settime(entry, 0); in fuse_invalidate_entry_cache() [all …]
|
/fs/efivarfs/ |
D | super.c | 111 struct efivar_entry *entry; in efivarfs_callback() local 120 entry = kzalloc(sizeof(*entry), GFP_KERNEL); in efivarfs_callback() 121 if (!entry) in efivarfs_callback() 124 memcpy(entry->var.VariableName, name16, name_size); in efivarfs_callback() 125 memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t)); in efivarfs_callback() 127 len = ucs2_utf8size(entry->var.VariableName); in efivarfs_callback() 134 ucs2_as_utf8(name, entry->var.VariableName, len); in efivarfs_callback() 136 if (efivar_variable_is_removable(entry->var.VendorGuid, name, len)) in efivarfs_callback() 141 efi_guid_to_str(&entry->var.VendorGuid, name + len + 1); in efivarfs_callback() 159 efivar_entry_size(entry, &size); in efivarfs_callback() [all …]
|
/fs/f2fs/ |
D | acl.c | 51 struct f2fs_acl_entry *entry = (struct f2fs_acl_entry *)(hdr + 1); in f2fs_acl_from_disk() local 72 if ((char *)entry > end) in f2fs_acl_from_disk() 75 acl->a_entries[i].e_tag = le16_to_cpu(entry->e_tag); in f2fs_acl_from_disk() 76 acl->a_entries[i].e_perm = le16_to_cpu(entry->e_perm); in f2fs_acl_from_disk() 83 entry = (struct f2fs_acl_entry *)((char *)entry + in f2fs_acl_from_disk() 90 le32_to_cpu(entry->e_id)); in f2fs_acl_from_disk() 91 entry = (struct f2fs_acl_entry *)((char *)entry + in f2fs_acl_from_disk() 97 le32_to_cpu(entry->e_id)); in f2fs_acl_from_disk() 98 entry = (struct f2fs_acl_entry *)((char *)entry + in f2fs_acl_from_disk() 105 if ((char *)entry != end) in f2fs_acl_from_disk() [all …]
|
D | xattr.h | 62 #define ENTRY_SIZE(entry) (XATTR_ALIGN(sizeof(struct f2fs_xattr_entry) + \ argument 63 (entry)->e_name_len + le16_to_cpu((entry)->e_value_size))) 65 #define XATTR_NEXT_ENTRY(entry) ((struct f2fs_xattr_entry *)((char *)(entry) +\ argument 66 ENTRY_SIZE(entry))) 68 #define IS_XATTR_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0) argument 70 #define list_for_each_xattr(entry, addr) \ argument 71 for (entry = XATTR_FIRST_ENTRY(addr);\ 72 !IS_XATTR_LAST_ENTRY(entry);\ 73 entry = XATTR_NEXT_ENTRY(entry))
|
D | recovery.c | 68 struct fsync_inode_entry *entry; in get_fsync_inode() local 70 list_for_each_entry(entry, head, list) in get_fsync_inode() 71 if (entry->inode->i_ino == ino) in get_fsync_inode() 72 return entry; in get_fsync_inode() 81 struct fsync_inode_entry *entry; in add_fsync_inode() local 98 entry = f2fs_kmem_cache_alloc(fsync_entry_slab, in add_fsync_inode() 100 entry->inode = inode; in add_fsync_inode() 101 list_add_tail(&entry->list, head); in add_fsync_inode() 103 return entry; in add_fsync_inode() 109 static void del_fsync_inode(struct fsync_inode_entry *entry, int drop) in del_fsync_inode() argument [all …]
|
D | xattr.c | 232 struct f2fs_xattr_entry *entry; in __find_xattr() local 234 list_for_each_xattr(entry, base_addr) { in __find_xattr() 235 if ((void *)(entry) + sizeof(__u32) > last_base_addr || in __find_xattr() 236 (void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) { in __find_xattr() 238 *last_addr = entry; in __find_xattr() 242 if (entry->e_name_index != index) in __find_xattr() 244 if (entry->e_name_len != len) in __find_xattr() 246 if (!memcmp(entry->e_name, name, len)) in __find_xattr() 249 return entry; in __find_xattr() 256 struct f2fs_xattr_entry *entry; in __find_inline_xattr() local [all …]
|
/fs/ext2/ |
D | xattr.c | 72 #define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0) argument 149 ext2_xattr_entry_valid(struct ext2_xattr_entry *entry, in ext2_xattr_entry_valid() argument 155 next = EXT2_XATTR_NEXT(entry); in ext2_xattr_entry_valid() 159 if (entry->e_value_block != 0) in ext2_xattr_entry_valid() 162 size = le32_to_cpu(entry->e_value_size); in ext2_xattr_entry_valid() 164 le16_to_cpu(entry->e_value_offs) + size > end_offs) in ext2_xattr_entry_valid() 172 struct ext2_xattr_entry *entry) in ext2_xattr_cmp_entry() argument 176 cmp = name_index - entry->e_name_index; in ext2_xattr_cmp_entry() 178 cmp = name_len - entry->e_name_len; in ext2_xattr_cmp_entry() 180 cmp = memcmp(name, entry->e_name, name_len); in ext2_xattr_cmp_entry() [all …]
|
/fs/fat/ |
D | fatent.c | 23 static void fat12_ent_blocknr(struct super_block *sb, int entry, in fat12_ent_blocknr() argument 27 int bytes = entry + (entry >> 1); in fat12_ent_blocknr() 28 WARN_ON(!fat_valid_entry(sbi, entry)); in fat12_ent_blocknr() 33 static void fat_ent_blocknr(struct super_block *sb, int entry, in fat_ent_blocknr() argument 37 int bytes = (entry << sbi->fatent_shift); in fat_ent_blocknr() 38 WARN_ON(!fat_valid_entry(sbi, entry)); in fat_ent_blocknr() 126 if (fatent->entry & 1) in fat12_ent_get() 164 if (fatent->entry & 1) { in fat12_ent_put() 199 u8 *nextp = ent12_p[1] + 1 + (fatent->entry & 1); in fat12_ent_next() 201 fatent->entry++; in fat12_ent_next() [all …]
|
/fs/nfs_common/ |
D | nfsacl.c | 55 struct posix_acl_entry *entry = in xdr_nfsace_encode() local 58 *p++ = htonl(entry->e_tag | nfsacl_desc->typeflag); in xdr_nfsace_encode() 59 switch(entry->e_tag) { in xdr_nfsace_encode() 67 *p++ = htonl(from_kuid(&init_user_ns, entry->e_uid)); in xdr_nfsace_encode() 70 *p++ = htonl(from_kgid(&init_user_ns, entry->e_gid)); in xdr_nfsace_encode() 76 *p++ = htonl(entry->e_perm & S_IRWXO); in xdr_nfsace_encode() 222 struct posix_acl_entry *entry; in xdr_nfsace_decode() local 234 entry = &nfsacl_desc->acl->a_entries[nfsacl_desc->count++]; in xdr_nfsace_decode() 235 entry->e_tag = ntohl(*p++) & ~NFS_ACL_DEFAULT; in xdr_nfsace_decode() 237 entry->e_perm = ntohl(*p++); in xdr_nfsace_decode() [all …]
|
/fs/omfs/ |
D | file.c | 35 struct omfs_extent_entry *entry; in omfs_shrink_inode() local 71 entry = &oe->e_entry; in omfs_shrink_inode() 76 start = be64_to_cpu(entry->e_cluster); in omfs_shrink_inode() 77 count = be64_to_cpu(entry->e_blocks); in omfs_shrink_inode() 80 entry++; in omfs_shrink_inode() 120 struct omfs_extent_entry *entry = &oe->e_entry; in omfs_grow_extent() local 142 terminator = entry + extent_count - 1; in omfs_grow_extent() 144 entry = terminator-1; in omfs_grow_extent() 145 new_block = be64_to_cpu(entry->e_cluster) + in omfs_grow_extent() 146 be64_to_cpu(entry->e_blocks); in omfs_grow_extent() [all …]
|
/fs/xfs/libxfs/ |
D | xfs_attr_leaf.c | 980 struct xfs_attr_leaf_entry *entry; in xfs_attr_shortform_allfit() local 989 entry = xfs_attr3_leaf_entryp(leaf); in xfs_attr_shortform_allfit() 992 for (i = 0; i < leafhdr.count; entry++, i++) { in xfs_attr_shortform_allfit() 993 if (entry->flags & XFS_ATTR_INCOMPLETE) in xfs_attr_shortform_allfit() 995 if (!(entry->flags & XFS_ATTR_LOCAL)) in xfs_attr_shortform_allfit() 1099 struct xfs_attr_leaf_entry *entry; in xfs_attr3_leaf_to_shortform() local 1117 entry = xfs_attr3_leaf_entryp(leaf); in xfs_attr3_leaf_to_shortform() 1149 for (i = 0; i < ichdr.count; entry++, i++) { in xfs_attr3_leaf_to_shortform() 1150 if (entry->flags & XFS_ATTR_INCOMPLETE) in xfs_attr3_leaf_to_shortform() 1152 if (!entry->nameidx) in xfs_attr3_leaf_to_shortform() [all …]
|
/fs/afs/ |
D | vlclient.c | 20 struct afs_vldb_entry *entry; in afs_deliver_vl_get_entry_by_name_u() local 33 entry = call->ret_vldb; in afs_deliver_vl_get_entry_by_name_u() 40 entry->name[i] = (u8)ntohl(uvldb->name[i]); in afs_deliver_vl_get_entry_by_name_u() 41 entry->name[i] = 0; in afs_deliver_vl_get_entry_by_name_u() 42 entry->name_len = strlen(entry->name); in afs_deliver_vl_get_entry_by_name_u() 59 int n = entry->nr_servers; in afs_deliver_vl_get_entry_by_name_u() 66 entry->fs_mask[n] |= AFS_VOL_VTM_RW; in afs_deliver_vl_get_entry_by_name_u() 68 entry->fs_mask[n] |= AFS_VOL_VTM_BAK; in afs_deliver_vl_get_entry_by_name_u() 71 entry->fs_mask[n] |= AFS_VOL_VTM_RO; in afs_deliver_vl_get_entry_by_name_u() 72 if (!entry->fs_mask[n]) in afs_deliver_vl_get_entry_by_name_u() [all …]
|