/fs/ntfs/ |
D | attrib.c | 76 ATTR_RECORD *a; in ntfs_map_runlist_nolock() local 103 a = ctx->attr; in ntfs_map_runlist_nolock() 104 BUG_ON(!a->non_resident); in ntfs_map_runlist_nolock() 106 end_vcn = sle64_to_cpu(a->data.non_resident.highest_vcn); in ntfs_map_runlist_nolock() 111 if (!a->data.non_resident.lowest_vcn && end_vcn <= 0) in ntfs_map_runlist_nolock() 120 if (vcn >= allocated_size_vcn || (a->type == ni->type && in ntfs_map_runlist_nolock() 121 a->name_length == ni->name_len && in ntfs_map_runlist_nolock() 122 !memcmp((u8*)a + le16_to_cpu(a->name_offset), in ntfs_map_runlist_nolock() 124 sle64_to_cpu(a->data.non_resident.lowest_vcn) in ntfs_map_runlist_nolock() 161 a = ctx->attr; in ntfs_map_runlist_nolock() [all …]
|
D | inode.c | 529 ATTR_RECORD *a; in ntfs_read_locked_inode() local 629 a = ctx->attr; in ntfs_read_locked_inode() 631 si = (STANDARD_INFORMATION*)((u8*)a + in ntfs_read_locked_inode() 632 le16_to_cpu(a->data.resident.value_offset)); in ntfs_read_locked_inode() 671 a = ctx->attr; in ntfs_read_locked_inode() 672 if (a->flags & ATTR_COMPRESSION_MASK) { in ntfs_read_locked_inode() 677 if (a->flags & ATTR_IS_ENCRYPTED || in ntfs_read_locked_inode() 678 a->flags & ATTR_IS_SPARSE) { in ntfs_read_locked_inode() 679 if (a->non_resident) { in ntfs_read_locked_inode() 695 ni->attr_list_size = (u32)ntfs_attr_size(a); in ntfs_read_locked_inode() [all …]
|
D | file.c | 104 ATTR_RECORD *a; in ntfs_attr_extend_initialized() local 150 a = ctx->attr; in ntfs_attr_extend_initialized() 151 BUG_ON(a->non_resident); in ntfs_attr_extend_initialized() 153 attr_len = le32_to_cpu(a->data.resident.value_length); in ntfs_attr_extend_initialized() 159 kattr = (u8*)a + le16_to_cpu(a->data.resident.value_offset); in ntfs_attr_extend_initialized() 161 a->data.resident.value_length = cpu_to_le32((u32)new_init_size); in ntfs_attr_extend_initialized() 194 a = ctx->attr; in ntfs_attr_extend_initialized() 195 BUG_ON(!a->non_resident); in ntfs_attr_extend_initialized() 197 sle64_to_cpu(a->data.non_resident.data_size)); in ntfs_attr_extend_initialized() 198 a->data.non_resident.data_size = cpu_to_sle64(new_init_size); in ntfs_attr_extend_initialized() [all …]
|
D | debug.h | 30 #define ntfs_debug(f, a...) \ argument 31 __ntfs_debug(__FILE__, __LINE__, __func__, f, ##a) 50 #define ntfs_warning(sb, f, a...) __ntfs_warning(__func__, sb, f, ##a) argument 55 #define ntfs_error(sb, f, a...) __ntfs_error(__func__, sb, f, ##a) argument
|
/fs/reiserfs/ |
D | hashes.c | 36 b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); \ 49 u32 a, b, c, d; in keyed_hash() local 59 a = (u32) msg[0] | in keyed_hash() 77 a = (u32) msg[0] | in keyed_hash() 91 a = (u32) msg[0] | in keyed_hash() 102 a = (u32) msg[0] | in keyed_hash() 111 a = b = c = d = pad; in keyed_hash() 113 a <<= 8; in keyed_hash() 114 a |= msg[i]; in keyed_hash() 131 u32 a, c; in yura_hash() local [all …]
|
/fs/hpfs/ |
D | name.c | 23 static inline unsigned char upcase(unsigned char *dir, unsigned char a) in upcase() argument 25 if (a<128 || a==255) return a>='a' && a<='z' ? a - 0x20 : a; in upcase() 26 if (!dir) return a; in upcase() 27 return dir[a-128]; in upcase() 30 unsigned char hpfs_upcase(unsigned char *dir, unsigned char a) in hpfs_upcase() argument 32 return upcase(dir, a); in hpfs_upcase() 35 static inline unsigned char locase(unsigned char *dir, unsigned char a) in locase() argument 37 if (a<128 || a==255) return a>='A' && a<='Z' ? a + 0x20 : a; in locase() 38 if (!dir) return a; in locase() 39 return dir[a]; in locase()
|
D | ea.c | 15 void hpfs_ea_ext_remove(struct super_block *s, secno a, int ano, unsigned len) in hpfs_ea_ext_remove() argument 23 ano ? "anode" : "sectors", a, len); in hpfs_ea_ext_remove() 26 if (hpfs_ea_read(s, a, ano, pos, 4, ex)) return; in hpfs_ea_ext_remove() 30 ano ? "anode" : "sectors", a, pos); in hpfs_ea_ext_remove() 33 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 9, ex+4)) in hpfs_ea_ext_remove() 39 if (!ano) hpfs_free_sectors(s, a, (len+511) >> 9); in hpfs_ea_ext_remove() 43 if ((anode = hpfs_map_anode(s, a, &bh))) { in hpfs_ea_ext_remove() 46 hpfs_free_sectors(s, a, 1); in hpfs_ea_ext_remove() 51 static char *get_indirect_ea(struct super_block *s, int ano, secno a, int size) in get_indirect_ea() argument 58 if (hpfs_ea_read(s, a, ano, 0, size, ret)) { in get_indirect_ea() [all …]
|
D | anode.c | 18 anode_secno a = -1; in hpfs_bplus_lookup() local 23 if (hpfs_sb(s)->sb_chk) if (hpfs_stop_cycles(s, a, &c1, &c2, "hpfs_bplus_lookup")) return -1; in hpfs_bplus_lookup() 27 a = le32_to_cpu(btree->u.internal[i].down); in hpfs_bplus_lookup() 29 if (!(anode = hpfs_map_anode(s, a, &bh))) return -1; in hpfs_bplus_lookup() 33 hpfs_error(s, "sector %08x not found in internal anode %08x", sec, a); in hpfs_bplus_lookup() 40 …a = le32_to_cpu(btree->u.external[i].disk_secno) + sec - le32_to_cpu(btree->u.external[i].file_sec… in hpfs_bplus_lookup() 41 if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, a, 1, "data")) { in hpfs_bplus_lookup() 52 return a; in hpfs_bplus_lookup() 54 hpfs_error(s, "sector %08x not found in external anode %08x", sec, a); in hpfs_bplus_lookup() 66 anode_secno a, na = -1, ra, up = -1; in hpfs_add_sector_to_btree() local [all …]
|
/fs/ext4/ |
D | hash.c | 20 __u32 a = in[0], b = in[1], c = in[2], d = in[3]; in TEA_transform() local 25 b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); in TEA_transform() 44 #define ROUND(f, a, b, c, d, x, s) \ argument 45 (a += f(b, c, d) + x, a = rol32(a, s)) 55 __u32 a = buf[0], b = buf[1], c = buf[2], d = buf[3]; in half_md4_transform() local 58 ROUND(F, a, b, c, d, in[0] + K1, 3); in half_md4_transform() 59 ROUND(F, d, a, b, c, in[1] + K1, 7); in half_md4_transform() 60 ROUND(F, c, d, a, b, in[2] + K1, 11); in half_md4_transform() 61 ROUND(F, b, c, d, a, in[3] + K1, 19); in half_md4_transform() 62 ROUND(F, a, b, c, d, in[4] + K1, 3); in half_md4_transform() [all …]
|
/fs/jfs/ |
D | jfs_extent.c | 24 #define DPD(a) (printk("(a): %d\n",(a))) argument 25 #define DPC(a) (printk("(a): %c\n",(a))) argument 26 #define DPL1(a) \ argument 28 if ((a) >> 32) \ 29 printk("(a): %x%08x ",(a)); \ 31 printk("(a): %x ",(a) << 32); \ 33 #define DPL(a) \ argument 35 if ((a) >> 32) \ 36 printk("(a): %x%08x\n",(a)); \ 38 printk("(a): %x\n",(a) << 32); \ [all …]
|
/fs/jffs2/ |
D | summary.h | 196 #define jffs2_sum_init(a) (0) argument 197 #define jffs2_sum_exit(a) argument 198 #define jffs2_sum_disable_collecting(a) argument 199 #define jffs2_sum_is_disabled(a) (0) argument 200 #define jffs2_sum_reset_collected(a) argument 201 #define jffs2_sum_add_kvec(a,b,c,d) (0) argument 202 #define jffs2_sum_move_collected(a,b) argument 203 #define jffs2_sum_write_sumnode(a) (0) argument 204 #define jffs2_sum_add_padding_mem(a,b) argument 205 #define jffs2_sum_add_inode_mem(a,b,c) argument [all …]
|
/fs/xfs/libxfs/ |
D | xfs_rmap_btree.c | 54 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_rmapbt_dup_cursor() 63 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_set_root() 86 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_alloc_block() 92 error = xfs_alloc_get_freelist(cur->bc_tp, cur->bc_private.a.agbp, in xfs_rmapbt_alloc_block() 97 trace_xfs_rmapbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_rmapbt_alloc_block() 104 xfs_extent_busy_reuse(cur->bc_mp, cur->bc_private.a.agno, bno, 1, in xfs_rmapbt_alloc_block() 112 xfs_ag_resv_rmapbt_alloc(cur->bc_mp, cur->bc_private.a.agno); in xfs_rmapbt_alloc_block() 123 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_free_block() 129 trace_xfs_rmapbt_free_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_rmapbt_free_block() 141 xfs_ag_resv_rmapbt_free(cur->bc_mp, cur->bc_private.a.agno); in xfs_rmapbt_free_block() [all …]
|
D | xfs_refcount.c | 49 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_le() 66 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_ge() 83 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_eq() 111 xfs_agnumber_t agno = cur->bc_private.a.agno; in xfs_refcount_get_rec() 122 agno = cur->bc_private.a.agno; in xfs_refcount_get_rec() 147 trace_xfs_refcount_get(cur->bc_mp, cur->bc_private.a.agno, irec); in xfs_refcount_get_rec() 172 trace_xfs_refcount_update(cur->bc_mp, cur->bc_private.a.agno, irec); in xfs_refcount_update() 179 cur->bc_private.a.agno, error, _RET_IP_); in xfs_refcount_update() 196 trace_xfs_refcount_insert(cur->bc_mp, cur->bc_private.a.agno, irec); in xfs_refcount_insert() 208 cur->bc_private.a.agno, error, _RET_IP_); in xfs_refcount_insert() [all …]
|
D | xfs_btree.h | 173 struct xfs_alloc_rec_incore a; member 213 } a; member 447 #define XFS_EXTLEN_MIN(a,b) min_t(xfs_extlen_t, (a), (b)) argument 448 #define XFS_EXTLEN_MAX(a,b) max_t(xfs_extlen_t, (a), (b)) argument 449 #define XFS_AGBLOCK_MIN(a,b) min_t(xfs_agblock_t, (a), (b)) argument 450 #define XFS_AGBLOCK_MAX(a,b) max_t(xfs_agblock_t, (a), (b)) argument 451 #define XFS_FILEOFF_MIN(a,b) min_t(xfs_fileoff_t, (a), (b)) argument 452 #define XFS_FILEOFF_MAX(a,b) max_t(xfs_fileoff_t, (a), (b)) argument 453 #define XFS_FILBLKS_MIN(a,b) min_t(xfs_filblks_t, (a), (b)) argument 454 #define XFS_FILBLKS_MAX(a,b) max_t(xfs_filblks_t, (a), (b)) argument [all …]
|
D | xfs_refcount_btree.c | 28 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_refcountbt_dup_cursor() 37 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_set_root() 60 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_alloc_block() 69 args.fsbno = XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcountbt_alloc_block() 78 trace_xfs_refcountbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcountbt_alloc_block() 84 ASSERT(args.agno == cur->bc_private.a.agno); in xfs_refcountbt_alloc_block() 104 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_free_block() 109 trace_xfs_refcountbt_free_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcountbt_free_block() 172 struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); in xfs_refcountbt_init_ptr_from_cur() 174 ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno)); in xfs_refcountbt_init_ptr_from_cur() [all …]
|
D | xfs_rmap.c | 82 trace_xfs_rmap_update(cur->bc_mp, cur->bc_private.a.agno, in xfs_rmap_update() 94 cur->bc_private.a.agno, error, _RET_IP_); in xfs_rmap_update() 110 trace_xfs_rmap_insert(rcur->bc_mp, rcur->bc_private.a.agno, agbno, in xfs_rmap_insert() 130 rcur->bc_private.a.agno, error, _RET_IP_); in xfs_rmap_insert() 146 trace_xfs_rmap_delete(rcur->bc_mp, rcur->bc_private.a.agno, agbno, in xfs_rmap_delete() 161 rcur->bc_private.a.agno, error, _RET_IP_); in xfs_rmap_delete() 188 xfs_agnumber_t agno = cur->bc_private.a.agno; in xfs_rmap_get_rec() 251 cur->bc_private.a.agno, rec->rm_startblock, in xfs_rmap_find_left_neighbor_helper() 303 cur->bc_private.a.agno, bno, 0, owner, offset, flags); in xfs_rmap_find_left_neighbor() 311 cur->bc_private.a.agno, irec->rm_startblock, in xfs_rmap_find_left_neighbor() [all …]
|
/fs/f2fs/ |
D | sysfs.c | 65 static ssize_t dirty_segments_show(struct f2fs_attr *a, in dirty_segments_show() argument 72 static ssize_t unusable_show(struct f2fs_attr *a, in unusable_show() argument 85 static ssize_t encoding_show(struct f2fs_attr *a, in encoding_show() argument 99 static ssize_t lifetime_write_kbytes_show(struct f2fs_attr *a, in lifetime_write_kbytes_show() argument 112 static ssize_t features_show(struct f2fs_attr *a, in features_show() argument 163 static ssize_t current_reserved_blocks_show(struct f2fs_attr *a, in current_reserved_blocks_show() argument 169 static ssize_t f2fs_sbi_show(struct f2fs_attr *a, in f2fs_sbi_show() argument 175 ptr = __struct_ptr(sbi, a->struct_type); in f2fs_sbi_show() 179 if (!strcmp(a->attr.name, "extension_list")) { in f2fs_sbi_show() 200 ui = (unsigned int *)(ptr + a->offset); in f2fs_sbi_show() [all …]
|
/fs/proc/ |
D | consoles.c | 30 unsigned int a; in show_console_dev() local 43 for (a = 0; a < ARRAY_SIZE(con_flags); a++) in show_console_dev() 44 flags[a] = (con->flags & con_flags[a].flag) ? in show_console_dev() 45 con_flags[a].name : ' '; in show_console_dev() 46 flags[a] = 0; in show_console_dev()
|
/fs/cramfs/ |
D | Kconfig | 7 System). CramFs is designed to be a simple, small, and compressed 15 To compile this as a module, choose M here: the module will be called 17 directory /) cannot be compiled as a module. 22 consider a more capable compressed filesystem such as SquashFS 28 bool "Support CramFs image over a regular block device" if EXPERT 32 This option allows the CramFs driver to load data from a regular 33 block device such a disk partition or a ramdisk. 41 a linear adressed memory range (usually non volatile memory 46 The location of the CramFs image is determined by a 48 the 'physmap' map driver or a resulting MTD partition.
|
D | README | 5 a bit looser, e.g. it doesn't care if the <file_data> items are 7 a given directory are contiguous, as this is used by readdir). 23 null-padded to a multiple of 4 bytes. 27 a directory's entries before recursing down its subdirectories: the 33 allows cramfs_lookup to return more quickly when a filename does not 37 One <file_data> for each file that's either a symlink or a 46 The i'th <block_pointer> for a file stores the byte offset of the 61 aligned to a 4-byte boundary. The block size is either blksize 68 The order of <file_data>'s is a depth-first descent of the directory 78 Each <block> may be a different size. (See <block_pointer> above.) [all …]
|
/fs/nfs_common/ |
D | nfsacl.c | 199 const struct posix_acl_entry *a = x, *b = y; in cmp_acl_entry() local 201 if (a->e_tag != b->e_tag) in cmp_acl_entry() 202 return a->e_tag - b->e_tag; in cmp_acl_entry() 203 else if ((a->e_tag == ACL_USER) && uid_gt(a->e_uid, b->e_uid)) in cmp_acl_entry() 205 else if ((a->e_tag == ACL_USER) && uid_lt(a->e_uid, b->e_uid)) in cmp_acl_entry() 207 else if ((a->e_tag == ACL_GROUP) && gid_gt(a->e_gid, b->e_gid)) in cmp_acl_entry() 209 else if ((a->e_tag == ACL_GROUP) && gid_lt(a->e_gid, b->e_gid)) in cmp_acl_entry()
|
/fs/xfs/ |
D | Kconfig | 9 XFS is a high performance journaling filesystem which originated 20 To compile this file system support as a module, choose M here: the 22 system of your root partition is compiled as a module, you'll need 31 a per user and/or a per group basis under XFS. XFS considers quota 32 information as filesystem metadata and uses journaling to provide a 34 quota is also compatible with the IRIX version of XFS, allowing a 58 which contain a realtime subvolume. The realtime subvolume is a 61 for media streaming applications, but is also useful as a generic 63 separated. Regular file I/Os are isolated to a separate device 76 If you say Y here you will be able to check metadata on a [all …]
|
/fs/overlayfs/ |
D | Kconfig | 7 and a 'lower' filesystem. When a name exists in both filesystems, the 21 "redirect_dir=off" module option or on a filesystem instance basis 25 an overlay which has redirects on a kernel that doesn't support this 35 Disable this to get a possibly more secure configuration, but that 52 "index=off" module option or on a filesystem instance basis with the 59 That is, mounting an overlay which has an inodes index on a kernel 73 globally with the "nfs_export=off" module option or on a filesystem 84 That is, mounting an overlay which has a full index on a kernel 87 Most users should say N here and enable this feature on a case-by- 99 inodes to a unified address space. The mapped 64bit inode numbers [all …]
|
/fs/affs/ |
D | Changes | 21 might leave a trashed file system with the 24 - When a file is truncated to a size that is not 25 a multiple of the blocksize, the rest of the 34 - fix for a possible bitmap corruption 112 - truncate() zeroes the unused remainder of a 113 partially used last block when a file is truncated. 118 - Added a few comments, fixed some typos (and 120 more consistent. Changed a bad example in the 148 when it allocated a new block. 173 up if a file was enlarged via truncate() and there [all …]
|
/fs/fscache/ |
D | Kconfig | 6 This option enables a generic filesystem caching manager that can be 22 The gathering of statistics adds a certain amount of overhead to 23 execution as there are a quite a few stats gathered, and on a 39 The generation of this histogram adds a certain amount of overhead to 40 execution as there are a number of points at which data is gathered, 41 and on a multi-CPU system these may be on cachelines that keep 61 Maintain a global list of active fscache objects that can be
|