Home
last modified time | relevance | path

Searched refs:a (Results 1 – 25 of 194) sorted by relevance

12345678

/fs/reiserfs/
Dhashes.c36 b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); \
49 u32 a, b, c, d; in keyed_hash() local
59 a = (u32) msg[0] | in keyed_hash()
77 a = (u32) msg[0] | in keyed_hash()
91 a = (u32) msg[0] | in keyed_hash()
102 a = (u32) msg[0] | in keyed_hash()
111 a = b = c = d = pad; in keyed_hash()
113 a <<= 8; in keyed_hash()
114 a |= msg[i]; in keyed_hash()
131 u32 a, c; in yura_hash() local
[all …]
/fs/ntfs/
Dattrib.c90 ATTR_RECORD *a; in ntfs_map_runlist_nolock() local
117 a = ctx->attr; in ntfs_map_runlist_nolock()
118 BUG_ON(!a->non_resident); in ntfs_map_runlist_nolock()
120 end_vcn = sle64_to_cpu(a->data.non_resident.highest_vcn); in ntfs_map_runlist_nolock()
125 if (!a->data.non_resident.lowest_vcn && end_vcn <= 0) in ntfs_map_runlist_nolock()
134 if (vcn >= allocated_size_vcn || (a->type == ni->type && in ntfs_map_runlist_nolock()
135 a->name_length == ni->name_len && in ntfs_map_runlist_nolock()
136 !memcmp((u8*)a + le16_to_cpu(a->name_offset), in ntfs_map_runlist_nolock()
138 sle64_to_cpu(a->data.non_resident.lowest_vcn) in ntfs_map_runlist_nolock()
175 a = ctx->attr; in ntfs_map_runlist_nolock()
[all …]
Dinode.c555 ATTR_RECORD *a; in ntfs_read_locked_inode() local
662 a = ctx->attr; in ntfs_read_locked_inode()
664 si = (STANDARD_INFORMATION*)((u8*)a + in ntfs_read_locked_inode()
665 le16_to_cpu(a->data.resident.value_offset)); in ntfs_read_locked_inode()
704 a = ctx->attr; in ntfs_read_locked_inode()
705 if (a->flags & ATTR_COMPRESSION_MASK) { in ntfs_read_locked_inode()
710 if (a->flags & ATTR_IS_ENCRYPTED || in ntfs_read_locked_inode()
711 a->flags & ATTR_IS_SPARSE) { in ntfs_read_locked_inode()
712 if (a->non_resident) { in ntfs_read_locked_inode()
728 ni->attr_list_size = (u32)ntfs_attr_size(a); in ntfs_read_locked_inode()
[all …]
Ddebug.h44 #define ntfs_debug(f, a...) \ argument
45 __ntfs_debug(__FILE__, __LINE__, __func__, f, ##a)
64 #define ntfs_warning(sb, f, a...) __ntfs_warning(__func__, sb, f, ##a) argument
69 #define ntfs_error(sb, f, a...) __ntfs_error(__func__, sb, f, ##a) argument
Dfile.c118 ATTR_RECORD *a; in ntfs_attr_extend_initialized() local
164 a = ctx->attr; in ntfs_attr_extend_initialized()
165 BUG_ON(a->non_resident); in ntfs_attr_extend_initialized()
167 attr_len = le32_to_cpu(a->data.resident.value_length); in ntfs_attr_extend_initialized()
173 kattr = (u8*)a + le16_to_cpu(a->data.resident.value_offset); in ntfs_attr_extend_initialized()
175 a->data.resident.value_length = cpu_to_le32((u32)new_init_size); in ntfs_attr_extend_initialized()
208 a = ctx->attr; in ntfs_attr_extend_initialized()
209 BUG_ON(!a->non_resident); in ntfs_attr_extend_initialized()
211 sle64_to_cpu(a->data.non_resident.data_size)); in ntfs_attr_extend_initialized()
212 a->data.non_resident.data_size = cpu_to_sle64(new_init_size); in ntfs_attr_extend_initialized()
[all …]
/fs/hpfs/
Dname.c22 static inline unsigned char upcase(unsigned char *dir, unsigned char a) in upcase() argument
24 if (a<128 || a==255) return a>='a' && a<='z' ? a - 0x20 : a; in upcase()
25 if (!dir) return a; in upcase()
26 return dir[a-128]; in upcase()
29 unsigned char hpfs_upcase(unsigned char *dir, unsigned char a) in hpfs_upcase() argument
31 return upcase(dir, a); in hpfs_upcase()
34 static inline unsigned char locase(unsigned char *dir, unsigned char a) in locase() argument
36 if (a<128 || a==255) return a>='A' && a<='Z' ? a + 0x20 : a; in locase()
37 if (!dir) return a; in locase()
38 return dir[a]; in locase()
Dea.c14 void hpfs_ea_ext_remove(struct super_block *s, secno a, int ano, unsigned len) in hpfs_ea_ext_remove() argument
22 ano ? "anode" : "sectors", a, len); in hpfs_ea_ext_remove()
25 if (hpfs_ea_read(s, a, ano, pos, 4, ex)) return; in hpfs_ea_ext_remove()
29 ano ? "anode" : "sectors", a, pos); in hpfs_ea_ext_remove()
32 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 9, ex+4)) in hpfs_ea_ext_remove()
38 if (!ano) hpfs_free_sectors(s, a, (len+511) >> 9); in hpfs_ea_ext_remove()
42 if ((anode = hpfs_map_anode(s, a, &bh))) { in hpfs_ea_ext_remove()
45 hpfs_free_sectors(s, a, 1); in hpfs_ea_ext_remove()
50 static char *get_indirect_ea(struct super_block *s, int ano, secno a, int size) in get_indirect_ea() argument
57 if (hpfs_ea_read(s, a, ano, 0, size, ret)) { in get_indirect_ea()
[all …]
Danode.c17 anode_secno a = -1; in hpfs_bplus_lookup() local
22 if (hpfs_sb(s)->sb_chk) if (hpfs_stop_cycles(s, a, &c1, &c2, "hpfs_bplus_lookup")) return -1; in hpfs_bplus_lookup()
26 a = le32_to_cpu(btree->u.internal[i].down); in hpfs_bplus_lookup()
28 if (!(anode = hpfs_map_anode(s, a, &bh))) return -1; in hpfs_bplus_lookup()
32 hpfs_error(s, "sector %08x not found in internal anode %08x", sec, a); in hpfs_bplus_lookup()
39a = le32_to_cpu(btree->u.external[i].disk_secno) + sec - le32_to_cpu(btree->u.external[i].file_sec… in hpfs_bplus_lookup()
40 if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, a, 1, "data")) { in hpfs_bplus_lookup()
51 return a; in hpfs_bplus_lookup()
53 hpfs_error(s, "sector %08x not found in external anode %08x", sec, a); in hpfs_bplus_lookup()
65 anode_secno a, na = -1, ra, up = -1; in hpfs_add_sector_to_btree() local
[all …]
/fs/xfs/
Dxfs_linux.h155 #define MIN(a,b) (min(a,b)) argument
156 #define MAX(a,b) (max(a,b)) argument
211 #define xfs_sort(a,n,s,fn) sort(a,n,s,fn,NULL) argument
222 static inline __u32 xfs_do_div(void *a, __u32 b, int n) in xfs_do_div() argument
228 mod = *(__u32 *)a % b; in xfs_do_div()
229 *(__u32 *)a = *(__u32 *)a / b; in xfs_do_div()
234 __u64 c = *(__u64 *)a; in xfs_do_div()
243 *(__u64 *)a = c; in xfs_do_div()
253 static inline __u32 xfs_do_mod(void *a, __u32 b, int n) in xfs_do_mod() argument
257 return *(__u32 *)a % b; in xfs_do_mod()
[all …]
/fs/jfs/
Djfs_extent.c37 #define DPD(a) (printk("(a): %d\n",(a))) argument
38 #define DPC(a) (printk("(a): %c\n",(a))) argument
39 #define DPL1(a) \ argument
41 if ((a) >> 32) \
42 printk("(a): %x%08x ",(a)); \
44 printk("(a): %x ",(a) << 32); \
46 #define DPL(a) \ argument
48 if ((a) >> 32) \
49 printk("(a): %x%08x\n",(a)); \
51 printk("(a): %x\n",(a) << 32); \
[all …]
/fs/jffs2/
Dsummary.h196 #define jffs2_sum_init(a) (0) argument
197 #define jffs2_sum_exit(a) argument
198 #define jffs2_sum_disable_collecting(a) argument
199 #define jffs2_sum_is_disabled(a) (0) argument
200 #define jffs2_sum_reset_collected(a) argument
201 #define jffs2_sum_add_kvec(a,b,c,d) (0) argument
202 #define jffs2_sum_move_collected(a,b) argument
203 #define jffs2_sum_write_sumnode(a) (0) argument
204 #define jffs2_sum_add_padding_mem(a,b) argument
205 #define jffs2_sum_add_inode_mem(a,b,c) argument
[all …]
/fs/xfs/libxfs/
Dxfs_rmap_btree.c70 cur->bc_private.a.agbp, cur->bc_private.a.agno); in xfs_rmapbt_dup_cursor()
79 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_set_root()
102 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_alloc_block()
110 error = xfs_alloc_get_freelist(cur->bc_tp, cur->bc_private.a.agbp, in xfs_rmapbt_alloc_block()
117 trace_xfs_rmapbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_rmapbt_alloc_block()
125 xfs_extent_busy_reuse(cur->bc_mp, cur->bc_private.a.agno, bno, 1, in xfs_rmapbt_alloc_block()
143 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_rmapbt_free_block()
149 trace_xfs_rmapbt_free_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_rmapbt_free_block()
236 struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); in xfs_rmapbt_init_ptr_from_cur()
238 ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno)); in xfs_rmapbt_init_ptr_from_cur()
[all …]
Dxfs_refcount.c66 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_le()
83 trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, in xfs_refcount_lookup_ge()
116 trace_xfs_refcount_get(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcount_get_rec()
135 trace_xfs_refcount_update(cur->bc_mp, cur->bc_private.a.agno, irec); in xfs_refcount_update()
142 cur->bc_private.a.agno, error, _RET_IP_); in xfs_refcount_update()
159 trace_xfs_refcount_insert(cur->bc_mp, cur->bc_private.a.agno, irec); in xfs_refcount_insert()
168 cur->bc_private.a.agno, error, _RET_IP_); in xfs_refcount_insert()
191 trace_xfs_refcount_delete(cur->bc_mp, cur->bc_private.a.agno, &irec); in xfs_refcount_delete()
200 cur->bc_private.a.agno, error, _RET_IP_); in xfs_refcount_delete()
317 trace_xfs_refcount_split_extent(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcount_split_extent()
[all …]
Dxfs_btree.h222 struct xfs_alloc_rec_incore a; member
262 } a; member
485 #define XFS_EXTLEN_MIN(a,b) min_t(xfs_extlen_t, (a), (b)) argument
486 #define XFS_EXTLEN_MAX(a,b) max_t(xfs_extlen_t, (a), (b)) argument
487 #define XFS_AGBLOCK_MIN(a,b) min_t(xfs_agblock_t, (a), (b)) argument
488 #define XFS_AGBLOCK_MAX(a,b) max_t(xfs_agblock_t, (a), (b)) argument
489 #define XFS_FILEOFF_MIN(a,b) min_t(xfs_fileoff_t, (a), (b)) argument
490 #define XFS_FILEOFF_MAX(a,b) max_t(xfs_fileoff_t, (a), (b)) argument
491 #define XFS_FILBLKS_MIN(a,b) min_t(xfs_filblks_t, (a), (b)) argument
492 #define XFS_FILBLKS_MAX(a,b) max_t(xfs_filblks_t, (a), (b)) argument
Dxfs_refcount_btree.c44 cur->bc_private.a.agbp, cur->bc_private.a.agno, in xfs_refcountbt_dup_cursor()
45 cur->bc_private.a.dfops); in xfs_refcountbt_dup_cursor()
54 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_set_root()
77 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_alloc_block()
88 args.fsbno = XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcountbt_alloc_block()
98 trace_xfs_refcountbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcountbt_alloc_block()
105 ASSERT(args.agno == cur->bc_private.a.agno); in xfs_refcountbt_alloc_block()
127 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_refcountbt_free_block()
133 trace_xfs_refcountbt_free_block(cur->bc_mp, cur->bc_private.a.agno, in xfs_refcountbt_free_block()
197 struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); in xfs_refcountbt_init_ptr_from_cur()
[all …]
Dxfs_alloc_btree.c41 cur->bc_private.a.agbp, cur->bc_private.a.agno, in xfs_allocbt_dup_cursor()
51 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_allocbt_set_root()
80 error = xfs_alloc_get_freelist(cur->bc_tp, cur->bc_private.a.agbp, in xfs_allocbt_alloc_block()
93 xfs_extent_busy_reuse(cur->bc_mp, cur->bc_private.a.agno, bno, 1, false); in xfs_allocbt_alloc_block()
108 struct xfs_buf *agbp = cur->bc_private.a.agbp; in xfs_allocbt_free_block()
135 struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); in xfs_allocbt_update_lastrec()
184 xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp, XFS_AGF_LONGEST); in xfs_allocbt_update_lastrec()
219 ASSERT(cur->bc_rec.a.ar_startblock != 0); in xfs_allocbt_init_rec_from_cur()
221 rec->alloc.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock); in xfs_allocbt_init_rec_from_cur()
222 rec->alloc.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount); in xfs_allocbt_init_rec_from_cur()
[all …]
/fs/ext4/
Dsysfs.c50 static ssize_t session_write_kbytes_show(struct ext4_attr *a, in session_write_kbytes_show() argument
62 static ssize_t lifetime_write_kbytes_show(struct ext4_attr *a, in lifetime_write_kbytes_show() argument
75 static ssize_t inode_readahead_blks_store(struct ext4_attr *a, in inode_readahead_blks_store() argument
93 static ssize_t reserved_clusters_store(struct ext4_attr *a, in reserved_clusters_store() argument
110 static ssize_t trigger_test_error(struct ext4_attr *a, in trigger_test_error() argument
242 static void *calc_ptr(struct ext4_attr *a, struct ext4_sb_info *sbi) in calc_ptr() argument
244 switch (a->attr_ptr) { in calc_ptr()
246 return a->u.explicit_ptr; in calc_ptr()
248 return (void *) (((char *) sbi) + a->u.offset); in calc_ptr()
250 return (void *) (((char *) sbi->s_es) + a->u.offset); in calc_ptr()
[all …]
/fs/f2fs/
Dsysfs.c66 static ssize_t dirty_segments_show(struct f2fs_attr *a, in dirty_segments_show() argument
73 static ssize_t lifetime_write_kbytes_show(struct f2fs_attr *a, in lifetime_write_kbytes_show() argument
86 static ssize_t features_show(struct f2fs_attr *a, in features_show() argument
126 static ssize_t current_reserved_blocks_show(struct f2fs_attr *a, in current_reserved_blocks_show() argument
132 static ssize_t f2fs_sbi_show(struct f2fs_attr *a, in f2fs_sbi_show() argument
138 ptr = __struct_ptr(sbi, a->struct_type); in f2fs_sbi_show()
142 if (!strcmp(a->attr.name, "extension_list")) { in f2fs_sbi_show()
163 ui = (unsigned int *)(ptr + a->offset); in f2fs_sbi_show()
168 static ssize_t f2fs_sbi_store(struct f2fs_attr *a, in f2fs_sbi_store() argument
177 ptr = __struct_ptr(sbi, a->struct_type); in f2fs_sbi_store()
[all …]
/fs/proc/
Dconsoles.c31 unsigned int a; in show_console_dev() local
44 for (a = 0; a < ARRAY_SIZE(con_flags); a++) in show_console_dev()
45 flags[a] = (con->flags & con_flags[a].flag) ? in show_console_dev()
46 con_flags[a].name : ' '; in show_console_dev()
47 flags[a] = 0; in show_console_dev()
DKconfig5 This is a virtual file system providing information about the status
13 (there is a small number of Interrupt ReQuest lines in your computer
15 often a source of trouble if two devices are mistakenly configured
20 meaning it has to be given a location in the directory hierarchy.
35 Provides a virtual ELF core file of the live kernel. This can
52 The sysctl interface provides a means of dynamically changing
54 a recompile of the kernel or reboot of the system. The primary
55 interface is through /proc/sys. If you say Y here a tree of
61 As it is generally a good thing, you should say Y here unless
62 building a kernel for install/rescue disks or your system is very
[all …]
/fs/cramfs/
DREADME5 a bit looser, e.g. it doesn't care if the <file_data> items are
7 a given directory are contiguous, as this is used by readdir).
23 null-padded to a multiple of 4 bytes.
27 a directory's entries before recursing down its subdirectories: the
33 allows cramfs_lookup to return more quickly when a filename does not
37 One <file_data> for each file that's either a symlink or a
46 The i'th <block_pointer> for a file stores the byte offset of the
52 The order of <file_data>'s is a depth-first descent of the directory
60 Each <block> may be a different size. (See <block_pointer> above.)
88 compressed at a time. It's intended to be somewhere around
[all …]
/fs/nfs_common/
Dnfsacl.c198 const struct posix_acl_entry *a = x, *b = y; in cmp_acl_entry() local
200 if (a->e_tag != b->e_tag) in cmp_acl_entry()
201 return a->e_tag - b->e_tag; in cmp_acl_entry()
202 else if ((a->e_tag == ACL_USER) && uid_gt(a->e_uid, b->e_uid)) in cmp_acl_entry()
204 else if ((a->e_tag == ACL_USER) && uid_lt(a->e_uid, b->e_uid)) in cmp_acl_entry()
206 else if ((a->e_tag == ACL_GROUP) && gid_gt(a->e_gid, b->e_gid)) in cmp_acl_entry()
208 else if ((a->e_tag == ACL_GROUP) && gid_lt(a->e_gid, b->e_gid)) in cmp_acl_entry()
/fs/exofs/
Dore_raid.h18 #define ORE_ERR(fmt, a...) printk(KERN_ERR "ore: " fmt, ##a) argument
21 #define ORE_DBGMSG(fmt, a...) \ argument
22 printk(KERN_NOTICE "ore @%s:%d: " fmt, __func__, __LINE__, ##a)
24 #define ORE_DBGMSG(fmt, a...) \ argument
25 do { if (0) printk(fmt, ##a); } while (0)
/fs/fscache/
DKconfig5 This option enables a generic filesystem caching manager that can be
21 The gathering of statistics adds a certain amount of overhead to
22 execution as there are a quite a few stats gathered, and on a
38 The generation of this histogram adds a certain amount of overhead to
39 execution as there are a number of points at which data is gathered,
40 and on a multi-CPU system these may be on cachelines that keep
60 Maintain a global list of active fscache objects that can be
/fs/affs/
DChanges21 might leave a trashed file system with the
24 - When a file is truncated to a size that is not
25 a multiple of the blocksize, the rest of the
34 - fix for a possible bitmap corruption
112 - truncate() zeroes the unused remainder of a
113 partially used last block when a file is truncated.
118 - Added a few comments, fixed some typos (and
120 more consistent. Changed a bad example in the
148 when it allocated a new block.
173 up if a file was enlarged via truncate() and there
[all …]

12345678