Home
last modified time | relevance | path

Searched refs:n (Results 1 – 25 of 273) sorted by relevance

1234567891011

/fs/befs/
Dendian.h15 fs64_to_cpu(const struct super_block *sb, fs64 n) in fs64_to_cpu() argument
18 return le64_to_cpu((__force __le64)n); in fs64_to_cpu()
20 return be64_to_cpu((__force __be64)n); in fs64_to_cpu()
24 cpu_to_fs64(const struct super_block *sb, u64 n) in cpu_to_fs64() argument
27 return (__force fs64)cpu_to_le64(n); in cpu_to_fs64()
29 return (__force fs64)cpu_to_be64(n); in cpu_to_fs64()
33 fs32_to_cpu(const struct super_block *sb, fs32 n) in fs32_to_cpu() argument
36 return le32_to_cpu((__force __le32)n); in fs32_to_cpu()
38 return be32_to_cpu((__force __be32)n); in fs32_to_cpu()
42 cpu_to_fs32(const struct super_block *sb, u32 n) in cpu_to_fs32() argument
[all …]
/fs/ufs/
Dswab.h26 fs64_to_cpu(struct super_block *sbp, __fs64 n) in fs64_to_cpu() argument
29 return le64_to_cpu((__force __le64)n); in fs64_to_cpu()
31 return be64_to_cpu((__force __be64)n); in fs64_to_cpu()
35 cpu_to_fs64(struct super_block *sbp, u64 n) in cpu_to_fs64() argument
38 return (__force __fs64)cpu_to_le64(n); in cpu_to_fs64()
40 return (__force __fs64)cpu_to_be64(n); in cpu_to_fs64()
44 fs32_to_cpu(struct super_block *sbp, __fs32 n) in fs32_to_cpu() argument
47 return le32_to_cpu((__force __le32)n); in fs32_to_cpu()
49 return be32_to_cpu((__force __be32)n); in fs32_to_cpu()
53 cpu_to_fs32(struct super_block *sbp, u32 n) in cpu_to_fs32() argument
[all …]
/fs/ubifs/
Dtnc.c186 struct ubifs_old_idx *old_idx, *n; in destroy_old_idx() local
188 rbtree_postorder_for_each_entry_safe(old_idx, n, &c->old_idx, rb) in destroy_old_idx()
219 const int n = zn->child_cnt; in copy_znode() local
222 for (i = 0; i < n; i++) { in copy_znode()
589 struct ubifs_znode *znode, int n) in get_znode() argument
593 zbr = &znode->zbranch[n]; in get_znode()
597 znode = ubifs_load_znode(c, zbr, znode, n); in get_znode()
610 static int tnc_next(struct ubifs_info *c, struct ubifs_znode **zn, int *n) in tnc_next() argument
613 int nn = *n; in tnc_next()
617 *n = nn; in tnc_next()
[all …]
/fs/fscache/
Dhistogram.c30 unsigned n[5], t; in fscache_histogram_show() local
41 n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]); in fscache_histogram_show()
42 n[1] = atomic_read(&fscache_ops_histogram[index]); in fscache_histogram_show()
43 n[2] = atomic_read(&fscache_objs_histogram[index]); in fscache_histogram_show()
44 n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]); in fscache_histogram_show()
45 n[4] = atomic_read(&fscache_retrieval_histogram[index]); in fscache_histogram_show()
46 if (!(n[0] | n[1] | n[2] | n[3] | n[4])) in fscache_histogram_show()
52 index, t, n[0], n[1], n[2], n[3], n[4]); in fscache_histogram_show()
/fs/sysv/
Dsysv.h190 static inline __u32 fs32_to_cpu(struct sysv_sb_info *sbi, __fs32 n) in fs32_to_cpu() argument
193 return PDP_swab((__force __u32)n); in fs32_to_cpu()
195 return le32_to_cpu((__force __le32)n); in fs32_to_cpu()
197 return be32_to_cpu((__force __be32)n); in fs32_to_cpu()
200 static inline __fs32 cpu_to_fs32(struct sysv_sb_info *sbi, __u32 n) in cpu_to_fs32() argument
203 return (__force __fs32)PDP_swab(n); in cpu_to_fs32()
205 return (__force __fs32)cpu_to_le32(n); in cpu_to_fs32()
207 return (__force __fs32)cpu_to_be32(n); in cpu_to_fs32()
210 static inline __fs32 fs32_add(struct sysv_sb_info *sbi, __fs32 *n, int d) in fs32_add() argument
213 *(__u32*)n = PDP_swab(PDP_swab(*(__u32*)n)+d); in fs32_add()
[all …]
Ditree.c29 int n = 0; in block_to_path() local
34 offsets[n++] = block; in block_to_path()
36 offsets[n++] = DIRECT; in block_to_path()
37 offsets[n++] = block; in block_to_path()
39 offsets[n++] = DIRECT+1; in block_to_path()
40 offsets[n++] = block >> ptrs_bits; in block_to_path()
41 offsets[n++] = block & (indirect_blocks - 1); in block_to_path()
43 offsets[n++] = DIRECT+2; in block_to_path()
44 offsets[n++] = block >> (ptrs_bits * 2); in block_to_path()
45 offsets[n++] = (block >> ptrs_bits) & (indirect_blocks - 1); in block_to_path()
[all …]
/fs/minix/
Ditree_v2.c8 static inline unsigned long block_to_cpu(block_t n) in block_to_cpu() argument
10 return n; in block_to_cpu()
13 static inline block_t cpu_to_block(unsigned long n) in cpu_to_block() argument
15 return n; in cpu_to_block()
28 int n = 0; in block_to_path() local
42 offsets[n++] = block; in block_to_path()
44 offsets[n++] = DIRCOUNT; in block_to_path()
45 offsets[n++] = block; in block_to_path()
47 offsets[n++] = DIRCOUNT + 1; in block_to_path()
48 offsets[n++] = block / INDIRCOUNT(sb); in block_to_path()
[all …]
Ditree_v1.c9 static inline unsigned long block_to_cpu(block_t n) in block_to_cpu() argument
11 return n; in block_to_cpu()
14 static inline block_t cpu_to_block(unsigned long n) in cpu_to_block() argument
16 return n; in cpu_to_block()
26 int n = 0; in block_to_path() local
38 offsets[n++] = block; in block_to_path()
40 offsets[n++] = 7; in block_to_path()
41 offsets[n++] = block; in block_to_path()
44 offsets[n++] = 8; in block_to_path()
45 offsets[n++] = block>>9; in block_to_path()
[all …]
Ditree_common.c74 int n = 0; in alloc_branch() local
80 if (parent) for (n = 1; n < num; n++) { in alloc_branch()
86 branch[n].key = cpu_to_block(nr); in alloc_branch()
95 branch[n].bh = bh; in alloc_branch()
96 branch[n].p = (block_t*) bh->b_data + offsets[n]; in alloc_branch()
97 *branch[n].p = branch[n].key; in alloc_branch()
103 if (n == num) in alloc_branch()
107 for (i = 1; i < n; i++) in alloc_branch()
109 for (i = 0; i < n; i++) in alloc_branch()
307 int n; in truncate() local
[all …]
/fs/qnx6/
Dqnx6.h77 static inline __u64 fs64_to_cpu(struct qnx6_sb_info *sbi, __fs64 n) in fs64_to_cpu() argument
80 return le64_to_cpu((__force __le64)n); in fs64_to_cpu()
82 return be64_to_cpu((__force __be64)n); in fs64_to_cpu()
85 static inline __fs64 cpu_to_fs64(struct qnx6_sb_info *sbi, __u64 n) in cpu_to_fs64() argument
88 return (__force __fs64)cpu_to_le64(n); in cpu_to_fs64()
90 return (__force __fs64)cpu_to_be64(n); in cpu_to_fs64()
93 static inline __u32 fs32_to_cpu(struct qnx6_sb_info *sbi, __fs32 n) in fs32_to_cpu() argument
96 return le32_to_cpu((__force __le32)n); in fs32_to_cpu()
98 return be32_to_cpu((__force __be32)n); in fs32_to_cpu()
101 static inline __fs32 cpu_to_fs32(struct qnx6_sb_info *sbi, __u32 n) in cpu_to_fs32() argument
[all …]
Ddir.c26 static struct page *qnx6_get_page(struct inode *dir, unsigned long n) in qnx6_get_page() argument
29 struct page *page = read_mapping_page(mapping, n, NULL); in qnx6_get_page()
50 u32 n = s >> (PAGE_CACHE_SHIFT - sb->s_blocksize_bits); /* in pages */ in qnx6_longname() local
54 struct page *page = read_mapping_page(mapping, n, NULL); in qnx6_longname()
118 unsigned long n = pos >> PAGE_CACHE_SHIFT; in qnx6_readdir() local
126 for ( ; !done && n < npages; n++, start = 0) { in qnx6_readdir()
127 struct page *page = qnx6_get_page(inode, n); in qnx6_readdir()
128 int limit = last_entry(inode, n); in qnx6_readdir()
134 ctx->pos = (n + 1) << PAGE_CACHE_SHIFT; in qnx6_readdir()
218 unsigned long start, n; in qnx6_find_entry() local
[all …]
/fs/jfs/
Djfs_unicode.h56 size_t n) in UniStrncpy_le() argument
60 while (n-- && *ucs2) /* Copy the strings */ in UniStrncpy_le()
63 n++; in UniStrncpy_le()
64 while (n--) /* Pad with nulls */ in UniStrncpy_le()
73 size_t n) in UniStrncmp_le() argument
75 if (!n) in UniStrncmp_le()
77 while ((*ucs1 == __le16_to_cpu(*ucs2)) && *ucs1 && --n) { in UniStrncmp_le()
88 size_t n) in UniStrncpy_to_le() argument
92 while (n-- && *ucs2) /* Copy the strings */ in UniStrncpy_to_le()
95 n++; in UniStrncpy_to_le()
[all …]
/fs/xfs/libxfs/
Dxfs_bit.h28 static inline __uint64_t xfs_mask64hi(int n) in xfs_mask64hi() argument
30 return (__uint64_t)-1 << (64 - (n)); in xfs_mask64hi()
32 static inline __uint32_t xfs_mask32lo(int n) in xfs_mask32lo() argument
34 return ((__uint32_t)1 << (n)) - 1; in xfs_mask32lo()
36 static inline __uint64_t xfs_mask64lo(int n) in xfs_mask64lo() argument
38 return ((__uint64_t)1 << (n)) - 1; in xfs_mask64lo()
63 int n = 0; in xfs_lowbit64() local
66 n = ffs(w); in xfs_lowbit64()
70 n = ffs(w); in xfs_lowbit64()
71 if (n) in xfs_lowbit64()
[all …]
/fs/ext4/
Dblock_validity.c63 struct rb_node **n = &sbi->system_blks.rb_node, *node; in add_system_zone() local
66 while (*n) { in add_system_zone()
67 parent = *n; in add_system_zone()
70 n = &(*n)->rb_left; in add_system_zone()
72 n = &(*n)->rb_right; in add_system_zone()
86 rb_link_node(new_node, parent, n); in add_system_zone()
138 int err = 0, n; in ext4_protect_reserved_inode() local
151 n = ext4_map_blocks(NULL, inode, &map, 0); in ext4_protect_reserved_inode()
152 if (n < 0) { in ext4_protect_reserved_inode()
153 err = n; in ext4_protect_reserved_inode()
[all …]
/fs/hpfs/
Dalloc.c114 static secno alloc_in_bmp(struct super_block *s, secno near, unsigned n, unsigned forward) in alloc_in_bmp() argument
119 unsigned nr = (near & 0x3fff) & ~(n - 1); in alloc_in_bmp()
124 if (n != 1 && n != 4) { in alloc_in_bmp()
125 hpfs_error(s, "Bad allocation size: %d", n); in alloc_in_bmp()
133 if (!tstbits(bmp, nr, n + forward)) { in alloc_in_bmp()
137 q = nr + n; b = 0; in alloc_in_bmp()
138 while ((a = tstbits(bmp, q, n + forward)) != 0) { in alloc_in_bmp()
140 if (n != 1) q = ((q-1)&~(n-1))+n; in alloc_in_bmp()
157 if (n + forward >= 0x3f && le32_to_cpu(bmp[i]) != 0xffffffff) goto cont; in alloc_in_bmp()
165 if (n != 1) q = ((q-1)&~(n-1))+n; in alloc_in_bmp()
[all …]
/fs/
Dselect.c337 #define FDS_IN(fds, n) (fds->in + n) argument
338 #define FDS_OUT(fds, n) (fds->out + n) argument
339 #define FDS_EX(fds, n) (fds->ex + n) argument
341 #define BITS(fds, n) (*FDS_IN(fds, n)|*FDS_OUT(fds, n)|*FDS_EX(fds, n)) argument
343 static int max_select_fd(unsigned long n, fd_set_bits *fds) in max_select_fd() argument
351 set = ~(~0UL << (n & (BITS_PER_LONG-1))); in max_select_fd()
352 n /= BITS_PER_LONG; in max_select_fd()
354 open_fds = fdt->open_fds + n; in max_select_fd()
357 set &= BITS(fds, n); in max_select_fd()
364 while (n) { in max_select_fd()
[all …]
Dchar_dev.c184 dev_t n, next; in register_chrdev_region() local
186 for (n = from; n < to; n = next) { in register_chrdev_region()
187 next = MKDEV(MAJOR(n)+1, 0); in register_chrdev_region()
190 cd = __register_chrdev_region(MAJOR(n), MINOR(n), in register_chrdev_region()
191 next - n, name); in register_chrdev_region()
197 to = n; in register_chrdev_region()
198 for (n = from; n < to; n = next) { in register_chrdev_region()
199 next = MKDEV(MAJOR(n)+1, 0); in register_chrdev_region()
200 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); in register_chrdev_region()
294 dev_t n, next; in unregister_chrdev_region() local
[all …]
/fs/ceph/
Dmdsmap.c21 int n = 0; in ceph_mdsmap_get_random_mds() local
31 n++; in ceph_mdsmap_get_random_mds()
32 if (n == 0) in ceph_mdsmap_get_random_mds()
36 n = prandom_u32() % n; in ceph_mdsmap_get_random_mds()
38 for (i = 0; n > 0; i++, n--) in ceph_mdsmap_get_random_mds()
55 int i, j, n; in ceph_mdsmap_decode() local
84 n = ceph_decode_32(p); in ceph_mdsmap_decode()
85 for (i = 0; i < n; i++) { in ceph_mdsmap_decode()
127 i+1, n, global_id, mds, inc, in ceph_mdsmap_decode()
155 ceph_decode_32_safe(p, end, n, bad); in ceph_mdsmap_decode()
[all …]
/fs/cifs/
Dcifs_unicode.h227 UniStrncat(wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncat() argument
233 while (n-- && (*ucs1 = *ucs2)) { /* copy s2 after s1 */ in UniStrncat()
245 UniStrncmp(const wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncmp() argument
247 if (!n) in UniStrncmp()
249 while ((*ucs1 == *ucs2) && *ucs1 && --n) { in UniStrncmp()
260 UniStrncmp_le(const wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncmp_le() argument
262 if (!n) in UniStrncmp_le()
264 while ((*ucs1 == __le16_to_cpu(*ucs2)) && *ucs1 && --n) { in UniStrncmp_le()
275 UniStrncpy(wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncpy() argument
279 while (n-- && *ucs2) /* Copy the strings */ in UniStrncpy()
[all …]
/fs/hfs/
Dbitmap.c32 u32 mask, start, len, n; in hfs_find_set_zero_bits() local
46 n = be32_to_cpu(val); in hfs_find_set_zero_bits()
50 if (!(n & mask)) in hfs_find_set_zero_bits()
59 n = be32_to_cpu(val); in hfs_find_set_zero_bits()
62 if (!(n & mask)) in hfs_find_set_zero_bits()
76 n |= mask; in hfs_find_set_zero_bits()
80 if (!--len || n & mask) in hfs_find_set_zero_bits()
85 *curr++ = cpu_to_be32(n); in hfs_find_set_zero_bits()
88 n = be32_to_cpu(*curr); in hfs_find_set_zero_bits()
91 if (n) { in hfs_find_set_zero_bits()
[all …]
/fs/nls/
Dnls_euc-jp.c353 int n; in euc2sjisibm() local
366 if ((n = euc2sjisibm_g3upper(sjis, euc_hi, euc_lo))) { in euc2sjisibm()
367 return n; in euc2sjisibm()
368 } else if ((n = euc2sjisibm_jisx0212(sjis, euc_hi, euc_lo))) { in euc2sjisibm()
369 return n; in euc2sjisibm()
412 int n; in uni2char() local
416 if ((n = p_nls->uni2char(uni, out, boundlen)) < 0) in uni2char()
417 return n; in uni2char()
420 if (n == 1) { in uni2char()
430 } else if (n == 2) { in uni2char()
[all …]
Dnls_utf8.c16 int n; in uni2char() local
21 n = utf32_to_utf8(uni, out, boundlen); in uni2char()
22 if (n < 0) { in uni2char()
26 return n; in uni2char()
31 int n; in char2uni() local
34 n = utf8_to_utf32(rawstring, boundlen, &u); in char2uni()
35 if (n < 0 || u > MAX_WCHAR_T) { in char2uni()
40 return n; in char2uni()
/fs/ext2/
Dacl.c22 int n, count; in ext2_acl_from_disk() local
41 for (n=0; n < count; n++) { in ext2_acl_from_disk()
46 acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); in ext2_acl_from_disk()
47 acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); in ext2_acl_from_disk()
48 switch(acl->a_entries[n].e_tag) { in ext2_acl_from_disk()
61 acl->a_entries[n].e_uid = in ext2_acl_from_disk()
69 acl->a_entries[n].e_gid = in ext2_acl_from_disk()
95 size_t n; in ext2_acl_to_disk() local
104 for (n=0; n < acl->a_count; n++) { in ext2_acl_to_disk()
105 const struct posix_acl_entry *acl_e = &acl->a_entries[n]; in ext2_acl_to_disk()
/fs/reiserfs/
Dxattr_acl.c64 int n, count; in reiserfs_posix_acl_from_disk() local
83 for (n = 0; n < count; n++) { in reiserfs_posix_acl_from_disk()
87 acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); in reiserfs_posix_acl_from_disk()
88 acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); in reiserfs_posix_acl_from_disk()
89 switch (acl->a_entries[n].e_tag) { in reiserfs_posix_acl_from_disk()
102 acl->a_entries[n].e_uid = in reiserfs_posix_acl_from_disk()
110 acl->a_entries[n].e_gid = in reiserfs_posix_acl_from_disk()
135 int n; in reiserfs_posix_acl_to_disk() local
146 for (n = 0; n < acl->a_count; n++) { in reiserfs_posix_acl_to_disk()
147 const struct posix_acl_entry *acl_e = &acl->a_entries[n]; in reiserfs_posix_acl_to_disk()
[all …]
/fs/ocfs2/
Dacl.c45 int n, count; in ocfs2_acl_from_xattr() local
58 for (n = 0; n < count; n++) { in ocfs2_acl_from_xattr()
62 acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); in ocfs2_acl_from_xattr()
63 acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); in ocfs2_acl_from_xattr()
64 switch(acl->a_entries[n].e_tag) { in ocfs2_acl_from_xattr()
66 acl->a_entries[n].e_uid = in ocfs2_acl_from_xattr()
71 acl->a_entries[n].e_gid = in ocfs2_acl_from_xattr()
91 size_t n; in ocfs2_acl_to_xattr() local
100 for (n = 0; n < acl->a_count; n++, entry++) { in ocfs2_acl_to_xattr()
101 entry->e_tag = cpu_to_le16(acl->a_entries[n].e_tag); in ocfs2_acl_to_xattr()
[all …]

1234567891011