/fs/befs/ |
D | endian.h | 15 fs64_to_cpu(const struct super_block *sb, fs64 n) in fs64_to_cpu() argument 18 return le64_to_cpu((__force __le64)n); in fs64_to_cpu() 20 return be64_to_cpu((__force __be64)n); in fs64_to_cpu() 24 cpu_to_fs64(const struct super_block *sb, u64 n) in cpu_to_fs64() argument 27 return (__force fs64)cpu_to_le64(n); in cpu_to_fs64() 29 return (__force fs64)cpu_to_be64(n); in cpu_to_fs64() 33 fs32_to_cpu(const struct super_block *sb, fs32 n) in fs32_to_cpu() argument 36 return le32_to_cpu((__force __le32)n); in fs32_to_cpu() 38 return be32_to_cpu((__force __be32)n); in fs32_to_cpu() 42 cpu_to_fs32(const struct super_block *sb, u32 n) in cpu_to_fs32() argument [all …]
|
/fs/ufs/ |
D | swab.h | 26 fs64_to_cpu(struct super_block *sbp, __fs64 n) in fs64_to_cpu() argument 29 return le64_to_cpu((__force __le64)n); in fs64_to_cpu() 31 return be64_to_cpu((__force __be64)n); in fs64_to_cpu() 35 cpu_to_fs64(struct super_block *sbp, u64 n) in cpu_to_fs64() argument 38 return (__force __fs64)cpu_to_le64(n); in cpu_to_fs64() 40 return (__force __fs64)cpu_to_be64(n); in cpu_to_fs64() 44 fs32_to_cpu(struct super_block *sbp, __fs32 n) in fs32_to_cpu() argument 47 return le32_to_cpu((__force __le32)n); in fs32_to_cpu() 49 return be32_to_cpu((__force __be32)n); in fs32_to_cpu() 53 cpu_to_fs32(struct super_block *sbp, u32 n) in cpu_to_fs32() argument [all …]
|
/fs/ubifs/ |
D | tnc.c | 186 struct ubifs_old_idx *old_idx, *n; in destroy_old_idx() local 188 rbtree_postorder_for_each_entry_safe(old_idx, n, &c->old_idx, rb) in destroy_old_idx() 219 const int n = zn->child_cnt; in copy_znode() local 222 for (i = 0; i < n; i++) { in copy_znode() 589 struct ubifs_znode *znode, int n) in get_znode() argument 593 zbr = &znode->zbranch[n]; in get_znode() 597 znode = ubifs_load_znode(c, zbr, znode, n); in get_znode() 610 static int tnc_next(struct ubifs_info *c, struct ubifs_znode **zn, int *n) in tnc_next() argument 613 int nn = *n; in tnc_next() 617 *n = nn; in tnc_next() [all …]
|
/fs/fscache/ |
D | histogram.c | 30 unsigned n[5], t; in fscache_histogram_show() local 41 n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]); in fscache_histogram_show() 42 n[1] = atomic_read(&fscache_ops_histogram[index]); in fscache_histogram_show() 43 n[2] = atomic_read(&fscache_objs_histogram[index]); in fscache_histogram_show() 44 n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]); in fscache_histogram_show() 45 n[4] = atomic_read(&fscache_retrieval_histogram[index]); in fscache_histogram_show() 46 if (!(n[0] | n[1] | n[2] | n[3] | n[4])) in fscache_histogram_show() 52 index, t, n[0], n[1], n[2], n[3], n[4]); in fscache_histogram_show()
|
/fs/sysv/ |
D | sysv.h | 190 static inline __u32 fs32_to_cpu(struct sysv_sb_info *sbi, __fs32 n) in fs32_to_cpu() argument 193 return PDP_swab((__force __u32)n); in fs32_to_cpu() 195 return le32_to_cpu((__force __le32)n); in fs32_to_cpu() 197 return be32_to_cpu((__force __be32)n); in fs32_to_cpu() 200 static inline __fs32 cpu_to_fs32(struct sysv_sb_info *sbi, __u32 n) in cpu_to_fs32() argument 203 return (__force __fs32)PDP_swab(n); in cpu_to_fs32() 205 return (__force __fs32)cpu_to_le32(n); in cpu_to_fs32() 207 return (__force __fs32)cpu_to_be32(n); in cpu_to_fs32() 210 static inline __fs32 fs32_add(struct sysv_sb_info *sbi, __fs32 *n, int d) in fs32_add() argument 213 *(__u32*)n = PDP_swab(PDP_swab(*(__u32*)n)+d); in fs32_add() [all …]
|
D | itree.c | 29 int n = 0; in block_to_path() local 34 offsets[n++] = block; in block_to_path() 36 offsets[n++] = DIRECT; in block_to_path() 37 offsets[n++] = block; in block_to_path() 39 offsets[n++] = DIRECT+1; in block_to_path() 40 offsets[n++] = block >> ptrs_bits; in block_to_path() 41 offsets[n++] = block & (indirect_blocks - 1); in block_to_path() 43 offsets[n++] = DIRECT+2; in block_to_path() 44 offsets[n++] = block >> (ptrs_bits * 2); in block_to_path() 45 offsets[n++] = (block >> ptrs_bits) & (indirect_blocks - 1); in block_to_path() [all …]
|
/fs/minix/ |
D | itree_v2.c | 8 static inline unsigned long block_to_cpu(block_t n) in block_to_cpu() argument 10 return n; in block_to_cpu() 13 static inline block_t cpu_to_block(unsigned long n) in cpu_to_block() argument 15 return n; in cpu_to_block() 28 int n = 0; in block_to_path() local 42 offsets[n++] = block; in block_to_path() 44 offsets[n++] = DIRCOUNT; in block_to_path() 45 offsets[n++] = block; in block_to_path() 47 offsets[n++] = DIRCOUNT + 1; in block_to_path() 48 offsets[n++] = block / INDIRCOUNT(sb); in block_to_path() [all …]
|
D | itree_v1.c | 9 static inline unsigned long block_to_cpu(block_t n) in block_to_cpu() argument 11 return n; in block_to_cpu() 14 static inline block_t cpu_to_block(unsigned long n) in cpu_to_block() argument 16 return n; in cpu_to_block() 26 int n = 0; in block_to_path() local 38 offsets[n++] = block; in block_to_path() 40 offsets[n++] = 7; in block_to_path() 41 offsets[n++] = block; in block_to_path() 44 offsets[n++] = 8; in block_to_path() 45 offsets[n++] = block>>9; in block_to_path() [all …]
|
D | itree_common.c | 74 int n = 0; in alloc_branch() local 80 if (parent) for (n = 1; n < num; n++) { in alloc_branch() 86 branch[n].key = cpu_to_block(nr); in alloc_branch() 95 branch[n].bh = bh; in alloc_branch() 96 branch[n].p = (block_t*) bh->b_data + offsets[n]; in alloc_branch() 97 *branch[n].p = branch[n].key; in alloc_branch() 103 if (n == num) in alloc_branch() 107 for (i = 1; i < n; i++) in alloc_branch() 109 for (i = 0; i < n; i++) in alloc_branch() 307 int n; in truncate() local [all …]
|
/fs/qnx6/ |
D | qnx6.h | 77 static inline __u64 fs64_to_cpu(struct qnx6_sb_info *sbi, __fs64 n) in fs64_to_cpu() argument 80 return le64_to_cpu((__force __le64)n); in fs64_to_cpu() 82 return be64_to_cpu((__force __be64)n); in fs64_to_cpu() 85 static inline __fs64 cpu_to_fs64(struct qnx6_sb_info *sbi, __u64 n) in cpu_to_fs64() argument 88 return (__force __fs64)cpu_to_le64(n); in cpu_to_fs64() 90 return (__force __fs64)cpu_to_be64(n); in cpu_to_fs64() 93 static inline __u32 fs32_to_cpu(struct qnx6_sb_info *sbi, __fs32 n) in fs32_to_cpu() argument 96 return le32_to_cpu((__force __le32)n); in fs32_to_cpu() 98 return be32_to_cpu((__force __be32)n); in fs32_to_cpu() 101 static inline __fs32 cpu_to_fs32(struct qnx6_sb_info *sbi, __u32 n) in cpu_to_fs32() argument [all …]
|
D | dir.c | 26 static struct page *qnx6_get_page(struct inode *dir, unsigned long n) in qnx6_get_page() argument 29 struct page *page = read_mapping_page(mapping, n, NULL); in qnx6_get_page() 50 u32 n = s >> (PAGE_CACHE_SHIFT - sb->s_blocksize_bits); /* in pages */ in qnx6_longname() local 54 struct page *page = read_mapping_page(mapping, n, NULL); in qnx6_longname() 118 unsigned long n = pos >> PAGE_CACHE_SHIFT; in qnx6_readdir() local 126 for ( ; !done && n < npages; n++, start = 0) { in qnx6_readdir() 127 struct page *page = qnx6_get_page(inode, n); in qnx6_readdir() 128 int limit = last_entry(inode, n); in qnx6_readdir() 134 ctx->pos = (n + 1) << PAGE_CACHE_SHIFT; in qnx6_readdir() 218 unsigned long start, n; in qnx6_find_entry() local [all …]
|
/fs/jfs/ |
D | jfs_unicode.h | 56 size_t n) in UniStrncpy_le() argument 60 while (n-- && *ucs2) /* Copy the strings */ in UniStrncpy_le() 63 n++; in UniStrncpy_le() 64 while (n--) /* Pad with nulls */ in UniStrncpy_le() 73 size_t n) in UniStrncmp_le() argument 75 if (!n) in UniStrncmp_le() 77 while ((*ucs1 == __le16_to_cpu(*ucs2)) && *ucs1 && --n) { in UniStrncmp_le() 88 size_t n) in UniStrncpy_to_le() argument 92 while (n-- && *ucs2) /* Copy the strings */ in UniStrncpy_to_le() 95 n++; in UniStrncpy_to_le() [all …]
|
/fs/xfs/libxfs/ |
D | xfs_bit.h | 28 static inline __uint64_t xfs_mask64hi(int n) in xfs_mask64hi() argument 30 return (__uint64_t)-1 << (64 - (n)); in xfs_mask64hi() 32 static inline __uint32_t xfs_mask32lo(int n) in xfs_mask32lo() argument 34 return ((__uint32_t)1 << (n)) - 1; in xfs_mask32lo() 36 static inline __uint64_t xfs_mask64lo(int n) in xfs_mask64lo() argument 38 return ((__uint64_t)1 << (n)) - 1; in xfs_mask64lo() 63 int n = 0; in xfs_lowbit64() local 66 n = ffs(w); in xfs_lowbit64() 70 n = ffs(w); in xfs_lowbit64() 71 if (n) in xfs_lowbit64() [all …]
|
/fs/ext4/ |
D | block_validity.c | 63 struct rb_node **n = &sbi->system_blks.rb_node, *node; in add_system_zone() local 66 while (*n) { in add_system_zone() 67 parent = *n; in add_system_zone() 70 n = &(*n)->rb_left; in add_system_zone() 72 n = &(*n)->rb_right; in add_system_zone() 86 rb_link_node(new_node, parent, n); in add_system_zone() 138 int err = 0, n; in ext4_protect_reserved_inode() local 151 n = ext4_map_blocks(NULL, inode, &map, 0); in ext4_protect_reserved_inode() 152 if (n < 0) { in ext4_protect_reserved_inode() 153 err = n; in ext4_protect_reserved_inode() [all …]
|
/fs/hpfs/ |
D | alloc.c | 114 static secno alloc_in_bmp(struct super_block *s, secno near, unsigned n, unsigned forward) in alloc_in_bmp() argument 119 unsigned nr = (near & 0x3fff) & ~(n - 1); in alloc_in_bmp() 124 if (n != 1 && n != 4) { in alloc_in_bmp() 125 hpfs_error(s, "Bad allocation size: %d", n); in alloc_in_bmp() 133 if (!tstbits(bmp, nr, n + forward)) { in alloc_in_bmp() 137 q = nr + n; b = 0; in alloc_in_bmp() 138 while ((a = tstbits(bmp, q, n + forward)) != 0) { in alloc_in_bmp() 140 if (n != 1) q = ((q-1)&~(n-1))+n; in alloc_in_bmp() 157 if (n + forward >= 0x3f && le32_to_cpu(bmp[i]) != 0xffffffff) goto cont; in alloc_in_bmp() 165 if (n != 1) q = ((q-1)&~(n-1))+n; in alloc_in_bmp() [all …]
|
/fs/ |
D | select.c | 337 #define FDS_IN(fds, n) (fds->in + n) argument 338 #define FDS_OUT(fds, n) (fds->out + n) argument 339 #define FDS_EX(fds, n) (fds->ex + n) argument 341 #define BITS(fds, n) (*FDS_IN(fds, n)|*FDS_OUT(fds, n)|*FDS_EX(fds, n)) argument 343 static int max_select_fd(unsigned long n, fd_set_bits *fds) in max_select_fd() argument 351 set = ~(~0UL << (n & (BITS_PER_LONG-1))); in max_select_fd() 352 n /= BITS_PER_LONG; in max_select_fd() 354 open_fds = fdt->open_fds + n; in max_select_fd() 357 set &= BITS(fds, n); in max_select_fd() 364 while (n) { in max_select_fd() [all …]
|
D | char_dev.c | 184 dev_t n, next; in register_chrdev_region() local 186 for (n = from; n < to; n = next) { in register_chrdev_region() 187 next = MKDEV(MAJOR(n)+1, 0); in register_chrdev_region() 190 cd = __register_chrdev_region(MAJOR(n), MINOR(n), in register_chrdev_region() 191 next - n, name); in register_chrdev_region() 197 to = n; in register_chrdev_region() 198 for (n = from; n < to; n = next) { in register_chrdev_region() 199 next = MKDEV(MAJOR(n)+1, 0); in register_chrdev_region() 200 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); in register_chrdev_region() 294 dev_t n, next; in unregister_chrdev_region() local [all …]
|
/fs/ceph/ |
D | mdsmap.c | 21 int n = 0; in ceph_mdsmap_get_random_mds() local 31 n++; in ceph_mdsmap_get_random_mds() 32 if (n == 0) in ceph_mdsmap_get_random_mds() 36 n = prandom_u32() % n; in ceph_mdsmap_get_random_mds() 38 for (i = 0; n > 0; i++, n--) in ceph_mdsmap_get_random_mds() 55 int i, j, n; in ceph_mdsmap_decode() local 84 n = ceph_decode_32(p); in ceph_mdsmap_decode() 85 for (i = 0; i < n; i++) { in ceph_mdsmap_decode() 127 i+1, n, global_id, mds, inc, in ceph_mdsmap_decode() 155 ceph_decode_32_safe(p, end, n, bad); in ceph_mdsmap_decode() [all …]
|
/fs/cifs/ |
D | cifs_unicode.h | 227 UniStrncat(wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncat() argument 233 while (n-- && (*ucs1 = *ucs2)) { /* copy s2 after s1 */ in UniStrncat() 245 UniStrncmp(const wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncmp() argument 247 if (!n) in UniStrncmp() 249 while ((*ucs1 == *ucs2) && *ucs1 && --n) { in UniStrncmp() 260 UniStrncmp_le(const wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncmp_le() argument 262 if (!n) in UniStrncmp_le() 264 while ((*ucs1 == __le16_to_cpu(*ucs2)) && *ucs1 && --n) { in UniStrncmp_le() 275 UniStrncpy(wchar_t *ucs1, const wchar_t *ucs2, size_t n) in UniStrncpy() argument 279 while (n-- && *ucs2) /* Copy the strings */ in UniStrncpy() [all …]
|
/fs/hfs/ |
D | bitmap.c | 32 u32 mask, start, len, n; in hfs_find_set_zero_bits() local 46 n = be32_to_cpu(val); in hfs_find_set_zero_bits() 50 if (!(n & mask)) in hfs_find_set_zero_bits() 59 n = be32_to_cpu(val); in hfs_find_set_zero_bits() 62 if (!(n & mask)) in hfs_find_set_zero_bits() 76 n |= mask; in hfs_find_set_zero_bits() 80 if (!--len || n & mask) in hfs_find_set_zero_bits() 85 *curr++ = cpu_to_be32(n); in hfs_find_set_zero_bits() 88 n = be32_to_cpu(*curr); in hfs_find_set_zero_bits() 91 if (n) { in hfs_find_set_zero_bits() [all …]
|
/fs/nls/ |
D | nls_euc-jp.c | 353 int n; in euc2sjisibm() local 366 if ((n = euc2sjisibm_g3upper(sjis, euc_hi, euc_lo))) { in euc2sjisibm() 367 return n; in euc2sjisibm() 368 } else if ((n = euc2sjisibm_jisx0212(sjis, euc_hi, euc_lo))) { in euc2sjisibm() 369 return n; in euc2sjisibm() 412 int n; in uni2char() local 416 if ((n = p_nls->uni2char(uni, out, boundlen)) < 0) in uni2char() 417 return n; in uni2char() 420 if (n == 1) { in uni2char() 430 } else if (n == 2) { in uni2char() [all …]
|
D | nls_utf8.c | 16 int n; in uni2char() local 21 n = utf32_to_utf8(uni, out, boundlen); in uni2char() 22 if (n < 0) { in uni2char() 26 return n; in uni2char() 31 int n; in char2uni() local 34 n = utf8_to_utf32(rawstring, boundlen, &u); in char2uni() 35 if (n < 0 || u > MAX_WCHAR_T) { in char2uni() 40 return n; in char2uni()
|
/fs/ext2/ |
D | acl.c | 22 int n, count; in ext2_acl_from_disk() local 41 for (n=0; n < count; n++) { in ext2_acl_from_disk() 46 acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); in ext2_acl_from_disk() 47 acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); in ext2_acl_from_disk() 48 switch(acl->a_entries[n].e_tag) { in ext2_acl_from_disk() 61 acl->a_entries[n].e_uid = in ext2_acl_from_disk() 69 acl->a_entries[n].e_gid = in ext2_acl_from_disk() 95 size_t n; in ext2_acl_to_disk() local 104 for (n=0; n < acl->a_count; n++) { in ext2_acl_to_disk() 105 const struct posix_acl_entry *acl_e = &acl->a_entries[n]; in ext2_acl_to_disk()
|
/fs/reiserfs/ |
D | xattr_acl.c | 64 int n, count; in reiserfs_posix_acl_from_disk() local 83 for (n = 0; n < count; n++) { in reiserfs_posix_acl_from_disk() 87 acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); in reiserfs_posix_acl_from_disk() 88 acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); in reiserfs_posix_acl_from_disk() 89 switch (acl->a_entries[n].e_tag) { in reiserfs_posix_acl_from_disk() 102 acl->a_entries[n].e_uid = in reiserfs_posix_acl_from_disk() 110 acl->a_entries[n].e_gid = in reiserfs_posix_acl_from_disk() 135 int n; in reiserfs_posix_acl_to_disk() local 146 for (n = 0; n < acl->a_count; n++) { in reiserfs_posix_acl_to_disk() 147 const struct posix_acl_entry *acl_e = &acl->a_entries[n]; in reiserfs_posix_acl_to_disk() [all …]
|
/fs/ocfs2/ |
D | acl.c | 45 int n, count; in ocfs2_acl_from_xattr() local 58 for (n = 0; n < count; n++) { in ocfs2_acl_from_xattr() 62 acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); in ocfs2_acl_from_xattr() 63 acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); in ocfs2_acl_from_xattr() 64 switch(acl->a_entries[n].e_tag) { in ocfs2_acl_from_xattr() 66 acl->a_entries[n].e_uid = in ocfs2_acl_from_xattr() 71 acl->a_entries[n].e_gid = in ocfs2_acl_from_xattr() 91 size_t n; in ocfs2_acl_to_xattr() local 100 for (n = 0; n < acl->a_count; n++, entry++) { in ocfs2_acl_to_xattr() 101 entry->e_tag = cpu_to_le16(acl->a_entries[n].e_tag); in ocfs2_acl_to_xattr() [all …]
|