/fs/xfs/support/ |
D | ktrace.c | 155 int index; in ktrace_enter() local 163 index = atomic_add_return(1, &ktp->kt_index); in ktrace_enter() 164 index = (index - 1) & ktp->kt_index_mask; in ktrace_enter() 165 if (!ktp->kt_rollover && index == ktp->kt_nentries - 1) in ktrace_enter() 168 ASSERT((index >= 0) && (index < ktp->kt_nentries)); in ktrace_enter() 170 ktep = &(ktp->kt_entries[index]); in ktrace_enter() 197 int index; in ktrace_nentries() local 201 index = atomic_read(&ktp->kt_index) & ktp->kt_index_mask; in ktrace_nentries() 202 return (ktp->kt_rollover ? ktp->kt_nentries : index); in ktrace_nentries() 223 int index; in ktrace_first() local [all …]
|
/fs/jfs/ |
D | jfs_dtree.c | 116 s16 index; member 184 static void dtInsertEntry(dtpage_t * p, int index, struct component_name * key, 248 static struct dir_table_slot *find_index(struct inode *ip, u32 index, in find_index() argument 258 if (index < 2) { in find_index() 260 jfs_warn("find_entry called with index = %d", index); in find_index() 266 if (index >= jfs_ip->next_index) { in find_index() 276 slot = &jfs_ip->i_dirtable[index - 2]; in find_index() 278 offset = (index - 2) * sizeof(struct dir_table_slot); in find_index() 304 u32 index) in lock_index() argument 313 if (llck->index >= llck->maxcnt) in lock_index() [all …]
|
D | jfs_xtree.c | 92 s16 index; member 154 int index; in xtLookup() local 190 XT_GETSEARCH(ip, btstack.top, bn, mp, p, index); in xtLookup() 205 xad = &p->xad[index]; in xtLookup() 256 int index; in xtLookupList() local 296 XT_GETSEARCH(ip, btstack.top, bn, mp, p, index); in xtLookupList() 299 if (index == le16_to_cpu(p->header.nextindex)) { in xtLookupList() 313 index = XTENTRYSTART; in xtLookupList() 316 xad = &p->xad[index]; in xtLookupList() 382 if (index == le16_to_cpu(p->header.nextindex) - 1) { in xtLookupList() [all …]
|
D | ioctl.c | 38 int index=0; in jfs_map_ext2() local 41 while (jfs_map[index].jfs_flag) { in jfs_map_ext2() 43 if (jfs_map[index].ext2_flag & flags) in jfs_map_ext2() 44 mapped |= jfs_map[index].jfs_flag; in jfs_map_ext2() 46 if (jfs_map[index].jfs_flag & flags) in jfs_map_ext2() 47 mapped |= jfs_map[index].ext2_flag; in jfs_map_ext2() 49 index++; in jfs_map_ext2()
|
/fs/nls/ |
D | nls_euc-jp.c | 287 int index; in sjisibm2euc() local 289 index = ((sjis_hi - 0xFA) * (0xFD - 0x40)) + (sjis_lo - 0x40); in sjisibm2euc() 290 if (IS_EUC_IBM2JISX0208(sjisibm2euc_map[index][0], in sjisibm2euc() 291 sjisibm2euc_map[index][1])) { in sjisibm2euc() 292 euc[0] = sjisibm2euc_map[index][0]; in sjisibm2euc() 293 euc[1] = sjisibm2euc_map[index][1]; in sjisibm2euc() 297 euc[1] = sjisibm2euc_map[index][0]; in sjisibm2euc() 298 euc[2] = sjisibm2euc_map[index][1]; in sjisibm2euc() 307 int index, min_index, max_index; in euc2sjisibm_jisx0212() local 315 index = (min_index + max_index) / 2; in euc2sjisibm_jisx0212() [all …]
|
/fs/udf/ |
D | unicode.c | 390 int index, newIndex = 0, needsCRC = 0; in udf_translate_to_linux() local 402 for (index = 0; index < udfLen; index++) { in udf_translate_to_linux() 403 curr = udfName[index]; in udf_translate_to_linux() 407 while (index + 1 < udfLen && in udf_translate_to_linux() 408 (udfName[index + 1] == '/' || in udf_translate_to_linux() 409 udfName[index + 1] == 0)) in udf_translate_to_linux() 410 index++; in udf_translate_to_linux() 413 (udfLen - index - 1) <= EXT_SIZE) { in udf_translate_to_linux() 414 if (udfLen == index + 1) in udf_translate_to_linux() 418 extIndex = index; in udf_translate_to_linux() [all …]
|
/fs/squashfs/ |
D | namei.c | 82 struct squashfs_dir_index *index; in get_dir_index_using_name() local 87 index = kmalloc(sizeof(*index) + SQUASHFS_NAME_LEN * 2 + 2, GFP_KERNEL); in get_dir_index_using_name() 88 if (index == NULL) { in get_dir_index_using_name() 93 str = &index->name[SQUASHFS_NAME_LEN + 1]; in get_dir_index_using_name() 98 err = squashfs_read_metadata(sb, index, &index_start, in get_dir_index_using_name() 99 &index_offset, sizeof(*index)); in get_dir_index_using_name() 104 size = le32_to_cpu(index->size) + 1; in get_dir_index_using_name() 106 err = squashfs_read_metadata(sb, index->name, &index_start, in get_dir_index_using_name() 111 index->name[size] = '\0'; in get_dir_index_using_name() 113 if (strcmp(index->name, str) > 0) in get_dir_index_using_name() [all …]
|
D | file.c | 62 int index) in locate_meta_index() argument 70 TRACE("locate_meta_index: index %d, offset %d\n", index, offset); in locate_meta_index() 78 msblk->meta_index[i].offset <= index && in locate_meta_index() 237 static int fill_meta_index(struct inode *inode, int index, in fill_meta_index() argument 253 index /= SQUASHFS_META_INDEXES * skip; in fill_meta_index() 255 while (offset < index) { in fill_meta_index() 256 meta = locate_meta_index(inode, offset + 1, index); in fill_meta_index() 263 offset = index < meta->offset + meta->entries ? index : in fill_meta_index() 283 for (i = meta->offset + meta->entries; i <= index && in fill_meta_index() 336 static int read_blocklist(struct inode *inode, int index, u64 *block) in read_blocklist() argument [all …]
|
D | block.c | 82 int squashfs_read_data(struct super_block *sb, void **buffer, u64 index, in squashfs_read_data() argument 87 int offset = index & ((1 << msblk->devblksize_log2) - 1); in squashfs_read_data() 88 u64 cur_index = index >> msblk->devblksize_log2; in squashfs_read_data() 105 *next_index = index + length; in squashfs_read_data() 108 index, compressed ? "" : "un", length, srclength); in squashfs_read_data() 111 (index + length) > msblk->bytes_used) in squashfs_read_data() 125 if ((index + 2) > msblk->bytes_used) in squashfs_read_data() 137 *next_index = index + length + 2; in squashfs_read_data() 139 TRACE("Block @ 0x%llx, %scompressed size %d\n", index, in squashfs_read_data() 143 (index + length) > msblk->bytes_used) in squashfs_read_data() [all …]
|
D | symlink.c | 51 int index = page->index << PAGE_CACHE_SHIFT; in squashfs_symlink_readpage() local 54 int length = min_t(int, i_size_read(inode) - index, PAGE_CACHE_SIZE); in squashfs_symlink_readpage() 60 "%llx, offset %x\n", page->index, block, offset); in squashfs_symlink_readpage() 65 if (index) { in squashfs_symlink_readpage() 67 index); in squashfs_symlink_readpage()
|
/fs/ntfs/ |
D | dir.c | 127 index_end = (u8*)&ir->index + le32_to_cpu(ir->index.index_length); in ntfs_lookup_inode_by_name() 129 ie = (INDEX_ENTRY*)((u8*)&ir->index + in ntfs_lookup_inode_by_name() 130 le32_to_cpu(ir->index.entries_offset)); in ntfs_lookup_inode_by_name() 321 dir_ni->itype.index.vcn_size_bits >> PAGE_CACHE_SHIFT); in ntfs_lookup_inode_by_name() 333 dir_ni->itype.index.vcn_size_bits) & ~PAGE_CACHE_MASK)); in ntfs_lookup_inode_by_name() 356 if (le32_to_cpu(ia->index.allocated_size) + 0x18 != in ntfs_lookup_inode_by_name() 357 dir_ni->itype.index.block_size) { in ntfs_lookup_inode_by_name() 363 le32_to_cpu(ia->index.allocated_size) + 0x18, in ntfs_lookup_inode_by_name() 364 dir_ni->itype.index.block_size); in ntfs_lookup_inode_by_name() 367 index_end = (u8*)ia + dir_ni->itype.index.block_size; in ntfs_lookup_inode_by_name() [all …]
|
D | index.c | 143 idx_ni->itype.index.collation_rule)) { in ntfs_index_lookup() 146 idx_ni->itype.index.collation_rule)); in ntfs_index_lookup() 175 index_end = (u8*)&ir->index + le32_to_cpu(ir->index.index_length); in ntfs_index_lookup() 177 ie = (INDEX_ENTRY*)((u8*)&ir->index + in ntfs_index_lookup() 178 le32_to_cpu(ir->index.entries_offset)); in ntfs_index_lookup() 225 rc = ntfs_collate(vol, idx_ni->itype.index.collation_rule, key, in ntfs_index_lookup() 277 idx_ni->itype.index.vcn_size_bits >> PAGE_CACHE_SHIFT); in ntfs_index_lookup() 289 idx_ni->itype.index.vcn_size_bits) & ~PAGE_CACHE_MASK)); in ntfs_index_lookup() 312 if (le32_to_cpu(ia->index.allocated_size) + 0x18 != in ntfs_index_lookup() 313 idx_ni->itype.index.block_size) { in ntfs_index_lookup() [all …]
|
/fs/xfs/ |
D | xfs_dir2_node.c | 47 static int xfs_dir2_leafn_add(xfs_dabuf_t *bp, xfs_da_args_t *args, int index); 60 int index, xfs_da_state_blk_t *dblk, 186 int index) /* insertion pt for new entry */ in xfs_dir2_leafn_add() argument 199 xfs_dir2_trace_args_sb("leafn_add", args, index, bp); in xfs_dir2_leafn_add() 209 if (index < 0) in xfs_dir2_leafn_add() 225 ASSERT(index == 0 || be32_to_cpu(leaf->ents[index - 1].hashval) <= args->hashval); in xfs_dir2_leafn_add() 226 ASSERT(index == be16_to_cpu(leaf->hdr.count) || in xfs_dir2_leafn_add() 227 be32_to_cpu(leaf->ents[index].hashval) >= args->hashval); in xfs_dir2_leafn_add() 237 xfs_dir2_leaf_compact_x1(bp, &index, &lowstale, &highstale, in xfs_dir2_leafn_add() 251 lep = &leaf->ents[index]; in xfs_dir2_leafn_add() [all …]
|
D | xfs_bmap_btree.h | 152 #define XFS_BMBT_REC_ADDR(mp, block, index) \ argument 156 ((index) - 1) * sizeof(xfs_bmbt_rec_t))) 158 #define XFS_BMBT_KEY_ADDR(mp, block, index) \ argument 162 ((index) - 1) * sizeof(xfs_bmbt_key_t))) 164 #define XFS_BMBT_PTR_ADDR(mp, block, index, maxrecs) \ argument 169 ((index) - 1) * sizeof(xfs_bmbt_ptr_t))) 171 #define XFS_BMDR_REC_ADDR(block, index) \ argument 175 ((index) - 1) * sizeof(xfs_bmdr_rec_t))) 177 #define XFS_BMDR_KEY_ADDR(block, index) \ argument 181 ((index) - 1) * sizeof(xfs_bmdr_key_t))) [all …]
|
D | xfs_dir2_leaf.c | 174 int index; /* leaf table position */ in xfs_dir2_leaf_addname() local 210 index = xfs_dir2_leaf_search_hash(args, lbp); in xfs_dir2_leaf_addname() 221 for (use_block = -1, lep = &leaf->ents[index]; in xfs_dir2_leaf_addname() 222 index < be16_to_cpu(leaf->hdr.count) && be32_to_cpu(lep->hashval) == args->hashval; in xfs_dir2_leaf_addname() 223 index++, lep++) { in xfs_dir2_leaf_addname() 324 xfs_dir2_leaf_compact_x1(lbp, &index, &lowstale, &highstale, in xfs_dir2_leaf_addname() 441 if (index < be16_to_cpu(leaf->hdr.count)) in xfs_dir2_leaf_addname() 443 (be16_to_cpu(leaf->hdr.count) - index) * sizeof(*lep)); in xfs_dir2_leaf_addname() 447 lfloglow = index; in xfs_dir2_leaf_addname() 467 for (lowstale = index - 1; in xfs_dir2_leaf_addname() [all …]
|
D | xfs_alloc_btree.h | 86 #define XFS_ALLOC_REC_ADDR(mp, block, index) \ argument 90 (((index) - 1) * sizeof(xfs_alloc_rec_t)))) 92 #define XFS_ALLOC_KEY_ADDR(mp, block, index) \ argument 96 ((index) - 1) * sizeof(xfs_alloc_key_t))) 98 #define XFS_ALLOC_PTR_ADDR(mp, block, index, maxrecs) \ argument 103 ((index) - 1) * sizeof(xfs_alloc_ptr_t)))
|
D | xfs_ialloc_btree.h | 103 #define XFS_INOBT_REC_ADDR(mp, block, index) \ argument 107 (((index) - 1) * sizeof(xfs_inobt_rec_t)))) 109 #define XFS_INOBT_KEY_ADDR(mp, block, index) \ argument 113 ((index) - 1) * sizeof(xfs_inobt_key_t))) 115 #define XFS_INOBT_PTR_ADDR(mp, block, index, maxrecs) \ argument 120 ((index) - 1) * sizeof(xfs_inobt_ptr_t)))
|
D | xfs_attr_leaf.c | 91 STATIC int xfs_attr_leaf_entsize(xfs_attr_leafblock_t *leaf, int index); 1010 ASSERT((args->index >= 0) in xfs_attr_leaf_add() 1011 && (args->index <= be16_to_cpu(leaf->hdr.count))); in xfs_attr_leaf_add() 1085 ASSERT((args->index >= 0) && (args->index <= be16_to_cpu(hdr->count))); in xfs_attr_leaf_add_work() 1090 entry = &leaf->entries[args->index]; in xfs_attr_leaf_add_work() 1091 if (args->index < be16_to_cpu(hdr->count)) { in xfs_attr_leaf_add_work() 1092 tmp = be16_to_cpu(hdr->count) - args->index; in xfs_attr_leaf_add_work() 1123 (args->index2 <= args->index)) { in xfs_attr_leaf_add_work() 1129 ASSERT((args->index == 0) || in xfs_attr_leaf_add_work() 1131 ASSERT((args->index == be16_to_cpu(hdr->count)-1) || in xfs_attr_leaf_add_work() [all …]
|
/fs/ecryptfs/ |
D | mmap.c | 46 struct page *ecryptfs_get_locked_page(struct file *file, loff_t index) in ecryptfs_get_locked_page() argument 56 page = read_mapping_page(mapping, index, (void *)file); in ecryptfs_get_locked_page() 75 "page (upper index [0x%.16x])\n", page->index); in ecryptfs_writepage() 134 loff_t view_extent_num = ((((loff_t)page->index) in ecryptfs_copy_up_encrypted_with_header() 204 rc = ecryptfs_read_lower_page_segment(page, page->index, 0, in ecryptfs_readpage() 222 page, page->index, 0, PAGE_CACHE_SIZE, in ecryptfs_readpage() 244 page->index); in ecryptfs_readpage() 257 if ((i_size_read(inode) / PAGE_CACHE_SIZE) != page->index) in fill_zeros_to_end_of_page() 286 pgoff_t index = pos >> PAGE_CACHE_SHIFT; in ecryptfs_write_begin() local 291 page = grab_cache_page_write_begin(mapping, index, flags); in ecryptfs_write_begin() [all …]
|
/fs/ocfs2/ |
D | uptodate.c | 188 int index = -1; in ocfs2_buffer_cached() local 199 index = ocfs2_search_cache_array(&oi->ip_metadata_cache, in ocfs2_buffer_cached() 207 mlog(0, "index = %d, item = %p\n", index, item); in ocfs2_buffer_cached() 209 return (index != -1) || (item != NULL); in ocfs2_buffer_cached() 485 int index) in ocfs2_remove_metadata_array() argument 490 BUG_ON(index < 0 || index >= OCFS2_INODE_MAX_CACHE_ARRAY); in ocfs2_remove_metadata_array() 491 BUG_ON(index >= ci->ci_num_cached); in ocfs2_remove_metadata_array() 494 mlog(0, "remove index %d (num_cached = %u\n", index, in ocfs2_remove_metadata_array() 501 if (ci->ci_num_cached && index < ci->ci_num_cached) { in ocfs2_remove_metadata_array() 502 bytes = sizeof(sector_t) * (ci->ci_num_cached - index); in ocfs2_remove_metadata_array() [all …]
|
/fs/ |
D | filesystems.c | 127 int err, index; in fs_index() local 136 for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) { in fs_index() 138 err = index; in fs_index() 147 static int fs_name(unsigned int index, char __user * buf) in fs_name() argument 153 for (tmp = file_systems; tmp; tmp = tmp->next, index--) in fs_name() 154 if (index <= 0 && try_module_get(tmp->owner)) in fs_name() 170 int index; in fs_maxindex() local 173 for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index++) in fs_maxindex() 176 return index; in fs_maxindex()
|
D | seq_file.c | 67 loff_t pos = 0, index; in traverse() local 72 index = 0; in traverse() 75 m->index = index; in traverse() 83 p = m->op->start(m, &index); in traverse() 100 m->index = index; in traverse() 106 index++; in traverse() 107 m->index = index; in traverse() 110 p = m->op->next(m, p, &index); in traverse() 113 m->index = index; in traverse() 152 m->index = 0; in seq_read() [all …]
|
/fs/afs/ |
D | write.c | 130 pgoff_t index = pos >> PAGE_CACHE_SHIFT; in afs_write_begin() local 134 vnode->fid.vid, vnode->fid.vnode, index, from, to); in afs_write_begin() 140 candidate->first = candidate->last = index; in afs_write_begin() 147 page = grab_cache_page_write_begin(mapping, index, flags); in afs_write_begin() 178 if (index > 0) { in afs_write_begin() 182 if (wb->last == index - 1 && wb->key == key && in afs_write_begin() 198 ASSERTRANGE(wb->first, <=, index, <=, wb->last); in afs_write_begin() 199 if (index == wb->first && from < wb->offset_first) in afs_write_begin() 201 if (index == wb->last && to > wb->to_last) in afs_write_begin() 254 vnode->fid.vid, vnode->fid.vnode, page->index); in afs_write_end() [all …]
|
/fs/btrfs/ |
D | compression.c | 171 unsigned long index; in end_compressed_bio_read() local 202 index = 0; in end_compressed_bio_read() 203 for (index = 0; index < cb->nr_pages; index++) { in end_compressed_bio_read() 204 page = cb->compressed_pages[index]; in end_compressed_bio_read() 242 unsigned long index = start >> PAGE_CACHE_SHIFT; in end_compressed_writeback() local 245 unsigned long nr_pages = end_index - index + 1; in end_compressed_writeback() 250 ret = find_get_pages_contig(inode->i_mapping, index, in end_compressed_writeback() 255 index += 1; in end_compressed_writeback() 263 index += ret; in end_compressed_writeback() 283 unsigned long index; in end_compressed_bio_write() local [all …]
|
D | extent_io.c | 959 unsigned long index = start >> PAGE_CACHE_SHIFT; in set_range_dirty() local 963 while (index <= end_index) { in set_range_dirty() 964 page = find_get_page(tree->mapping, index); in set_range_dirty() 968 index++; in set_range_dirty() 979 unsigned long index = start >> PAGE_CACHE_SHIFT; in set_range_writeback() local 983 while (index <= end_index) { in set_range_writeback() 984 page = find_get_page(tree->mapping, index); in set_range_writeback() 988 index++; in set_range_writeback() 1127 unsigned long index = start >> PAGE_CACHE_SHIFT; in __unlock_for_delalloc() local 1129 unsigned long nr_pages = end_index - index + 1; in __unlock_for_delalloc() [all …]
|