/fs/ntfs/ |
D | logfile.c | 147 RESTART_AREA *ra; in ntfs_check_restart_area() local 153 ra = (RESTART_AREA*)((u8*)rp + ra_ofs); in ntfs_check_restart_area() 172 ca_ofs = le16_to_cpu(ra->client_array_offset); in ntfs_check_restart_area() 184 ra_len = ca_ofs + le16_to_cpu(ra->log_clients) * in ntfs_check_restart_area() 187 ra_ofs + le16_to_cpu(ra->restart_area_length) > in ntfs_check_restart_area() 189 ra_len > le16_to_cpu(ra->restart_area_length)) { in ntfs_check_restart_area() 201 if ((ra->client_free_list != LOGFILE_NO_CLIENT && in ntfs_check_restart_area() 202 le16_to_cpu(ra->client_free_list) >= in ntfs_check_restart_area() 203 le16_to_cpu(ra->log_clients)) || in ntfs_check_restart_area() 204 (ra->client_in_use_list != LOGFILE_NO_CLIENT && in ntfs_check_restart_area() [all …]
|
/fs/btrfs/ |
D | ref-verify.c | 206 static void __save_stack_trace(struct ref_action *ra) in __save_stack_trace() argument 208 ra->trace_len = stack_trace_save(ra->trace, MAX_TRACE, 2); in __save_stack_trace() 212 struct ref_action *ra) in __print_stack_trace() argument 214 if (ra->trace_len == 0) { in __print_stack_trace() 218 stack_trace_print(ra->trace, ra->trace_len, 2); in __print_stack_trace() 221 static inline void __save_stack_trace(struct ref_action *ra) in __save_stack_trace() argument 226 struct ref_action *ra) in __print_stack_trace() argument 236 struct ref_action *ra; in free_block_entry() local 252 ra = list_first_entry(&be->actions, struct ref_action, in free_block_entry() 254 list_del(&ra->list); in free_block_entry() [all …]
|
D | relocation.c | 2965 static int relocate_one_page(struct inode *inode, struct file_ra_state *ra, in relocate_one_page() argument 2982 page_cache_sync_readahead(inode->i_mapping, ra, NULL, in relocate_one_page() 2990 page_cache_async_readahead(inode->i_mapping, ra, NULL, in relocate_one_page() 3102 struct file_ra_state *ra; in relocate_file_extent_cluster() local 3109 ra = kzalloc(sizeof(*ra), GFP_NOFS); in relocate_file_extent_cluster() 3110 if (!ra) in relocate_file_extent_cluster() 3117 file_ra_state_init(ra, inode->i_mapping); in relocate_file_extent_cluster() 3127 ret = relocate_one_page(inode, ra, cluster, &cluster_nr, index); in relocate_file_extent_cluster() 3131 kfree(ra); in relocate_file_extent_cluster()
|
D | ioctl.c | 1717 struct file_ra_state *ra, in defrag_one_cluster() argument 1758 if (ra) in defrag_one_cluster() 1760 ra, NULL, entry->start >> PAGE_SHIFT, in defrag_one_cluster() 1803 int btrfs_defrag_file(struct inode *inode, struct file_ra_state *ra, in btrfs_defrag_file() argument 1852 if (!ra) { in btrfs_defrag_file() 1854 ra = kzalloc(sizeof(*ra), GFP_KERNEL); in btrfs_defrag_file() 1855 if (ra) in btrfs_defrag_file() 1856 file_ra_state_init(ra, inode->i_mapping); in btrfs_defrag_file() 1894 ret = defrag_one_cluster(BTRFS_I(inode), ra, cur, in btrfs_defrag_file() 1914 kfree(ra); in btrfs_defrag_file()
|
D | raid56.c | 1693 const struct btrfs_raid_bio *ra = container_of(a, struct btrfs_raid_bio, in plug_cmp() local 1697 u64 a_sector = ra->bio_list.head->bi_iter.bi_sector; in plug_cmp()
|
D | send.c | 148 struct file_ra_state ra; member 5036 &sctx->ra, NULL, index, in put_file_data() 5049 &sctx->ra, NULL, page_folio(page), in put_file_data() 5496 memset(&sctx->ra, 0, sizeof(struct file_ra_state)); in send_extent_data() 5497 file_ra_state_init(&sctx->ra, sctx->cur_inode->i_mapping); in send_extent_data()
|
D | free-space-cache.c | 380 struct file_ra_state ra; in readahead_cache() local 383 file_ra_state_init(&ra, inode->i_mapping); in readahead_cache() 386 page_cache_sync_readahead(inode->i_mapping, &ra, NULL, 0, last_index); in readahead_cache()
|
D | ctree.h | 3497 int btrfs_defrag_file(struct inode *inode, struct file_ra_state *ra,
|
/fs/fat/ |
D | fatent.c | 647 static void fat_ra_init(struct super_block *sb, struct fatent_ra *ra, in fat_ra_init() argument 672 ra->cur = 0; in fat_ra_init() 673 ra->limit = (block_end + 1) - blocknr; in fat_ra_init() 676 ra->ra_blocks = reada_blocks >> 1; in fat_ra_init() 677 ra->ra_advance = ra->cur; in fat_ra_init() 678 ra->ra_next = ra->cur; in fat_ra_init() 679 ra->ra_limit = ra->cur + min_t(sector_t, reada_blocks, ra->limit); in fat_ra_init() 683 static void fat_ent_reada(struct super_block *sb, struct fatent_ra *ra, in fat_ent_reada() argument 686 if (ra->ra_next >= ra->ra_limit) in fat_ent_reada() 689 if (ra->cur >= ra->ra_advance) { in fat_ent_reada() [all …]
|
/fs/ntfs3/ |
D | fslog.c | 466 const struct RESTART_AREA *ra; in is_rst_area_valid() local 476 ra = Add2Ptr(rhdr, ro); in is_rst_area_valid() 477 cl = le16_to_cpu(ra->log_clients); in is_rst_area_valid() 482 off = le16_to_cpu(ra->client_off); in is_rst_area_valid() 496 if (le16_to_cpu(rhdr->ra_off) + le16_to_cpu(ra->ra_len) > sys_page || in is_rst_area_valid() 497 off > le16_to_cpu(ra->ra_len)) { in is_rst_area_valid() 505 fl = le16_to_cpu(ra->client_idx[0]); in is_rst_area_valid() 506 ul = le16_to_cpu(ra->client_idx[1]); in is_rst_area_valid() 512 l_size = le64_to_cpu(ra->l_size); in is_rst_area_valid() 520 if (seq_bits != ra->seq_num_bits) in is_rst_area_valid() [all …]
|
/fs/xfs/scrub/ |
D | agheader_repair.c | 461 struct xrep_agfl *ra = priv; in xrep_agfl_walk_rmap() local 465 if (xchk_should_terminate(ra->sc, &error)) in xrep_agfl_walk_rmap() 472 error = xbitmap_set(ra->freesp, fsb, rec->rm_blockcount); in xrep_agfl_walk_rmap() 477 return xbitmap_set_btcur_path(&ra->agmetablocks, cur); in xrep_agfl_walk_rmap() 496 struct xrep_agfl ra; in xrep_agfl_collect_blocks() local 501 ra.sc = sc; in xrep_agfl_collect_blocks() 502 ra.freesp = agfl_extents; in xrep_agfl_collect_blocks() 503 xbitmap_init(&ra.agmetablocks); in xrep_agfl_collect_blocks() 507 error = xfs_rmap_query_all(cur, xrep_agfl_walk_rmap, &ra); in xrep_agfl_collect_blocks() 515 error = xbitmap_set_btblocks(&ra.agmetablocks, cur); in xrep_agfl_collect_blocks() [all …]
|
/fs/fuse/ |
D | dev.c | 1661 struct fuse_retrieve_args *ra = in fuse_retrieve_end() local 1662 container_of(args, typeof(*ra), ap.args); in fuse_retrieve_end() 1664 release_pages(ra->ap.pages, ra->ap.num_pages); in fuse_retrieve_end() 1665 kfree(ra); in fuse_retrieve_end() 1680 struct fuse_retrieve_args *ra; in fuse_retrieve() local 1681 size_t args_size = sizeof(*ra); in fuse_retrieve() 1699 ra = kzalloc(args_size, GFP_KERNEL); in fuse_retrieve() 1700 if (!ra) in fuse_retrieve() 1703 ap = &ra->ap; in fuse_retrieve() 1704 ap->pages = (void *) (ra + 1); in fuse_retrieve() [all …]
|
D | file.c | 101 struct fuse_release_args *ra = container_of(args, typeof(*ra), args); in fuse_release_end() local 103 iput(ra->inode); in fuse_release_end() 104 kfree(ra); in fuse_release_end() 307 struct fuse_release_args *ra = ff->release_args; in fuse_prepare_release() local 322 ra->inarg.fh = ff->fh; in fuse_prepare_release() 323 ra->inarg.flags = flags; in fuse_prepare_release() 324 ra->args.in_numargs = 1; in fuse_prepare_release() 325 ra->args.in_args[0].size = sizeof(struct fuse_release_in); in fuse_prepare_release() 326 ra->args.in_args[0].value = &ra->inarg; in fuse_prepare_release() 327 ra->args.opcode = opcode; in fuse_prepare_release() [all …]
|
/fs/hpfs/ |
D | anode.c | 66 anode_secno a, na = -1, ra, up = -1; in hpfs_add_sector_to_btree() local 145 } else if (!(ranode = hpfs_alloc_anode(s, /*a*/0, &ra, &bh2))) { in hpfs_add_sector_to_btree() 185 hpfs_free_sectors(s, ra, 1); in hpfs_add_sector_to_btree() 248 unode->up = cpu_to_le32(ra); in hpfs_add_sector_to_btree() 259 btree->u.internal[0].down = cpu_to_le32(ra); in hpfs_add_sector_to_btree()
|
/fs/xfs/ |
D | xfs_refcount_item.c | 288 struct xfs_refcount_intent *ra; in xfs_refcount_update_diff_items() local 291 ra = container_of(a, struct xfs_refcount_intent, ri_list); in xfs_refcount_update_diff_items() 293 return XFS_FSB_TO_AGNO(mp, ra->ri_startblock) - in xfs_refcount_update_diff_items()
|
D | xfs_rmap_item.c | 331 struct xfs_rmap_intent *ra; in xfs_rmap_update_diff_items() local 334 ra = container_of(a, struct xfs_rmap_intent, ri_list); in xfs_rmap_update_diff_items() 336 return XFS_FSB_TO_AGNO(mp, ra->ri_bmap.br_startblock) - in xfs_rmap_update_diff_items()
|
D | xfs_extfree_item.c | 393 struct xfs_extent_free_item *ra; in xfs_extent_free_diff_items() local 396 ra = container_of(a, struct xfs_extent_free_item, xefi_list); in xfs_extent_free_diff_items() 398 return XFS_FSB_TO_AGNO(mp, ra->xefi_startblock) - in xfs_extent_free_diff_items()
|
/fs/nilfs2/ |
D | btree.c | 472 const struct nilfs_btree_readahead_info *ra) in __nilfs_btree_get_block() argument 497 if (ra) { in __nilfs_btree_get_block() 502 for (n = ra->max_ra_blocks, i = ra->index + 1; in __nilfs_btree_get_block() 503 n > 0 && i < ra->ncmax; n--, i++) { in __nilfs_btree_get_block() 504 ptr2 = nilfs_btree_node_get_ptr(ra->node, i, ra->ncmax); in __nilfs_btree_get_block() 552 struct nilfs_btree_readahead_info p, *ra; in nilfs_btree_do_lookup() local 570 ra = NULL; in nilfs_btree_do_lookup() 576 ra = &p; in nilfs_btree_do_lookup() 579 ra); in nilfs_btree_do_lookup()
|
/fs/ubifs/ |
D | replay.c | 306 struct replay_entry *ra, *rb; in replay_entries_cmp() local 312 ra = list_entry(a, struct replay_entry, list); in replay_entries_cmp() 314 ubifs_assert(c, ra->sqnum != rb->sqnum); in replay_entries_cmp() 315 if (ra->sqnum > rb->sqnum) in replay_entries_cmp()
|
/fs/xfs/libxfs/ |
D | xfs_btree.c | 494 new->bc_levels[i].ra = cur->bc_levels[i].ra; in xfs_btree_dup_cursor() 1000 if ((cur->bc_levels[lev].ra | lr) == cur->bc_levels[lev].ra) in xfs_btree_readahead() 1003 cur->bc_levels[lev].ra |= lr; in xfs_btree_readahead() 1072 cur->bc_levels[lev].ra = 0; in xfs_btree_setbuf() 1077 cur->bc_levels[lev].ra |= XFS_BTCUR_LEFTRA; in xfs_btree_setbuf() 1079 cur->bc_levels[lev].ra |= XFS_BTCUR_RIGHTRA; in xfs_btree_setbuf() 1082 cur->bc_levels[lev].ra |= XFS_BTCUR_LEFTRA; in xfs_btree_setbuf() 1084 cur->bc_levels[lev].ra |= XFS_BTCUR_RIGHTRA; in xfs_btree_setbuf() 3677 cur->bc_levels[level].ra = 0; in xfs_btree_kill_root() 4185 cur->bc_levels[level].ra = 0; in xfs_btree_delrec() [all …]
|
D | xfs_btree.h | 221 uint16_t ra; member
|
/fs/cramfs/ |
D | inode.c | 186 struct file_ra_state ra = {}; in cramfs_blkdev_read() local 216 file_ra_state_init(&ra, mapping); in cramfs_blkdev_read() 217 page_cache_sync_readahead(mapping, &ra, NULL, blocknr, BLKS_PER_BUF); in cramfs_blkdev_read()
|
/fs/f2fs/ |
D | dir.c | 1063 struct file_ra_state *ra = &file->f_ra; in f2fs_readdir() local 1096 if (npages - n > 1 && !ra_has_index(ra, n)) in f2fs_readdir() 1097 page_cache_sync_readahead(inode->i_mapping, ra, file, n, in f2fs_readdir()
|
/fs/kernfs/ |
D | dir.c | 78 struct kernfs_root *ra = kernfs_root(a), *rb = kernfs_root(b); in kernfs_common_ancestor() local 80 if (ra != rb) in kernfs_common_ancestor() 83 da = kernfs_depth(ra->kn, a); in kernfs_common_ancestor()
|
/fs/erofs/ |
D | zdata.c | 1750 bool force_fg, bool ra) in z_erofs_runqueue() argument 1756 z_erofs_submit_queue(f, io, &force_fg, ra); in z_erofs_runqueue()
|