Home
last modified time | relevance | path

Searched refs:ra (Results 1 – 25 of 25) sorted by relevance

/fs/ntfs/
Dlogfile.c147 RESTART_AREA *ra; in ntfs_check_restart_area() local
153 ra = (RESTART_AREA*)((u8*)rp + ra_ofs); in ntfs_check_restart_area()
172 ca_ofs = le16_to_cpu(ra->client_array_offset); in ntfs_check_restart_area()
184 ra_len = ca_ofs + le16_to_cpu(ra->log_clients) * in ntfs_check_restart_area()
187 ra_ofs + le16_to_cpu(ra->restart_area_length) > in ntfs_check_restart_area()
189 ra_len > le16_to_cpu(ra->restart_area_length)) { in ntfs_check_restart_area()
201 if ((ra->client_free_list != LOGFILE_NO_CLIENT && in ntfs_check_restart_area()
202 le16_to_cpu(ra->client_free_list) >= in ntfs_check_restart_area()
203 le16_to_cpu(ra->log_clients)) || in ntfs_check_restart_area()
204 (ra->client_in_use_list != LOGFILE_NO_CLIENT && in ntfs_check_restart_area()
[all …]
/fs/btrfs/
Dref-verify.c206 static void __save_stack_trace(struct ref_action *ra) in __save_stack_trace() argument
208 ra->trace_len = stack_trace_save(ra->trace, MAX_TRACE, 2); in __save_stack_trace()
212 struct ref_action *ra) in __print_stack_trace() argument
214 if (ra->trace_len == 0) { in __print_stack_trace()
218 stack_trace_print(ra->trace, ra->trace_len, 2); in __print_stack_trace()
221 static inline void __save_stack_trace(struct ref_action *ra) in __save_stack_trace() argument
226 struct ref_action *ra) in __print_stack_trace() argument
236 struct ref_action *ra; in free_block_entry() local
252 ra = list_first_entry(&be->actions, struct ref_action, in free_block_entry()
254 list_del(&ra->list); in free_block_entry()
[all …]
Drelocation.c2965 static int relocate_one_page(struct inode *inode, struct file_ra_state *ra, in relocate_one_page() argument
2982 page_cache_sync_readahead(inode->i_mapping, ra, NULL, in relocate_one_page()
2990 page_cache_async_readahead(inode->i_mapping, ra, NULL, in relocate_one_page()
3102 struct file_ra_state *ra; in relocate_file_extent_cluster() local
3109 ra = kzalloc(sizeof(*ra), GFP_NOFS); in relocate_file_extent_cluster()
3110 if (!ra) in relocate_file_extent_cluster()
3117 file_ra_state_init(ra, inode->i_mapping); in relocate_file_extent_cluster()
3127 ret = relocate_one_page(inode, ra, cluster, &cluster_nr, index); in relocate_file_extent_cluster()
3131 kfree(ra); in relocate_file_extent_cluster()
Dioctl.c1717 struct file_ra_state *ra, in defrag_one_cluster() argument
1758 if (ra) in defrag_one_cluster()
1760 ra, NULL, entry->start >> PAGE_SHIFT, in defrag_one_cluster()
1803 int btrfs_defrag_file(struct inode *inode, struct file_ra_state *ra, in btrfs_defrag_file() argument
1852 if (!ra) { in btrfs_defrag_file()
1854 ra = kzalloc(sizeof(*ra), GFP_KERNEL); in btrfs_defrag_file()
1855 if (ra) in btrfs_defrag_file()
1856 file_ra_state_init(ra, inode->i_mapping); in btrfs_defrag_file()
1894 ret = defrag_one_cluster(BTRFS_I(inode), ra, cur, in btrfs_defrag_file()
1914 kfree(ra); in btrfs_defrag_file()
Draid56.c1693 const struct btrfs_raid_bio *ra = container_of(a, struct btrfs_raid_bio, in plug_cmp() local
1697 u64 a_sector = ra->bio_list.head->bi_iter.bi_sector; in plug_cmp()
Dsend.c148 struct file_ra_state ra; member
5036 &sctx->ra, NULL, index, in put_file_data()
5049 &sctx->ra, NULL, page_folio(page), in put_file_data()
5496 memset(&sctx->ra, 0, sizeof(struct file_ra_state)); in send_extent_data()
5497 file_ra_state_init(&sctx->ra, sctx->cur_inode->i_mapping); in send_extent_data()
Dfree-space-cache.c380 struct file_ra_state ra; in readahead_cache() local
383 file_ra_state_init(&ra, inode->i_mapping); in readahead_cache()
386 page_cache_sync_readahead(inode->i_mapping, &ra, NULL, 0, last_index); in readahead_cache()
Dctree.h3497 int btrfs_defrag_file(struct inode *inode, struct file_ra_state *ra,
/fs/fat/
Dfatent.c647 static void fat_ra_init(struct super_block *sb, struct fatent_ra *ra, in fat_ra_init() argument
672 ra->cur = 0; in fat_ra_init()
673 ra->limit = (block_end + 1) - blocknr; in fat_ra_init()
676 ra->ra_blocks = reada_blocks >> 1; in fat_ra_init()
677 ra->ra_advance = ra->cur; in fat_ra_init()
678 ra->ra_next = ra->cur; in fat_ra_init()
679 ra->ra_limit = ra->cur + min_t(sector_t, reada_blocks, ra->limit); in fat_ra_init()
683 static void fat_ent_reada(struct super_block *sb, struct fatent_ra *ra, in fat_ent_reada() argument
686 if (ra->ra_next >= ra->ra_limit) in fat_ent_reada()
689 if (ra->cur >= ra->ra_advance) { in fat_ent_reada()
[all …]
/fs/ntfs3/
Dfslog.c466 const struct RESTART_AREA *ra; in is_rst_area_valid() local
476 ra = Add2Ptr(rhdr, ro); in is_rst_area_valid()
477 cl = le16_to_cpu(ra->log_clients); in is_rst_area_valid()
482 off = le16_to_cpu(ra->client_off); in is_rst_area_valid()
496 if (le16_to_cpu(rhdr->ra_off) + le16_to_cpu(ra->ra_len) > sys_page || in is_rst_area_valid()
497 off > le16_to_cpu(ra->ra_len)) { in is_rst_area_valid()
505 fl = le16_to_cpu(ra->client_idx[0]); in is_rst_area_valid()
506 ul = le16_to_cpu(ra->client_idx[1]); in is_rst_area_valid()
512 l_size = le64_to_cpu(ra->l_size); in is_rst_area_valid()
520 if (seq_bits != ra->seq_num_bits) in is_rst_area_valid()
[all …]
/fs/xfs/scrub/
Dagheader_repair.c461 struct xrep_agfl *ra = priv; in xrep_agfl_walk_rmap() local
465 if (xchk_should_terminate(ra->sc, &error)) in xrep_agfl_walk_rmap()
472 error = xbitmap_set(ra->freesp, fsb, rec->rm_blockcount); in xrep_agfl_walk_rmap()
477 return xbitmap_set_btcur_path(&ra->agmetablocks, cur); in xrep_agfl_walk_rmap()
496 struct xrep_agfl ra; in xrep_agfl_collect_blocks() local
501 ra.sc = sc; in xrep_agfl_collect_blocks()
502 ra.freesp = agfl_extents; in xrep_agfl_collect_blocks()
503 xbitmap_init(&ra.agmetablocks); in xrep_agfl_collect_blocks()
507 error = xfs_rmap_query_all(cur, xrep_agfl_walk_rmap, &ra); in xrep_agfl_collect_blocks()
515 error = xbitmap_set_btblocks(&ra.agmetablocks, cur); in xrep_agfl_collect_blocks()
[all …]
/fs/fuse/
Ddev.c1661 struct fuse_retrieve_args *ra = in fuse_retrieve_end() local
1662 container_of(args, typeof(*ra), ap.args); in fuse_retrieve_end()
1664 release_pages(ra->ap.pages, ra->ap.num_pages); in fuse_retrieve_end()
1665 kfree(ra); in fuse_retrieve_end()
1680 struct fuse_retrieve_args *ra; in fuse_retrieve() local
1681 size_t args_size = sizeof(*ra); in fuse_retrieve()
1699 ra = kzalloc(args_size, GFP_KERNEL); in fuse_retrieve()
1700 if (!ra) in fuse_retrieve()
1703 ap = &ra->ap; in fuse_retrieve()
1704 ap->pages = (void *) (ra + 1); in fuse_retrieve()
[all …]
Dfile.c101 struct fuse_release_args *ra = container_of(args, typeof(*ra), args); in fuse_release_end() local
103 iput(ra->inode); in fuse_release_end()
104 kfree(ra); in fuse_release_end()
307 struct fuse_release_args *ra = ff->release_args; in fuse_prepare_release() local
322 ra->inarg.fh = ff->fh; in fuse_prepare_release()
323 ra->inarg.flags = flags; in fuse_prepare_release()
324 ra->args.in_numargs = 1; in fuse_prepare_release()
325 ra->args.in_args[0].size = sizeof(struct fuse_release_in); in fuse_prepare_release()
326 ra->args.in_args[0].value = &ra->inarg; in fuse_prepare_release()
327 ra->args.opcode = opcode; in fuse_prepare_release()
[all …]
/fs/hpfs/
Danode.c66 anode_secno a, na = -1, ra, up = -1; in hpfs_add_sector_to_btree() local
145 } else if (!(ranode = hpfs_alloc_anode(s, /*a*/0, &ra, &bh2))) { in hpfs_add_sector_to_btree()
185 hpfs_free_sectors(s, ra, 1); in hpfs_add_sector_to_btree()
248 unode->up = cpu_to_le32(ra); in hpfs_add_sector_to_btree()
259 btree->u.internal[0].down = cpu_to_le32(ra); in hpfs_add_sector_to_btree()
/fs/xfs/
Dxfs_refcount_item.c288 struct xfs_refcount_intent *ra; in xfs_refcount_update_diff_items() local
291 ra = container_of(a, struct xfs_refcount_intent, ri_list); in xfs_refcount_update_diff_items()
293 return XFS_FSB_TO_AGNO(mp, ra->ri_startblock) - in xfs_refcount_update_diff_items()
Dxfs_rmap_item.c331 struct xfs_rmap_intent *ra; in xfs_rmap_update_diff_items() local
334 ra = container_of(a, struct xfs_rmap_intent, ri_list); in xfs_rmap_update_diff_items()
336 return XFS_FSB_TO_AGNO(mp, ra->ri_bmap.br_startblock) - in xfs_rmap_update_diff_items()
Dxfs_extfree_item.c393 struct xfs_extent_free_item *ra; in xfs_extent_free_diff_items() local
396 ra = container_of(a, struct xfs_extent_free_item, xefi_list); in xfs_extent_free_diff_items()
398 return XFS_FSB_TO_AGNO(mp, ra->xefi_startblock) - in xfs_extent_free_diff_items()
/fs/nilfs2/
Dbtree.c472 const struct nilfs_btree_readahead_info *ra) in __nilfs_btree_get_block() argument
497 if (ra) { in __nilfs_btree_get_block()
502 for (n = ra->max_ra_blocks, i = ra->index + 1; in __nilfs_btree_get_block()
503 n > 0 && i < ra->ncmax; n--, i++) { in __nilfs_btree_get_block()
504 ptr2 = nilfs_btree_node_get_ptr(ra->node, i, ra->ncmax); in __nilfs_btree_get_block()
552 struct nilfs_btree_readahead_info p, *ra; in nilfs_btree_do_lookup() local
570 ra = NULL; in nilfs_btree_do_lookup()
576 ra = &p; in nilfs_btree_do_lookup()
579 ra); in nilfs_btree_do_lookup()
/fs/ubifs/
Dreplay.c306 struct replay_entry *ra, *rb; in replay_entries_cmp() local
312 ra = list_entry(a, struct replay_entry, list); in replay_entries_cmp()
314 ubifs_assert(c, ra->sqnum != rb->sqnum); in replay_entries_cmp()
315 if (ra->sqnum > rb->sqnum) in replay_entries_cmp()
/fs/xfs/libxfs/
Dxfs_btree.c494 new->bc_levels[i].ra = cur->bc_levels[i].ra; in xfs_btree_dup_cursor()
1000 if ((cur->bc_levels[lev].ra | lr) == cur->bc_levels[lev].ra) in xfs_btree_readahead()
1003 cur->bc_levels[lev].ra |= lr; in xfs_btree_readahead()
1072 cur->bc_levels[lev].ra = 0; in xfs_btree_setbuf()
1077 cur->bc_levels[lev].ra |= XFS_BTCUR_LEFTRA; in xfs_btree_setbuf()
1079 cur->bc_levels[lev].ra |= XFS_BTCUR_RIGHTRA; in xfs_btree_setbuf()
1082 cur->bc_levels[lev].ra |= XFS_BTCUR_LEFTRA; in xfs_btree_setbuf()
1084 cur->bc_levels[lev].ra |= XFS_BTCUR_RIGHTRA; in xfs_btree_setbuf()
3677 cur->bc_levels[level].ra = 0; in xfs_btree_kill_root()
4185 cur->bc_levels[level].ra = 0; in xfs_btree_delrec()
[all …]
Dxfs_btree.h221 uint16_t ra; member
/fs/cramfs/
Dinode.c186 struct file_ra_state ra = {}; in cramfs_blkdev_read() local
216 file_ra_state_init(&ra, mapping); in cramfs_blkdev_read()
217 page_cache_sync_readahead(mapping, &ra, NULL, blocknr, BLKS_PER_BUF); in cramfs_blkdev_read()
/fs/f2fs/
Ddir.c1063 struct file_ra_state *ra = &file->f_ra; in f2fs_readdir() local
1096 if (npages - n > 1 && !ra_has_index(ra, n)) in f2fs_readdir()
1097 page_cache_sync_readahead(inode->i_mapping, ra, file, n, in f2fs_readdir()
/fs/kernfs/
Ddir.c78 struct kernfs_root *ra = kernfs_root(a), *rb = kernfs_root(b); in kernfs_common_ancestor() local
80 if (ra != rb) in kernfs_common_ancestor()
83 da = kernfs_depth(ra->kn, a); in kernfs_common_ancestor()
/fs/erofs/
Dzdata.c1750 bool force_fg, bool ra) in z_erofs_runqueue() argument
1756 z_erofs_submit_queue(f, io, &force_fg, ra); in z_erofs_runqueue()