Searched refs:XFS_INODES_PER_CHUNK (Results 1 – 11 of 11) sorted by relevance
224 nr_inodes = min_t(unsigned int, XFS_INODES_PER_CHUNK, in xchk_iallocbt_check_cluster()326 cluster_base < XFS_INODES_PER_CHUNK; in xchk_iallocbt_check_clusters()366 imask = min_t(unsigned int, XFS_INODES_PER_CHUNK, in xchk_iallocbt_rec_alignment()384 iabt->next_startino += XFS_INODES_PER_CHUNK; in xchk_iallocbt_rec_alignment()405 if (igeo->inodes_per_cluster <= XFS_INODES_PER_CHUNK) in xchk_iallocbt_rec_alignment()413 iabt->next_startino = irec->ir_startino + XFS_INODES_PER_CHUNK; in xchk_iallocbt_rec_alignment()438 if (irec.ir_count > XFS_INODES_PER_CHUNK || in xchk_iallocbt_rec()439 irec.ir_freecount > XFS_INODES_PER_CHUNK) in xchk_iallocbt_rec()443 (XFS_INODES_PER_CHUNK - irec.ir_count); in xchk_iallocbt_rec()450 !xfs_verify_agino(mp, agno, agino + XFS_INODES_PER_CHUNK - 1)) { in xchk_iallocbt_rec()[all …]
476 rec.ir_startino + XFS_INODES_PER_CHUNK <= agino) in xchk_inode_xref_finobt()
256 XFS_INODES_PER_CHUNK); in xrep_calc_ag_resblks()
109 for (i = 0; i < XFS_INODES_PER_CHUNK; i += igeo->inodes_per_cluster) { in xfs_iwalk_ichunk_ra()209 for (j = 0; j < XFS_INODES_PER_CHUNK; j++) { in xfs_iwalk_ag_recs()313 irec->ir_startino + XFS_INODES_PER_CHUNK - 1); in xfs_iwalk_ag_start()319 if (irec->ir_startino + XFS_INODES_PER_CHUNK <= agino) in xfs_iwalk_ag_start()369 ASSERT(next_agino >= irec->ir_startino + XFS_INODES_PER_CHUNK); in xfs_iwalk_run_callbacks()440 iwag->lastino = rec_fsino + XFS_INODES_PER_CHUNK - 1; in xfs_iwalk_ag()518 inodes = round_up(inodes, XFS_INODES_PER_CHUNK); in xfs_iwalk_prefetch()533 inobt_records = (inodes * 5) / (4 * XFS_INODES_PER_CHUNK); in xfs_iwalk_prefetch()
382 XFS_INODES_PER_CHUNK; in xfs_inumbers_walk()
2016 #define XFS_INODEGC_MAX_BACKLOG (4 * XFS_INODES_PER_CHUNK)
91 irec->ir_count = XFS_INODES_PER_CHUNK; in xfs_inobt_btrec_to_irec()122 irec->ir_count > XFS_INODES_PER_CHUNK) in xfs_inobt_get_rec()124 if (irec->ir_freecount > XFS_INODES_PER_CHUNK) in xfs_inobt_get_rec()189 thisino += XFS_INODES_PER_CHUNK) { in xfs_inobt_insert()198 XFS_INODES_PER_CHUNK, in xfs_inobt_insert()199 XFS_INODES_PER_CHUNK, in xfs_inobt_insert()465 if (trec->ir_count + srec->ir_count > XFS_INODES_PER_CHUNK) in __xfs_inobt_can_merge()789 ASSERT(newlen <= XFS_INODES_PER_CHUNK); in xfs_ialloc_ag_alloc()1097 (trec.ir_startino + XFS_INODES_PER_CHUNK - 1) < in xfs_dialloc_ag_inobt()1222 ASSERT(offset < XFS_INODES_PER_CHUNK); in xfs_dialloc_ag_inobt()[all …]
205 x += XFS_INODES_PER_CHUNK - 1; in xfs_inobt_init_high_key_from_rec()377 return be32_to_cpu(r1->inobt.ir_startino) + XFS_INODES_PER_CHUNK <= in xfs_inobt_recs_inorder()653 XFS_INODES_PER_CHUNK); in xfs_inobt_max_size()
230 *min = XFS_INODES_PER_CHUNK; in xfs_icount_range()
1326 #define XFS_INODES_PER_CHUNK (NBBY * sizeof(xfs_inofree_t)) macro1334 (XFS_INODES_PER_CHUNK / (NBBY * sizeof(uint16_t)))1338 return ((n >= XFS_INODES_PER_CHUNK ? 0 : XFS_INOBT_MASK(n)) - 1) << i; in xfs_inobt_maskn()
351 align = XFS_INODES_PER_CHUNK * sbp->sb_inodesize in xfs_validate_sb_common()