• Home
  • Raw
  • Download

Lines Matching refs:count

138 		to->count = be16_to_cpu(from3->hdr.__count);  in xfs_da3_node_hdr_from_disk()
146 to->count = be16_to_cpu(from->hdr.__count); in xfs_da3_node_hdr_from_disk()
166 to3->hdr.__count = cpu_to_be16(from->count); in xfs_da3_node_hdr_to_disk()
173 to->hdr.__count = cpu_to_be16(from->count); in xfs_da3_node_hdr_to_disk()
225 if (ichdr.count == 0) in xfs_da3_node_verify()
232 if (ichdr.count > mp->m_dir_geo->node_ents && in xfs_da3_node_verify()
233 ichdr.count > mp->m_attr_geo->node_ents) in xfs_da3_node_verify()
676 size = (int)((char *)&btree[icnodehdr.count] - (char *)oldroot); in xfs_da3_root_split()
692 size = (int)((char *)&leafhdr.ents[leafhdr.count] - in xfs_da3_root_split()
739 nodehdr.count = 2; in xfs_da3_root_split()
792 if (nodehdr.count + newcount > state->args->geo->node_ents) { in xfs_da3_node_split()
830 if (oldblk->index <= nodehdr.count) { in xfs_da3_node_split()
875 int count; in xfs_da3_node_rebalance() local
893 if (nodehdr1.count > 0 && nodehdr2.count > 0 && in xfs_da3_node_rebalance()
895 (be32_to_cpu(btree2[nodehdr2.count - 1].hashval) < in xfs_da3_node_rebalance()
896 be32_to_cpu(btree1[nodehdr1.count - 1].hashval)))) { in xfs_da3_node_rebalance()
907 count = (nodehdr1.count - nodehdr2.count) / 2; in xfs_da3_node_rebalance()
908 if (count == 0) in xfs_da3_node_rebalance()
914 if (count > 0) { in xfs_da3_node_rebalance()
918 tmp = nodehdr2.count; in xfs_da3_node_rebalance()
922 btree_d = &btree2[count]; in xfs_da3_node_rebalance()
930 nodehdr2.count += count; in xfs_da3_node_rebalance()
931 tmp = count * (uint)sizeof(xfs_da_node_entry_t); in xfs_da3_node_rebalance()
932 btree_s = &btree1[nodehdr1.count - count]; in xfs_da3_node_rebalance()
935 nodehdr1.count -= count; in xfs_da3_node_rebalance()
941 count = -count; in xfs_da3_node_rebalance()
942 tmp = count * (uint)sizeof(xfs_da_node_entry_t); in xfs_da3_node_rebalance()
944 btree_d = &btree1[nodehdr1.count]; in xfs_da3_node_rebalance()
946 nodehdr1.count += count; in xfs_da3_node_rebalance()
954 tmp = nodehdr2.count - count; in xfs_da3_node_rebalance()
956 btree_s = &btree2[count]; in xfs_da3_node_rebalance()
959 nodehdr2.count -= count; in xfs_da3_node_rebalance()
974 (sizeof(btree2[0]) * nodehdr2.count))); in xfs_da3_node_rebalance()
988 blk1->hashval = be32_to_cpu(btree1[nodehdr1.count - 1].hashval); in xfs_da3_node_rebalance()
989 blk2->hashval = be32_to_cpu(btree2[nodehdr2.count - 1].hashval); in xfs_da3_node_rebalance()
994 if (blk1->index >= nodehdr1.count) { in xfs_da3_node_rebalance()
995 blk2->index = blk1->index - nodehdr1.count; in xfs_da3_node_rebalance()
996 blk1->index = nodehdr1.count + 1; /* make it invalid */ in xfs_da3_node_rebalance()
1021 ASSERT(oldblk->index >= 0 && oldblk->index <= nodehdr.count); in xfs_da3_node_add()
1031 if (oldblk->index < nodehdr.count) { in xfs_da3_node_add()
1032 tmp = (nodehdr.count - oldblk->index) * (uint)sizeof(*btree); in xfs_da3_node_add()
1041 nodehdr.count += 1; in xfs_da3_node_add()
1050 oldblk->hashval = be32_to_cpu(btree[nodehdr.count - 1].hashval); in xfs_da3_node_add()
1196 if (oldroothdr.count > 1) in xfs_da3_root_join()
1250 int count; in xfs_da3_node_toosmall() local
1268 if (nodehdr.count > (state->args->geo->node_ents >> 1)) { in xfs_da3_node_toosmall()
1279 if (nodehdr.count == 0) { in xfs_da3_node_toosmall()
1305 count = state->args->geo->node_ents; in xfs_da3_node_toosmall()
1306 count -= state->args->geo->node_ents >> 2; in xfs_da3_node_toosmall()
1307 count -= nodehdr.count; in xfs_da3_node_toosmall()
1328 if (count - thdr.count >= 0) in xfs_da3_node_toosmall()
1365 int *count) in xfs_da3_node_lasthash() argument
1370 if (count) in xfs_da3_node_lasthash()
1371 *count = nodehdr.count; in xfs_da3_node_lasthash()
1372 if (!nodehdr.count) in xfs_da3_node_lasthash()
1374 return be32_to_cpu(nodehdr.btree[nodehdr.count - 1].hashval); in xfs_da3_node_lasthash()
1391 int count; in xfs_da3_fixhashpath() local
1400 lasthash = xfs_attr_leaf_lasthash(blk->bp, &count); in xfs_da3_fixhashpath()
1401 if (count == 0) in xfs_da3_fixhashpath()
1405 lasthash = xfs_dir2_leaf_lasthash(dp, blk->bp, &count); in xfs_da3_fixhashpath()
1406 if (count == 0) in xfs_da3_fixhashpath()
1410 lasthash = xfs_da3_node_lasthash(dp, blk->bp, &count); in xfs_da3_fixhashpath()
1411 if (count == 0) in xfs_da3_fixhashpath()
1429 lasthash = be32_to_cpu(btree[nodehdr.count - 1].hashval); in xfs_da3_fixhashpath()
1452 ASSERT(drop_blk->index < nodehdr.count); in xfs_da3_node_remove()
1460 if (index < nodehdr.count - 1) { in xfs_da3_node_remove()
1461 tmp = nodehdr.count - index - 1; in xfs_da3_node_remove()
1466 index = nodehdr.count - 1; in xfs_da3_node_remove()
1471 nodehdr.count -= 1; in xfs_da3_node_remove()
1519 (be32_to_cpu(drop_btree[drop_hdr.count - 1].hashval) < in xfs_da3_node_unbalance()
1520 be32_to_cpu(save_btree[save_hdr.count - 1].hashval))) { in xfs_da3_node_unbalance()
1522 tmp = save_hdr.count * sizeof(xfs_da_node_entry_t); in xfs_da3_node_unbalance()
1523 memmove(&save_btree[drop_hdr.count], &save_btree[0], tmp); in xfs_da3_node_unbalance()
1528 (save_hdr.count + drop_hdr.count) * in xfs_da3_node_unbalance()
1531 sindex = save_hdr.count; in xfs_da3_node_unbalance()
1534 drop_hdr.count * sizeof(xfs_da_node_entry_t))); in xfs_da3_node_unbalance()
1540 tmp = drop_hdr.count * (uint)sizeof(xfs_da_node_entry_t); in xfs_da3_node_unbalance()
1542 save_hdr.count += drop_hdr.count; in xfs_da3_node_unbalance()
1552 save_blk->hashval = be32_to_cpu(save_btree[save_hdr.count - 1].hashval); in xfs_da3_node_unbalance()
1661 max = nodehdr.count; in xfs_da3_node_lookup_int()
1779 if (node1hdr.count > 0 && node2hdr.count > 0 && in xfs_da3_node_order()
1781 (be32_to_cpu(btree2[node2hdr.count - 1].hashval) < in xfs_da3_node_order()
1782 be32_to_cpu(btree1[node1hdr.count - 1].hashval)))) { in xfs_da3_node_order()
2000 if (forward && (blk->index < nodehdr.count - 1)) { in xfs_da3_path_shift()
2060 blk->hashval = be32_to_cpu(btree[nodehdr.count - 1].hashval); in xfs_da3_path_shift()
2064 blk->index = nodehdr.count - 1; in xfs_da3_path_shift()
2143 int count) in xfs_da_grow_inode_int() argument
2155 error = xfs_bmap_first_unused(tp, dp, count, bno, w); in xfs_da_grow_inode_int()
2163 error = xfs_bmapi_write(tp, dp, *bno, count, in xfs_da_grow_inode_int()
2173 } else if (nmap == 0 && count > 1) { in xfs_da_grow_inode_int()
2181 mapp = kmem_alloc(sizeof(*mapp) * count, 0); in xfs_da_grow_inode_int()
2182 for (b = *bno, mapi = 0; b < *bno + count; ) { in xfs_da_grow_inode_int()
2183 nmap = min(XFS_BMAP_MAX_NMAP, count); in xfs_da_grow_inode_int()
2184 c = (int)(*bno + count - b); in xfs_da_grow_inode_int()
2206 if (got != count || mapp[0].br_startoff != *bno || in xfs_da_grow_inode_int()
2208 *bno + count) { in xfs_da_grow_inode_int()
2324 dead_hash = be32_to_cpu(ents[leafhdr.count - 1].hashval); in xfs_da3_swap_lastblock()
2332 dead_hash = be32_to_cpu(btree[deadhdr.count - 1].hashval); in xfs_da3_swap_lastblock()
2394 entno < par_hdr.count && in xfs_da3_swap_lastblock()
2398 if (XFS_IS_CORRUPT(mp, entno == par_hdr.count)) { in xfs_da3_swap_lastblock()
2414 entno < par_hdr.count && in xfs_da3_swap_lastblock()
2418 if (entno < par_hdr.count) in xfs_da3_swap_lastblock()
2468 int done, error, w, count; in xfs_da_shrink_inode() local
2476 count = args->geo->fsbcount; in xfs_da_shrink_inode()
2482 error = xfs_bunmapi(tp, dp, dead_blkno, count, in xfs_da_shrink_inode()