Searched refs:NAT_ENTRY_PER_BLOCK (Results 1 – 6 of 6) sorted by relevance
49 NAT_ENTRY_PER_BLOCK); in get_new_sb()264 new_max_nid = NAT_ENTRY_PER_BLOCK * nat_blocks; in shrink_nats()266 for (nid = nm_i->max_nid - 1; nid > new_max_nid; nid -= NAT_ENTRY_PER_BLOCK) { in shrink_nats()267 block_off = nid / NAT_ENTRY_PER_BLOCK; in shrink_nats()309 for (nid = nm_i->max_nid - 1; nid >= 0; nid -= NAT_ENTRY_PER_BLOCK) { in migrate_nat()310 block_off = nid / NAT_ENTRY_PER_BLOCK; in migrate_nat()334 new_max_nid = NAT_ENTRY_PER_BLOCK * nat_blocks; in migrate_nat()342 nid += NAT_ENTRY_PER_BLOCK) { in migrate_nat()343 block_off = nid / NAT_ENTRY_PER_BLOCK; in migrate_nat()
678 sbi->blocks_per_seg * NAT_ENTRY_PER_BLOCK; in init_sb_info()929 if (!(nid % NAT_ENTRY_PER_BLOCK)) { in f2fs_init_nid_bitmap()937 if (nat_block->entries[nid % NAT_ENTRY_PER_BLOCK].block_addr) in f2fs_init_nid_bitmap()1017 for (j = 0; j < NAT_ENTRY_PER_BLOCK; j++) { in write_nat_bits()1024 else if (valid == NAT_ENTRY_PER_BLOCK) in write_nat_bits()1057 nm_i->max_nid = NAT_ENTRY_PER_BLOCK * nm_i->nat_blocks; in init_node_manager()1522 entry_off = nid % NAT_ENTRY_PER_BLOCK; in get_nat_entry()1596 entry_off = nid % NAT_ENTRY_PER_BLOCK; in update_nat_blkaddr()1834 entry_off = nid % NAT_ENTRY_PER_BLOCK; in flush_nat_journal_entries()2055 entry_off = nid % NAT_ENTRY_PER_BLOCK; in nullify_nat_entry()[all …]
72 nid = block_off * NAT_ENTRY_PER_BLOCK; in nat_dump()73 for (i = 0; i < NAT_ENTRY_PER_BLOCK; i++) { in nat_dump()
360 return (nid <= (NAT_ENTRY_PER_BLOCK * in IS_VALID_NID()
826 #define NAT_ENTRY_PER_BLOCK (PAGE_CACHE_SIZE / sizeof(struct f2fs_nat_entry)) macro827 #define NAT_BLOCK_OFFSET(start_nid) (start_nid / NAT_ENTRY_PER_BLOCK)836 struct f2fs_nat_entry entries[NAT_ENTRY_PER_BLOCK];
273 NAT_ENTRY_PER_BLOCK); in f2fs_prepare_super_block()