Searched refs:from_cblock (Results 1 – 4 of 4) sorted by relevance
/kernel/linux/linux-5.10/drivers/md/ |
D | dm-cache-metadata.c | 705 disk_super->cache_blocks = cpu_to_le32(from_cblock(cmd->cache_blocks)); in __commit_transaction() 895 r = dm_array_get_value(&cmd->info, cmd->root, from_cblock(b), &value); in block_clean_combined_dirty() 921 (unsigned long long) from_cblock(begin)); in blocks_are_clean_combined_dirty() 925 begin = to_cblock(from_cblock(begin) + 1); in blocks_are_clean_combined_dirty() 939 if (from_cblock(cmd->cache_blocks) == 0) in blocks_are_clean_separate_dirty() 944 from_cblock(cmd->cache_blocks), &cmd->dirty_cursor); in blocks_are_clean_separate_dirty() 950 r = dm_bitset_cursor_skip(&cmd->dirty_cursor, from_cblock(begin)); in blocks_are_clean_separate_dirty() 965 (unsigned long long) from_cblock(begin)); in blocks_are_clean_separate_dirty() 971 begin = to_cblock(from_cblock(begin) + 1); in blocks_are_clean_separate_dirty() 1057 if (from_cblock(new_cache_size) < from_cblock(cmd->cache_blocks)) { in dm_cache_resize() [all …]
|
D | dm-cache-policy-smq.c | 999 if (!test_and_set_bit(from_cblock(infer_cblock(mq, e)), mq->cache_hit_bits)) { in requeue() 1107 clear_bitset(mq->cache_hit_bits, from_cblock(mq->cache_size)); in end_cache_period() 1127 return from_cblock(mq->cache_size) * p / 100u; in percent_to_target() 1153 nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; in free_target_met() 1463 from_cblock(work->cblock)); in __complete_background_work() 1519 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in __smq_set_clear_dirty() 1552 return hash_32(from_cblock(cblock), 9) & (NR_CACHE_LEVELS - 1); in random_level() 1562 e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_load_mapping() 1580 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_invalidate_mapping() 1595 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_get_hint() [all …]
|
D | dm-cache-target.c | 676 return test_bit(from_cblock(b), cache->dirty_bitset); in is_dirty() 681 if (!test_and_set_bit(from_cblock(cblock), cache->dirty_bitset)) { in set_dirty() 693 if (!test_and_set_bit(from_cblock(cblock), cache->dirty_bitset)) in force_set_dirty() 700 if (test_and_clear_bit(from_cblock(cblock), cache->dirty_bitset)) { in force_clear_dirty() 790 sector_t block = from_cblock(cblock); in remap_to_cache() 1170 c_region.sector = from_cblock(mg->op->cblock) * cache->sectors_per_block; in copy() 2410 dm_block_t nr_blocks = from_cblock(size); in set_cache_size() 2538 cache->dirty_bitset = alloc_bitset(from_cblock(cache->cache_size)); in cache_create() 2543 clear_bitset(cache->dirty_bitset, from_cblock(cache->cache_size)); in cache_create() 2737 r = dm_cache_set_dirty_bits(cache->cmd, from_cblock(cache->cache_size), cache->dirty_bitset); in write_dirty_bitset() [all …]
|
D | dm-cache-block-types.h | 39 static inline uint32_t from_cblock(dm_cblock_t b) in from_cblock() function
|