Lines Matching refs:cache_alloc
811 struct entry_alloc cache_alloc; member
994 return to_cblock(get_index(&mq->cache_alloc, e)); in infer_cblock()
1049 unsigned threshold_level = allocator_empty(&mq->cache_alloc) ? in update_promote_levels()
1159 nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; in free_target_met()
1242 if (allocator_empty(&mq->cache_alloc)) { in queue_promotion()
1259 e = alloc_entry(&mq->cache_alloc); in queue_promotion()
1267 free_entry(&mq->cache_alloc, e); in queue_promotion()
1290 if (!allocator_empty(&mq->cache_alloc) && fast_promote) in should_promote()
1468 struct entry *e = get_entry(&mq->cache_alloc, in __complete_background_work()
1481 free_entry(&mq->cache_alloc, e); in __complete_background_work()
1490 free_entry(&mq->cache_alloc, e); in __complete_background_work()
1525 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in __smq_set_clear_dirty()
1568 e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_load_mapping()
1586 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_invalidate_mapping()
1594 free_entry(&mq->cache_alloc, e); in smq_invalidate_mapping()
1601 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_get_hint()
1616 r = to_cblock(mq->cache_alloc.nr_allocated); in smq_residency()
1762 init_allocator(&mq->cache_alloc, &mq->es, in __smq_create()