Lines Matching refs:cache_alloc
811 struct entry_alloc cache_alloc; member
988 return to_cblock(get_index(&mq->cache_alloc, e)); in infer_cblock()
1043 unsigned threshold_level = allocator_empty(&mq->cache_alloc) ? in update_promote_levels()
1153 nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; in free_target_met()
1236 if (allocator_empty(&mq->cache_alloc)) { in queue_promotion()
1253 e = alloc_entry(&mq->cache_alloc); in queue_promotion()
1261 free_entry(&mq->cache_alloc, e); in queue_promotion()
1284 if (!allocator_empty(&mq->cache_alloc) && fast_promote) in should_promote()
1462 struct entry *e = get_entry(&mq->cache_alloc, in __complete_background_work()
1475 free_entry(&mq->cache_alloc, e); in __complete_background_work()
1484 free_entry(&mq->cache_alloc, e); in __complete_background_work()
1519 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in __smq_set_clear_dirty()
1562 e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_load_mapping()
1580 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_invalidate_mapping()
1588 free_entry(&mq->cache_alloc, e); in smq_invalidate_mapping()
1595 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_get_hint()
1610 r = to_cblock(mq->cache_alloc.nr_allocated); in smq_residency()
1758 init_allocator(&mq->cache_alloc, &mq->es, in __smq_create()