cache_alloc
ret = cache_alloc(ca);
unsigned int threshold_level = allocator_empty(&mq->cache_alloc) ?
nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated;
if (allocator_empty(&mq->cache_alloc)) {
e = alloc_entry(&mq->cache_alloc);
free_entry(&mq->cache_alloc, e);
if (!allocator_empty(&mq->cache_alloc) && fast_promote)
struct entry *e = get_entry(&mq->cache_alloc,
free_entry(&mq->cache_alloc, e);
free_entry(&mq->cache_alloc, e);
struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock));
struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
free_entry(&mq->cache_alloc, e);
struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
r = to_cblock(mq->cache_alloc.nr_allocated);
init_allocator(&mq->cache_alloc, &mq->es,
struct entry_alloc cache_alloc;
return to_cblock(get_index(&mq->cache_alloc, e));