bpf_mem_cache
struct bpf_mem_cache __percpu *cache;
struct bpf_mem_cache;
*(struct bpf_mem_cache **)ret = c;
struct bpf_mem_cache *c;
struct bpf_mem_cache *tgt;
struct bpf_mem_cache cache[NUM_CACHES];
static void *__alloc(struct bpf_mem_cache *c, int node, gfp_t flags)
static struct mem_cgroup *get_memcg(const struct bpf_mem_cache *c)
static void inc_active(struct bpf_mem_cache *c, unsigned long *flags)
static void dec_active(struct bpf_mem_cache *c, unsigned long *flags)
static void add_obj_to_free_list(struct bpf_mem_cache *c, void *obj)
static void alloc_bulk(struct bpf_mem_cache *c, int cnt, int node, bool atomic)
static int free_all(struct bpf_mem_cache *c, struct llist_node *llnode, bool percpu)
struct bpf_mem_cache *c = container_of(head, struct bpf_mem_cache, rcu_ttrace);
static void enque_to_free(struct bpf_mem_cache *c, void *obj)
static void do_call_rcu_ttrace(struct bpf_mem_cache *c)
static void free_bulk(struct bpf_mem_cache *c)
struct bpf_mem_cache *tgt = c->tgt;
struct bpf_mem_cache *c = container_of(head, struct bpf_mem_cache, rcu);
struct bpf_mem_cache *tgt = c->tgt;
static void check_free_by_rcu(struct bpf_mem_cache *c)
struct bpf_mem_cache *c = container_of(work, struct bpf_mem_cache, refill_work);
static void notrace irq_work_raise(struct bpf_mem_cache *c)
static void init_refill_work(struct bpf_mem_cache *c)
static void prefill_mem_cache(struct bpf_mem_cache *c, int cpu)
struct bpf_mem_cache *c; struct bpf_mem_cache __percpu *pc;
struct bpf_mem_cache *c;
static void drain_mem_cache(struct bpf_mem_cache *c)
static void check_mem_cache(struct bpf_mem_cache *c)
struct bpf_mem_cache *c;
struct bpf_mem_cache *c;
static void notrace *unit_alloc(struct bpf_mem_cache *c)
*(struct bpf_mem_cache **)llnode = c;
static void notrace unit_free(struct bpf_mem_cache *c, void *ptr)
c->tgt = *(struct bpf_mem_cache **)llnode;
static void notrace unit_free_rcu(struct bpf_mem_cache *c, void *ptr)
c->tgt = *(struct bpf_mem_cache **)llnode;
struct bpf_mem_cache *c;
struct bpf_mem_cache *c;
struct bpf_mem_cache *c;