root/fs/btrfs/lru_cache.h
/* SPDX-License-Identifier: GPL-2.0 */

#ifndef BTRFS_LRU_CACHE_H
#define BTRFS_LRU_CACHE_H

#include <linux/types.h>
#include <linux/maple_tree.h>
#include <linux/list.h>

/*
 * A cache entry. This is meant to be embedded in a structure of a user of
 * this module. Similar to how struct list_head and struct rb_node are used.
 *
 * Note: it should be embedded as the first element in a struct (offset 0), and
 * this module assumes it was allocated with kmalloc(), so it calls kfree() when
 * it needs to free an entry.
 */
struct btrfs_lru_cache_entry {
        struct list_head lru_list;
        u64 key;
        /*
         * Optional generation associated to a key. Use 0 if not needed/used.
         * Entries with the same key and different generations are stored in a
         * linked list, so use this only for cases where there's a small number
         * of different generations.
         */
        u64 gen;
        /*
         * The maple tree uses unsigned long type for the keys, which is 32 bits
         * on 32 bits systems, and 64 bits on 64 bits systems. So if we want to
         * use something like inode numbers as keys, which are always a u64, we
         * have to deal with this in a special way - we store the key in the
         * entry itself, as a u64, and the values inserted into the maple tree
         * are linked lists of entries - so in case we are on a 64 bits system,
         * that list always has a single entry, while on 32 bits systems it
         * may have more than one, with each entry having the same value for
         * their lower 32 bits of the u64 key.
         */
        struct list_head list;
};

struct btrfs_lru_cache {
        struct list_head lru_list;
        struct maple_tree entries;
        /* Number of entries stored in the cache. */
        unsigned int size;
        /* Maximum number of entries the cache can have. */
        unsigned int max_size;
};

#define btrfs_lru_cache_for_each_entry_safe(cache, entry, tmp)          \
        list_for_each_entry_safe_reverse((entry), (tmp), &(cache)->lru_list, lru_list)

static inline struct btrfs_lru_cache_entry *btrfs_lru_cache_lru_entry(
                                              struct btrfs_lru_cache *cache)
{
        return list_first_entry_or_null(&cache->lru_list,
                                        struct btrfs_lru_cache_entry, lru_list);
}

void btrfs_lru_cache_init(struct btrfs_lru_cache *cache, unsigned int max_size);
struct btrfs_lru_cache_entry *btrfs_lru_cache_lookup(struct btrfs_lru_cache *cache,
                                                     u64 key, u64 gen);
int btrfs_lru_cache_store(struct btrfs_lru_cache *cache,
                          struct btrfs_lru_cache_entry *new_entry,
                          gfp_t gfp);
void btrfs_lru_cache_remove(struct btrfs_lru_cache *cache,
                            struct btrfs_lru_cache_entry *entry);
void btrfs_lru_cache_clear(struct btrfs_lru_cache *cache);

#endif