root/fs/cachefiles/cache.c
// SPDX-License-Identifier: GPL-2.0-or-later
/* Manage high-level VFS aspects of a cache.
 *
 * Copyright (C) 2007, 2021 Red Hat, Inc. All Rights Reserved.
 * Written by David Howells (dhowells@redhat.com)
 */

#include <linux/slab.h>
#include <linux/statfs.h>
#include <linux/namei.h>
#include <trace/events/fscache.h>
#include "internal.h"

/*
 * Bring a cache online.
 */
int cachefiles_add_cache(struct cachefiles_cache *cache)
{
        struct fscache_cache *cache_cookie;
        struct path path;
        struct kstatfs stats;
        struct dentry *graveyard, *cachedir, *root;
        const struct cred *saved_cred;
        int ret;

        _enter("");

        cache_cookie = fscache_acquire_cache(cache->tag);
        if (IS_ERR(cache_cookie))
                return PTR_ERR(cache_cookie);

        /* we want to work under the module's security ID */
        ret = cachefiles_get_security_ID(cache);
        if (ret < 0)
                goto error_getsec;

        cachefiles_begin_secure(cache, &saved_cred);

        /* look up the directory at the root of the cache */
        ret = kern_path(cache->rootdirname, LOOKUP_DIRECTORY, &path);
        if (ret < 0)
                goto error_open_root;

        cache->mnt = path.mnt;
        root = path.dentry;

        ret = -EINVAL;
        if (is_idmapped_mnt(path.mnt)) {
                pr_warn("File cache on idmapped mounts not supported");
                goto error_unsupported;
        }

        /* Check features of the backing filesystem:
         * - Directories must support looking up and directory creation
         * - We create tmpfiles to handle invalidation
         * - We use xattrs to store metadata
         * - We need to be able to query the amount of space available
         * - We want to be able to sync the filesystem when stopping the cache
         * - We use DIO to/from pages, so the blocksize mustn't be too big.
         */
        ret = -EOPNOTSUPP;
        if (d_is_negative(root) ||
            !d_backing_inode(root)->i_op->lookup ||
            !d_backing_inode(root)->i_op->mkdir ||
            !d_backing_inode(root)->i_op->tmpfile ||
            !(d_backing_inode(root)->i_opflags & IOP_XATTR) ||
            !root->d_sb->s_op->statfs ||
            !root->d_sb->s_op->sync_fs ||
            root->d_sb->s_blocksize > PAGE_SIZE)
                goto error_unsupported;

        ret = -EROFS;
        if (sb_rdonly(root->d_sb))
                goto error_unsupported;

        /* determine the security of the on-disk cache as this governs
         * security ID of files we create */
        ret = cachefiles_determine_cache_security(cache, root, &saved_cred);
        if (ret < 0)
                goto error_unsupported;

        /* get the cache size and blocksize */
        ret = vfs_statfs(&path, &stats);
        if (ret < 0)
                goto error_unsupported;

        ret = -ERANGE;
        if (stats.f_bsize <= 0)
                goto error_unsupported;

        ret = -EOPNOTSUPP;
        if (stats.f_bsize > PAGE_SIZE)
                goto error_unsupported;

        cache->bsize = stats.f_bsize;
        cache->bshift = ilog2(stats.f_bsize);

        _debug("blksize %u (shift %u)",
               cache->bsize, cache->bshift);

        _debug("size %llu, avail %llu",
               (unsigned long long) stats.f_blocks,
               (unsigned long long) stats.f_bavail);

        /* set up caching limits */
        do_div(stats.f_files, 100);
        cache->fstop = stats.f_files * cache->fstop_percent;
        cache->fcull = stats.f_files * cache->fcull_percent;
        cache->frun  = stats.f_files * cache->frun_percent;

        _debug("limits {%llu,%llu,%llu} files",
               (unsigned long long) cache->frun,
               (unsigned long long) cache->fcull,
               (unsigned long long) cache->fstop);

        do_div(stats.f_blocks, 100);
        cache->bstop = stats.f_blocks * cache->bstop_percent;
        cache->bcull = stats.f_blocks * cache->bcull_percent;
        cache->brun  = stats.f_blocks * cache->brun_percent;

        _debug("limits {%llu,%llu,%llu} blocks",
               (unsigned long long) cache->brun,
               (unsigned long long) cache->bcull,
               (unsigned long long) cache->bstop);

        /* get the cache directory and check its type */
        cachedir = cachefiles_get_directory(cache, root, "cache", NULL);
        if (IS_ERR(cachedir)) {
                ret = PTR_ERR(cachedir);
                goto error_unsupported;
        }

        cache->store = cachedir;

        /* get the graveyard directory */
        graveyard = cachefiles_get_directory(cache, root, "graveyard", NULL);
        if (IS_ERR(graveyard)) {
                ret = PTR_ERR(graveyard);
                goto error_unsupported;
        }

        cache->graveyard = graveyard;
        cache->cache = cache_cookie;

        ret = fscache_add_cache(cache_cookie, &cachefiles_cache_ops, cache);
        if (ret < 0)
                goto error_add_cache;

        /* done */
        set_bit(CACHEFILES_READY, &cache->flags);
        dput(root);

        pr_info("File cache on %s registered\n", cache_cookie->name);

        /* check how much space the cache has */
        cachefiles_has_space(cache, 0, 0, cachefiles_has_space_check);
        cachefiles_end_secure(cache, saved_cred);
        _leave(" = 0 [%px]", cache->cache);
        return 0;

error_add_cache:
        cachefiles_put_directory(cache->graveyard);
        cache->graveyard = NULL;
error_unsupported:
        cachefiles_put_directory(cache->store);
        cache->store = NULL;
        mntput(cache->mnt);
        cache->mnt = NULL;
        dput(root);
error_open_root:
        cachefiles_end_secure(cache, saved_cred);
        put_cred(cache->cache_cred);
        cache->cache_cred = NULL;
error_getsec:
        fscache_relinquish_cache(cache_cookie);
        cache->cache = NULL;
        pr_err("Failed to register: %d\n", ret);
        return ret;
}

/*
 * See if we have space for a number of pages and/or a number of files in the
 * cache
 */
int cachefiles_has_space(struct cachefiles_cache *cache,
                         unsigned fnr, unsigned bnr,
                         enum cachefiles_has_space_for reason)
{
        struct kstatfs stats;
        u64 b_avail, b_writing;
        int ret;

        struct path path = {
                .mnt    = cache->mnt,
                .dentry = cache->mnt->mnt_root,
        };

        //_enter("{%llu,%llu,%llu,%llu,%llu,%llu},%u,%u",
        //       (unsigned long long) cache->frun,
        //       (unsigned long long) cache->fcull,
        //       (unsigned long long) cache->fstop,
        //       (unsigned long long) cache->brun,
        //       (unsigned long long) cache->bcull,
        //       (unsigned long long) cache->bstop,
        //       fnr, bnr);

        /* find out how many pages of blockdev are available */
        memset(&stats, 0, sizeof(stats));

        ret = vfs_statfs(&path, &stats);
        if (ret < 0) {
                trace_cachefiles_vfs_error(NULL, d_inode(path.dentry), ret,
                                           cachefiles_trace_statfs_error);
                if (ret == -EIO)
                        cachefiles_io_error(cache, "statfs failed");
                _leave(" = %d", ret);
                return ret;
        }

        b_avail = stats.f_bavail;
        b_writing = atomic_long_read(&cache->b_writing);
        if (b_avail > b_writing)
                b_avail -= b_writing;
        else
                b_avail = 0;

        //_debug("avail %llu,%llu",
        //       (unsigned long long)stats.f_ffree,
        //       (unsigned long long)b_avail);

        /* see if there is sufficient space */
        if (stats.f_ffree > fnr)
                stats.f_ffree -= fnr;
        else
                stats.f_ffree = 0;

        if (b_avail > bnr)
                b_avail -= bnr;
        else
                b_avail = 0;

        ret = -ENOBUFS;
        if (stats.f_ffree < cache->fstop ||
            b_avail < cache->bstop)
                goto stop_and_begin_cull;

        ret = 0;
        if (stats.f_ffree < cache->fcull ||
            b_avail < cache->bcull)
                goto begin_cull;

        if (test_bit(CACHEFILES_CULLING, &cache->flags) &&
            stats.f_ffree >= cache->frun &&
            b_avail >= cache->brun &&
            test_and_clear_bit(CACHEFILES_CULLING, &cache->flags)
            ) {
                _debug("cease culling");
                cachefiles_state_changed(cache);
        }

        //_leave(" = 0");
        return 0;

stop_and_begin_cull:
        switch (reason) {
        case cachefiles_has_space_for_write:
                fscache_count_no_write_space();
                break;
        case cachefiles_has_space_for_create:
                fscache_count_no_create_space();
                break;
        default:
                break;
        }
begin_cull:
        if (!test_and_set_bit(CACHEFILES_CULLING, &cache->flags)) {
                _debug("### CULL CACHE ###");
                cachefiles_state_changed(cache);
        }

        _leave(" = %d", ret);
        return ret;
}

/*
 * Mark all the objects as being out of service and queue them all for cleanup.
 */
static void cachefiles_withdraw_objects(struct cachefiles_cache *cache)
{
        struct cachefiles_object *object;
        unsigned int count = 0;

        _enter("");

        spin_lock(&cache->object_list_lock);

        while (!list_empty(&cache->object_list)) {
                object = list_first_entry(&cache->object_list,
                                          struct cachefiles_object, cache_link);
                cachefiles_see_object(object, cachefiles_obj_see_withdrawal);
                list_del_init(&object->cache_link);
                fscache_withdraw_cookie(object->cookie);
                count++;
                if ((count & 63) == 0) {
                        spin_unlock(&cache->object_list_lock);
                        cond_resched();
                        spin_lock(&cache->object_list_lock);
                }
        }

        spin_unlock(&cache->object_list_lock);
        _leave(" [%u objs]", count);
}

/*
 * Withdraw fscache volumes.
 */
static void cachefiles_withdraw_fscache_volumes(struct cachefiles_cache *cache)
{
        struct list_head *cur;
        struct cachefiles_volume *volume;
        struct fscache_volume *vcookie;

        _enter("");
retry:
        spin_lock(&cache->object_list_lock);
        list_for_each(cur, &cache->volumes) {
                volume = list_entry(cur, struct cachefiles_volume, cache_link);

                if (atomic_read(&volume->vcookie->n_accesses) == 0)
                        continue;

                vcookie = fscache_try_get_volume(volume->vcookie,
                                                 fscache_volume_get_withdraw);
                if (vcookie) {
                        spin_unlock(&cache->object_list_lock);
                        fscache_withdraw_volume(vcookie);
                        fscache_put_volume(vcookie, fscache_volume_put_withdraw);
                        goto retry;
                }
        }
        spin_unlock(&cache->object_list_lock);

        _leave("");
}

/*
 * Withdraw cachefiles volumes.
 */
static void cachefiles_withdraw_volumes(struct cachefiles_cache *cache)
{
        _enter("");

        for (;;) {
                struct fscache_volume *vcookie = NULL;
                struct cachefiles_volume *volume = NULL;

                spin_lock(&cache->object_list_lock);
                if (!list_empty(&cache->volumes)) {
                        volume = list_first_entry(&cache->volumes,
                                                  struct cachefiles_volume, cache_link);
                        vcookie = fscache_try_get_volume(volume->vcookie,
                                                         fscache_volume_get_withdraw);
                        if (!vcookie) {
                                spin_unlock(&cache->object_list_lock);
                                cpu_relax();
                                continue;
                        }
                        list_del_init(&volume->cache_link);
                }
                spin_unlock(&cache->object_list_lock);
                if (!volume)
                        break;

                cachefiles_withdraw_volume(volume);
                fscache_put_volume(vcookie, fscache_volume_put_withdraw);
        }

        _leave("");
}

/*
 * Sync a cache to backing disk.
 */
static void cachefiles_sync_cache(struct cachefiles_cache *cache)
{
        const struct cred *saved_cred;
        int ret;

        _enter("%s", cache->cache->name);

        /* make sure all pages pinned by operations on behalf of the netfs are
         * written to disc */
        cachefiles_begin_secure(cache, &saved_cred);
        down_read(&cache->mnt->mnt_sb->s_umount);
        ret = sync_filesystem(cache->mnt->mnt_sb);
        up_read(&cache->mnt->mnt_sb->s_umount);
        cachefiles_end_secure(cache, saved_cred);

        if (ret == -EIO)
                cachefiles_io_error(cache,
                                    "Attempt to sync backing fs superblock returned error %d",
                                    ret);
}

/*
 * Withdraw cache objects.
 */
void cachefiles_withdraw_cache(struct cachefiles_cache *cache)
{
        struct fscache_cache *fscache = cache->cache;

        pr_info("File cache on %s unregistering\n", fscache->name);

        fscache_withdraw_cache(fscache);
        cachefiles_withdraw_fscache_volumes(cache);

        /* we now have to destroy all the active objects pertaining to this
         * cache - which we do by passing them off to thread pool to be
         * disposed of */
        cachefiles_withdraw_objects(cache);
        fscache_wait_for_objects(fscache);

        cachefiles_withdraw_volumes(cache);
        cachefiles_sync_cache(cache);
        cache->cache = NULL;
        fscache_relinquish_cache(fscache);
}