root/fs/nfs/fscache.c
// SPDX-License-Identifier: GPL-2.0-or-later
/* NFS filesystem cache interface
 *
 * Copyright (C) 2008 Red Hat, Inc. All Rights Reserved.
 * Written by David Howells (dhowells@redhat.com)
 */

#include <linux/init.h>
#include <linux/kernel.h>
#include <linux/sched.h>
#include <linux/mm.h>
#include <linux/nfs_fs.h>
#include <linux/nfs_fs_sb.h>
#include <linux/in6.h>
#include <linux/seq_file.h>
#include <linux/slab.h>
#include <linux/iversion.h>
#include <linux/xarray.h>
#include <linux/fscache.h>
#include <linux/netfs.h>

#include "internal.h"
#include "iostat.h"
#include "fscache.h"
#include "nfstrace.h"

#define NFS_MAX_KEY_LEN 1000

static bool nfs_append_int(char *key, int *_len, unsigned long long x)
{
        if (*_len > NFS_MAX_KEY_LEN)
                return false;
        if (x == 0)
                key[(*_len)++] = ',';
        else
                *_len += sprintf(key + *_len, ",%llx", x);
        return true;
}

/*
 * Get the per-client index cookie for an NFS client if the appropriate mount
 * flag was set
 * - We always try and get an index cookie for the client, but get filehandle
 *   cookies on a per-superblock basis, depending on the mount flags
 */
static bool nfs_fscache_get_client_key(struct nfs_client *clp,
                                       char *key, int *_len)
{
        const struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *) &clp->cl_addr;
        const struct sockaddr_in *sin = (struct sockaddr_in *) &clp->cl_addr;

        *_len += snprintf(key + *_len, NFS_MAX_KEY_LEN - *_len,
                          ",%u.%u,%x",
                          clp->rpc_ops->version,
                          clp->cl_minorversion,
                          clp->cl_addr.ss_family);

        switch (clp->cl_addr.ss_family) {
        case AF_INET:
                if (!nfs_append_int(key, _len, sin->sin_port) ||
                    !nfs_append_int(key, _len, sin->sin_addr.s_addr))
                        return false;
                return true;

        case AF_INET6:
                if (!nfs_append_int(key, _len, sin6->sin6_port) ||
                    !nfs_append_int(key, _len, sin6->sin6_addr.s6_addr32[0]) ||
                    !nfs_append_int(key, _len, sin6->sin6_addr.s6_addr32[1]) ||
                    !nfs_append_int(key, _len, sin6->sin6_addr.s6_addr32[2]) ||
                    !nfs_append_int(key, _len, sin6->sin6_addr.s6_addr32[3]))
                        return false;
                return true;

        default:
                printk(KERN_WARNING "NFS: Unknown network family '%d'\n",
                       clp->cl_addr.ss_family);
                return false;
        }
}

/*
 * Get the cache cookie for an NFS superblock.
 *
 * The default uniquifier is just an empty string, but it may be overridden
 * either by the 'fsc=xxx' option to mount, or by inheriting it from the parent
 * superblock across an automount point of some nature.
 */
int nfs_fscache_get_super_cookie(struct super_block *sb, const char *uniq, int ulen)
{
        struct fscache_volume *vcookie;
        struct nfs_server *nfss = NFS_SB(sb);
        unsigned int len = 3;
        char *key;

        if (uniq) {
                nfss->fscache_uniq = kmemdup_nul(uniq, ulen, GFP_KERNEL);
                if (!nfss->fscache_uniq)
                        return -ENOMEM;
        }

        key = kmalloc(NFS_MAX_KEY_LEN + 24, GFP_KERNEL);
        if (!key)
                return -ENOMEM;

        memcpy(key, "nfs", 3);
        if (!nfs_fscache_get_client_key(nfss->nfs_client, key, &len) ||
            !nfs_append_int(key, &len, nfss->fsid.major) ||
            !nfs_append_int(key, &len, nfss->fsid.minor) ||
            !nfs_append_int(key, &len, sb->s_flags & NFS_SB_MASK) ||
            !nfs_append_int(key, &len, nfss->flags) ||
            !nfs_append_int(key, &len, nfss->rsize) ||
            !nfs_append_int(key, &len, nfss->wsize) ||
            !nfs_append_int(key, &len, nfss->acregmin) ||
            !nfs_append_int(key, &len, nfss->acregmax) ||
            !nfs_append_int(key, &len, nfss->acdirmin) ||
            !nfs_append_int(key, &len, nfss->acdirmax) ||
            !nfs_append_int(key, &len, nfss->client->cl_auth->au_flavor))
                goto out;

        if (ulen > 0) {
                if (ulen > NFS_MAX_KEY_LEN - len)
                        goto out;
                key[len++] = ',';
                memcpy(key + len, uniq, ulen);
                len += ulen;
        }
        key[len] = 0;

        /* create a cache index for looking up filehandles */
        vcookie = fscache_acquire_volume(key,
                                         NULL, /* preferred_cache */
                                         NULL, 0 /* coherency_data */);
        if (IS_ERR(vcookie)) {
                if (vcookie != ERR_PTR(-EBUSY)) {
                        kfree(key);
                        return PTR_ERR(vcookie);
                }
                pr_err("NFS: Cache volume key already in use (%s)\n", key);
                vcookie = NULL;
        }
        nfss->fscache = vcookie;

out:
        kfree(key);
        return 0;
}

/*
 * release a per-superblock cookie
 */
void nfs_fscache_release_super_cookie(struct super_block *sb)
{
        struct nfs_server *nfss = NFS_SB(sb);

        fscache_relinquish_volume(nfss->fscache, NULL, false);
        nfss->fscache = NULL;
        kfree(nfss->fscache_uniq);
}

/*
 * Initialise the per-inode cache cookie pointer for an NFS inode.
 */
void nfs_fscache_init_inode(struct inode *inode)
{
        struct nfs_fscache_inode_auxdata auxdata;
        struct nfs_server *nfss = NFS_SERVER(inode);
        struct nfs_inode *nfsi = NFS_I(inode);

        netfs_inode(inode)->cache = NULL;
        if (!(nfss->fscache && S_ISREG(inode->i_mode)))
                return;

        nfs_fscache_update_auxdata(&auxdata, inode);

        netfs_inode(inode)->cache = fscache_acquire_cookie(
                                               nfss->fscache,
                                               0,
                                               nfsi->fh.data, /* index_key */
                                               nfsi->fh.size,
                                               &auxdata,      /* aux_data */
                                               sizeof(auxdata),
                                               i_size_read(inode));

        if (netfs_inode(inode)->cache)
                mapping_set_release_always(inode->i_mapping);
}

/*
 * Release a per-inode cookie.
 */
void nfs_fscache_clear_inode(struct inode *inode)
{
        fscache_relinquish_cookie(netfs_i_cookie(netfs_inode(inode)), false);
        netfs_inode(inode)->cache = NULL;
}

/*
 * Enable or disable caching for a file that is being opened as appropriate.
 * The cookie is allocated when the inode is initialised, but is not enabled at
 * that time.  Enablement is deferred to file-open time to avoid stat() and
 * access() thrashing the cache.
 *
 * For now, with NFS, only regular files that are open read-only will be able
 * to use the cache.
 *
 * We enable the cache for an inode if we open it read-only and it isn't
 * currently open for writing.  We disable the cache if the inode is open
 * write-only.
 *
 * The caller uses the file struct to pin i_writecount on the inode before
 * calling us when a file is opened for writing, so we can make use of that.
 *
 * Note that this may be invoked multiple times in parallel by parallel
 * nfs_open() functions.
 */
void nfs_fscache_open_file(struct inode *inode, struct file *filp)
{
        struct nfs_fscache_inode_auxdata auxdata;
        struct fscache_cookie *cookie = netfs_i_cookie(netfs_inode(inode));
        bool open_for_write = inode_is_open_for_write(inode);

        if (!fscache_cookie_valid(cookie))
                return;

        fscache_use_cookie(cookie, open_for_write);
        if (open_for_write) {
                nfs_fscache_update_auxdata(&auxdata, inode);
                fscache_invalidate(cookie, &auxdata, i_size_read(inode),
                                   FSCACHE_INVAL_DIO_WRITE);
        }
}
EXPORT_SYMBOL_GPL(nfs_fscache_open_file);

void nfs_fscache_release_file(struct inode *inode, struct file *filp)
{
        struct nfs_fscache_inode_auxdata auxdata;
        struct fscache_cookie *cookie = netfs_i_cookie(netfs_inode(inode));
        loff_t i_size = i_size_read(inode);

        nfs_fscache_update_auxdata(&auxdata, inode);
        fscache_unuse_cookie(cookie, &auxdata, &i_size);
}

int nfs_netfs_read_folio(struct file *file, struct folio *folio)
{
        if (!netfs_inode(folio_inode(folio))->cache)
                return -ENOBUFS;

        return netfs_read_folio(file, folio);
}

int nfs_netfs_readahead(struct readahead_control *ractl)
{
        struct inode *inode = ractl->mapping->host;

        if (!netfs_inode(inode)->cache)
                return -ENOBUFS;

        netfs_readahead(ractl);
        return 0;
}

static atomic_t nfs_netfs_debug_id;
static int nfs_netfs_init_request(struct netfs_io_request *rreq, struct file *file)
{
        if (!file) {
                if (WARN_ON_ONCE(rreq->origin != NETFS_PGPRIV2_COPY_TO_CACHE))
                        return -EIO;
                return 0;
        }

        rreq->netfs_priv = get_nfs_open_context(nfs_file_open_context(file));
        rreq->debug_id = atomic_inc_return(&nfs_netfs_debug_id);
        /* [DEPRECATED] Use PG_private_2 to mark folio being written to the cache. */
        __set_bit(NETFS_RREQ_USE_PGPRIV2, &rreq->flags);
        rreq->io_streams[0].sreq_max_len = NFS_SB(rreq->inode->i_sb)->rsize;

        return 0;
}

static void nfs_netfs_free_request(struct netfs_io_request *rreq)
{
        if (rreq->netfs_priv)
                put_nfs_open_context(rreq->netfs_priv);
}

static struct nfs_netfs_io_data *nfs_netfs_alloc(struct netfs_io_subrequest *sreq)
{
        struct nfs_netfs_io_data *netfs;

        netfs = kzalloc_obj(*netfs, GFP_KERNEL_ACCOUNT);
        if (!netfs)
                return NULL;
        netfs->sreq = sreq;
        refcount_set(&netfs->refcount, 1);
        return netfs;
}

static void nfs_netfs_issue_read(struct netfs_io_subrequest *sreq)
{
        struct nfs_netfs_io_data        *netfs;
        struct nfs_pageio_descriptor    pgio;
        struct inode *inode = sreq->rreq->inode;
        struct nfs_open_context *ctx = sreq->rreq->netfs_priv;
        struct page *page;
        unsigned long idx;
        pgoff_t start, last;
        int err;

        start = (sreq->start + sreq->transferred) >> PAGE_SHIFT;
        last = ((sreq->start + sreq->len - sreq->transferred - 1) >> PAGE_SHIFT);

        nfs_pageio_init_read(&pgio, inode, false,
                             &nfs_async_read_completion_ops);

        netfs = nfs_netfs_alloc(sreq);
        if (!netfs) {
                sreq->error = -ENOMEM;
                return netfs_read_subreq_terminated(sreq);
        }

        pgio.pg_netfs = netfs; /* used in completion */

        xa_for_each_range(&sreq->rreq->mapping->i_pages, idx, page, start, last) {
                /* nfs_read_add_folio() may schedule() due to pNFS layout and other RPCs  */
                err = nfs_read_add_folio(&pgio, ctx, page_folio(page));
                if (err < 0) {
                        netfs->error = err;
                        goto out;
                }
        }
out:
        nfs_pageio_complete_read(&pgio);
        nfs_netfs_put(netfs);
}

void nfs_netfs_initiate_read(struct nfs_pgio_header *hdr)
{
        struct nfs_netfs_io_data        *netfs = hdr->netfs;

        if (!netfs)
                return;

        nfs_netfs_get(netfs);
}

int nfs_netfs_folio_unlock(struct folio *folio)
{
        struct inode *inode = folio->mapping->host;

        /*
         * If fscache is enabled, netfs will unlock pages.
         */
        if (netfs_inode(inode)->cache)
                return 0;

        return 1;
}

void nfs_netfs_read_completion(struct nfs_pgio_header *hdr)
{
        struct nfs_netfs_io_data        *netfs = hdr->netfs;
        struct netfs_io_subrequest      *sreq;

        if (!netfs)
                return;

        sreq = netfs->sreq;
        if (test_bit(NFS_IOHDR_EOF, &hdr->flags) &&
            sreq->rreq->origin != NETFS_UNBUFFERED_READ &&
            sreq->rreq->origin != NETFS_DIO_READ)
                __set_bit(NETFS_SREQ_CLEAR_TAIL, &sreq->flags);

        if (hdr->error)
                netfs->error = hdr->error;
        else
                atomic64_add(hdr->res.count, &netfs->transferred);

        nfs_netfs_put(netfs);
        hdr->netfs = NULL;
}

const struct netfs_request_ops nfs_netfs_ops = {
        .init_request           = nfs_netfs_init_request,
        .free_request           = nfs_netfs_free_request,
        .issue_read             = nfs_netfs_issue_read,
};