root/fs/smb/server/vfs_cache.c
// SPDX-License-Identifier: GPL-2.0-or-later
/*
 * Copyright (C) 2016 Namjae Jeon <linkinjeon@kernel.org>
 * Copyright (C) 2019 Samsung Electronics Co., Ltd.
 */

#include <linux/fs.h>
#include <linux/filelock.h>
#include <linux/slab.h>
#include <linux/vmalloc.h>
#include <linux/kthread.h>
#include <linux/freezer.h>

#include "glob.h"
#include "vfs_cache.h"
#include "oplock.h"
#include "vfs.h"
#include "connection.h"
#include "misc.h"
#include "mgmt/tree_connect.h"
#include "mgmt/user_session.h"
#include "smb_common.h"
#include "server.h"
#include "smb2pdu.h"

#define S_DEL_PENDING                   1
#define S_DEL_ON_CLS                    2
#define S_DEL_ON_CLS_STREAM             8

static unsigned int inode_hash_mask __read_mostly;
static unsigned int inode_hash_shift __read_mostly;
static struct hlist_head *inode_hashtable __read_mostly;
static DEFINE_RWLOCK(inode_hash_lock);

static struct ksmbd_file_table global_ft;
static atomic_long_t fd_limit;
static struct kmem_cache *filp_cache;

#define OPLOCK_NONE      0
#define OPLOCK_EXCLUSIVE 1
#define OPLOCK_BATCH     2
#define OPLOCK_READ      3  /* level 2 oplock */

#ifdef CONFIG_PROC_FS

static const struct ksmbd_const_name ksmbd_lease_const_names[] = {
        {le32_to_cpu(SMB2_LEASE_NONE_LE), "LEASE_NONE"},
        {le32_to_cpu(SMB2_LEASE_READ_CACHING_LE), "LEASE_R"},
        {le32_to_cpu(SMB2_LEASE_HANDLE_CACHING_LE), "LEASE_H"},
        {le32_to_cpu(SMB2_LEASE_WRITE_CACHING_LE), "LEASE_W"},
        {le32_to_cpu(SMB2_LEASE_READ_CACHING_LE |
                     SMB2_LEASE_HANDLE_CACHING_LE), "LEASE_RH"},
        {le32_to_cpu(SMB2_LEASE_READ_CACHING_LE |
                     SMB2_LEASE_WRITE_CACHING_LE), "LEASE_RW"},
        {le32_to_cpu(SMB2_LEASE_HANDLE_CACHING_LE |
                     SMB2_LEASE_WRITE_CACHING_LE), "LEASE_WH"},
        {le32_to_cpu(SMB2_LEASE_READ_CACHING_LE |
                     SMB2_LEASE_HANDLE_CACHING_LE |
                     SMB2_LEASE_WRITE_CACHING_LE), "LEASE_RWH"},
};

static const struct ksmbd_const_name ksmbd_oplock_const_names[] = {
        {SMB2_OPLOCK_LEVEL_NONE, "OPLOCK_NONE"},
        {SMB2_OPLOCK_LEVEL_II, "OPLOCK_II"},
        {SMB2_OPLOCK_LEVEL_EXCLUSIVE, "OPLOCK_EXECL"},
        {SMB2_OPLOCK_LEVEL_BATCH, "OPLOCK_BATCH"},
};

static int proc_show_files(struct seq_file *m, void *v)
{
        struct ksmbd_file *fp = NULL;
        unsigned int id;
        struct oplock_info *opinfo;

        seq_printf(m, "#%-10s %-10s %-10s %-10s %-15s %-10s %-10s %s\n",
                   "<tree id>", "<pid>", "<vid>", "<refcnt>",
                   "<oplock>", "<daccess>", "<saccess>",
                   "<name>");

        read_lock(&global_ft.lock);
        idr_for_each_entry(global_ft.idr, fp, id) {
                seq_printf(m, "%#-10x %#-10llx %#-10llx %#-10x",
                           fp->tcon->id,
                           fp->persistent_id,
                           fp->volatile_id,
                           atomic_read(&fp->refcount));

                rcu_read_lock();
                opinfo = rcu_dereference(fp->f_opinfo);
                if (opinfo) {
                        const struct ksmbd_const_name *const_names;
                        int count;
                        unsigned int level;

                        if (opinfo->is_lease) {
                                const_names = ksmbd_lease_const_names;
                                count = ARRAY_SIZE(ksmbd_lease_const_names);
                                level = le32_to_cpu(opinfo->o_lease->state);
                        } else {
                                const_names = ksmbd_oplock_const_names;
                                count = ARRAY_SIZE(ksmbd_oplock_const_names);
                                level = opinfo->level;
                        }
                        rcu_read_unlock();
                        ksmbd_proc_show_const_name(m, " %-15s",
                                                   const_names, count, level);
                } else {
                        rcu_read_unlock();
                        seq_printf(m, " %-15s", " ");
                }

                seq_printf(m, " %#010x %#010x %s\n",
                           le32_to_cpu(fp->daccess),
                           le32_to_cpu(fp->saccess),
                           fp->filp->f_path.dentry->d_name.name);
        }
        read_unlock(&global_ft.lock);
        return 0;
}

static int create_proc_files(void)
{
        ksmbd_proc_create("files", proc_show_files, NULL);
        return 0;
}
#else
static int create_proc_files(void) { return 0; }
#endif

static bool durable_scavenger_running;
static DEFINE_MUTEX(durable_scavenger_lock);
static wait_queue_head_t dh_wq;

void ksmbd_set_fd_limit(unsigned long limit)
{
        limit = min(limit, get_max_files());
        atomic_long_set(&fd_limit, limit);
}

static bool fd_limit_depleted(void)
{
        long v = atomic_long_dec_return(&fd_limit);

        if (v >= 0)
                return false;
        atomic_long_inc(&fd_limit);
        return true;
}

static void fd_limit_close(void)
{
        atomic_long_inc(&fd_limit);
}

/*
 * INODE hash
 */

static unsigned long inode_hash(struct super_block *sb, unsigned long hashval)
{
        unsigned long tmp;

        tmp = (hashval * (unsigned long)sb) ^ (GOLDEN_RATIO_PRIME + hashval) /
                L1_CACHE_BYTES;
        tmp = tmp ^ ((tmp ^ GOLDEN_RATIO_PRIME) >> inode_hash_shift);
        return tmp & inode_hash_mask;
}

static struct ksmbd_inode *__ksmbd_inode_lookup(struct dentry *de)
{
        struct hlist_head *head = inode_hashtable +
                inode_hash(d_inode(de)->i_sb, (unsigned long)de);
        struct ksmbd_inode *ci = NULL, *ret_ci = NULL;

        hlist_for_each_entry(ci, head, m_hash) {
                if (ci->m_de == de) {
                        if (atomic_inc_not_zero(&ci->m_count))
                                ret_ci = ci;
                        break;
                }
        }
        return ret_ci;
}

static struct ksmbd_inode *ksmbd_inode_lookup(struct ksmbd_file *fp)
{
        return __ksmbd_inode_lookup(fp->filp->f_path.dentry);
}

struct ksmbd_inode *ksmbd_inode_lookup_lock(struct dentry *d)
{
        struct ksmbd_inode *ci;

        read_lock(&inode_hash_lock);
        ci = __ksmbd_inode_lookup(d);
        read_unlock(&inode_hash_lock);

        return ci;
}

int ksmbd_query_inode_status(struct dentry *dentry)
{
        struct ksmbd_inode *ci;
        int ret = KSMBD_INODE_STATUS_UNKNOWN;

        read_lock(&inode_hash_lock);
        ci = __ksmbd_inode_lookup(dentry);
        read_unlock(&inode_hash_lock);
        if (!ci)
                return ret;

        down_read(&ci->m_lock);
        if (ci->m_flags & (S_DEL_PENDING | S_DEL_ON_CLS))
                ret = KSMBD_INODE_STATUS_PENDING_DELETE;
        else
                ret = KSMBD_INODE_STATUS_OK;
        up_read(&ci->m_lock);

        atomic_dec(&ci->m_count);
        return ret;
}

bool ksmbd_inode_pending_delete(struct ksmbd_file *fp)
{
        struct ksmbd_inode *ci = fp->f_ci;
        int ret;

        down_read(&ci->m_lock);
        ret = (ci->m_flags & (S_DEL_PENDING | S_DEL_ON_CLS));
        up_read(&ci->m_lock);

        return ret;
}

void ksmbd_set_inode_pending_delete(struct ksmbd_file *fp)
{
        struct ksmbd_inode *ci = fp->f_ci;

        down_write(&ci->m_lock);
        ci->m_flags |= S_DEL_PENDING;
        up_write(&ci->m_lock);
}

void ksmbd_clear_inode_pending_delete(struct ksmbd_file *fp)
{
        struct ksmbd_inode *ci = fp->f_ci;

        down_write(&ci->m_lock);
        ci->m_flags &= ~S_DEL_PENDING;
        up_write(&ci->m_lock);
}

void ksmbd_fd_set_delete_on_close(struct ksmbd_file *fp,
                                  int file_info)
{
        struct ksmbd_inode *ci = fp->f_ci;

        down_write(&ci->m_lock);
        if (ksmbd_stream_fd(fp))
                ci->m_flags |= S_DEL_ON_CLS_STREAM;
        else
                ci->m_flags |= S_DEL_ON_CLS;
        up_write(&ci->m_lock);
}

static void ksmbd_inode_hash(struct ksmbd_inode *ci)
{
        struct hlist_head *b = inode_hashtable +
                inode_hash(d_inode(ci->m_de)->i_sb, (unsigned long)ci->m_de);

        hlist_add_head(&ci->m_hash, b);
}

static void ksmbd_inode_unhash(struct ksmbd_inode *ci)
{
        write_lock(&inode_hash_lock);
        hlist_del_init(&ci->m_hash);
        write_unlock(&inode_hash_lock);
}

static int ksmbd_inode_init(struct ksmbd_inode *ci, struct ksmbd_file *fp)
{
        atomic_set(&ci->m_count, 1);
        atomic_set(&ci->op_count, 0);
        atomic_set(&ci->sop_count, 0);
        ci->m_flags = 0;
        ci->m_fattr = 0;
        INIT_LIST_HEAD(&ci->m_fp_list);
        INIT_LIST_HEAD(&ci->m_op_list);
        init_rwsem(&ci->m_lock);
        ci->m_de = fp->filp->f_path.dentry;
        return 0;
}

static struct ksmbd_inode *ksmbd_inode_get(struct ksmbd_file *fp)
{
        struct ksmbd_inode *ci, *tmpci;
        int rc;

        read_lock(&inode_hash_lock);
        ci = ksmbd_inode_lookup(fp);
        read_unlock(&inode_hash_lock);
        if (ci)
                return ci;

        ci = kmalloc_obj(struct ksmbd_inode, KSMBD_DEFAULT_GFP);
        if (!ci)
                return NULL;

        rc = ksmbd_inode_init(ci, fp);
        if (rc) {
                pr_err("inode initialized failed\n");
                kfree(ci);
                return NULL;
        }

        write_lock(&inode_hash_lock);
        tmpci = ksmbd_inode_lookup(fp);
        if (!tmpci) {
                ksmbd_inode_hash(ci);
        } else {
                kfree(ci);
                ci = tmpci;
        }
        write_unlock(&inode_hash_lock);
        return ci;
}

static void ksmbd_inode_free(struct ksmbd_inode *ci)
{
        ksmbd_inode_unhash(ci);
        kfree(ci);
}

void ksmbd_inode_put(struct ksmbd_inode *ci)
{
        if (atomic_dec_and_test(&ci->m_count))
                ksmbd_inode_free(ci);
}

int __init ksmbd_inode_hash_init(void)
{
        unsigned int loop;
        unsigned long numentries = 16384;
        unsigned long bucketsize = sizeof(struct hlist_head);
        unsigned long size;

        inode_hash_shift = ilog2(numentries);
        inode_hash_mask = (1 << inode_hash_shift) - 1;

        size = bucketsize << inode_hash_shift;

        /* init master fp hash table */
        inode_hashtable = vmalloc(size);
        if (!inode_hashtable)
                return -ENOMEM;

        for (loop = 0; loop < (1U << inode_hash_shift); loop++)
                INIT_HLIST_HEAD(&inode_hashtable[loop]);
        return 0;
}

void ksmbd_release_inode_hash(void)
{
        vfree(inode_hashtable);
}

static void __ksmbd_inode_close(struct ksmbd_file *fp)
{
        struct ksmbd_inode *ci = fp->f_ci;
        int err;
        struct file *filp;

        filp = fp->filp;

        if (ksmbd_stream_fd(fp)) {
                bool remove_stream_xattr = false;

                down_write(&ci->m_lock);
                if (ci->m_flags & S_DEL_ON_CLS_STREAM) {
                        ci->m_flags &= ~S_DEL_ON_CLS_STREAM;
                        remove_stream_xattr = true;
                }
                up_write(&ci->m_lock);

                if (remove_stream_xattr) {
                        err = ksmbd_vfs_remove_xattr(file_mnt_idmap(filp),
                                                     &filp->f_path,
                                                     fp->stream.name,
                                                     true);
                        if (err)
                                pr_err("remove xattr failed : %s\n",
                                       fp->stream.name);
                }
        }

        if (atomic_dec_and_test(&ci->m_count)) {
                bool do_unlink = false;

                down_write(&ci->m_lock);
                if (ci->m_flags & (S_DEL_ON_CLS | S_DEL_PENDING)) {
                        ci->m_flags &= ~(S_DEL_ON_CLS | S_DEL_PENDING);
                        do_unlink = true;
                }
                up_write(&ci->m_lock);

                if (do_unlink)
                        ksmbd_vfs_unlink(filp);

                ksmbd_inode_free(ci);
        }
}

static void __ksmbd_remove_durable_fd(struct ksmbd_file *fp)
{
        if (!has_file_id(fp->persistent_id))
                return;

        idr_remove(global_ft.idr, fp->persistent_id);
}

static void ksmbd_remove_durable_fd(struct ksmbd_file *fp)
{
        write_lock(&global_ft.lock);
        __ksmbd_remove_durable_fd(fp);
        write_unlock(&global_ft.lock);
        if (waitqueue_active(&dh_wq))
                wake_up(&dh_wq);
}

static void __ksmbd_remove_fd(struct ksmbd_file_table *ft, struct ksmbd_file *fp)
{
        if (!has_file_id(fp->volatile_id))
                return;

        down_write(&fp->f_ci->m_lock);
        list_del_init(&fp->node);
        up_write(&fp->f_ci->m_lock);

        write_lock(&ft->lock);
        idr_remove(ft->idr, fp->volatile_id);
        write_unlock(&ft->lock);
}

static void __ksmbd_close_fd(struct ksmbd_file_table *ft, struct ksmbd_file *fp)
{
        struct file *filp;
        struct ksmbd_lock *smb_lock, *tmp_lock;

        fd_limit_close();
        ksmbd_remove_durable_fd(fp);
        if (ft)
                __ksmbd_remove_fd(ft, fp);

        close_id_del_oplock(fp);
        filp = fp->filp;

        __ksmbd_inode_close(fp);
        if (!IS_ERR_OR_NULL(filp))
                fput(filp);

        /* because the reference count of fp is 0, it is guaranteed that
         * there are not accesses to fp->lock_list.
         */
        list_for_each_entry_safe(smb_lock, tmp_lock, &fp->lock_list, flist) {
                spin_lock(&fp->conn->llist_lock);
                list_del(&smb_lock->clist);
                spin_unlock(&fp->conn->llist_lock);

                list_del(&smb_lock->flist);
                locks_free_lock(smb_lock->fl);
                kfree(smb_lock);
        }

        if (ksmbd_stream_fd(fp))
                kfree(fp->stream.name);
        kmem_cache_free(filp_cache, fp);
}

static struct ksmbd_file *ksmbd_fp_get(struct ksmbd_file *fp)
{
        if (fp->f_state != FP_INITED)
                return NULL;

        if (!atomic_inc_not_zero(&fp->refcount))
                return NULL;
        return fp;
}

static struct ksmbd_file *__ksmbd_lookup_fd(struct ksmbd_file_table *ft,
                                            u64 id)
{
        struct ksmbd_file *fp;

        if (!has_file_id(id))
                return NULL;

        read_lock(&ft->lock);
        fp = idr_find(ft->idr, id);
        if (fp)
                fp = ksmbd_fp_get(fp);
        read_unlock(&ft->lock);
        return fp;
}

static void __put_fd_final(struct ksmbd_work *work, struct ksmbd_file *fp)
{
        __ksmbd_close_fd(&work->sess->file_table, fp);
        atomic_dec(&work->conn->stats.open_files_count);
}

static void set_close_state_blocked_works(struct ksmbd_file *fp)
{
        struct ksmbd_work *cancel_work;

        spin_lock(&fp->f_lock);
        list_for_each_entry(cancel_work, &fp->blocked_works,
                                 fp_entry) {
                cancel_work->state = KSMBD_WORK_CLOSED;
                cancel_work->cancel_fn(cancel_work->cancel_argv);
        }
        spin_unlock(&fp->f_lock);
}

int ksmbd_close_fd(struct ksmbd_work *work, u64 id)
{
        struct ksmbd_file       *fp;
        struct ksmbd_file_table *ft;

        if (!has_file_id(id))
                return 0;

        ft = &work->sess->file_table;
        write_lock(&ft->lock);
        fp = idr_find(ft->idr, id);
        if (fp) {
                set_close_state_blocked_works(fp);

                if (fp->f_state != FP_INITED)
                        fp = NULL;
                else {
                        fp->f_state = FP_CLOSED;
                        if (!atomic_dec_and_test(&fp->refcount))
                                fp = NULL;
                }
        }
        write_unlock(&ft->lock);

        if (!fp)
                return -EINVAL;

        __put_fd_final(work, fp);
        return 0;
}

void ksmbd_fd_put(struct ksmbd_work *work, struct ksmbd_file *fp)
{
        if (!fp)
                return;

        if (!atomic_dec_and_test(&fp->refcount))
                return;
        __put_fd_final(work, fp);
}

static bool __sanity_check(struct ksmbd_tree_connect *tcon, struct ksmbd_file *fp)
{
        if (!fp)
                return false;
        if (fp->tcon != tcon)
                return false;
        return true;
}

struct ksmbd_file *ksmbd_lookup_foreign_fd(struct ksmbd_work *work, u64 id)
{
        return __ksmbd_lookup_fd(&work->sess->file_table, id);
}

struct ksmbd_file *ksmbd_lookup_fd_fast(struct ksmbd_work *work, u64 id)
{
        struct ksmbd_file *fp = __ksmbd_lookup_fd(&work->sess->file_table, id);

        if (__sanity_check(work->tcon, fp))
                return fp;

        ksmbd_fd_put(work, fp);
        return NULL;
}

struct ksmbd_file *ksmbd_lookup_fd_slow(struct ksmbd_work *work, u64 id,
                                        u64 pid)
{
        struct ksmbd_file *fp;

        if (!has_file_id(id)) {
                id = work->compound_fid;
                pid = work->compound_pfid;
        }

        fp = __ksmbd_lookup_fd(&work->sess->file_table, id);
        if (!__sanity_check(work->tcon, fp)) {
                ksmbd_fd_put(work, fp);
                return NULL;
        }
        if (fp->persistent_id != pid) {
                ksmbd_fd_put(work, fp);
                return NULL;
        }
        return fp;
}

struct ksmbd_file *ksmbd_lookup_global_fd(unsigned long long id)
{
        return __ksmbd_lookup_fd(&global_ft, id);
}

struct ksmbd_file *ksmbd_lookup_durable_fd(unsigned long long id)
{
        struct ksmbd_file *fp;

        fp = __ksmbd_lookup_fd(&global_ft, id);
        if (fp && (fp->conn ||
                   (fp->durable_scavenger_timeout &&
                    (fp->durable_scavenger_timeout <
                     jiffies_to_msecs(jiffies))))) {
                ksmbd_put_durable_fd(fp);
                fp = NULL;
        }

        return fp;
}

void ksmbd_put_durable_fd(struct ksmbd_file *fp)
{
        if (!atomic_dec_and_test(&fp->refcount))
                return;

        __ksmbd_close_fd(NULL, fp);
}

struct ksmbd_file *ksmbd_lookup_fd_cguid(char *cguid)
{
        struct ksmbd_file       *fp = NULL;
        unsigned int            id;

        read_lock(&global_ft.lock);
        idr_for_each_entry(global_ft.idr, fp, id) {
                if (!memcmp(fp->create_guid,
                            cguid,
                            SMB2_CREATE_GUID_SIZE)) {
                        fp = ksmbd_fp_get(fp);
                        break;
                }
        }
        read_unlock(&global_ft.lock);

        return fp;
}

struct ksmbd_file *ksmbd_lookup_fd_inode(struct dentry *dentry)
{
        struct ksmbd_file       *lfp;
        struct ksmbd_inode      *ci;
        struct inode            *inode = d_inode(dentry);

        read_lock(&inode_hash_lock);
        ci = __ksmbd_inode_lookup(dentry);
        read_unlock(&inode_hash_lock);
        if (!ci)
                return NULL;

        down_read(&ci->m_lock);
        list_for_each_entry(lfp, &ci->m_fp_list, node) {
                if (inode == file_inode(lfp->filp)) {
                        atomic_dec(&ci->m_count);
                        lfp = ksmbd_fp_get(lfp);
                        up_read(&ci->m_lock);
                        return lfp;
                }
        }
        atomic_dec(&ci->m_count);
        up_read(&ci->m_lock);
        return NULL;
}

#define OPEN_ID_TYPE_VOLATILE_ID        (0)
#define OPEN_ID_TYPE_PERSISTENT_ID      (1)

static void __open_id_set(struct ksmbd_file *fp, u64 id, int type)
{
        if (type == OPEN_ID_TYPE_VOLATILE_ID)
                fp->volatile_id = id;
        if (type == OPEN_ID_TYPE_PERSISTENT_ID)
                fp->persistent_id = id;
}

static int __open_id(struct ksmbd_file_table *ft, struct ksmbd_file *fp,
                     int type)
{
        u64                     id = 0;
        int                     ret;

        if (type == OPEN_ID_TYPE_VOLATILE_ID && fd_limit_depleted()) {
                __open_id_set(fp, KSMBD_NO_FID, type);
                return -EMFILE;
        }

        idr_preload(KSMBD_DEFAULT_GFP);
        write_lock(&ft->lock);
        ret = idr_alloc_cyclic(ft->idr, fp, 0, INT_MAX - 1, GFP_NOWAIT);
        if (ret >= 0) {
                id = ret;
                ret = 0;
        } else {
                id = KSMBD_NO_FID;
                fd_limit_close();
        }

        __open_id_set(fp, id, type);
        write_unlock(&ft->lock);
        idr_preload_end();
        return ret;
}

unsigned int ksmbd_open_durable_fd(struct ksmbd_file *fp)
{
        __open_id(&global_ft, fp, OPEN_ID_TYPE_PERSISTENT_ID);
        return fp->persistent_id;
}

struct ksmbd_file *ksmbd_open_fd(struct ksmbd_work *work, struct file *filp)
{
        struct ksmbd_file *fp;
        int ret;

        fp = kmem_cache_zalloc(filp_cache, KSMBD_DEFAULT_GFP);
        if (!fp) {
                pr_err("Failed to allocate memory\n");
                return ERR_PTR(-ENOMEM);
        }

        INIT_LIST_HEAD(&fp->blocked_works);
        INIT_LIST_HEAD(&fp->node);
        INIT_LIST_HEAD(&fp->lock_list);
        spin_lock_init(&fp->f_lock);
        atomic_set(&fp->refcount, 1);

        fp->filp                = filp;
        fp->conn                = work->conn;
        fp->tcon                = work->tcon;
        fp->volatile_id         = KSMBD_NO_FID;
        fp->persistent_id       = KSMBD_NO_FID;
        fp->f_state             = FP_NEW;
        fp->f_ci                = ksmbd_inode_get(fp);

        if (!fp->f_ci) {
                ret = -ENOMEM;
                goto err_out;
        }

        ret = __open_id(&work->sess->file_table, fp, OPEN_ID_TYPE_VOLATILE_ID);
        if (ret) {
                ksmbd_inode_put(fp->f_ci);
                goto err_out;
        }

        atomic_inc(&work->conn->stats.open_files_count);
        return fp;

err_out:
        kmem_cache_free(filp_cache, fp);
        return ERR_PTR(ret);
}

void ksmbd_update_fstate(struct ksmbd_file_table *ft, struct ksmbd_file *fp,
                         unsigned int state)
{
        if (!fp)
                return;

        write_lock(&ft->lock);
        fp->f_state = state;
        write_unlock(&ft->lock);
}

static int
__close_file_table_ids(struct ksmbd_file_table *ft,
                       struct ksmbd_tree_connect *tcon,
                       bool (*skip)(struct ksmbd_tree_connect *tcon,
                                    struct ksmbd_file *fp))
{
        struct ksmbd_file *fp;
        unsigned int id = 0;
        int num = 0;

        while (1) {
                write_lock(&ft->lock);
                fp = idr_get_next(ft->idr, &id);
                if (!fp) {
                        write_unlock(&ft->lock);
                        break;
                }

                if (skip(tcon, fp) ||
                    !atomic_dec_and_test(&fp->refcount)) {
                        id++;
                        write_unlock(&ft->lock);
                        continue;
                }

                set_close_state_blocked_works(fp);
                idr_remove(ft->idr, fp->volatile_id);
                fp->volatile_id = KSMBD_NO_FID;
                write_unlock(&ft->lock);

                down_write(&fp->f_ci->m_lock);
                list_del_init(&fp->node);
                up_write(&fp->f_ci->m_lock);

                __ksmbd_close_fd(ft, fp);

                num++;
                id++;
        }

        return num;
}

static inline bool is_reconnectable(struct ksmbd_file *fp)
{
        struct oplock_info *opinfo = opinfo_get(fp);
        bool reconn = false;

        if (!opinfo)
                return false;

        if (opinfo->op_state != OPLOCK_STATE_NONE) {
                opinfo_put(opinfo);
                return false;
        }

        if (fp->is_resilient || fp->is_persistent)
                reconn = true;
        else if (fp->is_durable && opinfo->is_lease &&
                 opinfo->o_lease->state & SMB2_LEASE_HANDLE_CACHING_LE)
                reconn = true;

        else if (fp->is_durable && opinfo->level == SMB2_OPLOCK_LEVEL_BATCH)
                reconn = true;

        opinfo_put(opinfo);
        return reconn;
}

static bool tree_conn_fd_check(struct ksmbd_tree_connect *tcon,
                               struct ksmbd_file *fp)
{
        return fp->tcon != tcon;
}

static bool ksmbd_durable_scavenger_alive(void)
{
        if (!durable_scavenger_running)
                return false;

        if (kthread_should_stop())
                return false;

        if (idr_is_empty(global_ft.idr))
                return false;

        return true;
}

static void ksmbd_scavenger_dispose_dh(struct list_head *head)
{
        while (!list_empty(head)) {
                struct ksmbd_file *fp;

                fp = list_first_entry(head, struct ksmbd_file, node);
                list_del_init(&fp->node);
                __ksmbd_close_fd(NULL, fp);
        }
}

static int ksmbd_durable_scavenger(void *dummy)
{
        struct ksmbd_file *fp = NULL;
        unsigned int id;
        unsigned int min_timeout = 1;
        bool found_fp_timeout;
        LIST_HEAD(scavenger_list);
        unsigned long remaining_jiffies;

        __module_get(THIS_MODULE);

        set_freezable();
        while (ksmbd_durable_scavenger_alive()) {
                if (try_to_freeze())
                        continue;

                found_fp_timeout = false;

                remaining_jiffies = wait_event_timeout(dh_wq,
                                   ksmbd_durable_scavenger_alive() == false,
                                   __msecs_to_jiffies(min_timeout));
                if (remaining_jiffies)
                        min_timeout = jiffies_to_msecs(remaining_jiffies);
                else
                        min_timeout = DURABLE_HANDLE_MAX_TIMEOUT;

                write_lock(&global_ft.lock);
                idr_for_each_entry(global_ft.idr, fp, id) {
                        if (!fp->durable_timeout)
                                continue;

                        if (atomic_read(&fp->refcount) > 1 ||
                            fp->conn)
                                continue;

                        found_fp_timeout = true;
                        if (fp->durable_scavenger_timeout <=
                            jiffies_to_msecs(jiffies)) {
                                __ksmbd_remove_durable_fd(fp);
                                list_add(&fp->node, &scavenger_list);
                        } else {
                                unsigned long durable_timeout;

                                durable_timeout =
                                        fp->durable_scavenger_timeout -
                                                jiffies_to_msecs(jiffies);

                                if (min_timeout > durable_timeout)
                                        min_timeout = durable_timeout;
                        }
                }
                write_unlock(&global_ft.lock);

                ksmbd_scavenger_dispose_dh(&scavenger_list);

                if (found_fp_timeout == false)
                        break;
        }

        durable_scavenger_running = false;

        module_put(THIS_MODULE);

        return 0;
}

void ksmbd_launch_ksmbd_durable_scavenger(void)
{
        if (!(server_conf.flags & KSMBD_GLOBAL_FLAG_DURABLE_HANDLE))
                return;

        mutex_lock(&durable_scavenger_lock);
        if (durable_scavenger_running == true) {
                mutex_unlock(&durable_scavenger_lock);
                return;
        }

        durable_scavenger_running = true;

        server_conf.dh_task = kthread_run(ksmbd_durable_scavenger,
                                     (void *)NULL, "ksmbd-durable-scavenger");
        if (IS_ERR(server_conf.dh_task))
                pr_err("cannot start conn thread, err : %ld\n",
                       PTR_ERR(server_conf.dh_task));
        mutex_unlock(&durable_scavenger_lock);
}

void ksmbd_stop_durable_scavenger(void)
{
        if (!(server_conf.flags & KSMBD_GLOBAL_FLAG_DURABLE_HANDLE))
                return;

        mutex_lock(&durable_scavenger_lock);
        if (!durable_scavenger_running) {
                mutex_unlock(&durable_scavenger_lock);
                return;
        }

        durable_scavenger_running = false;
        if (waitqueue_active(&dh_wq))
                wake_up(&dh_wq);
        mutex_unlock(&durable_scavenger_lock);
        kthread_stop(server_conf.dh_task);
}

static bool session_fd_check(struct ksmbd_tree_connect *tcon,
                             struct ksmbd_file *fp)
{
        struct ksmbd_inode *ci;
        struct oplock_info *op;
        struct ksmbd_conn *conn;

        if (!is_reconnectable(fp))
                return false;

        conn = fp->conn;
        ci = fp->f_ci;
        down_write(&ci->m_lock);
        list_for_each_entry_rcu(op, &ci->m_op_list, op_entry) {
                if (op->conn != conn)
                        continue;
                if (op->conn && atomic_dec_and_test(&op->conn->refcnt))
                        kfree(op->conn);
                op->conn = NULL;
        }
        up_write(&ci->m_lock);

        fp->conn = NULL;
        fp->tcon = NULL;
        fp->volatile_id = KSMBD_NO_FID;

        if (fp->durable_timeout)
                fp->durable_scavenger_timeout =
                        jiffies_to_msecs(jiffies) + fp->durable_timeout;

        return true;
}

void ksmbd_close_tree_conn_fds(struct ksmbd_work *work)
{
        int num = __close_file_table_ids(&work->sess->file_table,
                                         work->tcon,
                                         tree_conn_fd_check);

        atomic_sub(num, &work->conn->stats.open_files_count);
}

void ksmbd_close_session_fds(struct ksmbd_work *work)
{
        int num = __close_file_table_ids(&work->sess->file_table,
                                         work->tcon,
                                         session_fd_check);

        atomic_sub(num, &work->conn->stats.open_files_count);
}

int ksmbd_init_global_file_table(void)
{
        create_proc_files();
        return ksmbd_init_file_table(&global_ft);
}

void ksmbd_free_global_file_table(void)
{
        struct ksmbd_file       *fp = NULL;
        unsigned int            id;

        idr_for_each_entry(global_ft.idr, fp, id) {
                ksmbd_remove_durable_fd(fp);
                __ksmbd_close_fd(NULL, fp);
        }

        idr_destroy(global_ft.idr);
        kfree(global_ft.idr);
}

int ksmbd_validate_name_reconnect(struct ksmbd_share_config *share,
                                  struct ksmbd_file *fp, char *name)
{
        char *pathname, *ab_pathname;
        int ret = 0;

        pathname = kmalloc(PATH_MAX, KSMBD_DEFAULT_GFP);
        if (!pathname)
                return -EACCES;

        ab_pathname = d_path(&fp->filp->f_path, pathname, PATH_MAX);
        if (IS_ERR(ab_pathname)) {
                kfree(pathname);
                return -EACCES;
        }

        if (name && strcmp(&ab_pathname[share->path_sz + 1], name)) {
                ksmbd_debug(SMB, "invalid name reconnect %s\n", name);
                ret = -EINVAL;
        }

        kfree(pathname);

        return ret;
}

int ksmbd_reopen_durable_fd(struct ksmbd_work *work, struct ksmbd_file *fp)
{
        struct ksmbd_inode *ci;
        struct oplock_info *op;

        if (!fp->is_durable || fp->conn || fp->tcon) {
                pr_err("Invalid durable fd [%p:%p]\n", fp->conn, fp->tcon);
                return -EBADF;
        }

        if (has_file_id(fp->volatile_id)) {
                pr_err("Still in use durable fd: %llu\n", fp->volatile_id);
                return -EBADF;
        }

        fp->conn = work->conn;
        fp->tcon = work->tcon;

        ci = fp->f_ci;
        down_write(&ci->m_lock);
        list_for_each_entry_rcu(op, &ci->m_op_list, op_entry) {
                if (op->conn)
                        continue;
                op->conn = fp->conn;
                atomic_inc(&op->conn->refcnt);
        }
        up_write(&ci->m_lock);

        fp->f_state = FP_NEW;
        __open_id(&work->sess->file_table, fp, OPEN_ID_TYPE_VOLATILE_ID);
        if (!has_file_id(fp->volatile_id)) {
                fp->conn = NULL;
                fp->tcon = NULL;
                return -EBADF;
        }
        return 0;
}

int ksmbd_init_file_table(struct ksmbd_file_table *ft)
{
        ft->idr = kzalloc_obj(struct idr, KSMBD_DEFAULT_GFP);
        if (!ft->idr)
                return -ENOMEM;

        idr_init(ft->idr);
        rwlock_init(&ft->lock);
        return 0;
}

void ksmbd_destroy_file_table(struct ksmbd_file_table *ft)
{
        if (!ft->idr)
                return;

        __close_file_table_ids(ft, NULL, session_fd_check);
        idr_destroy(ft->idr);
        kfree(ft->idr);
        ft->idr = NULL;
}

int ksmbd_init_file_cache(void)
{
        filp_cache = kmem_cache_create("ksmbd_file_cache",
                                       sizeof(struct ksmbd_file), 0,
                                       SLAB_HWCACHE_ALIGN, NULL);
        if (!filp_cache)
                goto out;

        init_waitqueue_head(&dh_wq);

        return 0;

out:
        pr_err("failed to allocate file cache\n");
        return -ENOMEM;
}

void ksmbd_exit_file_cache(void)
{
        kmem_cache_destroy(filp_cache);
}