root/fs/xfs/scrub/iscan.h
/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * Copyright (c) 2021-2024 Oracle.  All Rights Reserved.
 * Author: Darrick J. Wong <djwong@kernel.org>
 */
#ifndef __XFS_SCRUB_ISCAN_H__
#define __XFS_SCRUB_ISCAN_H__

struct xchk_iscan {
        struct xfs_scrub        *sc;

        /* Lock to protect the scan cursor. */
        struct mutex            lock;

        /*
         * This is the first inode in the inumber address space that we
         * examined.  When the scan wraps around back to here, the scan is
         * finished.
         */
        xfs_ino_t               scan_start_ino;

        /* This is the inode that will be examined next. */
        xfs_ino_t               cursor_ino;

        /* If nonzero and non-NULL, skip this inode when scanning. */
        xfs_ino_t               skip_ino;

        /*
         * This is the last inode that we've successfully scanned, either
         * because the caller scanned it, or we moved the cursor past an empty
         * part of the inode address space.  Scan callers should only use the
         * xchk_iscan_visit function to modify this.
         */
        xfs_ino_t               __visited_ino;

        /* Operational state of the livescan. */
        unsigned long           __opstate;

        /* Give up on iterating @cursor_ino if we can't iget it by this time. */
        unsigned long           __iget_deadline;

        /* Amount of time (in ms) that we will try to iget an inode. */
        unsigned int            iget_timeout;

        /* Wait this many ms to retry an iget. */
        unsigned int            iget_retry_delay;

        /*
         * The scan grabs batches of inodes and stashes them here before
         * handing them out with _iter.  Unallocated inodes are set in the
         * mask so that all updates to that inode are selected for live
         * update propagation.
         */
        xfs_ino_t               __batch_ino;
        xfs_inofree_t           __skipped_inomask;
        struct xfs_inode        *__inodes[XFS_INODES_PER_CHUNK];
};

/* Set if the scan has been aborted due to some event in the fs. */
#define XCHK_ISCAN_OPSTATE_ABORTED      (1)

/* Use trylock to acquire the AGI */
#define XCHK_ISCAN_OPSTATE_TRYLOCK_AGI  (2)

static inline bool
xchk_iscan_aborted(const struct xchk_iscan *iscan)
{
        return test_bit(XCHK_ISCAN_OPSTATE_ABORTED, &iscan->__opstate);
}

static inline void
xchk_iscan_abort(struct xchk_iscan *iscan)
{
        set_bit(XCHK_ISCAN_OPSTATE_ABORTED, &iscan->__opstate);
}

static inline bool
xchk_iscan_agi_needs_trylock(const struct xchk_iscan *iscan)
{
        return test_bit(XCHK_ISCAN_OPSTATE_TRYLOCK_AGI, &iscan->__opstate);
}

static inline void
xchk_iscan_set_agi_trylock(struct xchk_iscan *iscan)
{
        set_bit(XCHK_ISCAN_OPSTATE_TRYLOCK_AGI, &iscan->__opstate);
}

void xchk_iscan_start(struct xfs_scrub *sc, unsigned int iget_timeout,
                unsigned int iget_retry_delay, struct xchk_iscan *iscan);
void xchk_iscan_finish_early(struct xchk_iscan *iscan);
void xchk_iscan_teardown(struct xchk_iscan *iscan);

int xchk_iscan_iter(struct xchk_iscan *iscan, struct xfs_inode **ipp);
void xchk_iscan_iter_finish(struct xchk_iscan *iscan);

void xchk_iscan_mark_visited(struct xchk_iscan *iscan, struct xfs_inode *ip);
bool xchk_iscan_want_live_update(struct xchk_iscan *iscan, xfs_ino_t ino);

#endif /* __XFS_SCRUB_ISCAN_H__ */