root/include/linux/mm_types_task.h
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _LINUX_MM_TYPES_TASK_H
#define _LINUX_MM_TYPES_TASK_H

/*
 * Here are the definitions of the MM data types that are embedded in 'struct task_struct'.
 *
 * (These are defined separately to decouple sched.h from mm_types.h as much as possible.)
 */

#include <linux/align.h>
#include <linux/types.h>

#include <asm/page.h>

#ifdef CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH
#include <asm/tlbbatch.h>
#endif

#define ALLOC_SPLIT_PTLOCKS     (SPINLOCK_SIZE > BITS_PER_LONG/8)

/*
 * When updating this, please also update struct resident_page_types[] in
 * kernel/fork.c
 */
enum {
        MM_FILEPAGES,   /* Resident file mapping pages */
        MM_ANONPAGES,   /* Resident anonymous pages */
        MM_SWAPENTS,    /* Anonymous swap entries */
        MM_SHMEMPAGES,  /* Resident shared memory pages */
        NR_MM_COUNTERS
};

struct page;

struct page_frag {
        struct page *page;
#if (BITS_PER_LONG > 32) || (PAGE_SIZE >= 65536)
        __u32 offset;
        __u32 size;
#else
        __u16 offset;
        __u16 size;
#endif
};

#define PAGE_FRAG_CACHE_MAX_SIZE        __ALIGN_MASK(32768, ~PAGE_MASK)
#define PAGE_FRAG_CACHE_MAX_ORDER       get_order(PAGE_FRAG_CACHE_MAX_SIZE)
struct page_frag_cache {
        /* encoded_page consists of the virtual address, pfmemalloc bit and
         * order of a page.
         */
        unsigned long encoded_page;

        /* we maintain a pagecount bias, so that we dont dirty cache line
         * containing page->_refcount every time we allocate a fragment.
         */
#if (PAGE_SIZE < PAGE_FRAG_CACHE_MAX_SIZE) && (BITS_PER_LONG <= 32)
        __u16 offset;
        __u16 pagecnt_bias;
#else
        __u32 offset;
        __u32 pagecnt_bias;
#endif
};

/* Track pages that require TLB flushes */
struct tlbflush_unmap_batch {
#ifdef CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH
        /*
         * The arch code makes the following promise: generic code can modify a
         * PTE, then call arch_tlbbatch_add_pending() (which internally provides
         * all needed barriers), then call arch_tlbbatch_flush(), and the entries
         * will be flushed on all CPUs by the time that arch_tlbbatch_flush()
         * returns.
         */
        struct arch_tlbflush_unmap_batch arch;

        /* True if a flush is needed. */
        bool flush_required;

        /*
         * If true then the PTE was dirty when unmapped. The entry must be
         * flushed before IO is initiated or a stale TLB entry potentially
         * allows an update without redirtying the page.
         */
        bool writable;
#endif
};

struct lazy_mmu_state {
        u8 enable_count;
        u8 pause_count;
};

#endif /* _LINUX_MM_TYPES_TASK_H */