root/virt/kvm/kvm_mm.h
/* SPDX-License-Identifier: GPL-2.0-only */

#ifndef __KVM_MM_H__
#define __KVM_MM_H__ 1

/*
 * Architectures can choose whether to use an rwlock or spinlock
 * for the mmu_lock.  These macros, for use in common code
 * only, avoids using #ifdefs in places that must deal with
 * multiple architectures.
 */

#ifdef KVM_HAVE_MMU_RWLOCK
#define KVM_MMU_LOCK_INIT(kvm)          rwlock_init(&(kvm)->mmu_lock)
#define KVM_MMU_LOCK(kvm)               write_lock(&(kvm)->mmu_lock)
#define KVM_MMU_UNLOCK(kvm)             write_unlock(&(kvm)->mmu_lock)
#else
#define KVM_MMU_LOCK_INIT(kvm)          spin_lock_init(&(kvm)->mmu_lock)
#define KVM_MMU_LOCK(kvm)               spin_lock(&(kvm)->mmu_lock)
#define KVM_MMU_UNLOCK(kvm)             spin_unlock(&(kvm)->mmu_lock)
#endif /* KVM_HAVE_MMU_RWLOCK */


struct kvm_follow_pfn {
        const struct kvm_memory_slot *slot;
        const gfn_t gfn;

        unsigned long hva;

        /* FOLL_* flags modifying lookup behavior, e.g. FOLL_WRITE. */
        unsigned int flags;

        /*
         * Pin the page (effectively FOLL_PIN, which is an mm/ internal flag).
         * The page *must* be pinned if KVM will write to the page via a kernel
         * mapping, e.g. via kmap(), mremap(), etc.
         */
        bool pin;

        /*
         * If non-NULL, try to get a writable mapping even for a read fault.
         * Set to true if a writable mapping was obtained.
         */
        bool *map_writable;

        /*
         * Optional output.  Set to a valid "struct page" if the returned pfn
         * is for a refcounted or pinned struct page, NULL if the returned pfn
         * has no struct page or if the struct page is not being refcounted
         * (e.g. tail pages of non-compound higher order allocations from
         * IO/PFNMAP mappings).
         */
        struct page **refcounted_page;
};

kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp);

#ifdef CONFIG_HAVE_KVM_PFNCACHE
void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
                                       unsigned long start,
                                       unsigned long end);
#else
static inline void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
                                                     unsigned long start,
                                                     unsigned long end)
{
}
#endif /* HAVE_KVM_PFNCACHE */

#ifdef CONFIG_KVM_GUEST_MEMFD
int kvm_gmem_init(struct module *module);
void kvm_gmem_exit(void);
int kvm_gmem_create(struct kvm *kvm, struct kvm_create_guest_memfd *args);
int kvm_gmem_bind(struct kvm *kvm, struct kvm_memory_slot *slot,
                  unsigned int fd, loff_t offset);
void kvm_gmem_unbind(struct kvm_memory_slot *slot);
#else
static inline int kvm_gmem_init(struct module *module)
{
        return 0;
}
static inline void kvm_gmem_exit(void) {};
static inline int kvm_gmem_bind(struct kvm *kvm,
                                         struct kvm_memory_slot *slot,
                                         unsigned int fd, loff_t offset)
{
        WARN_ON_ONCE(1);
        return -EIO;
}

static inline void kvm_gmem_unbind(struct kvm_memory_slot *slot)
{
        WARN_ON_ONCE(1);
}
#endif /* CONFIG_KVM_GUEST_MEMFD */

#endif /* __KVM_MM_H__ */