root/lib/libvmmapi/vmmapi.h
/*-
 * SPDX-License-Identifier: BSD-2-Clause
 *
 * Copyright (c) 2011 NetApp, Inc.
 * All rights reserved.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 *
 * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 * ARE DISCLAIMED.  IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 */

#ifndef _VMMAPI_H_
#define _VMMAPI_H_

#include <sys/param.h>
#include <sys/cpuset.h>
#include <machine/vmm.h>
#include <machine/vmm_dev.h>

#include <stdbool.h>

/*
 * API version for out-of-tree consumers like grub-bhyve for making compile
 * time decisions.
 */
#define VMMAPI_VERSION  0300    /* 2 digit major followed by 2 digit minor */

struct iovec;
struct vcpu;
struct vmctx;
struct vm_snapshot_meta;
enum x2apic_state;

/*
 * Different styles of mapping the memory assigned to a VM into the address
 * space of the controlling process.
 */
enum vm_mmap_style {
        VM_MMAP_NONE,           /* no mapping */
        VM_MMAP_ALL,            /* fully and statically mapped */
        VM_MMAP_SPARSE,         /* mappings created on-demand */
};

/*
 * 'flags' value passed to 'vm_set_memflags()'.
 */
#define VM_MEM_F_INCORE 0x01    /* include guest memory in core file */
#define VM_MEM_F_WIRED  0x02    /* guest memory is wired */

/* Memory size and allocation policy for a single NUMA domain. */
struct vm_mem_domain {
        size_t size;
        int ds_policy;
        domainset_t *ds_mask;
        size_t ds_size;
};

__BEGIN_DECLS
/*
 * Get the length and name of the memory segment identified by 'segid'.
 * Note that system memory segments are identified with a nul name.
 *
 * Returns 0 on success and non-zero otherwise.
 */
int     vm_get_memseg(struct vmctx *ctx, int ident, size_t *lenp, char *name,
            size_t namesiz);

/*
 * Iterate over the guest address space. This function finds an address range
 * that starts at an address >= *gpa.
 *
 * Returns 0 if the next address range was found and non-zero otherwise.
 */
int     vm_mmap_getnext(struct vmctx *ctx, vm_paddr_t *gpa, int *segid,
            vm_ooffset_t *segoff, size_t *len, int *prot, int *flags);

int     vm_get_guestmem_from_ctx(struct vmctx *ctx, char **guest_baseaddr,
                                 size_t *lowmem_size, size_t *highmem_size);

/*
 * Create a device memory segment identified by 'segid'.
 *
 * Returns a pointer to the memory segment on success and MAP_FAILED otherwise.
 */
void    *vm_create_devmem(struct vmctx *ctx, int segid, const char *name,
            size_t len);

/*
 * Map the memory segment identified by 'segid' into the guest address space
 * at [gpa,gpa+len) with protection 'prot'.
 */
int     vm_mmap_memseg(struct vmctx *ctx, vm_paddr_t gpa, int segid,
            vm_ooffset_t segoff, size_t len, int prot);

int     vm_munmap_memseg(struct vmctx *ctx, vm_paddr_t gpa, size_t len);

int     vm_create(const char *name);
struct vmctx *vm_open(const char *name);
#define VMMAPI_OPEN_CREATE      0x01    /* create if the VM does not exist */
#define VMMAPI_OPEN_REINIT      0x02    /* reinitialize the VM if it exists */
#define VMMAPI_OPEN_CREATE_DESTROY_ON_CLOSE     0x04    /* Destroy the VM when closing vmm_ctl */
struct vmctx *vm_openf(const char *name, int flags);
void    vm_close(struct vmctx *ctx);
void    vm_destroy(struct vmctx *ctx);
int     vm_limit_rights(struct vmctx *ctx);
struct vcpu *vm_vcpu_open(struct vmctx *ctx, int vcpuid);
void    vm_vcpu_close(struct vcpu *vcpu);
int     vcpu_id(struct vcpu *vcpu);
int     vm_parse_memsize(const char *optarg, size_t *memsize);
int vm_setup_memory(struct vmctx *ctx, size_t len, enum vm_mmap_style s);
int vm_setup_memory_domains(struct vmctx *ctx, enum vm_mmap_style s,
                            struct vm_mem_domain *doms, int ndoms);
void    *vm_map_gpa(struct vmctx *ctx, vm_paddr_t gaddr, size_t len);
/* inverse operation to vm_map_gpa - extract guest address from host pointer */
vm_paddr_t vm_rev_map_gpa(struct vmctx *ctx, void *addr);
#ifdef __amd64__
int     vm_get_gpa_pmap(struct vmctx *, uint64_t gpa, uint64_t *pte, int *num);
int     vm_gla2gpa(struct vcpu *vcpu, struct vm_guest_paging *paging,
                   uint64_t gla, int prot, uint64_t *gpa, int *fault);
#endif
int     vm_gla2gpa_nofault(struct vcpu *vcpu,
                   struct vm_guest_paging *paging, uint64_t gla, int prot,
                   uint64_t *gpa, int *fault);
uint32_t vm_get_lowmem_limit(struct vmctx *ctx);
void    vm_set_memflags(struct vmctx *ctx, int flags);
int     vm_get_memflags(struct vmctx *ctx);
const char *vm_get_name(struct vmctx *ctx);
size_t  vm_get_lowmem_size(struct vmctx *ctx);
vm_paddr_t vm_get_highmem_base(struct vmctx *ctx);
size_t  vm_get_highmem_size(struct vmctx *ctx);
#ifdef __amd64__
int     vm_set_desc(struct vcpu *vcpu, int reg,
                    uint64_t base, uint32_t limit, uint32_t access);
int     vm_get_desc(struct vcpu *vcpu, int reg,
                    uint64_t *base, uint32_t *limit, uint32_t *access);
int     vm_get_seg_desc(struct vcpu *vcpu, int reg, struct seg_desc *seg_desc);
#endif
int     vm_set_register(struct vcpu *vcpu, int reg, uint64_t val);
int     vm_get_register(struct vcpu *vcpu, int reg, uint64_t *retval);
int     vm_set_register_set(struct vcpu *vcpu, unsigned int count,
    const int *regnums, uint64_t *regvals);
int     vm_get_register_set(struct vcpu *vcpu, unsigned int count,
    const int *regnums, uint64_t *regvals);
int     vm_run(struct vcpu *vcpu, struct vm_run *vmrun);
int     vm_suspend(struct vmctx *ctx, enum vm_suspend_how how);
int     vm_reinit(struct vmctx *ctx);
int     vm_raise_msi(struct vmctx *ctx, uint64_t addr, uint64_t msg,
    int bus, int slot, int func);
#if defined(__aarch64__)
int     vm_attach_vgic(struct vmctx *ctx, uint64_t dist_start, size_t dist_size,
    uint64_t redist_start, size_t redist_size);
int     vm_inject_exception(struct vcpu *vcpu, uint64_t esr, uint64_t far);
#elif defined(__riscv)
int     vm_attach_aplic(struct vmctx *ctx, uint64_t mem_start, size_t mem_size);
int     vm_inject_exception(struct vcpu *vcpu, uint64_t scause);
#endif
#if defined(__aarch64__) || defined(__riscv)
int     vm_assert_irq(struct vmctx *ctx, uint32_t irq);
int     vm_deassert_irq(struct vmctx *ctx, uint32_t irq);
#endif
#ifdef __amd64__
int     vm_apicid2vcpu(struct vmctx *ctx, int apicid);
int     vm_inject_exception(struct vcpu *vcpu, int vector,
    int errcode_valid, uint32_t errcode, int restart_instruction);
int     vm_lapic_irq(struct vcpu *vcpu, int vector);
int     vm_lapic_local_irq(struct vcpu *vcpu, int vector);
int     vm_lapic_msi(struct vmctx *ctx, uint64_t addr, uint64_t msg);
int     vm_ioapic_assert_irq(struct vmctx *ctx, int irq);
int     vm_ioapic_deassert_irq(struct vmctx *ctx, int irq);
int     vm_ioapic_pulse_irq(struct vmctx *ctx, int irq);
int     vm_ioapic_pincount(struct vmctx *ctx, int *pincount);
int     vm_isa_assert_irq(struct vmctx *ctx, int atpic_irq, int ioapic_irq);
int     vm_isa_deassert_irq(struct vmctx *ctx, int atpic_irq, int ioapic_irq);
int     vm_isa_pulse_irq(struct vmctx *ctx, int atpic_irq, int ioapic_irq);
int     vm_isa_set_irq_trigger(struct vmctx *ctx, int atpic_irq,
            enum vm_intr_trigger trigger);
int     vm_inject_nmi(struct vcpu *vcpu);
int     vm_readwrite_kernemu_device(struct vcpu *vcpu,
            vm_paddr_t gpa, bool write, int size, uint64_t *value);
#endif
int     vm_capability_name2type(const char *capname);
const char *vm_capability_type2name(int type);
int     vm_get_capability(struct vcpu *vcpu, enum vm_cap_type cap,
                          int *retval);
int     vm_set_capability(struct vcpu *vcpu, enum vm_cap_type cap,
                          int val);
int     vm_assign_pptdev(struct vmctx *ctx, int bus, int slot, int func);
int     vm_unassign_pptdev(struct vmctx *ctx, int bus, int slot, int func);
int     vm_map_pptdev_mmio(struct vmctx *ctx, int bus, int slot, int func,
                           vm_paddr_t gpa, size_t len, vm_paddr_t hpa);
int     vm_unmap_pptdev_mmio(struct vmctx *ctx, int bus, int slot, int func,
                             vm_paddr_t gpa, size_t len);
int     vm_setup_pptdev_msi(struct vmctx *ctx, int bus, int slot,
            int func, uint64_t addr, uint64_t msg, int numvec);
int     vm_setup_pptdev_msix(struct vmctx *ctx, int bus, int slot,
            int func, int idx, uint64_t addr, uint64_t msg,
            uint32_t vector_control);
int     vm_disable_pptdev_msix(struct vmctx *ctx, int bus, int slot, int func);

int     vm_get_intinfo(struct vcpu *vcpu, uint64_t *i1, uint64_t *i2);
int     vm_set_intinfo(struct vcpu *vcpu, uint64_t exit_intinfo);

/*
 * Return a pointer to the statistics buffer. Note that this is not MT-safe.
 */
uint64_t *vm_get_stats(struct vcpu *vcpu, struct timeval *ret_tv,
                       int *ret_entries);
const char *vm_get_stat_desc(struct vmctx *ctx, int index);

#ifdef __amd64__
int     vm_get_x2apic_state(struct vcpu *vcpu, enum x2apic_state *s);
int     vm_set_x2apic_state(struct vcpu *vcpu, enum x2apic_state s);

int     vm_get_hpet_capabilities(struct vmctx *ctx, uint32_t *capabilities);

/*
 * Translate the GLA range [gla,gla+len) into GPA segments in 'iov'.
 * The 'iovcnt' should be big enough to accommodate all GPA segments.
 *
 * retval       fault           Interpretation
 *   0            0             Success
 *   0            1             An exception was injected into the guest
 * EFAULT        N/A            Error
 */
int     vm_copy_setup(struct vcpu *vcpu, struct vm_guest_paging *pg,
            uint64_t gla, size_t len, int prot, struct iovec *iov, int iovcnt,
            int *fault);
#endif
void    vm_copyin(struct iovec *guest_iov, void *host_dst, size_t len);
void    vm_copyout(const void *host_src, struct iovec *guest_iov, size_t len);
void    vm_copy_teardown(struct iovec *iov, int iovcnt);

#ifdef __amd64__
/* RTC */
int     vm_rtc_write(struct vmctx *ctx, int offset, uint8_t value);
int     vm_rtc_read(struct vmctx *ctx, int offset, uint8_t *retval);
int     vm_rtc_settime(struct vmctx *ctx, time_t secs);
int     vm_rtc_gettime(struct vmctx *ctx, time_t *secs);
#endif

/* Reset vcpu register state */
int     vcpu_reset(struct vcpu *vcpu);

int     vm_active_cpus(struct vmctx *ctx, cpuset_t *cpus);
int     vm_suspended_cpus(struct vmctx *ctx, cpuset_t *cpus);
int     vm_debug_cpus(struct vmctx *ctx, cpuset_t *cpus);
int     vm_activate_cpu(struct vcpu *vcpu);
int     vm_suspend_all_cpus(struct vmctx *ctx);
int     vm_suspend_cpu(struct vcpu *vcpu);
int     vm_resume_all_cpus(struct vmctx *ctx);
int     vm_resume_cpu(struct vcpu *vcpu);
int     vm_restart_instruction(struct vcpu *vcpu);

/* CPU topology */
int     vm_set_topology(struct vmctx *ctx, uint16_t sockets, uint16_t cores,
            uint16_t threads, uint16_t maxcpus);
int     vm_get_topology(struct vmctx *ctx, uint16_t *sockets, uint16_t *cores,
            uint16_t *threads, uint16_t *maxcpus);

/*
 * FreeBSD specific APIs
 */
int     vm_setup_freebsd_registers(struct vcpu *vcpu,
                                uint64_t rip, uint64_t cr3, uint64_t gdtbase,
                                uint64_t rsp);
int     vm_setup_freebsd_registers_i386(struct vcpu *vcpu,
                                        uint32_t eip, uint32_t gdtbase,
                                        uint32_t esp);
void    vm_setup_freebsd_gdt(uint64_t *gdtr);

/*
 * Save and restore
 */
int     vm_snapshot_req(struct vmctx *ctx, struct vm_snapshot_meta *meta);
int     vm_restore_time(struct vmctx *ctx);

/*
 * Deprecated interfaces, do not use them in new code.
 */
int     vm_get_device_fd(struct vmctx *ctx);
const cap_ioctl_t *vm_get_ioctls(size_t *len);

__END_DECLS

#endif  /* _VMMAPI_H_ */