root/tools/testing/selftests/kvm/arm64/vgic_irq.c
// SPDX-License-Identifier: GPL-2.0
/*
 * vgic_irq.c - Test userspace injection of IRQs
 *
 * This test validates the injection of IRQs from userspace using various
 * methods (e.g., KVM_IRQ_LINE) and modes (e.g., EOI). The guest "asks" the
 * host to inject a specific intid via a GUEST_SYNC call, and then checks that
 * it received it.
 */
#include <asm/kvm.h>
#include <asm/kvm_para.h>
#include <sys/eventfd.h>
#include <linux/sizes.h>

#include "processor.h"
#include "test_util.h"
#include "kvm_util.h"
#include "gic.h"
#include "gic_v3.h"
#include "vgic.h"

/*
 * Stores the user specified args; it's passed to the guest and to every test
 * function.
 */
struct test_args {
        uint32_t nr_irqs; /* number of KVM supported IRQs. */
        bool eoi_split; /* 1 is eoir+dir, 0 is eoir only */
        bool level_sensitive; /* 1 is level, 0 is edge */
        int kvm_max_routes; /* output of KVM_CAP_IRQ_ROUTING */
        bool kvm_supports_irqfd; /* output of KVM_CAP_IRQFD */
        uint32_t shared_data;
};

/*
 * KVM implements 32 priority levels:
 * 0x00 (highest priority) - 0xF8 (lowest priority), in steps of 8
 *
 * Note that these macros will still be correct in the case that KVM implements
 * more priority levels. Also note that 32 is the minimum for GICv3 and GICv2.
 */
#define KVM_NUM_PRIOS           32
#define KVM_PRIO_SHIFT          3 /* steps of 8 = 1 << 3 */
#define KVM_PRIO_STEPS          (1 << KVM_PRIO_SHIFT) /* 8 */
#define LOWEST_PRIO             (KVM_NUM_PRIOS - 1)
#define CPU_PRIO_MASK           (LOWEST_PRIO << KVM_PRIO_SHIFT) /* 0xf8 */
#define IRQ_DEFAULT_PRIO        (LOWEST_PRIO - 1)
#define IRQ_DEFAULT_PRIO_REG    (IRQ_DEFAULT_PRIO << KVM_PRIO_SHIFT) /* 0xf0 */

/*
 * The kvm_inject_* utilities are used by the guest to ask the host to inject
 * interrupts (e.g., using the KVM_IRQ_LINE ioctl).
 */

typedef enum {
        KVM_INJECT_EDGE_IRQ_LINE = 1,
        KVM_SET_IRQ_LINE,
        KVM_SET_IRQ_LINE_HIGH,
        KVM_SET_LEVEL_INFO_HIGH,
        KVM_INJECT_IRQFD,
        KVM_WRITE_ISPENDR,
        KVM_WRITE_ISACTIVER,
} kvm_inject_cmd;

struct kvm_inject_args {
        kvm_inject_cmd cmd;
        uint32_t first_intid;
        uint32_t num;
        int level;
        bool expect_failure;
};

/* Used on the guest side to perform the hypercall. */
static void kvm_inject_call(kvm_inject_cmd cmd, uint32_t first_intid,
                uint32_t num, int level, bool expect_failure);

/* Used on the host side to get the hypercall info. */
static void kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc,
                struct kvm_inject_args *args);

#define _KVM_INJECT_MULTI(cmd, intid, num, expect_failure)                      \
        kvm_inject_call(cmd, intid, num, -1 /* not used */, expect_failure)

#define KVM_INJECT_MULTI(cmd, intid, num)                                       \
        _KVM_INJECT_MULTI(cmd, intid, num, false)

#define _KVM_INJECT(cmd, intid, expect_failure)                                 \
        _KVM_INJECT_MULTI(cmd, intid, 1, expect_failure)

#define KVM_INJECT(cmd, intid)                                                  \
        _KVM_INJECT_MULTI(cmd, intid, 1, false)

#define KVM_ACTIVATE(cmd, intid)                                                \
        kvm_inject_call(cmd, intid, 1, 1, false);

struct kvm_inject_desc {
        kvm_inject_cmd cmd;
        /* can inject PPIs, PPIs, and/or SPIs. */
        bool sgi, ppi, spi;
};

static struct kvm_inject_desc inject_edge_fns[] = {
        /*                                      sgi    ppi    spi */
        { KVM_INJECT_EDGE_IRQ_LINE,             false, false, true },
        { KVM_INJECT_IRQFD,                     false, false, true },
        { KVM_WRITE_ISPENDR,                    true,  false, true },
        { 0, },
};

static struct kvm_inject_desc inject_level_fns[] = {
        /*                                      sgi    ppi    spi */
        { KVM_SET_IRQ_LINE_HIGH,                false, true,  true },
        { KVM_SET_LEVEL_INFO_HIGH,              false, true,  true },
        { KVM_INJECT_IRQFD,                     false, false, true },
        { KVM_WRITE_ISPENDR,                    false, true,  true },
        { 0, },
};

static struct kvm_inject_desc set_active_fns[] = {
        /*                                      sgi    ppi    spi */
        { KVM_WRITE_ISACTIVER,                  true,  true,  true },
        { 0, },
};

#define for_each_inject_fn(t, f)                                                \
        for ((f) = (t); (f)->cmd; (f)++)

#define for_each_supported_inject_fn(args, t, f)                                \
        for_each_inject_fn(t, f)                                                \
                if ((args)->kvm_supports_irqfd || (f)->cmd != KVM_INJECT_IRQFD)

#define for_each_supported_activate_fn(args, t, f)                              \
        for_each_supported_inject_fn((args), (t), (f))

/* Shared between the guest main thread and the IRQ handlers. */
volatile uint64_t irq_handled;
volatile uint32_t irqnr_received[MAX_SPI + 1];

static void reset_stats(void)
{
        int i;

        irq_handled = 0;
        for (i = 0; i <= MAX_SPI; i++)
                irqnr_received[i] = 0;
}

static uint64_t gic_read_ap1r0(void)
{
        uint64_t reg = read_sysreg_s(SYS_ICC_AP1R0_EL1);

        dsb(sy);
        return reg;
}

static void gic_write_ap1r0(uint64_t val)
{
        write_sysreg_s(val, SYS_ICC_AP1R0_EL1);
        isb();
}

static void guest_set_irq_line(uint32_t intid, uint32_t level);

static void guest_irq_generic_handler(bool eoi_split, bool level_sensitive)
{
        uint32_t intid = gic_get_and_ack_irq();

        if (intid == IAR_SPURIOUS)
                return;

        GUEST_ASSERT(gic_irq_get_active(intid));

        if (!level_sensitive)
                GUEST_ASSERT(!gic_irq_get_pending(intid));

        if (level_sensitive)
                guest_set_irq_line(intid, 0);

        GUEST_ASSERT(intid < MAX_SPI);
        irqnr_received[intid] += 1;
        irq_handled += 1;

        gic_set_eoi(intid);
        GUEST_ASSERT_EQ(gic_read_ap1r0(), 0);
        if (eoi_split)
                gic_set_dir(intid);

        GUEST_ASSERT(!gic_irq_get_active(intid));
        GUEST_ASSERT(!gic_irq_get_pending(intid));
}

static void kvm_inject_call(kvm_inject_cmd cmd, uint32_t first_intid,
                uint32_t num, int level, bool expect_failure)
{
        struct kvm_inject_args args = {
                .cmd = cmd,
                .first_intid = first_intid,
                .num = num,
                .level = level,
                .expect_failure = expect_failure,
        };
        GUEST_SYNC(&args);
}

#define GUEST_ASSERT_IAR_EMPTY()                                                \
do {                                                                            \
        uint32_t _intid;                                                        \
        _intid = gic_get_and_ack_irq();                                         \
        GUEST_ASSERT(_intid == IAR_SPURIOUS);                                   \
} while (0)

#define CAT_HELPER(a, b) a ## b
#define CAT(a, b) CAT_HELPER(a, b)
#define PREFIX guest_irq_handler_
#define GUEST_IRQ_HANDLER_NAME(split, lev) CAT(PREFIX, CAT(split, lev))
#define GENERATE_GUEST_IRQ_HANDLER(split, lev)                                  \
static void CAT(PREFIX, CAT(split, lev))(struct ex_regs *regs)                  \
{                                                                               \
        guest_irq_generic_handler(split, lev);                                  \
}

GENERATE_GUEST_IRQ_HANDLER(0, 0);
GENERATE_GUEST_IRQ_HANDLER(0, 1);
GENERATE_GUEST_IRQ_HANDLER(1, 0);
GENERATE_GUEST_IRQ_HANDLER(1, 1);

static void (*guest_irq_handlers[2][2])(struct ex_regs *) = {
        {GUEST_IRQ_HANDLER_NAME(0, 0), GUEST_IRQ_HANDLER_NAME(0, 1),},
        {GUEST_IRQ_HANDLER_NAME(1, 0), GUEST_IRQ_HANDLER_NAME(1, 1),},
};

static void reset_priorities(struct test_args *args)
{
        int i;

        for (i = 0; i < args->nr_irqs; i++)
                gic_set_priority(i, IRQ_DEFAULT_PRIO_REG);
}

static void guest_set_irq_line(uint32_t intid, uint32_t level)
{
        kvm_inject_call(KVM_SET_IRQ_LINE, intid, 1, level, false);
}

static void test_inject_fail(struct test_args *args,
                uint32_t intid, kvm_inject_cmd cmd)
{
        reset_stats();

        _KVM_INJECT(cmd, intid, true);
        /* no IRQ to handle on entry */

        GUEST_ASSERT_EQ(irq_handled, 0);
        GUEST_ASSERT_IAR_EMPTY();
}

static void guest_inject(struct test_args *args,
                uint32_t first_intid, uint32_t num,
                kvm_inject_cmd cmd)
{
        uint32_t i;

        reset_stats();

        /* Cycle over all priorities to make things more interesting. */
        for (i = first_intid; i < num + first_intid; i++)
                gic_set_priority(i, (i % (KVM_NUM_PRIOS - 1)) << 3);

        asm volatile("msr daifset, #2" : : : "memory");
        KVM_INJECT_MULTI(cmd, first_intid, num);

        while (irq_handled < num) {
                wfi();
                local_irq_enable();
                isb(); /* handle IRQ */
                local_irq_disable();
        }
        local_irq_enable();

        GUEST_ASSERT_EQ(irq_handled, num);
        for (i = first_intid; i < num + first_intid; i++)
                GUEST_ASSERT_EQ(irqnr_received[i], 1);
        GUEST_ASSERT_IAR_EMPTY();

        reset_priorities(args);
}

/*
 * Restore the active state of multiple concurrent IRQs (given by
 * concurrent_irqs).  This does what a live-migration would do on the
 * destination side assuming there are some active IRQs that were not
 * deactivated yet.
 */
static void guest_restore_active(struct test_args *args,
                uint32_t first_intid, uint32_t num,
                kvm_inject_cmd cmd)
{
        uint32_t prio, intid, ap1r;
        int i;

        /*
         * Set the priorities of the first (KVM_NUM_PRIOS - 1) IRQs
         * in descending order, so intid+1 can preempt intid.
         */
        for (i = 0, prio = (num - 1) * 8; i < num; i++, prio -= 8) {
                GUEST_ASSERT(prio >= 0);
                intid = i + first_intid;
                gic_set_priority(intid, prio);
        }

        /*
         * In a real migration, KVM would restore all GIC state before running
         * guest code.
         */
        for (i = 0; i < num; i++) {
                intid = i + first_intid;
                KVM_ACTIVATE(cmd, intid);
                ap1r = gic_read_ap1r0();
                ap1r |= 1U << i;
                gic_write_ap1r0(ap1r);
        }

        /* This is where the "migration" would occur. */

        /* finish handling the IRQs starting with the highest priority one. */
        for (i = 0; i < num; i++) {
                intid = num - i - 1 + first_intid;
                gic_set_eoi(intid);
                if (args->eoi_split)
                        gic_set_dir(intid);
        }

        for (i = 0; i < num; i++)
                GUEST_ASSERT(!gic_irq_get_active(i + first_intid));
        GUEST_ASSERT_EQ(gic_read_ap1r0(), 0);
        GUEST_ASSERT_IAR_EMPTY();
}

/*
 * Polls the IAR until it's not a spurious interrupt.
 *
 * This function should only be used in test_inject_preemption (with IRQs
 * masked).
 */
static uint32_t wait_for_and_activate_irq(void)
{
        uint32_t intid;

        do {
                asm volatile("wfi" : : : "memory");
                intid = gic_get_and_ack_irq();
        } while (intid == IAR_SPURIOUS);

        return intid;
}

/*
 * Inject multiple concurrent IRQs (num IRQs starting at first_intid) and
 * handle them without handling the actual exceptions.  This is done by masking
 * interrupts for the whole test.
 */
static void test_inject_preemption(struct test_args *args,
                                   uint32_t first_intid, int num,
                                   const unsigned long *exclude,
                                   kvm_inject_cmd cmd)
{
        uint32_t intid, prio, step = KVM_PRIO_STEPS;
        int i;

        /* Set the priorities of the first (KVM_NUM_PRIOS - 1) IRQs
         * in descending order, so intid+1 can preempt intid.
         */
        for (i = 0, prio = (num - 1) * step; i < num; i++, prio -= step) {
                GUEST_ASSERT(prio >= 0);
                intid = i + first_intid;
                gic_set_priority(intid, prio);
        }

        local_irq_disable();

        for (i = 0; i < num; i++) {
                uint32_t tmp;
                intid = i + first_intid;

                if (exclude && test_bit(i, exclude))
                        continue;

                KVM_INJECT(cmd, intid);
                /* Each successive IRQ will preempt the previous one. */
                tmp = wait_for_and_activate_irq();
                GUEST_ASSERT_EQ(tmp, intid);
                if (args->level_sensitive)
                        guest_set_irq_line(intid, 0);
        }

        /* finish handling the IRQs starting with the highest priority one. */
        for (i = 0; i < num; i++) {
                intid = num - i - 1 + first_intid;

                if (exclude && test_bit(intid - first_intid, exclude))
                        continue;

                gic_set_eoi(intid);
        }

        if (args->eoi_split) {
                for (i = 0; i < num; i++) {
                        intid = i + first_intid;

                        if (exclude && test_bit(i, exclude))
                                continue;

                        if (args->eoi_split)
                                gic_set_dir(intid);
                }
        }

        local_irq_enable();

        for (i = 0; i < num; i++) {
                if (exclude && test_bit(i, exclude))
                        continue;

                GUEST_ASSERT(!gic_irq_get_active(i + first_intid));
        }
        GUEST_ASSERT_EQ(gic_read_ap1r0(), 0);
        GUEST_ASSERT_IAR_EMPTY();

        reset_priorities(args);
}

static void test_injection(struct test_args *args, struct kvm_inject_desc *f)
{
        uint32_t nr_irqs = args->nr_irqs;

        if (f->sgi) {
                guest_inject(args, MIN_SGI, 1, f->cmd);
                guest_inject(args, 0, 16, f->cmd);
        }

        if (f->ppi)
                guest_inject(args, MIN_PPI, 1, f->cmd);

        if (f->spi) {
                guest_inject(args, MIN_SPI, 1, f->cmd);
                guest_inject(args, nr_irqs - 1, 1, f->cmd);
                guest_inject(args, MIN_SPI, nr_irqs - MIN_SPI, f->cmd);
        }
}

static void test_injection_failure(struct test_args *args,
                struct kvm_inject_desc *f)
{
        uint32_t bad_intid[] = { args->nr_irqs, 1020, 1024, 1120, 5120, ~0U, };
        int i;

        for (i = 0; i < ARRAY_SIZE(bad_intid); i++)
                test_inject_fail(args, bad_intid[i], f->cmd);
}

static void test_preemption(struct test_args *args, struct kvm_inject_desc *f)
{
        /* Timer PPIs cannot be injected from userspace */
        static const unsigned long ppi_exclude = (BIT(27 - MIN_PPI) |
                                                  BIT(30 - MIN_PPI) |
                                                  BIT(28 - MIN_PPI) |
                                                  BIT(26 - MIN_PPI));

        if (f->sgi)
                test_inject_preemption(args, MIN_SGI, 16, NULL, f->cmd);

        if (f->ppi)
                test_inject_preemption(args, MIN_PPI, 16, &ppi_exclude, f->cmd);

        if (f->spi)
                test_inject_preemption(args, MIN_SPI, 31, NULL, f->cmd);
}

static void test_restore_active(struct test_args *args, struct kvm_inject_desc *f)
{
        if (f->sgi)
                guest_restore_active(args, MIN_SGI, 16, f->cmd);

        if (f->ppi)
                guest_restore_active(args, MIN_PPI, 16, f->cmd);

        if (f->spi)
                guest_restore_active(args, MIN_SPI, 31, f->cmd);
}

static void guest_code(struct test_args *args)
{
        uint32_t i, nr_irqs = args->nr_irqs;
        bool level_sensitive = args->level_sensitive;
        struct kvm_inject_desc *f, *inject_fns;

        gic_init(GIC_V3, 1);

        for (i = MIN_SPI; i < nr_irqs; i++)
                gic_irq_set_config(i, !level_sensitive);

        for (i = 0; i < nr_irqs; i++)
                gic_irq_enable(i);

        gic_set_eoi_split(args->eoi_split);

        reset_priorities(args);
        gic_set_priority_mask(CPU_PRIO_MASK);

        inject_fns  = level_sensitive ? inject_level_fns
                                      : inject_edge_fns;

        local_irq_enable();

        /* Start the tests. */
        for_each_supported_inject_fn(args, inject_fns, f) {
                test_injection(args, f);
                test_preemption(args, f);
                test_injection_failure(args, f);
        }

        /*
         * Restore the active state of IRQs. This would happen when live
         * migrating IRQs in the middle of being handled.
         */
        for_each_supported_activate_fn(args, set_active_fns, f)
                test_restore_active(args, f);

        GUEST_DONE();
}

static void kvm_irq_line_check(struct kvm_vm *vm, uint32_t intid, int level,
                        struct test_args *test_args, bool expect_failure)
{
        int ret;

        if (!expect_failure) {
                kvm_arm_irq_line(vm, intid, level);
        } else {
                /* The interface doesn't allow larger intid's. */
                if (intid > KVM_ARM_IRQ_NUM_MASK)
                        return;

                ret = _kvm_arm_irq_line(vm, intid, level);
                TEST_ASSERT(ret != 0 && errno == EINVAL,
                                "Bad intid %i did not cause KVM_IRQ_LINE "
                                "error: rc: %i errno: %i", intid, ret, errno);
        }
}

void kvm_irq_set_level_info_check(int gic_fd, uint32_t intid, int level,
                        bool expect_failure)
{
        if (!expect_failure) {
                kvm_irq_set_level_info(gic_fd, intid, level);
        } else {
                int ret = _kvm_irq_set_level_info(gic_fd, intid, level);
                /*
                 * The kernel silently fails for invalid SPIs and SGIs (which
                 * are not level-sensitive). It only checks for intid to not
                 * spill over 1U << 10 (the max reserved SPI). Also, callers
                 * are supposed to mask the intid with 0x3ff (1023).
                 */
                if (intid > VGIC_MAX_RESERVED)
                        TEST_ASSERT(ret != 0 && errno == EINVAL,
                                "Bad intid %i did not cause VGIC_GRP_LEVEL_INFO "
                                "error: rc: %i errno: %i", intid, ret, errno);
                else
                        TEST_ASSERT(!ret, "KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO "
                                "for intid %i failed, rc: %i errno: %i",
                                intid, ret, errno);
        }
}

static void kvm_set_gsi_routing_irqchip_check(struct kvm_vm *vm,
                uint32_t intid, uint32_t num, uint32_t kvm_max_routes,
                bool expect_failure)
{
        struct kvm_irq_routing *routing;
        int ret;
        uint64_t i;

        assert(num <= kvm_max_routes && kvm_max_routes <= KVM_MAX_IRQ_ROUTES);

        routing = kvm_gsi_routing_create();
        for (i = intid; i < (uint64_t)intid + num; i++)
                kvm_gsi_routing_irqchip_add(routing, i - MIN_SPI, i - MIN_SPI);

        if (!expect_failure) {
                kvm_gsi_routing_write(vm, routing);
        } else {
                ret = _kvm_gsi_routing_write(vm, routing);
                /* The kernel only checks e->irqchip.pin >= KVM_IRQCHIP_NUM_PINS */
                if (((uint64_t)intid + num - 1 - MIN_SPI) >= KVM_IRQCHIP_NUM_PINS)
                        TEST_ASSERT(ret != 0 && errno == EINVAL,
                                "Bad intid %u did not cause KVM_SET_GSI_ROUTING "
                                "error: rc: %i errno: %i", intid, ret, errno);
                else
                        TEST_ASSERT(ret == 0, "KVM_SET_GSI_ROUTING "
                                "for intid %i failed, rc: %i errno: %i",
                                intid, ret, errno);
        }
}

static void kvm_irq_write_ispendr_check(int gic_fd, uint32_t intid,
                                        struct kvm_vcpu *vcpu,
                                        bool expect_failure)
{
        /*
         * Ignore this when expecting failure as invalid intids will lead to
         * either trying to inject SGIs when we configured the test to be
         * level_sensitive (or the reverse), or inject large intids which
         * will lead to writing above the ISPENDR register space (and we
         * don't want to do that either).
         */
        if (!expect_failure)
                kvm_irq_write_ispendr(gic_fd, intid, vcpu);
}

static void kvm_routing_and_irqfd_check(struct kvm_vm *vm,
                uint32_t intid, uint32_t num, uint32_t kvm_max_routes,
                bool expect_failure)
{
        int fd[MAX_SPI];
        uint64_t val;
        int ret, f;
        uint64_t i;

        /*
         * There is no way to try injecting an SGI or PPI as the interface
         * starts counting from the first SPI (above the private ones), so just
         * exit.
         */
        if (INTID_IS_SGI(intid) || INTID_IS_PPI(intid))
                return;

        kvm_set_gsi_routing_irqchip_check(vm, intid, num,
                        kvm_max_routes, expect_failure);

        /*
         * If expect_failure, then just to inject anyway. These
         * will silently fail. And in any case, the guest will check
         * that no actual interrupt was injected for those cases.
         */

        for (f = 0, i = intid; i < (uint64_t)intid + num; i++, f++)
                fd[f] = kvm_new_eventfd();

        for (f = 0, i = intid; i < (uint64_t)intid + num; i++, f++) {
                assert(i <= (uint64_t)UINT_MAX);
                kvm_assign_irqfd(vm, i - MIN_SPI, fd[f]);
        }

        for (f = 0, i = intid; i < (uint64_t)intid + num; i++, f++) {
                val = 1;
                ret = write(fd[f], &val, sizeof(uint64_t));
                TEST_ASSERT(ret == sizeof(uint64_t),
                            __KVM_SYSCALL_ERROR("write()", ret));
        }

        for (f = 0, i = intid; i < (uint64_t)intid + num; i++, f++)
                kvm_close(fd[f]);
}

/* handles the valid case: intid=0xffffffff num=1 */
#define for_each_intid(first, num, tmp, i)                                      \
        for ((tmp) = (i) = (first);                                             \
                (tmp) < (uint64_t)(first) + (uint64_t)(num);                    \
                (tmp)++, (i)++)

static void run_guest_cmd(struct kvm_vcpu *vcpu, int gic_fd,
                          struct kvm_inject_args *inject_args,
                          struct test_args *test_args)
{
        kvm_inject_cmd cmd = inject_args->cmd;
        uint32_t intid = inject_args->first_intid;
        uint32_t num = inject_args->num;
        int level = inject_args->level;
        bool expect_failure = inject_args->expect_failure;
        struct kvm_vm *vm = vcpu->vm;
        uint64_t tmp;
        uint32_t i;

        /* handles the valid case: intid=0xffffffff num=1 */
        assert(intid < UINT_MAX - num || num == 1);

        switch (cmd) {
        case KVM_INJECT_EDGE_IRQ_LINE:
                for_each_intid(intid, num, tmp, i)
                        kvm_irq_line_check(vm, i, 1, test_args,
                                        expect_failure);
                for_each_intid(intid, num, tmp, i)
                        kvm_irq_line_check(vm, i, 0, test_args,
                                        expect_failure);
                break;
        case KVM_SET_IRQ_LINE:
                for_each_intid(intid, num, tmp, i)
                        kvm_irq_line_check(vm, i, level, test_args,
                                        expect_failure);
                break;
        case KVM_SET_IRQ_LINE_HIGH:
                for_each_intid(intid, num, tmp, i)
                        kvm_irq_line_check(vm, i, 1, test_args,
                                        expect_failure);
                break;
        case KVM_SET_LEVEL_INFO_HIGH:
                for_each_intid(intid, num, tmp, i)
                        kvm_irq_set_level_info_check(gic_fd, i, 1,
                                        expect_failure);
                break;
        case KVM_INJECT_IRQFD:
                kvm_routing_and_irqfd_check(vm, intid, num,
                                        test_args->kvm_max_routes,
                                        expect_failure);
                break;
        case KVM_WRITE_ISPENDR:
                for (i = intid; i < intid + num; i++)
                        kvm_irq_write_ispendr_check(gic_fd, i, vcpu,
                                                    expect_failure);
                break;
        case KVM_WRITE_ISACTIVER:
                for (i = intid; i < intid + num; i++)
                        kvm_irq_write_isactiver(gic_fd, i, vcpu);
                break;
        default:
                break;
        }
}

static void kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc,
                struct kvm_inject_args *args)
{
        struct kvm_inject_args *kvm_args_hva;
        vm_vaddr_t kvm_args_gva;

        kvm_args_gva = uc->args[1];
        kvm_args_hva = (struct kvm_inject_args *)addr_gva2hva(vm, kvm_args_gva);
        memcpy(args, kvm_args_hva, sizeof(struct kvm_inject_args));
}

static void print_args(struct test_args *args)
{
        printf("nr-irqs=%d level-sensitive=%d eoi-split=%d\n",
                        args->nr_irqs, args->level_sensitive,
                        args->eoi_split);
}

static void test_vgic(uint32_t nr_irqs, bool level_sensitive, bool eoi_split)
{
        struct ucall uc;
        int gic_fd;
        struct kvm_vcpu *vcpu;
        struct kvm_vm *vm;
        struct kvm_inject_args inject_args;
        vm_vaddr_t args_gva;

        struct test_args args = {
                .nr_irqs = nr_irqs,
                .level_sensitive = level_sensitive,
                .eoi_split = eoi_split,
                .kvm_max_routes = kvm_check_cap(KVM_CAP_IRQ_ROUTING),
                .kvm_supports_irqfd = kvm_check_cap(KVM_CAP_IRQFD),
        };

        print_args(&args);

        vm = vm_create_with_one_vcpu(&vcpu, guest_code);

        vm_init_descriptor_tables(vm);
        vcpu_init_descriptor_tables(vcpu);

        /* Setup the guest args page (so it gets the args). */
        args_gva = vm_vaddr_alloc_page(vm);
        memcpy(addr_gva2hva(vm, args_gva), &args, sizeof(args));
        vcpu_args_set(vcpu, 1, args_gva);

        gic_fd = vgic_v3_setup(vm, 1, nr_irqs);

        vm_install_exception_handler(vm, VECTOR_IRQ_CURRENT,
                guest_irq_handlers[args.eoi_split][args.level_sensitive]);

        while (1) {
                vcpu_run(vcpu);

                switch (get_ucall(vcpu, &uc)) {
                case UCALL_SYNC:
                        kvm_inject_get_call(vm, &uc, &inject_args);
                        run_guest_cmd(vcpu, gic_fd, &inject_args, &args);
                        break;
                case UCALL_ABORT:
                        REPORT_GUEST_ASSERT(uc);
                        break;
                case UCALL_DONE:
                        goto done;
                default:
                        TEST_FAIL("Unknown ucall %lu", uc.cmd);
                }
        }

done:
        close(gic_fd);
        kvm_vm_free(vm);
}

static void guest_code_asym_dir(struct test_args *args, int cpuid)
{
        gic_init(GIC_V3, 2);

        gic_set_eoi_split(1);
        gic_set_priority_mask(CPU_PRIO_MASK);

        if (cpuid == 0) {
                uint32_t intid;

                local_irq_disable();

                gic_set_priority(MIN_PPI, IRQ_DEFAULT_PRIO);
                gic_irq_enable(MIN_SPI);
                gic_irq_set_pending(MIN_SPI);

                intid = wait_for_and_activate_irq();
                GUEST_ASSERT_EQ(intid, MIN_SPI);

                gic_set_eoi(intid);
                isb();

                WRITE_ONCE(args->shared_data, MIN_SPI);
                dsb(ishst);

                do {
                        dsb(ishld);
                } while (READ_ONCE(args->shared_data) == MIN_SPI);
                GUEST_ASSERT(!gic_irq_get_active(MIN_SPI));
        } else {
                do {
                        dsb(ishld);
                } while (READ_ONCE(args->shared_data) != MIN_SPI);

                gic_set_dir(MIN_SPI);
                isb();

                WRITE_ONCE(args->shared_data, 0);
                dsb(ishst);
        }

        GUEST_DONE();
}

static void guest_code_group_en(struct test_args *args, int cpuid)
{
        uint32_t intid;

        gic_init(GIC_V3, 2);

        gic_set_eoi_split(0);
        gic_set_priority_mask(CPU_PRIO_MASK);
        /* SGI0 is G0, which is disabled */
        gic_irq_set_group(0, 0);

        /* Configure all SGIs with decreasing priority */
        for (intid = 0; intid < MIN_PPI; intid++) {
                gic_set_priority(intid, (intid + 1) * 8);
                gic_irq_enable(intid);
                gic_irq_set_pending(intid);
        }

        /* Ack and EOI all G1 interrupts */
        for (int i = 1; i < MIN_PPI; i++) {
                intid = wait_for_and_activate_irq();

                GUEST_ASSERT(intid < MIN_PPI);
                gic_set_eoi(intid);
                isb();
        }

        /*
         * Check that SGI0 is still pending, inactive, and that we cannot
         * ack anything.
         */
        GUEST_ASSERT(gic_irq_get_pending(0));
        GUEST_ASSERT(!gic_irq_get_active(0));
        GUEST_ASSERT_IAR_EMPTY();
        GUEST_ASSERT(read_sysreg_s(SYS_ICC_IAR0_EL1) == IAR_SPURIOUS);

        /* Open the G0 gates, and verify we can ack SGI0 */
        write_sysreg_s(1, SYS_ICC_IGRPEN0_EL1);
        isb();

        do {
                intid = read_sysreg_s(SYS_ICC_IAR0_EL1);
        } while (intid == IAR_SPURIOUS);

        GUEST_ASSERT(intid == 0);
        GUEST_DONE();
}

static void guest_code_timer_spi(struct test_args *args, int cpuid)
{
        uint32_t intid;
        u64 val;

        gic_init(GIC_V3, 2);

        gic_set_eoi_split(1);
        gic_set_priority_mask(CPU_PRIO_MASK);

        /* Add a pending SPI so that KVM starts trapping DIR */
        gic_set_priority(MIN_SPI + cpuid, IRQ_DEFAULT_PRIO);
        gic_irq_set_pending(MIN_SPI + cpuid);

        /* Configure the timer with a higher priority, make it pending */
        gic_set_priority(27, IRQ_DEFAULT_PRIO - 8);

        isb();
        val = read_sysreg(cntvct_el0);
        write_sysreg(val, cntv_cval_el0);
        write_sysreg(1, cntv_ctl_el0);
        isb();

        GUEST_ASSERT(gic_irq_get_pending(27));

        /* Enable both interrupts */
        gic_irq_enable(MIN_SPI + cpuid);
        gic_irq_enable(27);

        /* The timer must fire */
        intid = wait_for_and_activate_irq();
        GUEST_ASSERT(intid == 27);

        /* Check that we can deassert it */
        write_sysreg(0, cntv_ctl_el0);
        isb();

        GUEST_ASSERT(!gic_irq_get_pending(27));

        /*
         * Priority drop, deactivation -- we expect that the host
         * deactivation will have been effective
         */
        gic_set_eoi(27);
        gic_set_dir(27);

        GUEST_ASSERT(!gic_irq_get_active(27));

        /* Do it one more time */
        isb();
        val = read_sysreg(cntvct_el0);
        write_sysreg(val, cntv_cval_el0);
        write_sysreg(1, cntv_ctl_el0);
        isb();

        GUEST_ASSERT(gic_irq_get_pending(27));

        /* The timer must fire again */
        intid = wait_for_and_activate_irq();
        GUEST_ASSERT(intid == 27);

        GUEST_DONE();
}

static void *test_vcpu_run(void *arg)
{
        struct kvm_vcpu *vcpu = arg;
        struct ucall uc;

        while (1) {
                vcpu_run(vcpu);

                switch (get_ucall(vcpu, &uc)) {
                case UCALL_ABORT:
                        REPORT_GUEST_ASSERT(uc);
                        break;
                case UCALL_DONE:
                        return NULL;
                default:
                        TEST_FAIL("Unknown ucall %lu", uc.cmd);
                }
        }

        return NULL;
}

static void test_vgic_two_cpus(void *gcode)
{
        pthread_t thr[2];
        struct kvm_vcpu *vcpus[2];
        struct test_args args = {};
        struct kvm_vm *vm;
        vm_vaddr_t args_gva;
        int gic_fd, ret;

        vm = vm_create_with_vcpus(2, gcode, vcpus);

        vm_init_descriptor_tables(vm);
        vcpu_init_descriptor_tables(vcpus[0]);
        vcpu_init_descriptor_tables(vcpus[1]);

        /* Setup the guest args page (so it gets the args). */
        args_gva = vm_vaddr_alloc_page(vm);
        memcpy(addr_gva2hva(vm, args_gva), &args, sizeof(args));
        vcpu_args_set(vcpus[0], 2, args_gva, 0);
        vcpu_args_set(vcpus[1], 2, args_gva, 1);

        gic_fd = vgic_v3_setup(vm, 2, 64);

        ret = pthread_create(&thr[0], NULL, test_vcpu_run, vcpus[0]);
        if (ret)
                TEST_FAIL("Can't create thread for vcpu 0 (%d)\n", ret);
        ret = pthread_create(&thr[1], NULL, test_vcpu_run, vcpus[1]);
        if (ret)
                TEST_FAIL("Can't create thread for vcpu 1 (%d)\n", ret);

        pthread_join(thr[0], NULL);
        pthread_join(thr[1], NULL);

        close(gic_fd);
        kvm_vm_free(vm);
}

static void help(const char *name)
{
        printf(
        "\n"
        "usage: %s [-n num_irqs] [-e eoi_split] [-l level_sensitive]\n", name);
        printf(" -n: specify number of IRQs to setup the vgic with. "
                "It has to be a multiple of 32 and between 64 and 1024.\n");
        printf(" -e: if 1 then EOI is split into a write to DIR on top "
                "of writing EOI.\n");
        printf(" -l: specify whether the IRQs are level-sensitive (1) or not (0).");
        puts("");
        exit(1);
}

int main(int argc, char **argv)
{
        uint32_t nr_irqs = 64;
        bool default_args = true;
        bool level_sensitive = false;
        int opt;
        bool eoi_split = false;

        TEST_REQUIRE(kvm_supports_vgic_v3());
        test_disable_default_vgic();

        while ((opt = getopt(argc, argv, "hn:e:l:")) != -1) {
                switch (opt) {
                case 'n':
                        nr_irqs = atoi_non_negative("Number of IRQs", optarg);
                        if (nr_irqs > 1024 || nr_irqs % 32)
                                help(argv[0]);
                        break;
                case 'e':
                        eoi_split = (bool)atoi_paranoid(optarg);
                        default_args = false;
                        break;
                case 'l':
                        level_sensitive = (bool)atoi_paranoid(optarg);
                        default_args = false;
                        break;
                case 'h':
                default:
                        help(argv[0]);
                        break;
                }
        }

        /*
         * If the user just specified nr_irqs and/or gic_version, then run all
         * combinations.
         */
        if (default_args) {
                test_vgic(nr_irqs, false /* level */, false /* eoi_split */);
                test_vgic(nr_irqs, false /* level */, true /* eoi_split */);
                test_vgic(nr_irqs, true /* level */, false /* eoi_split */);
                test_vgic(nr_irqs, true /* level */, true /* eoi_split */);
                test_vgic_two_cpus(guest_code_asym_dir);
                test_vgic_two_cpus(guest_code_group_en);
                test_vgic_two_cpus(guest_code_timer_spi);
        } else {
                test_vgic(nr_irqs, level_sensitive, eoi_split);
        }

        return 0;
}