guest_code
u16 guest_code;
u16 guest_code;
struct block_data guest_code;
blkcount->guest_code = partition_length;
phyaddr->guest_code = physical_address;
__func__, blkcount->guest_code);
return rmi_f34v7_write_f34v7_blocks(f34, f34->v7.img.guest_code.data,
f34->v7.img.guest_code.size /
f34->v7.img.guest_code.data = content;
f34->v7.img.guest_code.size = length;
OPT_BOOLEAN(0, "guest-code", &symbol_conf.guest_code,
OPT_BOOLEAN(0, "guest-code", &symbol_conf.guest_code,
symbol_conf.guest_code) {
if (symbol_conf.guest_code && !machine__is_host(machine))
if (symbol_conf.guest_code && (state->from_nr || state->to_nr))
if (symbol_conf.guest_code) {
} else if ((!symbol_conf.guest_code && cpumode != PERF_RECORD_MISC_GUEST_KERNEL) ||
if (symbol_conf.guest_code)
guest_code,
vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus);
*vm = vm_create_with_one_vcpu(vcpu, guest_code);
vcpu = aarch64_vcpu_add(vm, 0, &init, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
static struct kvm_vm *vm_create_with_dabt_handler(struct kvm_vcpu **vcpu, void *guest_code,
struct kvm_vm *vm = vm_create_with_one_vcpu(vcpu, guest_code);
vcpu = aarch64_vcpu_add(vm, 0, &init, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vcpu = vm_vcpu_add(vm, 0, guest_code);
static struct kvm_vm *setup_vm(void *guest_code, struct kvm_vcpu **source,
*source = aarch64_vcpu_add(vm, 0, &init, guest_code);
*target = aarch64_vcpu_add(vm, 1, &init, guest_code);
vm = __vm_create_with_one_vcpu(vcpu, num_guest_pages, guest_code);
vcpu = vm_vcpu_add(vm, 0, guest_code);
vcpus[i] = vm_vcpu_add(v.vm, i, guest_code);
(void)vm_vcpu_add(v.vm, 0, guest_code);
(void)vm_vcpu_add(v.vm, 3, guest_code);
(void)vm_vcpu_add(v.vm, 1, guest_code);
(void)vm_vcpu_add(v.vm, 2, guest_code);
vm_vcpu_add(v.vm, vcpuids[i], guest_code);
vcpus[i] = vm_vcpu_add(v.vm, i, guest_code);
v.vm = vm_create_with_vcpus(NR_VCPUS, guest_code, vcpus);
v.vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_vcpus(test_data.nr_cpus, guest_code, vcpus);
static void create_vpmu_vm(void *guest_code)
vpmu_vm.vcpu = aarch64_vcpu_add(vpmu_vm.vm, 0, &init, guest_code);
create_vpmu_vm(guest_code);
vcpu_set_reg(vcpu, ARM64_CORE_REG(regs.pc), (uint64_t)guest_code);
create_vpmu_vm(guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
uint64_t extra_mem_pages, void *guest_code)
*vcpu = vm_vcpu_add(vm, 0, guest_code);
2ul << (DIRTY_MEM_BITS - PAGE_SHIFT_4K), guest_code);
vm = __vm_create_shape_with_one_vcpu(VM_SHAPE_DEFAULT, &vcpu, 1, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vcpu = vm_vcpu_add(vm, i, guest_code);
struct kvm_vcpu_init *init, void *guest_code);
void *guest_code, struct kvm_vcpu *vcpus[]);
void *guest_code,
guest_code, vcpus);
void *guest_code);
void *guest_code)
extra_mem_pages, guest_code);
void *guest_code)
return __vm_create_with_one_vcpu(vcpu, 0, guest_code);
void *guest_code)
return __vm_create_shape_with_one_vcpu(shape, vcpu, 0, guest_code);
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code);
void *guest_code)
vcpu_arch_set_entry_point(vcpu, guest_code);
struct kvm_vm *vm_sev_create_with_one_vcpu(uint32_t type, void *guest_code,
guest_code, test_args.vcpus);
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
vcpu_set_reg(vcpu, ARM64_CORE_REG(regs.pc), (uint64_t)guest_code);
struct kvm_vcpu_init *init, void *guest_code)
vcpu_arch_set_entry_point(vcpu, guest_code);
void *guest_code, struct kvm_vcpu *vcpus[])
vcpus[i] = vm_vcpu_add(vm, i, guest_code);
void *guest_code)
vm = __vm_create_with_vcpus(shape, 1, extra_mem_pages, guest_code, vcpus);
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
regs.pc = (uint64_t)guest_code;
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
vcpu_set_reg(vcpu, RISCV_CORE_REG(regs.pc), (unsigned long)guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
vcpu->run->psw_addr = (uintptr_t)guest_code;
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
regs.rip = (unsigned long) guest_code;
struct kvm_vm *vm_sev_create_with_one_vcpu(uint32_t type, void *guest_code,
vm = __vm_create_with_vcpus(shape, 1, 0, guest_code, cpus);
vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus);
void *guest_code, uint64_t mem_size,
data->vm = __vm_create_with_one_vcpu(&data->vcpu, mempages, guest_code);
void (*guest_code)(void);
if (!prepare_vm(data, nslots, maxslots, tdata->guest_code,
.guest_code = guest_code_test_memslot_map,
.guest_code = guest_code_test_memslot_unmap,
.guest_code = guest_code_test_memslot_unmap,
.guest_code = guest_code_test_memslot_move,
.guest_code = guest_code_test_memslot_move,
.guest_code = guest_code_test_memslot_rw,
guest_code, vcpus);
vm = vm_create_shape_with_one_vcpu(shape, &vcpu, guest_code);
vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
static void test_vm_basic_test(void *guest_code)
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
static void test_vm_events_test(void *guest_code)
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
static void test_vm_events_snapshot_test(void *guest_code)
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
static void test_vm_events_overflow(void *guest_code)
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
static struct kvm_vm *test_step_int_1(struct kvm_vcpu **vcpu, void *guest_code,
vm = vm_create_with_one_vcpu(vcpu, guest_code);
static void test_step_int(void *guest_code, size_t new_psw_off)
vm = test_step_int_1(&vcpu, guest_code, new_psw_off, new_psw);
extern char guest_code[];
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
static struct test_default test_default_init(void *guest_code)
t.kvm_vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
void *guest_code)
vm = vm_create_with_one_vcpu(vcpu, guest_code);
vm = vm_create_with_vcpus(NR_VCPUS, guest_code, vcpus);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vcpu = vm_vcpu_add(vm, 0, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vcpu = vm_vcpu_add(vm, 0, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vcpu = vm_vcpu_add(vm, 0, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
void *guest_code,
vm = vm_create_with_one_vcpu(vcpu, guest_code);
static void test_pmu_config_disable(void (*guest_code)(void))
vcpu = vm_vcpu_add(vm, 0, guest_code);
void (*guest_code)(void);
guest_code = use_intel_pmu() ? intel_guest_code : amd_guest_code;
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
test_pmu_config_disable(guest_code);
vm = __vm_create_with_vcpus(shape, nr_vcpus, 0, guest_code, vcpus);
static void test_sev(void *guest_code, uint32_t type, uint64_t policy)
vm = vm_sev_create_with_one_vcpu(type, guest_code, &vcpu);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
KVM_ONE_VCPU_TEST(sync_regs_test, read_invalid, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, set_invalid, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, req_and_verify_all_valid, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, set_and_verify_various, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, clear_kvm_dirty_regs_bits, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, clear_kvm_valid_and_dirty_regs, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, clear_kvm_valid_regs_bits, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, race_cr4, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, race_exc, guest_code)
KVM_ONE_VCPU_TEST(sync_regs_test, race_inj_pen, guest_code)
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vcpu = vm_vcpu_add(vm, vcpu_id, guest_code);
bool enable_cmci_p, void *guest_code)
struct kvm_vcpu *vcpu = vm_vcpu_add(vm, vcpuid, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
KVM_ONE_VCPU_TEST(vmx_pmu_caps, basic_perf_capabilities, guest_code)
KVM_ONE_VCPU_TEST(vmx_pmu_caps, fungible_perf_capabilities, guest_code)
KVM_ONE_VCPU_TEST(vmx_pmu_caps, immutable_perf_capabilities, guest_code)
KVM_ONE_VCPU_TEST(vmx_pmu_caps, lbr_perf_capabilities, guest_code)
KVM_ONE_VCPU_TEST(vmx_pmu_caps, perf_capabilities_unsupported, guest_code)
KVM_ONE_VCPU_TEST(vmx_pmu_caps, guest_wrmsr_perf_capabilities, guest_code)
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);
vm = vm_create_with_one_vcpu(&vcpu, guest_code);