memstress_args
uint64_t gva = base_gva + page * memstress_args.guest_page_size;
if (READ_ONCE(memstress_args.stop_vcpus))
struct kvm_vm *vm = memstress_args.vm;
max(memstress_args.guest_page_size,
vcpu_args = &memstress_args.vcpu_args[i];
vcpu_args->pages * memstress_args.guest_page_size);
vcpu_args = &memstress_args.vcpu_args[i];
vcpu_paging_rate = memstress_args.vcpu_args[0].pages /
kvm_parse_vcpu_pinning(cpulist, memstress_args.vcpu_to_pcpu,
memstress_args.pin_vcpus = true;
memstress_args.nested = true;
kvm_parse_vcpu_pinning(pcpu_list, memstress_args.vcpu_to_pcpu,
memstress_args.pin_vcpus = true;
extern struct memstress_args memstress_args;
struct memstress_args *args = &memstress_args;
struct memstress_args memstress_args;
sync_global_to_guest(vm, memstress_args);
memstress_args.write_percent = write_percent;
sync_global_to_guest(vm, memstress_args.write_percent);
memstress_args.random_access = random_access;
sync_global_to_guest(vm, memstress_args.random_access);
if (memstress_args.pin_vcpus)
pin_self_to_cpu(memstress_args.vcpu_to_pcpu[vcpu_idx]);
vcpu_thread_fn(&memstress_args.vcpu_args[vcpu_idx]);
WRITE_ONCE(memstress_args.stop_vcpus, false);
WRITE_ONCE(memstress_args.stop_vcpus, true);
struct memstress_args *args = &memstress_args;
for (i = 0; i < sizeof(memstress_args); i += args->guest_page_size)
struct memstress_args *args = &memstress_args;
end = align_up(memstress_args.gpa + memstress_args.size, PG_SIZE_1G);
start = align_down(memstress_args.gpa, PG_SIZE_1G);
while (!READ_ONCE(memstress_args.stop_vcpus)) {
gpa = memstress_args.gpa - pages * vm->page_size;