KVM_MEM_GUEST_MEMFD
return slot && (slot->flags & KVM_MEM_GUEST_MEMFD);
vm_set_user_memory_region2(vm, slot, KVM_MEM_GUEST_MEMFD, gpa, size, NULL, fd, 0);
if (flags & KVM_MEM_GUEST_MEMFD) {
TEST_ASSERT(region && region->region.flags & KVM_MEM_GUEST_MEMFD,
flags |= KVM_MEM_GUEST_MEMFD;
TEST_NPAGES, private ? KVM_MEM_GUEST_MEMFD : 0);
.flags = private ? KVM_MEM_GUEST_MEMFD : 0,
const uint32_t v2_only_flags = KVM_MEM_GUEST_MEMFD;
supported_flags |= KVM_MEM_GUEST_MEMFD;
if (supported_flags & KVM_MEM_GUEST_MEMFD) {
KVM_MEM_LOG_DIRTY_PAGES | KVM_MEM_GUEST_MEMFD,
KVM_MEM_READONLY | KVM_MEM_GUEST_MEMFD,
int r = __vm_set_user_memory_region2(vm, MEM_REGION_SLOT, KVM_MEM_GUEST_MEMFD,
vm_set_user_memory_region2(vm2, MEM_REGION_SLOT, KVM_MEM_GUEST_MEMFD,
vm_set_user_memory_region2(vm, MEM_REGION_SLOT, KVM_MEM_GUEST_MEMFD,
vm_set_user_memory_region2(vm, MEM_REGION_SLOT, KVM_MEM_GUEST_MEMFD,
vm_set_user_memory_region2(vm, MEM_REGION_SLOT + 1, KVM_MEM_GUEST_MEMFD,
vm_set_user_memory_region2(vm, MEM_REGION_SLOT, KVM_MEM_GUEST_MEMFD,
r = __vm_set_user_memory_region2(vm, MEM_REGION_SLOT, KVM_MEM_GUEST_MEMFD,
r = __vm_set_user_memory_region2(vm, MEM_REGION_SLOT, KVM_MEM_GUEST_MEMFD,
KVM_MEM_GUEST_MEMFD, memfd, slot_size * i);
KVM_MEM_GUEST_MEMFD);
valid_flags |= KVM_MEM_GUEST_MEMFD;
if (mem->flags & KVM_MEM_GUEST_MEMFD)
!(mem->flags & KVM_MEM_GUEST_MEMFD))
WARN_ON_ONCE(old->flags & KVM_MEM_GUEST_MEMFD);
if (WARN_ON_ONCE(old->flags & KVM_MEM_GUEST_MEMFD))
if (mem->flags & KVM_MEM_GUEST_MEMFD &&
if (mem->flags & KVM_MEM_GUEST_MEMFD)
((mem->flags ^ old->flags) & (KVM_MEM_READONLY | KVM_MEM_GUEST_MEMFD)))
if (mem->flags & KVM_MEM_GUEST_MEMFD) {
if (mem->flags & KVM_MEM_GUEST_MEMFD)
if (slot->flags & KVM_MEM_GUEST_MEMFD)