Symbol: kvm_svm
arch/x86/kvm/svm/avic.c
1004
struct kvm_svm *kvm_svm = to_kvm_svm(vcpu->kvm);
arch/x86/kvm/svm/avic.c
1047
WRITE_ONCE(kvm_svm->avic_physical_id_table[vcpu->vcpu_id], entry);
arch/x86/kvm/svm/avic.c
1071
struct kvm_svm *kvm_svm = to_kvm_svm(vcpu->kvm);
arch/x86/kvm/svm/avic.c
1103
WRITE_ONCE(kvm_svm->avic_physical_id_table[vcpu->vcpu_id], entry);
arch/x86/kvm/svm/avic.c
251
struct kvm_svm *kvm_svm;
arch/x86/kvm/svm/avic.c
260
hash_for_each_possible(svm_vm_data_hash, kvm_svm, hnode, vm_id) {
arch/x86/kvm/svm/avic.c
261
if (kvm_svm->avic_vm_id != vm_id)
arch/x86/kvm/svm/avic.c
263
vcpu = kvm_get_vcpu(&kvm_svm->kvm, vcpu_idx);
arch/x86/kvm/svm/avic.c
287
struct kvm_svm *kvm_svm = to_kvm_svm(kvm);
arch/x86/kvm/svm/avic.c
292
if (kvm_svm->avic_physical_id_table)
arch/x86/kvm/svm/avic.c
295
kvm_svm->avic_physical_id_table = (void *)__get_free_pages(GFP_KERNEL_ACCOUNT | __GFP_ZERO,
arch/x86/kvm/svm/avic.c
297
if (!kvm_svm->avic_physical_id_table)
arch/x86/kvm/svm/avic.c
306
struct kvm_svm *kvm_svm = to_kvm_svm(kvm);
arch/x86/kvm/svm/avic.c
311
free_page((unsigned long)kvm_svm->avic_logical_id_table);
arch/x86/kvm/svm/avic.c
312
free_pages((unsigned long)kvm_svm->avic_physical_id_table,
arch/x86/kvm/svm/avic.c
316
hash_del(&kvm_svm->hnode);
arch/x86/kvm/svm/avic.c
324
struct kvm_svm *kvm_svm = to_kvm_svm(kvm);
arch/x86/kvm/svm/avic.c
325
struct kvm_svm *k2;
arch/x86/kvm/svm/avic.c
331
kvm_svm->avic_logical_id_table = (void *)get_zeroed_page(GFP_KERNEL_ACCOUNT);
arch/x86/kvm/svm/avic.c
332
if (!kvm_svm->avic_logical_id_table)
arch/x86/kvm/svm/avic.c
349
kvm_svm->avic_vm_id = vm_id;
arch/x86/kvm/svm/avic.c
350
hash_add(svm_vm_data_hash, &kvm_svm->hnode, kvm_svm->avic_vm_id);
arch/x86/kvm/svm/avic.c
367
struct kvm_svm *kvm_svm = to_kvm_svm(svm->vcpu.kvm);
arch/x86/kvm/svm/avic.c
370
vmcb->control.avic_logical_id = __sme_set(__pa(kvm_svm->avic_logical_id_table));
arch/x86/kvm/svm/avic.c
371
vmcb->control.avic_physical_id = __sme_set(__pa(kvm_svm->avic_physical_id_table));
arch/x86/kvm/svm/avic.c
383
struct kvm_svm *kvm_svm = to_kvm_svm(vcpu->kvm);
arch/x86/kvm/svm/avic.c
435
WRITE_ONCE(kvm_svm->avic_physical_id_table[id], new_entry);
arch/x86/kvm/svm/avic.c
521
struct kvm_svm *kvm_svm = to_kvm_svm(kvm);
arch/x86/kvm/svm/avic.c
569
avic_logical_id_table = kvm_svm->avic_logical_id_table;
arch/x86/kvm/svm/avic.c
670
struct kvm_svm *kvm_svm = to_kvm_svm(vcpu->kvm);
arch/x86/kvm/svm/avic.c
691
return &kvm_svm->avic_logical_id_table[index];
arch/x86/kvm/svm/svm.c
5144
.vm_size = sizeof(struct kvm_svm),
arch/x86/kvm/svm/svm.c
5589
KVM_SANITY_CHECK_VM_STRUCT_SIZE(kvm_svm);
arch/x86/kvm/svm/svm.h
362
static __always_inline struct kvm_svm *to_kvm_svm(struct kvm *kvm)
arch/x86/kvm/svm/svm.h
364
return container_of(kvm, struct kvm_svm, kvm);