Symbol: EFER_LMA
arch/x86/kvm/emulate.c
1453
if (!(efer & EFER_LMA))
arch/x86/kvm/emulate.c
1502
if (!(efer & EFER_LMA))
arch/x86/kvm/emulate.c
1658
if (efer & EFER_LMA)
arch/x86/kvm/emulate.c
2373
if (efer & EFER_LMA) {
arch/x86/kvm/emulate.c
2381
if (efer & EFER_LMA) {
arch/x86/kvm/emulate.c
2423
if ((ctxt->mode != X86EMUL_MODE_PROT64) && (efer & EFER_LMA) &&
arch/x86/kvm/emulate.c
2439
if (efer & EFER_LMA) {
arch/x86/kvm/emulate.c
2448
ctxt->_eip = (efer & EFER_LMA) ? msr_data : (u32)msr_data;
arch/x86/kvm/emulate.c
2451
*reg_write(ctxt, VCPU_REGS_RSP) = (efer & EFER_LMA) ? msr_data :
arch/x86/kvm/emulate.c
2453
if (efer & EFER_LMA)
arch/x86/kvm/emulate.c
742
if (efer & EFER_LMA)
arch/x86/kvm/emulate.c
750
if (efer & EFER_LMA)
arch/x86/kvm/emulate.c
759
if (efer & EFER_LMA) {
arch/x86/kvm/mmu/mmu.c
215
BUILD_MMU_ROLE_REGS_ACCESSOR(efer, lma, EFER_LMA);
arch/x86/kvm/smm.c
536
if (__kvm_emulate_msr_write(vcpu, MSR_EFER, smstate->efer & ~EFER_LMA))
arch/x86/kvm/svm/svm.c
1728
vcpu->arch.efer |= EFER_LMA;
arch/x86/kvm/svm/svm.c
1730
svm->vmcb->save.efer |= EFER_LMA | EFER_LME;
arch/x86/kvm/svm/svm.c
1734
vcpu->arch.efer &= ~EFER_LMA;
arch/x86/kvm/svm/svm.c
1736
svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME);
arch/x86/kvm/svm/svm.c
213
if (!(efer & EFER_LMA))
arch/x86/kvm/vmx/nested.c
2280
return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/nested.c
2282
return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/nested.c
2492
if (guest_efer & EFER_LMA)
arch/x86/kvm/vmx/nested.c
3125
!!(vcpu->arch.efer & EFER_LMA)))
arch/x86/kvm/vmx/nested.c
3216
CC(ia32e != !!(vmcs12->host_ia32_efer & EFER_LMA)) ||
arch/x86/kvm/vmx/nested.c
3355
CC(ia32e != !!(vmcs12->guest_ia32_efer & EFER_LMA)) ||
arch/x86/kvm/vmx/nested.c
4784
vcpu->arch.efer |= (EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/nested.c
4786
vcpu->arch.efer &= ~(EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/tdx.c
696
vcpu->arch.efer = EFER_SCE | EFER_LME | EFER_LMA | EFER_NX;
arch/x86/kvm/vmx/vmx.c
1170
ignore_bits |= EFER_LMA | EFER_LME;
arch/x86/kvm/vmx/vmx.c
1172
if (guest_efer & EFER_LMA)
arch/x86/kvm/vmx/vmx.c
1183
if (!(guest_efer & EFER_LMA))
arch/x86/kvm/vmx/vmx.c
3319
if (efer & EFER_LMA)
arch/x86/kvm/vmx/vmx.c
3324
if (KVM_BUG_ON(efer & EFER_LMA, vcpu->kvm))
arch/x86/kvm/vmx/vmx.c
3348
vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA);
arch/x86/kvm/vmx/vmx.c
3353
vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA);
arch/x86/kvm/vmx/vmx.c
6646
vcpu->arch.efer | (EFER_LMA | EFER_LME));
arch/x86/kvm/vmx/vmx.c
6649
vcpu->arch.efer & ~(EFER_LMA | EFER_LME));
arch/x86/kvm/x86.c
115
u64 __read_mostly efer_reserved_bits = ~((u64)(EFER_SCE | EFER_LME | EFER_LMA));
arch/x86/kvm/x86.c
12337
if (!(sregs->cr4 & X86_CR4_PAE) || !(sregs->efer & EFER_LMA))
arch/x86/kvm/x86.c
12346
if (sregs->efer & EFER_LMA || sregs->cs.l)
arch/x86/kvm/x86.c
12454
!(sregs2->efer & EFER_LMA);
arch/x86/kvm/x86.c
1737
if (efer & (EFER_LME | EFER_LMA) &&
arch/x86/kvm/x86.c
1774
efer &= ~EFER_LMA;
arch/x86/kvm/x86.c
1775
efer |= vcpu->arch.efer & EFER_LMA;
arch/x86/kvm/x86.h
244
return !!(vcpu->arch.efer & EFER_LMA);
arch/x86/realmode/init.c
152
trampoline_header->efer = efer & ~EFER_LMA;
tools/testing/selftests/kvm/lib/x86/processor.c
665
sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX);