MSR_EE
# define MSR_KERNEL (MSR_EE | MSR_VM)
# define MSR_KERNEL_VMS (MSR_EE | MSR_VMS)
childregs->msr |= MSR_EE; /* exceptions will be enabled*/
WARN_ON_ONCE(mfmsr() & MSR_EE);
WARN_ON_ONCE(mfmsr() & MSR_EE);
WARN_ON(mfmsr() & MSR_EE);
mtmsr(flags & ~MSR_EE);
return (flags & MSR_EE) == 0;
return !(regs->msr & MSR_EE);
msr &= ~MSR_EE;
wrtee(MSR_EE);
__mtmsrd(MSR_EE | MSR_RI, 1);
mtmsr(mfmsr() | MSR_EE);
mtmsr(mfmsr() & ~MSR_EE);
mtmsr(mfmsr() & ~(MSR_EE | MSR_RI));
INT_SOFT_MASK_BUG_ON(regs, !(regs->msr & MSR_EE));
!(regs->msr & MSR_EE));
if (!(regs->msr & MSR_EE) || is_implicit_soft_masked(regs)) {
asm volatile("wrteei %0" : : "i" ((val & MSR_EE) ? 1 : 0) : "memory");
#define MSR_USER32 (MSR_ | MSR_PR | MSR_EE)
#define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
#define MSR_USER32 (MSR_ | MSR_PR | MSR_EE)
#define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
WARN_ON_ONCE(!(regs->msr & MSR_EE));
WARN_ON_ONCE(mfmsr() & MSR_EE);
regs.msr |= MSR_EE;
WARN_ON_ONCE(!(mfmsr() & MSR_EE));
WARN_ON_ONCE(mfmsr() & MSR_EE);
if (!(mfmsr() & MSR_EE)) {
if (WARN_ON_ONCE(mfmsr() & MSR_EE))
KVM_INST_LI | __PPC_RT(R30) | MSR_EE;
{MSR_EE, "EE"},
WARN_ON(mfmsr() & MSR_EE);
WARN_ON_ONCE(mfmsr() & MSR_EE);
deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit;
deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit;
new_msr &= ~(MSR_RI | MSR_EE);
new_msr |= rs_val & (MSR_RI | MSR_EE);
__kvmppc_set_msr_hv(vcpu, __kvmppc_get_msr_hv(vcpu) | MSR_EE);
if (!kvmhv_on_pseries() && (__kvmppc_get_msr_hv(vcpu) & MSR_EE))
if (vcpu->arch.shregs.msr & MSR_EE) {
msr &= ~MSR_EE;
msr = mfmsr() & ~MSR_EE;
smsr |= MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_PR | MSR_EE;
kvmppc_set_msr_fast(vcpu, kvmppc_get_msr(vcpu) | MSR_EE);
allowed = vcpu->arch.shared->msr & MSR_EE;
vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE)
| (kvmppc_get_gpr(vcpu, rs) & MSR_EE);
vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE)
| (inst & MSR_EE);
WARN_ON_ONCE(mfmsr() & MSR_EE);
VM_WARN_ON(mfmsr() & MSR_EE);
mtmsr(msr_save|MSR_WE|MSR_EE|MSR_CE|MSR_DE);
mtmsr(msr_save|MSR_WE|MSR_EE|MSR_CE|MSR_DE);
mtmsr(mfmsr() & ~(MSR_ME | MSR_EE | MSR_IR | MSR_DR));
msr &= ~MSR_EE;
if (WARN_ON(!(mfmsr() & MSR_EE)))