MSR_LE
return (kvmppc_get_msr(vcpu) & MSR_LE) != (MSR_KERNEL & MSR_LE);
#define MSR_ (__MSR | MSR_LE)
#define MSR_IDLE (MSR_ME | MSR_SF | MSR_HV | MSR_LE)
if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) {
{MSR_LE, "LE"},
regs_set_return_msr(regs, regs->msr & ~MSR_LE);
regs_set_return_msr(regs, regs->msr | MSR_LE);
if (regs->msr & MSR_LE) {
regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE));
regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE));
regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (MSR_KERNEL & MSR_LE));
regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (MSR_KERNEL & MSR_LE));
regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE));
regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE));
regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (MSR_KERNEL & MSR_LE));
current->thread.regs->msr ^ MSR_LE);
swap = (msr & MSR_LE) != (MSR_KERNEL & MSR_LE);
if ((kvmppc_get_msr(vcpu) & MSR_LE) &&
vcpu->arch.intr_msr |= MSR_LE;
vcpu->arch.intr_msr &= ~MSR_LE;
newmsr = (newmsr & ~MSR_LE) | (msr & MSR_LE);
newmsr = (newmsr & ~MSR_LE) | (msr & MSR_LE);
if (vcpu->arch.intr_msr & MSR_LE)
vcpu->arch.intr_msr |= MSR_LE;
vcpu->arch.intr_msr &= ~MSR_LE;
vcpu->arch.shadow_msr = MSR_USER64 & ~MSR_LE;
smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE |
smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE;
if (vcpu->arch.intr_msr & MSR_LE)
op->val = 0xffffffff & ~(MSR_ME | MSR_LE);
cross_endian = (regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE);
(MSR_LE|MSR_RI|MSR_DR|MSR_IR|MSR_ME|MSR_PR|
thread_endianness = MSR_LE & ucp->uc_mcontext.gp_regs[PT_MSR];