X86_CR4_PKE
if (native_read_cr4() & X86_CR4_PKE)
| X86_CR4_SMAP | X86_CR4_PKE | X86_CR4_UMIP \
cr4_set_bits(X86_CR4_PKE);
if (cr4 & X86_CR4_PKE)
kvm_is_cr4_bit_set(vcpu, X86_CR4_PKE));
X86_CR4_SMEP | X86_CR4_SMAP | X86_CR4_PKE)
BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, pke, X86_CR4_PKE);
.cr4 = cr4 & ~X86_CR4_PKE,
cr4 &= ~(X86_CR4_SMEP | X86_CR4_SMAP | X86_CR4_PKE);
if ((cr4 ^ old_cr4) & (X86_CR4_OSXSAVE | X86_CR4_PKE))
hw_cr4 &= ~(X86_CR4_SMEP | X86_CR4_SMAP | X86_CR4_PKE);
if ((cr4 ^ old_cr4) & (X86_CR4_OSXSAVE | X86_CR4_PKE))
cr4_fixed1_update(X86_CR4_PKE, ecx, feature_bit(PKU));
kvm_is_cr4_bit_set(vcpu, X86_CR4_PKE)))
kvm_is_cr4_bit_set(vcpu, X86_CR4_PKE))) {
__reserved_bits |= X86_CR4_PKE; \
sregs.cr4 ^= X86_CR4_PKE;
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_PKE);
cr4 |= X86_CR4_PKE;
TEST_ASSERT(!!(sregs.cr4 & X86_CR4_PKE) ==
(sregs.cr4 & X86_CR4_PKE) ? "set" : "clear");
set_cr4(get_cr4() | X86_CR4_PKE);