Symbol: cr4
arch/x86/coco/sev/core.c
767
u64 cr4;
arch/x86/coco/sev/core.c
804
cr4 = native_read_cr4() & X86_CR4_MCE;
arch/x86/coco/sev/core.c
831
vmsa->cr4 = cr4;
arch/x86/coco/sev/vc-shared.c
465
u32 cr4 = native_read_cr4();
arch/x86/coco/sev/vc-shared.c
471
if (cr4 & X86_CR4_OSXSAVE)
arch/x86/coco/sev/vc-shared.c
533
u32 cr4 = native_read_cr4();
arch/x86/coco/sev/vc-shared.c
546
if (cr4 & X86_CR4_OSXSAVE)
arch/x86/hyperv/hv_crash.c
171
asm volatile("movq %0, %%cr4" : : "r"(hv_crash_ctxt.cr4));
arch/x86/hyperv/hv_crash.c
205
ctxt->cr4 = native_read_cr4();
arch/x86/hyperv/hv_crash.c
61
ulong cr4;
arch/x86/hyperv/hv_vtl.c
168
input->vp_context.cr4 = native_read_cr4();
arch/x86/hyperv/ivm.c
333
vmsa->cr4 = native_read_cr4();
arch/x86/include/asm/kvm_host.h
1779
bool (*is_valid_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
1780
void (*set_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
2233
void kvm_post_set_cr4(struct kvm_vcpu *vcpu, unsigned long old_cr4, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
2236
int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
809
unsigned long cr4;
arch/x86/include/asm/realmode.h
53
u32 cr4;
arch/x86/include/asm/suspend_32.h
16
unsigned long cr0, cr2, cr3, cr4;
arch/x86/include/asm/suspend_64.h
42
unsigned long cr0, cr2, cr3, cr4;
arch/x86/include/asm/svm.h
338
u64 cr4;
arch/x86/include/asm/svm.h
395
u64 cr4;
arch/x86/include/asm/tlbflush.h
130
unsigned long cr4;
arch/x86/include/asm/tlbflush.h
181
this_cpu_write(cpu_tlbstate.cr4, __read_cr4());
arch/x86/include/asm/tlbflush.h
485
static inline void __native_tlb_flush_global(unsigned long cr4)
arch/x86/include/asm/tlbflush.h
487
native_write_cr4(cr4 ^ X86_CR4_PGE);
arch/x86/include/asm/tlbflush.h
488
native_write_cr4(cr4);
arch/x86/include/uapi/asm/kvm.h
155
__u64 cr0, cr2, cr3, cr4, cr8;
arch/x86/include/uapi/asm/kvm.h
166
__u64 cr0, cr2, cr3, cr4, cr8;
arch/x86/kernel/asm-offsets_64.c
52
ENTRY(cr4);
arch/x86/kernel/cpu/common.c
496
unsigned long newval, cr4 = this_cpu_read(cpu_tlbstate.cr4);
arch/x86/kernel/cpu/common.c
500
newval = (cr4 & ~clear) | set;
arch/x86/kernel/cpu/common.c
501
if (newval != cr4) {
arch/x86/kernel/cpu/common.c
502
this_cpu_write(cpu_tlbstate.cr4, newval);
arch/x86/kernel/cpu/common.c
511
return this_cpu_read(cpu_tlbstate.cr4);
arch/x86/kernel/cpu/common.c
517
unsigned long cr4 = __read_cr4();
arch/x86/kernel/cpu/common.c
520
cr4 |= X86_CR4_PCIDE;
arch/x86/kernel/cpu/common.c
522
cr4 = (cr4 & ~cr4_pinned_mask) | cr4_pinned_bits;
arch/x86/kernel/cpu/common.c
524
__write_cr4(cr4);
arch/x86/kernel/cpu/common.c
527
this_cpu_write(cpu_tlbstate.cr4, cr4);
arch/x86/kernel/cpu/common.c
537
cr4_pinned_bits = this_cpu_read(cpu_tlbstate.cr4) & cr4_pinned_mask;
arch/x86/kernel/cpu/mtrr/cyrix.c
135
static u32 cr4, ccr3;
arch/x86/kernel/cpu/mtrr/cyrix.c
143
cr4 = __read_cr4();
arch/x86/kernel/cpu/mtrr/cyrix.c
144
__write_cr4(cr4 & ~X86_CR4_PGE);
arch/x86/kernel/cpu/mtrr/cyrix.c
176
__write_cr4(cr4);
arch/x86/kernel/head64.c
274
__native_tlb_flush_global(this_cpu_read(cpu_tlbstate.cr4));
arch/x86/kernel/process.c
708
unsigned long newval, cr4 = this_cpu_read(cpu_tlbstate.cr4);
arch/x86/kernel/process.c
710
newval = cr4 ^ mask;
arch/x86/kernel/process.c
711
if (newval != cr4) {
arch/x86/kernel/process.c
712
this_cpu_write(cpu_tlbstate.cr4, newval);
arch/x86/kernel/process_32.c
62
unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L;
arch/x86/kernel/process_32.c
83
cr4 = __read_cr4();
arch/x86/kernel/process_32.c
85
log_lvl, cr0, cr2, cr3, cr4);
arch/x86/kernel/process_64.c
119
cr4 = __read_cr4();
arch/x86/kernel/process_64.c
126
log_lvl, cr2, cr3, cr4);
arch/x86/kernel/process_64.c
144
if (cr4 & X86_CR4_PKE)
arch/x86/kernel/process_64.c
73
unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L, fs, gs, shadowgs;
arch/x86/kvm/emulate.c
3828
u64 cr4;
arch/x86/kvm/emulate.c
3833
cr4 = ctxt->ops->get_cr(ctxt, 4);
arch/x86/kvm/emulate.c
3834
if ((cr4 & X86_CR4_DE) && (dr == 4 || dr == 5))
arch/x86/kvm/emulate.c
3886
u64 cr4 = ctxt->ops->get_cr(ctxt, 4);
arch/x86/kvm/emulate.c
3888
if (cr4 & X86_CR4_TSD && ctxt->ops->cpl(ctxt))
arch/x86/kvm/emulate.c
3896
u64 cr4 = ctxt->ops->get_cr(ctxt, 4);
arch/x86/kvm/emulate.c
3911
if ((!(cr4 & X86_CR4_PCE) && ctxt->ops->cpl(ctxt)) ||
arch/x86/kvm/kvm_cache_regs.h
197
return vcpu->arch.cr4 & mask;
arch/x86/kvm/mmu.h
88
unsigned long cr4, u64 efer, gpa_t nested_cr3);
arch/x86/kvm/mmu/mmu.c
188
const unsigned long cr4;
arch/x86/kvm/mmu/mmu.c
208
BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, pse, X86_CR4_PSE);
arch/x86/kvm/mmu/mmu.c
209
BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, pae, X86_CR4_PAE);
arch/x86/kvm/mmu/mmu.c
210
BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, smep, X86_CR4_SMEP);
arch/x86/kvm/mmu/mmu.c
211
BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, smap, X86_CR4_SMAP);
arch/x86/kvm/mmu/mmu.c
212
BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, pke, X86_CR4_PKE);
arch/x86/kvm/mmu/mmu.c
213
BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, la57, X86_CR4_LA57);
arch/x86/kvm/mmu/mmu.c
229
BUILD_MMU_ROLE_ACCESSOR(ext, cr4, pse);
arch/x86/kvm/mmu/mmu.c
230
BUILD_MMU_ROLE_ACCESSOR(ext, cr4, smep);
arch/x86/kvm/mmu/mmu.c
231
BUILD_MMU_ROLE_ACCESSOR(ext, cr4, smap);
arch/x86/kvm/mmu/mmu.c
232
BUILD_MMU_ROLE_ACCESSOR(ext, cr4, pke);
arch/x86/kvm/mmu/mmu.c
233
BUILD_MMU_ROLE_ACCESSOR(ext, cr4, la57);
arch/x86/kvm/mmu/mmu.c
251
.cr4 = kvm_read_cr4_bits(vcpu, KVM_MMU_CR4_ROLE_BITS),
arch/x86/kvm/mmu/mmu.c
5874
unsigned long cr4, u64 efer, gpa_t nested_cr3)
arch/x86/kvm/mmu/mmu.c
5879
.cr4 = cr4 & ~X86_CR4_PKE,
arch/x86/kvm/smm.c
220
smram->cr4 = kvm_read_cr4(vcpu);
arch/x86/kvm/smm.c
245
smram->cr4 = kvm_read_cr4(vcpu);
arch/x86/kvm/smm.c
30
CHECK_SMRAM32_OFFSET(cr4, 0xFF14);
arch/x86/kvm/smm.c
420
u64 cr0, u64 cr3, u64 cr4)
arch/x86/kvm/smm.c
427
if (cr4 & X86_CR4_PCIDE) {
arch/x86/kvm/smm.c
441
bad = kvm_set_cr4(vcpu, cr4 & ~X86_CR4_PCIDE);
arch/x86/kvm/smm.c
449
if (cr4 & X86_CR4_PCIDE) {
arch/x86/kvm/smm.c
450
bad = kvm_set_cr4(vcpu, cr4);
arch/x86/kvm/smm.c
504
smstate->cr3, smstate->cr4);
arch/x86/kvm/smm.c
551
r = rsm_enter_protected_mode(vcpu, smstate->cr0, smstate->cr3, smstate->cr4);
arch/x86/kvm/smm.c
600
unsigned long cr4;
arch/x86/kvm/smm.c
603
cr4 = kvm_read_cr4(vcpu);
arch/x86/kvm/smm.c
604
if (cr4 & X86_CR4_PCIDE)
arch/x86/kvm/smm.c
605
kvm_set_cr4(vcpu, cr4 & ~X86_CR4_PCIDE);
arch/x86/kvm/smm.c
622
unsigned long cr4, efer;
arch/x86/kvm/smm.c
625
cr4 = kvm_read_cr4(vcpu);
arch/x86/kvm/smm.c
626
if (cr4 & X86_CR4_PAE)
arch/x86/kvm/smm.c
627
kvm_set_cr4(vcpu, cr4 & ~X86_CR4_PAE);
arch/x86/kvm/smm.c
96
CHECK_SMRAM64_OFFSET(cr4, 0xFF48);
arch/x86/kvm/smm.h
127
u64 cr4;
arch/x86/kvm/smm.h
31
u32 cr4;
arch/x86/kvm/svm/nested.c
1053
vmcb01->save.cr4 = vcpu->arch.cr4;
arch/x86/kvm/svm/nested.c
1099
to_save->cr4 = from_save->cr4;
arch/x86/kvm/svm/nested.c
1168
vmcb12->save.cr4 = svm->vcpu.arch.cr4;
arch/x86/kvm/svm/nested.c
1291
svm_set_cr4(vcpu, vmcb01->save.cr4);
arch/x86/kvm/svm/nested.c
389
if (CC(!(save->cr4 & X86_CR4_PAE)) ||
arch/x86/kvm/svm/nested.c
396
if (CC(!__kvm_is_valid_cr4(vcpu, save->cr4)))
arch/x86/kvm/svm/nested.c
512
to->cr4 = from->cr4;
arch/x86/kvm/svm/nested.c
696
svm_set_cr4(vcpu, svm->nested.save.cr4);
arch/x86/kvm/svm/nested.c
96
kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4,
arch/x86/kvm/svm/svm.c
1776
static bool svm_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/svm/svm.c
1781
void svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/svm/svm.c
1784
unsigned long old_cr4 = vcpu->arch.cr4;
arch/x86/kvm/svm/svm.c
1786
vcpu->arch.cr4 = cr4;
arch/x86/kvm/svm/svm.c
1788
cr4 |= X86_CR4_PAE;
arch/x86/kvm/svm/svm.c
1791
cr4 &= ~(X86_CR4_SMEP | X86_CR4_SMAP | X86_CR4_PKE);
arch/x86/kvm/svm/svm.c
1793
cr4 |= host_cr4_mce;
arch/x86/kvm/svm/svm.c
1794
to_svm(vcpu)->vmcb->save.cr4 = cr4;
arch/x86/kvm/svm/svm.c
1797
if ((cr4 ^ old_cr4) & (X86_CR4_OSXSAVE | X86_CR4_PKE))
arch/x86/kvm/svm/svm.c
3426
"cr3:", save->cr3, "cr4:", save->cr4);
arch/x86/kvm/svm/svm.h
144
u64 cr4;
arch/x86/kvm/svm/svm.h
723
void svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/vmx/main.c
400
static bool vt_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/main.c
405
return vmx_is_valid_cr4(vcpu, cr4);
arch/x86/kvm/vmx/main.c
408
static void vt_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/main.c
413
vmx_set_cr4(vcpu, cr4);
arch/x86/kvm/vmx/vmcs.h
45
unsigned long cr4; /* May not match real cr4 */
arch/x86/kvm/vmx/vmx.c
2667
vcpu->arch.cr4 &= ~guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
2668
vcpu->arch.cr4 |= vmcs_readl(GUEST_CR4) & guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
3627
bool vmx_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/vmx.c
3634
if ((cr4 & X86_CR4_VMXE) && is_smm(vcpu))
arch/x86/kvm/vmx/vmx.c
3637
if (to_vmx(vcpu)->nested.vmxon && !nested_cr4_valid(vcpu, cr4))
arch/x86/kvm/vmx/vmx.c
3643
void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/vmx.c
3654
hw_cr4 = (cr4_read_shadow() & X86_CR4_MCE) | (cr4 & ~X86_CR4_MCE);
arch/x86/kvm/vmx/vmx.c
3663
if (cr4 & X86_CR4_UMIP) {
arch/x86/kvm/vmx/vmx.c
3672
vcpu->arch.cr4 = cr4;
arch/x86/kvm/vmx/vmx.c
3680
} else if (!(cr4 & X86_CR4_PAE)) {
arch/x86/kvm/vmx/vmx.c
3700
vmcs_writel(CR4_READ_SHADOW, cr4);
arch/x86/kvm/vmx/vmx.c
3703
if ((cr4 ^ old_cr4) & (X86_CR4_OSXSAVE | X86_CR4_PKE))
arch/x86/kvm/vmx/vmx.c
4517
unsigned long cr0, cr3, cr4;
arch/x86/kvm/vmx/vmx.c
4532
cr4 = cr4_read_shadow();
arch/x86/kvm/vmx/vmx.c
4533
vmcs_writel(HOST_CR4, cr4); /* 22.2.3, 22.2.5 */
arch/x86/kvm/vmx/vmx.c
4534
vmx->loaded_vmcs->host_state.cr4 = cr4;
arch/x86/kvm/vmx/vmx.c
6582
unsigned long cr4;
arch/x86/kvm/vmx/vmx.c
6594
cr4 = vmcs_readl(GUEST_CR4);
arch/x86/kvm/vmx/vmx.c
6613
cr4, vmcs_readl(CR4_READ_SHADOW), vmcs_readl(CR4_GUEST_HOST_MASK));
arch/x86/kvm/vmx/vmx.c
7609
unsigned long cr3, cr4;
arch/x86/kvm/vmx/vmx.c
7671
cr4 = cr4_read_shadow();
arch/x86/kvm/vmx/vmx.c
7672
if (unlikely(cr4 != vmx->loaded_vmcs->host_state.cr4)) {
arch/x86/kvm/vmx/vmx.c
7673
vmcs_writel(HOST_CR4, cr4);
arch/x86/kvm/vmx/vmx.c
7674
vmx->loaded_vmcs->host_state.cr4 = cr4;
arch/x86/kvm/vmx/vmx.h
357
void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/vmx/x86_ops.h
69
void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/vmx/x86_ops.h
70
bool vmx_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/x86.c
12161
sregs->cr4 = kvm_read_cr4(vcpu);
arch/x86/kvm/x86.c
12337
if (!(sregs->cr4 & X86_CR4_PAE) || !(sregs->efer & EFER_LMA))
arch/x86/kvm/x86.c
12350
return kvm_is_valid_cr4(vcpu, sregs->cr4) &&
arch/x86/kvm/x86.c
12390
*mmu_reset_needed |= kvm_read_cr4(vcpu) != sregs->cr4;
arch/x86/kvm/x86.c
12391
kvm_x86_call(set_cr4)(vcpu, sregs->cr4);
arch/x86/kvm/x86.c
12453
bool pae = (sregs2->cr0 & X86_CR0_PG) && (sregs2->cr4 & X86_CR4_PAE) &&
arch/x86/kvm/x86.c
1323
static bool kvm_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/x86.c
1325
return __kvm_is_valid_cr4(vcpu, cr4) &&
arch/x86/kvm/x86.c
1326
kvm_x86_call(is_valid_cr4)(vcpu, cr4);
arch/x86/kvm/x86.c
1329
void kvm_post_set_cr4(struct kvm_vcpu *vcpu, unsigned long old_cr4, unsigned long cr4)
arch/x86/kvm/x86.c
1331
if ((cr4 ^ old_cr4) & KVM_MMU_CR4_ROLE_BITS)
arch/x86/kvm/x86.c
1343
(cr4 & X86_CR4_PCIDE) && !(old_cr4 & X86_CR4_PCIDE))
arch/x86/kvm/x86.c
1354
if (((cr4 ^ old_cr4) & X86_CR4_PGE) ||
arch/x86/kvm/x86.c
1355
(!(cr4 & X86_CR4_PCIDE) && (old_cr4 & X86_CR4_PCIDE)))
arch/x86/kvm/x86.c
1364
else if (((cr4 ^ old_cr4) & X86_CR4_PAE) ||
arch/x86/kvm/x86.c
1365
((cr4 & X86_CR4_SMEP) && !(old_cr4 & X86_CR4_SMEP)))
arch/x86/kvm/x86.c
1371
int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/x86.c
1375
if (!kvm_is_valid_cr4(vcpu, cr4))
arch/x86/kvm/x86.c
1379
if (!(cr4 & X86_CR4_PAE))
arch/x86/kvm/x86.c
1381
if ((cr4 ^ old_cr4) & X86_CR4_LA57)
arch/x86/kvm/x86.c
1383
} else if (is_paging(vcpu) && (cr4 & X86_CR4_PAE)
arch/x86/kvm/x86.c
1384
&& ((cr4 ^ old_cr4) & X86_CR4_PDPTR_BITS)
arch/x86/kvm/x86.c
1388
if ((cr4 & X86_CR4_PCIDE) && !(old_cr4 & X86_CR4_PCIDE)) {
arch/x86/kvm/x86.c
1394
if ((cr4 & X86_CR4_CET) && !kvm_is_cr0_bit_set(vcpu, X86_CR0_WP))
arch/x86/kvm/x86.c
1397
kvm_x86_call(set_cr4)(vcpu, cr4);
arch/x86/kvm/x86.c
1399
kvm_post_set_cr4(vcpu, old_cr4, cr4);
arch/x86/kvm/x86.h
680
static inline bool __kvm_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/x86.h
682
return !(cr4 & vcpu->arch.cr4_guest_rsvd_bits);
arch/x86/mm/init.c
1016
.cr4 = ~0UL, /* fail hard if we screw up cr4 shadow initialization */
arch/x86/mm/tlb.c
1636
cpu_pcide = this_cpu_read(cpu_tlbstate.cr4) & X86_CR4_PCIDE;
arch/x86/mm/tlb.c
1679
__native_tlb_flush_global(this_cpu_read(cpu_tlbstate.cr4));
arch/x86/power/cpu.c
128
ctxt->cr4 = __read_cr4();
arch/x86/power/cpu.c
208
if (ctxt->cr4)
arch/x86/power/cpu.c
209
__write_cr4(ctxt->cr4);
arch/x86/power/cpu.c
213
__write_cr4(ctxt->cr4);
arch/x86/power/cpu.c
245
if (ctxt->cr4 & X86_CR4_FRED) {
arch/x86/realmode/init.c
155
trampoline_cr4_features = &trampoline_header->cr4;
arch/x86/xen/enlighten_pv.c
1081
static void xen_write_cr4(unsigned long cr4)
arch/x86/xen/enlighten_pv.c
1083
cr4 &= ~(X86_CR4_PGE | X86_CR4_PSE | X86_CR4_PCE);
arch/x86/xen/enlighten_pv.c
1085
native_write_cr4(cr4);
drivers/misc/lkdtm/bugs.c
682
unsigned long cr4;
drivers/misc/lkdtm/bugs.c
685
cr4 = native_read_cr4();
drivers/misc/lkdtm/bugs.c
687
if ((cr4 & X86_CR4_SMEP) != X86_CR4_SMEP) {
drivers/misc/lkdtm/bugs.c
691
cr4 &= ~(X86_CR4_SMEP);
drivers/misc/lkdtm/bugs.c
694
native_write_cr4(cr4);
drivers/misc/lkdtm/bugs.c
695
if (cr4 == native_read_cr4()) {
drivers/misc/lkdtm/bugs.c
697
cr4 |= X86_CR4_SMEP;
drivers/misc/lkdtm/bugs.c
699
native_write_cr4(cr4);
drivers/misc/lkdtm/bugs.c
729
direct_write_cr4(cr4);
drivers/misc/lkdtm/bugs.c
734
cr4 |= X86_CR4_SMEP;
drivers/misc/lkdtm/bugs.c
736
native_write_cr4(cr4);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
680
static u32 brcmf_chip_tcm_ramsize(struct brcmf_core_priv *cr4)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
691
corecap = brcmf_chip_core_read32(cr4, ARMCR4_CAP);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
698
brcmf_chip_core_write32(cr4, ARMCR4_BANKIDX, idx);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
699
bxinfo = brcmf_chip_core_read32(cr4, ARMCR4_BANKINFO);
drivers/parport/parport_pc.c
1011
: modes[cr4 & 0x03],
drivers/parport/parport_pc.c
1012
(cr4 & 0x40) ? "1.7" : "1.9");
drivers/parport/parport_pc.c
969
int cr1, cr4, cra, cr23, cr26, cr27;
drivers/parport/parport_pc.c
983
cr4 = inb(io + 1);
drivers/parport/parport_pc.c
996
cr1, cr4, cra, cr23, cr26, cr27);
drivers/tty/serial/fsl_lpuart.c
1993
u8 cr1, old_cr1, old_cr2, cr3, cr4, bdh, modem;
drivers/tty/serial/fsl_lpuart.c
2001
cr4 = readb(port->membase + UARTCR4);
drivers/tty/serial/fsl_lpuart.c
2118
cr4 &= ~UARTCR4_BRFA_MASK;
drivers/tty/serial/fsl_lpuart.c
2120
writeb(cr4 | brfa, port->membase + UARTCR4);
drivers/video/fbdev/sis/init301.c
6363
unsigned char cr4, cr14, cr5, cr15;
drivers/video/fbdev/sis/init301.c
6365
cr4 = SiS_Pr->CCRT1CRTC[4];
drivers/video/fbdev/sis/init301.c
6370
cr4 = SiS_Pr->SiS_CRT1Table[CRT1Index].CR[4];
drivers/video/fbdev/sis/init301.c
6375
tempbx = ((cr4 | ((cr14 & 0xC0) << 2)) - 3) << 3; /* (VGAHRS-3)*8 */
include/hyperv/hvgdk_mini.h
876
u64 cr4;
include/hyperv/hvhdk.h
111
u64 cr4;
include/uapi/linux/acrn.h
289
__le64 cr4;
include/xen/interface/hvm/hvm_vcpu.h
25
uint32_t cr4;
include/xen/interface/hvm/hvm_vcpu.h
88
uint64_t cr4;
tools/arch/x86/include/uapi/asm/kvm.h
155
__u64 cr0, cr2, cr3, cr4, cr8;
tools/arch/x86/include/uapi/asm/kvm.h
166
__u64 cr0, cr2, cr3, cr4, cr8;
tools/testing/selftests/kvm/include/x86/processor.h
581
uint64_t cr4;
tools/testing/selftests/kvm/include/x86/processor.h
584
: /* output */ [cr4]"=r"(cr4));
tools/testing/selftests/kvm/include/x86/processor.h
585
return cr4;
tools/testing/selftests/kvm/include/x86/svm.h
204
u64 cr4;
tools/testing/selftests/kvm/lib/x86/processor.c
141
sregs->cr0, sregs->cr2, sregs->cr3, sregs->cr4);
tools/testing/selftests/kvm/lib/x86/processor.c
660
sregs.cr4 |= X86_CR4_PAE | X86_CR4_OSFXSR;
tools/testing/selftests/kvm/lib/x86/processor.c
662
sregs.cr4 |= X86_CR4_OSXSAVE;
tools/testing/selftests/kvm/lib/x86/processor.c
664
sregs.cr4 |= X86_CR4_LA57;
tools/testing/selftests/kvm/lib/x86/svm.c
112
asm volatile ("mov %%cr4, %0" : "=r"(save->cr4) : : "memory");
tools/testing/selftests/kvm/lib/x86/vmx.c
131
unsigned long cr4;
tools/testing/selftests/kvm/lib/x86/vmx.c
143
__asm__ __volatile__("mov %%cr4, %0" : "=r"(cr4) : : "memory");
tools/testing/selftests/kvm/lib/x86/vmx.c
144
cr4 &= rdmsr(MSR_IA32_VMX_CR4_FIXED1);
tools/testing/selftests/kvm/lib/x86/vmx.c
145
cr4 |= rdmsr(MSR_IA32_VMX_CR4_FIXED0);
tools/testing/selftests/kvm/lib/x86/vmx.c
147
cr4 |= X86_CR4_VMXE;
tools/testing/selftests/kvm/lib/x86/vmx.c
148
__asm__ __volatile__("mov %0, %%cr4" : : "r"(cr4) : "memory");
tools/testing/selftests/kvm/x86/cpuid_test.c
172
sregs.cr4 ^= X86_CR4_OSXSAVE;
tools/testing/selftests/kvm/x86/cpuid_test.c
174
sregs.cr4 ^= X86_CR4_PKE;
tools/testing/selftests/kvm/x86/cr4_cpuid_sync_test.c
81
sregs.cr4 &= ~X86_CR4_OSXSAVE;
tools/testing/selftests/kvm/x86/set_sregs_test.c
100
sregs.cr4, cr4);
tools/testing/selftests/kvm/x86/set_sregs_test.c
102
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_UMIP);
tools/testing/selftests/kvm/x86/set_sregs_test.c
103
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_LA57);
tools/testing/selftests/kvm/x86/set_sregs_test.c
104
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_VMXE);
tools/testing/selftests/kvm/x86/set_sregs_test.c
105
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_SMXE);
tools/testing/selftests/kvm/x86/set_sregs_test.c
106
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_FSGSBASE);
tools/testing/selftests/kvm/x86/set_sregs_test.c
107
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_PCIDE);
tools/testing/selftests/kvm/x86/set_sregs_test.c
108
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_OSXSAVE);
tools/testing/selftests/kvm/x86/set_sregs_test.c
109
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_SMEP);
tools/testing/selftests/kvm/x86/set_sregs_test.c
110
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_SMAP);
tools/testing/selftests/kvm/x86/set_sregs_test.c
111
TEST_INVALID_CR_BIT(vcpu, cr4, sregs, X86_CR4_PKE);
tools/testing/selftests/kvm/x86/set_sregs_test.c
51
uint64_t cr4 = KVM_ALWAYS_ALLOWED_CR4;
tools/testing/selftests/kvm/x86/set_sregs_test.c
54
cr4 |= X86_CR4_UMIP;
tools/testing/selftests/kvm/x86/set_sregs_test.c
56
cr4 |= X86_CR4_LA57;
tools/testing/selftests/kvm/x86/set_sregs_test.c
58
cr4 |= X86_CR4_VMXE;
tools/testing/selftests/kvm/x86/set_sregs_test.c
60
cr4 |= X86_CR4_SMXE;
tools/testing/selftests/kvm/x86/set_sregs_test.c
62
cr4 |= X86_CR4_FSGSBASE;
tools/testing/selftests/kvm/x86/set_sregs_test.c
64
cr4 |= X86_CR4_PCIDE;
tools/testing/selftests/kvm/x86/set_sregs_test.c
66
cr4 |= X86_CR4_OSXSAVE;
tools/testing/selftests/kvm/x86/set_sregs_test.c
68
cr4 |= X86_CR4_SMEP;
tools/testing/selftests/kvm/x86/set_sregs_test.c
70
cr4 |= X86_CR4_SMAP;
tools/testing/selftests/kvm/x86/set_sregs_test.c
72
cr4 |= X86_CR4_PKE;
tools/testing/selftests/kvm/x86/set_sregs_test.c
74
return cr4;
tools/testing/selftests/kvm/x86/set_sregs_test.c
77
static void test_cr_bits(struct kvm_vcpu *vcpu, uint64_t cr4)
tools/testing/selftests/kvm/x86/set_sregs_test.c
84
sregs.cr4 |= cr4;
tools/testing/selftests/kvm/x86/set_sregs_test.c
86
TEST_ASSERT(!rc, "Failed to set supported CR4 bits (0x%lx)", cr4);
tools/testing/selftests/kvm/x86/set_sregs_test.c
88
TEST_ASSERT(!!(sregs.cr4 & X86_CR4_OSXSAVE) ==
tools/testing/selftests/kvm/x86/set_sregs_test.c
91
(sregs.cr4 & X86_CR4_OSXSAVE) ? "set" : "clear");
tools/testing/selftests/kvm/x86/set_sregs_test.c
93
TEST_ASSERT(!!(sregs.cr4 & X86_CR4_PKE) ==
tools/testing/selftests/kvm/x86/set_sregs_test.c
96
(sregs.cr4 & X86_CR4_PKE) ? "set" : "clear");
tools/testing/selftests/kvm/x86/set_sregs_test.c
99
TEST_ASSERT(sregs.cr4 == cr4, "sregs.CR4 (0x%llx) != CR4 (0x%lx)",
tools/testing/selftests/kvm/x86/sev_smoke_test.c
37
guest_sev_test_reg(cr4);
tools/testing/selftests/kvm/x86/sync_regs_test.c
138
__u64 *cr4 = &run->s.regs.sregs.cr4;
tools/testing/selftests/kvm/x86/sync_regs_test.c
139
__u64 pae_enabled = *cr4;
tools/testing/selftests/kvm/x86/sync_regs_test.c
140
__u64 pae_disabled = *cr4 & ~X86_CR4_PAE;
tools/testing/selftests/kvm/x86/sync_regs_test.c
144
WRITE_ONCE(*cr4, pae_enabled);
tools/testing/selftests/kvm/x86/sync_regs_test.c
148
WRITE_ONCE(*cr4, pae_disabled);
tools/testing/selftests/kvm/x86/sync_regs_test.c
178
TEST_ASSERT((run->s.regs.sregs.cr4 & X86_CR4_PAE) &&
tools/testing/selftests/kvm/x86/sync_regs_test.c
181
!!(run->s.regs.sregs.cr4 & X86_CR4_PAE),