Symbol: arch
arch/alpha/kernel/audit.c
31
int audit_classify_arch(int arch)
arch/alpha/kernel/module.c
149
got = sechdrs[me->arch.gotsecindex].sh_addr;
arch/alpha/kernel/module.c
83
me->arch.gotsecindex = s - sechdrs;
arch/arc/kernel/module.c
114
if (strcmp(module->arch.secstr+sechdrs[tgtsec].sh_name, ".eh_frame") == 0)
arch/arc/kernel/module.c
115
module->arch.unw_sec_idx = tgtsec;
arch/arc/kernel/module.c
137
int unwsec = mod->arch.unw_sec_idx;
arch/arc/kernel/module.c
142
mod->arch.unw_info = unw;
arch/arc/kernel/module.c
30
mod->arch.unw_sec_idx = 0;
arch/arc/kernel/module.c
31
mod->arch.unw_info = NULL;
arch/arc/kernel/module.c
33
mod->arch.secstr = secstr;
arch/arc/kernel/module.c
40
if (mod->arch.unw_info)
arch/arc/kernel/module.c
41
unwind_remove_table(mod->arch.unw_info, 0);
arch/arc/kernel/module.c
67
module->arch.secstr + sechdrs[tgtsec].sh_name, tgt_addr);
arch/arc/kernel/module.c
86
s = module->arch.secstr + sechdrs[sym_entry->st_shndx].sh_name;
arch/arm/kernel/cacheinfo.c
155
unsigned int arch = cpu_architecture();
arch/arm/kernel/cacheinfo.c
158
if (arch < CPU_ARCH_ARMv7)
arch/arm/kernel/ftrace.c
153
mod = rec->arch.mod;
arch/arm/kernel/ftrace.c
198
mod = rec->arch.mod;
arch/arm/kernel/ftrace.c
200
rec->arch.mod = mod;
arch/arm/kernel/hw_breakpoint.c
149
u8 arch = get_debug_arch();
arch/arm/kernel/hw_breakpoint.c
152
return (arch >= ARM_DEBUG_ARCH_V6 && arch <= ARM_DEBUG_ARCH_V7_ECP14) ||
arch/arm/kernel/hw_breakpoint.c
153
arch >= ARM_DEBUG_ARCH_V7_1;
arch/arm/kernel/machine_kexec.c
169
data->kexec_r2 = image->arch.kernel_r2;
arch/arm/kernel/machine_kexec.c
39
image->arch.kernel_r2 = image->start - KEXEC_ARM_ZIMAGE_OFFSET
arch/arm/kernel/machine_kexec.c
67
image->arch.kernel_r2 = current_segment->mem;
arch/arm/kernel/module-plts.c
223
mod->arch.core.plt = s;
arch/arm/kernel/module-plts.c
225
mod->arch.init.plt = s;
arch/arm/kernel/module-plts.c
230
if (!mod->arch.core.plt || !mod->arch.init.plt) {
arch/arm/kernel/module-plts.c
262
mod->arch.core.plt->sh_type = SHT_NOBITS;
arch/arm/kernel/module-plts.c
263
mod->arch.core.plt->sh_flags = SHF_EXECINSTR | SHF_ALLOC;
arch/arm/kernel/module-plts.c
264
mod->arch.core.plt->sh_addralign = L1_CACHE_BYTES;
arch/arm/kernel/module-plts.c
265
mod->arch.core.plt->sh_size = round_up(core_plts * PLT_ENT_SIZE,
arch/arm/kernel/module-plts.c
267
mod->arch.core.plt_count = 0;
arch/arm/kernel/module-plts.c
268
mod->arch.core.plt_ent = NULL;
arch/arm/kernel/module-plts.c
270
mod->arch.init.plt->sh_type = SHT_NOBITS;
arch/arm/kernel/module-plts.c
271
mod->arch.init.plt->sh_flags = SHF_EXECINSTR | SHF_ALLOC;
arch/arm/kernel/module-plts.c
272
mod->arch.init.plt->sh_addralign = L1_CACHE_BYTES;
arch/arm/kernel/module-plts.c
273
mod->arch.init.plt->sh_size = round_up(init_plts * PLT_ENT_SIZE,
arch/arm/kernel/module-plts.c
275
mod->arch.init.plt_count = 0;
arch/arm/kernel/module-plts.c
276
mod->arch.init.plt_ent = NULL;
arch/arm/kernel/module-plts.c
279
mod->arch.core.plt->sh_size, mod->arch.init.plt->sh_size);
arch/arm/kernel/module-plts.c
290
ret = mod && (loc - (u32)mod->arch.core.plt_ent < mod->arch.core.plt_count * PLT_ENT_SIZE ||
arch/arm/kernel/module-plts.c
291
loc - (u32)mod->arch.init.plt_ent < mod->arch.init.plt_count * PLT_ENT_SIZE);
arch/arm/kernel/module-plts.c
49
&mod->arch.core : &mod->arch.init;
arch/arm/kernel/module.c
421
struct list_head *unwind_list = &mod->arch.unwind_list;
arch/arm/kernel/module.c
424
mod->arch.init_table = NULL;
arch/arm/kernel/module.c
452
mod->arch.init_table = table;
arch/arm/kernel/module.c
479
&mod->arch.unwind_list, mod_list) {
arch/arm/kernel/module.c
483
mod->arch.init_table = NULL;
arch/arm/kernel/module.c
490
struct unwind_table *init = mod->arch.init_table;
arch/arm/kernel/module.c
493
mod->arch.init_table = NULL;
arch/arm/kernel/setup.c
277
static int cpu_has_aliasing_icache(unsigned int arch)
arch/arm/kernel/setup.c
287
switch (arch) {
arch/arm/kernel/setup.c
309
unsigned int arch = cpu_architecture();
arch/arm/kernel/setup.c
311
if (arch >= CPU_ARCH_ARMv6) {
arch/arm/kernel/setup.c
314
if ((arch == CPU_ARCH_ARMv7M) && !(cachetype & 0xf000f)) {
arch/arm/kernel/setup.c
318
arch = CPU_ARCH_ARMv7;
arch/arm/kernel/setup.c
329
arch = CPU_ARCH_ARMv6;
arch/arm/kernel/setup.c
335
if (cpu_has_aliasing_icache(arch))
arch/arm/probes/decode.c
54
int arch = cpu_architecture();
arch/arm/probes/decode.c
55
BUG_ON(arch == CPU_ARCH_UNKNOWN);
arch/arm/probes/decode.c
56
load_write_pc_interworks = arch >= CPU_ARCH_ARMv5T;
arch/arm/probes/decode.c
68
int arch = cpu_architecture();
arch/arm/probes/decode.c
69
BUG_ON(arch == CPU_ARCH_UNKNOWN);
arch/arm/probes/decode.c
70
alu_write_pc_interworks = arch >= CPU_ARCH_ARMv7;
arch/arm64/include/asm/elf.h
262
struct arch_elf_state *arch)
arch/arm64/include/asm/elf.h
276
arch->flags |= ARM64_ELF_BTI;
arch/arm64/include/asm/kvm_define_hypevents.h
3
#define REMOTE_EVENT_INCLUDE_FILE arch/arm64/include/asm/kvm_hypevents.h
arch/arm64/include/asm/kvm_emulate.h
107
vcpu->arch.hcr_el2 = HCR_GUEST_FLAGS;
arch/arm64/include/asm/kvm_emulate.h
116
vcpu->arch.hcr_el2 |= HCR_TVM;
arch/arm64/include/asm/kvm_emulate.h
121
return (unsigned long *)&vcpu->arch.hcr_el2;
arch/arm64/include/asm/kvm_emulate.h
126
return vcpu->arch.vsesr_el2;
arch/arm64/include/asm/kvm_emulate.h
131
vcpu->arch.vsesr_el2 = vsesr;
arch/arm64/include/asm/kvm_emulate.h
193
return vcpu_is_el2_ctxt(&vcpu->arch.ctxt);
arch/arm64/include/asm/kvm_emulate.h
204
return ctxt_sys_reg(&vcpu->arch.ctxt, HCR_EL2) & HCR_TGE;
arch/arm64/include/asm/kvm_emulate.h
223
return ctxt_sys_reg(&vcpu->arch.ctxt, HCR_EL2) & HCR_AMO;
arch/arm64/include/asm/kvm_emulate.h
317
return vcpu->arch.fault.esr_el2;
arch/arm64/include/asm/kvm_emulate.h
345
return vcpu->arch.fault.far_el2;
arch/arm64/include/asm/kvm_emulate.h
350
u64 hpfar = vcpu->arch.fault.hpfar_el2;
arch/arm64/include/asm/kvm_emulate.h
360
return vcpu->arch.fault.disr_el1;
arch/arm64/include/asm/kvm_emulate.h
670
vcpu->arch.hcrx_el2 = HCRX_EL2_SMPME;
arch/arm64/include/asm/kvm_emulate.h
673
vcpu->arch.hcrx_el2 |= (HCRX_EL2_MSCEn | HCRX_EL2_MCE2);
arch/arm64/include/asm/kvm_emulate.h
676
vcpu->arch.hcrx_el2 |= HCRX_EL2_TCR2En;
arch/arm64/include/asm/kvm_emulate.h
679
vcpu->arch.hcrx_el2 |= HCRX_EL2_EnFPM;
arch/arm64/include/asm/kvm_emulate.h
682
vcpu->arch.hcrx_el2 |= HCRX_EL2_SCTLR2En;
arch/arm64/include/asm/kvm_emulate.h
685
vcpu->arch.hcrx_el2 |= HCRX_EL2_EnALS;
arch/arm64/include/asm/kvm_emulate.h
688
vcpu->arch.hcrx_el2 |= HCRX_EL2_EnASR;
arch/arm64/include/asm/kvm_emulate.h
95
return !(vcpu->arch.hcr_el2 & HCR_RW);
arch/arm64/include/asm/kvm_host.h
1000
READ_ONCE(v->arch.flagset) & (m); \
arch/arm64/include/asm/kvm_host.h
1018
typeof(v->arch.flagset) *fset; \
arch/arm64/include/asm/kvm_host.h
1022
fset = &v->arch.flagset; \
arch/arm64/include/asm/kvm_host.h
1032
typeof(v->arch.flagset) *fset; \
arch/arm64/include/asm/kvm_host.h
1036
fset = &v->arch.flagset; \
arch/arm64/include/asm/kvm_host.h
1044
typeof(v->arch.flagset) set; \
arch/arm64/include/asm/kvm_host.h
1121
#define vcpu_sve_pffr(vcpu) (kern_hyp_va((vcpu)->arch.sve_state) + \
arch/arm64/include/asm/kvm_host.h
1122
sve_ffr_offset((vcpu)->arch.sve_max_vl))
arch/arm64/include/asm/kvm_host.h
1124
#define vcpu_sve_max_vq(vcpu) sve_vq_from_vl((vcpu)->arch.sve_max_vl)
arch/arm64/include/asm/kvm_host.h
1143
#define vcpu_sve_state_size(vcpu) sve_state_size_from_vl((vcpu)->arch.sve_max_vl)
arch/arm64/include/asm/kvm_host.h
1151
test_bit(KVM_ARCH_FLAG_GUEST_HAS_SVE, &(kvm)->arch.flags))
arch/arm64/include/asm/kvm_host.h
1178
#define vcpu_gp_regs(v) (&(v)->arch.ctxt.regs)
arch/arm64/include/asm/kvm_host.h
1213
const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \
arch/arm64/include/asm/kvm_host.h
1223
const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \
arch/arm64/include/asm/kvm_host.h
1234
const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \
arch/arm64/include/asm/kvm_host.h
1460
((vcpu)->arch.debug_owner != VCPU_DEBUG_FREE)
arch/arm64/include/asm/kvm_host.h
1462
((vcpu)->arch.debug_owner == VCPU_DEBUG_HOST_OWNED)
arch/arm64/include/asm/kvm_host.h
1464
((vcpu)->arch.debug_owner == VCPU_DEBUG_GUEST_OWNED)
arch/arm64/include/asm/kvm_host.h
1523
#define kvm_vm_is_protected(kvm) (is_protected_kvm_enabled() && (kvm)->arch.pkvm.is_protected)
arch/arm64/include/asm/kvm_host.h
1534
test_bit(KVM_ARCH_FLAG_MTE_ENABLED, &(kvm)->arch.flags))
arch/arm64/include/asm/kvm_host.h
1541
(test_bit(KVM_ARCH_FLAG_HAS_RAN_ONCE, &(kvm)->arch.flags))
arch/arm64/include/asm/kvm_host.h
1548
#define kvm_vcpu_has_feature(k, f) __vcpu_has_feature(&(k)->arch, (f))
arch/arm64/include/asm/kvm_host.h
1549
#define vcpu_has_feature(v, f) __vcpu_has_feature(&(v)->kvm->arch, (f))
arch/arm64/include/asm/kvm_host.h
1585
({ u64 __val = *__vm_id_reg(&(kvm)->arch, reg); __val; })
arch/arm64/include/asm/kvm_host.h
1724
p = &(vcpu)->arch.fgt[id].w; \
arch/arm64/include/asm/kvm_host.h
1727
p = &(vcpu)->arch.fgt[id].r; \
arch/arm64/include/asm/kvm_host.h
197
struct kvm_arch *arch;
arch/arm64/include/asm/kvm_host.h
658
static inline struct resx __kvm_get_sysreg_resx(struct kvm_arch *arch,
arch/arm64/include/asm/kvm_host.h
663
masks = arch->sysreg_masks;
arch/arm64/include/asm/kvm_host.h
671
#define kvm_get_sysreg_resx(k, sr) __kvm_get_sysreg_resx(&(k)->arch, (sr))
arch/arm64/include/asm/kvm_host.h
673
static inline void __kvm_set_sysreg_resx(struct kvm_arch *arch,
arch/arm64/include/asm/kvm_host.h
676
arch->sysreg_masks->mask[sr - __SANITISED_REG_START__] = resx;
arch/arm64/include/asm/kvm_host.h
680
__kvm_set_sysreg_resx(&(k)->arch, (sr), (resx))
arch/arm64/include/asm/kvm_host.h
988
typeof(v->arch.flagset) *_fset; \
arch/arm64/include/asm/kvm_mmu.h
322
struct kvm_arch *arch)
arch/arm64/include/asm/kvm_mmu.h
337
return container_of(mmu->arch, struct kvm, arch);
arch/arm64/include/asm/kvm_mmu.h
357
return &kvm->arch.mmu != mmu;
arch/arm64/kernel/asm-offsets.c
105
DEFINE(VCPU_CONTEXT, offsetof(struct kvm_vcpu, arch.ctxt));
arch/arm64/kernel/asm-offsets.c
106
DEFINE(VCPU_FAULT_DISR, offsetof(struct kvm_vcpu, arch.fault.disr_el1));
arch/arm64/kernel/asm-offsets.c
107
DEFINE(VCPU_HCR_EL2, offsetof(struct kvm_vcpu, arch.hcr_el2));
arch/arm64/kernel/asm-offsets.c
170
DEFINE(KIMAGE_ARCH_DTB_MEM, offsetof(struct kimage, arch.dtb_mem));
arch/arm64/kernel/asm-offsets.c
171
DEFINE(KIMAGE_ARCH_EL2_VECTORS, offsetof(struct kimage, arch.el2_vectors));
arch/arm64/kernel/asm-offsets.c
172
DEFINE(KIMAGE_ARCH_ZERO_PAGE, offsetof(struct kimage, arch.zero_page));
arch/arm64/kernel/asm-offsets.c
173
DEFINE(KIMAGE_ARCH_PHYS_OFFSET, offsetof(struct kimage, arch.phys_offset));
arch/arm64/kernel/asm-offsets.c
174
DEFINE(KIMAGE_ARCH_TTBR1, offsetof(struct kimage, arch.ttbr1));
arch/arm64/kernel/ftrace.c
267
plt = mod->arch.init_ftrace_trampolines;
arch/arm64/kernel/ftrace.c
269
plt = mod->arch.ftrace_trampolines;
arch/arm64/kernel/machine_kexec.c
123
kimage->arch.el2_vectors = 0;
arch/arm64/kernel/machine_kexec.c
126
&kimage->arch.el2_vectors);
arch/arm64/kernel/machine_kexec.c
135
kimage->arch.ttbr1 = __pa(trans_pgd);
arch/arm64/kernel/machine_kexec.c
136
kimage->arch.zero_page = __pa_symbol(empty_zero_page);
arch/arm64/kernel/machine_kexec.c
140
kimage->arch.kern_reloc = __pa(reloc_code);
arch/arm64/kernel/machine_kexec.c
141
rc = trans_pgd_idmap_page(&info, &kimage->arch.ttbr0,
arch/arm64/kernel/machine_kexec.c
142
&kimage->arch.t0sz, reloc_code);
arch/arm64/kernel/machine_kexec.c
145
kimage->arch.phys_offset = virt_to_phys(kimage) - (long)kimage;
arch/arm64/kernel/machine_kexec.c
192
restart(is_hyp_nvhe(), kimage->start, kimage->arch.dtb_mem,
arch/arm64/kernel/machine_kexec.c
198
__hyp_set_vectors(kimage->arch.el2_vectors);
arch/arm64/kernel/machine_kexec.c
199
cpu_install_ttbr0(kimage->arch.ttbr0, kimage->arch.t0sz);
arch/arm64/kernel/machine_kexec.c
200
kernel_reloc = (void *)kimage->arch.kern_reloc;
arch/arm64/kernel/machine_kexec.c
39
kexec_dprintk(" kern_reloc: %pa\n", &kimage->arch.kern_reloc);
arch/arm64/kernel/machine_kexec.c
40
kexec_dprintk(" el2_vectors: %pa\n", &kimage->arch.el2_vectors);
arch/arm64/kernel/machine_kexec_file.c
189
image->arch.dtb = dtb;
arch/arm64/kernel/machine_kexec_file.c
190
image->arch.dtb_mem = kbuf.mem;
arch/arm64/kernel/machine_kexec_file.c
32
kvfree(image->arch.dtb);
arch/arm64/kernel/machine_kexec_file.c
33
image->arch.dtb = NULL;
arch/arm64/kernel/module-plts.c
105
&mod->arch.core : &mod->arch.init;
arch/arm64/kernel/module-plts.c
295
mod->arch.core.plt_shndx = i;
arch/arm64/kernel/module-plts.c
297
mod->arch.init.plt_shndx = i;
arch/arm64/kernel/module-plts.c
308
if (!mod->arch.core.plt_shndx || !mod->arch.init.plt_shndx) {
arch/arm64/kernel/module-plts.c
346
pltsec = sechdrs + mod->arch.core.plt_shndx;
arch/arm64/kernel/module-plts.c
351
mod->arch.core.plt_num_entries = 0;
arch/arm64/kernel/module-plts.c
352
mod->arch.core.plt_max_entries = core_plts;
arch/arm64/kernel/module-plts.c
354
pltsec = sechdrs + mod->arch.init.plt_shndx;
arch/arm64/kernel/module-plts.c
359
mod->arch.init.plt_num_entries = 0;
arch/arm64/kernel/module-plts.c
360
mod->arch.init.plt_max_entries = init_plts;
arch/arm64/kernel/module-plts.c
74
&mod->arch.core : &mod->arch.init;
arch/arm64/kernel/module.c
468
mod->arch.ftrace_trampolines = plts;
arch/arm64/kernel/module.c
478
mod->arch.init_ftrace_trampolines = plts;
arch/arm64/kvm/arch_timer.c
1053
offs->vcpu_offset = __ctxt_sys_reg(&vcpu->arch.ctxt, CNTVOFF_EL2);
arch/arm64/kvm/arch_timer.c
1054
offs->vm_offset = &vcpu->kvm->arch.timer_data.poffset;
arch/arm64/kvm/arch_timer.c
1084
ctxt->offset.vm_offset = &kvm->arch.timer_data.voffset;
arch/arm64/kvm/arch_timer.c
1086
ctxt->offset.vm_offset = &kvm->arch.timer_data.poffset;
arch/arm64/kvm/arch_timer.c
1114
!test_bit(KVM_ARCH_FLAG_VM_COUNTER_OFFSET, &vcpu->kvm->arch.flags)) {
arch/arm64/kvm/arch_timer.c
1135
kvm->arch.timer_data.ppi[i] = get_vgic_ppi(kvm, default_ppi[i]);
arch/arm64/kvm/arch_timer.c
1535
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/arch_timer.c
1563
set_bit(KVM_ARCH_FLAG_TIMER_PPIS_IMMUTABLE, &vcpu->kvm->arch.flags);
arch/arm64/kvm/arch_timer.c
1565
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/arch_timer.c
1661
guard(mutex)(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/arch_timer.c
1664
&vcpu->kvm->arch.flags)) {
arch/arm64/kvm/arch_timer.c
1690
vcpu->kvm->arch.timer_data.ppi[idx] = irq;
arch/arm64/kvm/arch_timer.c
1749
set_bit(KVM_ARCH_FLAG_VM_COUNTER_OFFSET, &kvm->arch.flags);
arch/arm64/kvm/arch_timer.c
1757
kvm->arch.timer_data.voffset = offset->counter_offset;
arch/arm64/kvm/arch_timer.c
1758
kvm->arch.timer_data.poffset = offset->counter_offset;
arch/arm64/kvm/arch_timer.c
294
struct arch_timer_context *ctx = &vcpu->arch.timer_cpu.timers[i];
arch/arm64/kvm/arch_timer.c
318
vcpu = container_of(timer, struct kvm_vcpu, arch.timer_cpu);
arch/arm64/kvm/arm.c
1008
(!kvm_arm_vcpu_stopped(vcpu)) && (!vcpu->arch.pause),
arch/arm64/kvm/arm.c
1011
if (kvm_arm_vcpu_stopped(vcpu) || vcpu->arch.pause) {
arch/arm64/kvm/arm.c
148
&kvm->arch.flags);
arch/arm64/kvm/arm.c
1507
mask = kvm->arch.vgic.gicv5_vm.userspace_ppis;
arch/arm64/kvm/arm.c
154
set_bit(KVM_ARCH_FLAG_MTE_ENABLED, &kvm->arch.flags);
arch/arm64/kvm/arm.c
160
set_bit(KVM_ARCH_FLAG_SYSTEM_SUSPEND_ENABLED, &kvm->arch.flags);
arch/arm64/kvm/arm.c
1605
return !bitmap_equal(vcpu->kvm->arch.vcpu_features, &features,
arch/arm64/kvm/arm.c
1618
if (kvm_vcpu_has_pmu(vcpu) && !kvm->arch.arm_pmu)
arch/arm64/kvm/arm.c
1635
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
1637
if (test_bit(KVM_ARCH_FLAG_VCPU_FEATURES_CONFIGURED, &kvm->arch.flags) &&
arch/arm64/kvm/arm.c
1641
bitmap_copy(kvm->arch.vcpu_features, &features, KVM_VCPU_MAX_FEATURES);
arch/arm64/kvm/arm.c
1650
set_bit(KVM_ARCH_FLAG_VCPU_FEATURES_CONFIGURED, &kvm->arch.flags);
arch/arm64/kvm/arm.c
1654
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
1722
spin_lock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/arm.c
1727
WRITE_ONCE(vcpu->arch.mp_state.mp_state, KVM_MP_STATE_RUNNABLE);
arch/arm64/kvm/arm.c
1729
spin_unlock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/arm.c
173
kvm->arch.mmu.split_page_chunk_size = new_cap;
arch/arm64/kvm/arm.c
182
set_bit(KVM_ARCH_FLAG_WRITABLE_IMP_ID_REGS, &kvm->arch.flags);
arch/arm64/kvm/arm.c
188
set_bit(KVM_ARCH_FLAG_EXIT_SEA, &kvm->arch.flags);
arch/arm64/kvm/arm.c
214
mutex_init(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
219
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
220
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
230
if (!zalloc_cpumask_var(&kvm->arch.supported_cpus, GFP_KERNEL_ACCOUNT)) {
arch/arm64/kvm/arm.c
234
cpumask_copy(kvm->arch.supported_cpus, cpu_possible_mask);
arch/arm64/kvm/arm.c
236
ret = kvm_init_stage2_mmu(kvm, &kvm->arch.mmu, type);
arch/arm64/kvm/arm.c
262
bitmap_zero(kvm->arch.vcpu_features, KVM_VCPU_MAX_FEATURES);
arch/arm64/kvm/arm.c
269
free_cpumask_var(kvm->arch.supported_cpus);
arch/arm64/kvm/arm.c
2861
data = rcu_dereference(kvm->arch.mpidr_data);
arch/arm64/kvm/arm.c
290
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
292
data = rcu_dereference_protected(kvm->arch.mpidr_data,
arch/arm64/kvm/arm.c
293
lockdep_is_held(&kvm->arch.config_lock));
arch/arm64/kvm/arm.c
295
rcu_assign_pointer(kvm->arch.mpidr_data, NULL);
arch/arm64/kvm/arm.c
300
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
309
bitmap_free(kvm->arch.pmu_filter);
arch/arm64/kvm/arm.c
310
free_cpumask_var(kvm->arch.supported_cpus);
arch/arm64/kvm/arm.c
320
kfree(kvm->arch.sysreg_masks);
arch/arm64/kvm/arm.c
422
r = kvm->arch.vgic.msis_require_devid;
arch/arm64/kvm/arm.c
470
r = kvm->arch.mmu.split_page_chunk_size;
arch/arm64/kvm/arm.c
525
spin_lock_init(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/arm.c
530
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/arm.c
531
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/arm.c
538
vcpu->arch.mmu_page_cache.gfp_zero = __GFP_ZERO;
arch/arm64/kvm/arm.c
545
kvm_arm_pvtime_vcpu_init(&vcpu->arch);
arch/arm64/kvm/arm.c
547
vcpu->arch.hw_mmu = &vcpu->kvm->arch.mmu;
arch/arm64/kvm/arm.c
574
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
arch/arm64/kvm/arm.c
576
free_hyp_memcache(&vcpu->arch.pkvm_memcache);
arch/arm64/kvm/arm.c
605
vcpu->arch.hcr_el2 &= ~(HCR_API | HCR_APK);
arch/arm64/kvm/arm.c
606
vcpu->arch.hcr_el2 |= val;
arch/arm64/kvm/arm.c
608
vcpu->arch.hcr_el2 |= (HCR_API | HCR_APK);
arch/arm64/kvm/arm.c
616
if (vcpu->arch.hcr_el2 & (HCR_API | HCR_APK)) {
arch/arm64/kvm/arm.c
634
vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3 &&
arch/arm64/kvm/arm.c
635
(atomic_read(&vcpu->arch.vgic_cpu.vgic_v3.its_vpe.vlpi_count) ||
arch/arm64/kvm/arm.c
636
vcpu->kvm->arch.vgic.nassgireq);
arch/arm64/kvm/arm.c
658
mmu = vcpu->arch.hw_mmu;
arch/arm64/kvm/arm.c
700
if (kvm_arm_is_pvtime_enabled(&vcpu->arch))
arch/arm64/kvm/arm.c
704
vcpu->arch.hcr_el2 &= ~HCR_TWE;
arch/arm64/kvm/arm.c
706
vcpu->arch.hcr_el2 |= HCR_TWE;
arch/arm64/kvm/arm.c
709
vcpu->arch.hcr_el2 &= ~HCR_TWI;
arch/arm64/kvm/arm.c
711
vcpu->arch.hcr_el2 |= HCR_TWI;
arch/arm64/kvm/arm.c
717
vcpu->kvm->arch.pkvm.handle,
arch/arm64/kvm/arm.c
718
vcpu->vcpu_idx, vcpu->arch.hcr_el2);
arch/arm64/kvm/arm.c
720
&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/arm.c
723
if (!cpumask_test_cpu(cpu, vcpu->kvm->arch.supported_cpus))
arch/arm64/kvm/arm.c
726
vcpu->arch.pid = pid_nr(vcpu->pid);
arch/arm64/kvm/arm.c
732
kvm_call_hyp(__vgic_v3_save_aprs, &vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/arm.c
753
WRITE_ONCE(vcpu->arch.mp_state.mp_state, KVM_MP_STATE_STOPPED);
arch/arm64/kvm/arm.c
760
spin_lock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/arm.c
762
spin_unlock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/arm.c
767
return READ_ONCE(vcpu->arch.mp_state.mp_state) == KVM_MP_STATE_STOPPED;
arch/arm64/kvm/arm.c
772
WRITE_ONCE(vcpu->arch.mp_state.mp_state, KVM_MP_STATE_SUSPENDED);
arch/arm64/kvm/arm.c
779
return READ_ONCE(vcpu->arch.mp_state.mp_state) == KVM_MP_STATE_SUSPENDED;
arch/arm64/kvm/arm.c
785
*mp_state = READ_ONCE(vcpu->arch.mp_state);
arch/arm64/kvm/arm.c
795
spin_lock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/arm.c
799
WRITE_ONCE(vcpu->arch.mp_state, *mp_state);
arch/arm64/kvm/arm.c
811
spin_unlock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/arm.c
828
&& !kvm_arm_vcpu_stopped(v) && !v->arch.pause);
arch/arm64/kvm/arm.c
850
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
852
if (rcu_access_pointer(kvm->arch.mpidr_data) ||
arch/arm64/kvm/arm.c
890
rcu_assign_pointer(kvm->arch.mpidr_data, data);
arch/arm64/kvm/arm.c
892
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
970
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
971
set_bit(KVM_ARCH_FLAG_HAS_RAN_ONCE, &kvm->arch.flags);
arch/arm64/kvm/arm.c
972
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/arm.c
988
vcpu->arch.pause = true;
arch/arm64/kvm/arm.c
998
vcpu->arch.pause = false;
arch/arm64/kvm/at.c
1383
__load_stage2(mmu, mmu->arch);
arch/arm64/kvm/at.c
1387
write_sysreg_hcr(vcpu->arch.hcr_el2);
arch/arm64/kvm/config.c
1391
u64 regval = kvm->arch.id_regs[map->regidx];
arch/arm64/kvm/config.c
1530
kvm->arch.fgu[fgt] = val;
arch/arm64/kvm/config.c
1653
u64 fgu = vcpu->kvm->arch.fgu[__fgt_reg_to_group_id(reg)];
arch/arm64/kvm/debug.c
149
vcpu->arch.external_mdscr_el1 = mdscr;
arch/arm64/kvm/debug.c
176
vcpu->arch.debug_owner = VCPU_DEBUG_HOST_OWNED;
arch/arm64/kvm/debug.c
198
vcpu->arch.debug_owner = VCPU_DEBUG_GUEST_OWNED;
arch/arm64/kvm/debug.c
200
vcpu->arch.debug_owner = VCPU_DEBUG_FREE;
arch/arm64/kvm/debug.c
240
vcpu->arch.debug_owner = VCPU_DEBUG_GUEST_OWNED;
arch/arm64/kvm/debug.c
46
vcpu->arch.mdcr_el2 = FIELD_PREP(MDCR_EL2_HPMN,
arch/arm64/kvm/debug.c
48
vcpu->arch.mdcr_el2 |= (MDCR_EL2_TPM |
arch/arm64/kvm/debug.c
58
vcpu->arch.mdcr_el2 |= MDCR_EL2_TDE;
arch/arm64/kvm/debug.c
64
vcpu->arch.mdcr_el2 |= MDCR_EL2_TDA;
arch/arm64/kvm/debug.c
71
write_sysreg(vcpu->arch.mdcr_el2, mdcr_el2);
arch/arm64/kvm/emulate-nested.c
2571
(vcpu->kvm->arch.fgu[tc.fgt] & BIT(tc.bit))) {
arch/arm64/kvm/fpsimd.c
76
fp_state.st = &vcpu->arch.ctxt.fp_regs;
arch/arm64/kvm/fpsimd.c
77
fp_state.sve_state = vcpu->arch.sve_state;
arch/arm64/kvm/fpsimd.c
78
fp_state.sve_vl = vcpu->arch.sve_max_vl;
arch/arm64/kvm/fpsimd.c
80
fp_state.svcr = __ctxt_sys_reg(&vcpu->arch.ctxt, SVCR);
arch/arm64/kvm/fpsimd.c
81
fp_state.fpmr = __ctxt_sys_reg(&vcpu->arch.ctxt, FPMR);
arch/arm64/kvm/fpsimd.c
82
fp_state.fp_type = &vcpu->arch.fp_type;
arch/arm64/kvm/guest.c
137
return &vcpu->arch.ctxt.regs.regs[off];
arch/arm64/kvm/guest.c
140
return &vcpu->arch.ctxt.regs.sp;
arch/arm64/kvm/guest.c
143
return &vcpu->arch.ctxt.regs.pc;
arch/arm64/kvm/guest.c
146
return &vcpu->arch.ctxt.regs.pstate;
arch/arm64/kvm/guest.c
149
return __ctxt_sys_reg(&vcpu->arch.ctxt, SP_EL1);
arch/arm64/kvm/guest.c
152
return __ctxt_sys_reg(&vcpu->arch.ctxt, ELR_EL1);
arch/arm64/kvm/guest.c
155
return __ctxt_sys_reg(&vcpu->arch.ctxt, SPSR_EL1);
arch/arm64/kvm/guest.c
158
return &vcpu->arch.ctxt.spsr_abt;
arch/arm64/kvm/guest.c
161
return &vcpu->arch.ctxt.spsr_und;
arch/arm64/kvm/guest.c
164
return &vcpu->arch.ctxt.spsr_irq;
arch/arm64/kvm/guest.c
167
return &vcpu->arch.ctxt.spsr_fiq;
arch/arm64/kvm/guest.c
173
return &vcpu->arch.ctxt.fp_regs.vregs[off];
arch/arm64/kvm/guest.c
176
return &vcpu->arch.ctxt.fp_regs.fpsr;
arch/arm64/kvm/guest.c
179
return &vcpu->arch.ctxt.fp_regs.fpcr;
arch/arm64/kvm/guest.c
321
if (WARN_ON(!sve_vl_valid(vcpu->arch.sve_max_vl)))
arch/arm64/kvm/guest.c
348
if (WARN_ON(vcpu->arch.sve_state))
arch/arm64/kvm/guest.c
378
vcpu->arch.sve_max_vl = sve_vl_from_vq(max_vq);
arch/arm64/kvm/guest.c
503
if (copy_to_user(uptr, vcpu->arch.sve_state + region.koffset,
arch/arm64/kvm/guest.c
529
if (copy_from_user(vcpu->arch.sve_state + region.koffset, uptr,
arch/arm64/kvm/guest.c
752
events->exception.serror_pending = (vcpu->arch.hcr_el2 & HCR_VSE) ||
arch/arm64/kvm/guest.c
899
vcpu->arch.external_debug_state = dbg->arch;
arch/arm64/kvm/guest.c
911
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/guest.c
913
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/handle_exit.c
200
run->debug.arch.hsr = lower_32_bits(esr);
arch/arm64/kvm/handle_exit.c
201
run->debug.arch.hsr_high = upper_32_bits(esr);
arch/arm64/kvm/handle_exit.c
206
run->debug.arch.far = vcpu->arch.fault.far_el2;
arch/arm64/kvm/hyp/exception.c
57
vcpu->arch.ctxt.spsr_abt = val;
arch/arm64/kvm/hyp/exception.c
65
vcpu->arch.ctxt.spsr_und = val;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
100
return &vcpu->arch.external_debug_state;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
145
guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
164
guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
93
switch (vcpu->arch.debug_owner) {
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
98
return &vcpu->arch.vcpu_debug_state;
arch/arm64/kvm/hyp/include/hyp/switch.h
348
u64 hcrx = vcpu->arch.hcrx_el2;
arch/arm64/kvm/hyp/include/hyp/switch.h
408
vsesr = vcpu->arch.vsesr_el2;
arch/arm64/kvm/hyp/include/hyp/switch.h
423
hcr = __ctxt_sys_reg(&vcpu->arch.ctxt, HCR_EL2);
arch/arm64/kvm/hyp/include/hyp/switch.h
425
hcr = &vcpu->arch.hcr_el2;
arch/arm64/kvm/hyp/include/hyp/switch.h
444
return __get_fault_info(vcpu->arch.fault.esr_el2, &vcpu->arch.fault);
arch/arm64/kvm/hyp/include/hyp/switch.h
450
arm64_mops_reset_regs(vcpu_gp_regs(vcpu), vcpu->arch.fault.esr_el2);
arch/arm64/kvm/hyp/include/hyp/switch.h
471
&vcpu->arch.ctxt.fp_regs.fpsr,
arch/arm64/kvm/hyp/include/hyp/switch.h
625
__fpsimd_restore_state(&vcpu->arch.ctxt.fp_regs);
arch/arm64/kvm/hyp/include/hyp/switch.h
656
if (vcpu->arch.hcr_el2 & HCR_TVM)
arch/arm64/kvm/hyp/include/hyp/switch.h
901
vcpu->arch.ctxt.regs.pstate = read_sysreg_el2(SYS_SPSR);
arch/arm64/kvm/hyp/include/hyp/switch.h
913
vcpu->arch.fault.esr_el2 = read_sysreg_el2(SYS_ESR);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
26
vcpu = container_of(ctxt, struct kvm_vcpu, arch.ctxt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
348
vcpu->arch.ctxt.spsr_abt = read_sysreg(spsr_abt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
349
vcpu->arch.ctxt.spsr_und = read_sysreg(spsr_und);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
350
vcpu->arch.ctxt.spsr_irq = read_sysreg(spsr_irq);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
351
vcpu->arch.ctxt.spsr_fiq = read_sysreg(spsr_fiq);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
365
write_sysreg(vcpu->arch.ctxt.spsr_abt, spsr_abt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
366
write_sysreg(vcpu->arch.ctxt.spsr_und, spsr_und);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
367
write_sysreg(vcpu->arch.ctxt.spsr_irq, spsr_irq);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
368
write_sysreg(vcpu->arch.ctxt.spsr_fiq, spsr_fiq);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
41
return &vcpu->arch.external_mdscr_el1;
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
51
test_bit(KVM_ARCH_FLAG_WRITABLE_IMP_ID_REGS, &kvm->arch.flags)))
arch/arm64/kvm/hyp/include/nvhe/mem_protect.h
19
struct kvm_arch arch;
arch/arm64/kvm/hyp/include/nvhe/mem_protect.h
70
__load_stage2(&host_mmu.arch.mmu, &host_mmu.arch);
arch/arm64/kvm/hyp/include/nvhe/trace.h
18
return vcpu ? vcpu->arch.pid : 0;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
104
hyp_vcpu->vcpu.arch.debug_owner = host_vcpu->arch.debug_owner;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
107
hyp_vcpu->vcpu.arch.vcpu_debug_state = host_vcpu->arch.vcpu_debug_state;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
109
hyp_vcpu->vcpu.arch.external_debug_state = host_vcpu->arch.external_debug_state;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
117
host_vcpu->arch.vcpu_debug_state = hyp_vcpu->vcpu.arch.vcpu_debug_state;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
119
host_vcpu->arch.external_debug_state = hyp_vcpu->vcpu.arch.external_debug_state;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
129
hyp_vcpu->vcpu.arch.ctxt = host_vcpu->arch.ctxt;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
131
hyp_vcpu->vcpu.arch.mdcr_el2 = host_vcpu->arch.mdcr_el2;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
132
hyp_vcpu->vcpu.arch.hcr_el2 &= ~(HCR_TWI | HCR_TWE);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
133
hyp_vcpu->vcpu.arch.hcr_el2 |= READ_ONCE(host_vcpu->arch.hcr_el2) &
arch/arm64/kvm/hyp/nvhe/hyp-main.c
136
hyp_vcpu->vcpu.arch.iflags = host_vcpu->arch.iflags;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
138
hyp_vcpu->vcpu.arch.vsesr_el2 = host_vcpu->arch.vsesr_el2;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
140
hyp_vcpu->vcpu.arch.vgic_cpu.vgic_v3 = host_vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
142
hyp_vcpu->vcpu.arch.pid = host_vcpu->arch.pid;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
148
struct vgic_v3_cpu_if *hyp_cpu_if = &hyp_vcpu->vcpu.arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
149
struct vgic_v3_cpu_if *host_cpu_if = &host_vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
155
host_vcpu->arch.ctxt = hyp_vcpu->vcpu.arch.ctxt;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
157
host_vcpu->arch.hcr_el2 = hyp_vcpu->vcpu.arch.hcr_el2;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
159
host_vcpu->arch.fault = hyp_vcpu->vcpu.arch.fault;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
161
host_vcpu->arch.iflags = hyp_vcpu->vcpu.arch.iflags;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
182
hyp_vcpu->vcpu.arch.hcr_el2 &= ~(HCR_TWE | HCR_TWI);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
183
hyp_vcpu->vcpu.arch.hcr_el2 |= hcr_el2 & (HCR_TWE | HCR_TWI);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
185
memcpy(&hyp_vcpu->vcpu.arch.fgt, hyp_vcpu->host_vcpu->arch.fgt,
arch/arm64/kvm/hyp/nvhe/hyp-main.c
186
sizeof(hyp_vcpu->vcpu.arch.fgt));
arch/arm64/kvm/hyp/nvhe/hyp-main.c
243
return refill_memcache(&hyp_vcpu->vcpu.arch.pkvm_memcache,
arch/arm64/kvm/hyp/nvhe/hyp-main.c
244
host_vcpu->arch.pkvm_memcache.nr_pages,
arch/arm64/kvm/hyp/nvhe/hyp-main.c
245
&host_vcpu->arch.pkvm_memcache);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
38
__sve_save_state(vcpu_sve_pffr(vcpu), &vcpu->arch.ctxt.fp_regs.fpsr, true);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
431
__kvm_tlb_flush_vmid(&hyp_vm->kvm.arch.mmu);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
83
__fpsimd_save_state(&vcpu->arch.ctxt.fp_regs);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1020
&vcpu->vcpu.arch.pkvm_memcache, 0));
arch/arm64/kvm/hyp/nvhe/mem_protect.c
138
host_mmu.arch.mmu.vtcr = kvm_get_vtcr(id_aa64mmfr0_el1_sys_val,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1396
&vcpu->vcpu.arch.pkvm_memcache, 0));
arch/arm64/kvm/hyp/nvhe/mem_protect.c
146
struct kvm_s2_mmu *mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1463
&vcpu->vcpu.arch.pkvm_memcache, 0));
arch/arm64/kvm/hyp/nvhe/mem_protect.c
151
mmu->arch = &host_mmu.arch;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1655
return BIT(selftest_vcpu->vcpu.arch.hw_mmu->pgt->ia_bits - 1);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
260
struct kvm_s2_mmu *mmu = &vm->kvm.arch.mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
289
vm->kvm.arch.mmu.pgd_phys = __hyp_pa(vm->pgt.pgd);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
302
vm->kvm.arch.mmu.pgd_phys = 0ULL;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
319
struct kvm_s2_mmu *mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
340
__load_stage2(&host_mmu.arch.mmu, &host_mmu.arch);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
381
u64 limit = BIT(kvm_phys_shift(&host_mmu.arch.mmu) - PAGE_SHIFT);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
610
pkvm_handle_t handle = vm->kvm.arch.pkvm.handle;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
985
&vcpu->vcpu.arch.pkvm_memcache, 0));
arch/arm64/kvm/hyp/nvhe/pkvm.c
1003
vcpu_mc = &hyp_vcpu->vcpu.arch.pkvm_memcache;
arch/arm64/kvm/hyp/nvhe/pkvm.c
101
u64 val = vcpu->arch.mdcr_el2;
arch/arm64/kvm/hyp/nvhe/pkvm.c
1030
vcpu->arch.fault.esr_el2 = (ESR_ELx_EC_DABT_LOW << ESR_ELx_EC_SHIFT) |
arch/arm64/kvm/hyp/nvhe/pkvm.c
1035
vcpu->arch.fault.hpfar_el2 = (HPFAR_EL2_NS | (ipa >> 8)) & HPFAR_MASK;
arch/arm64/kvm/hyp/nvhe/pkvm.c
1038
vcpu->arch.fault.far_el2 = 0;
arch/arm64/kvm/hyp/nvhe/pkvm.c
129
vcpu->arch.mdcr_el2 = val;
arch/arm64/kvm/hyp/nvhe/pkvm.c
168
vcpu->arch.mdcr_el2 = 0;
arch/arm64/kvm/hyp/nvhe/pkvm.c
176
vcpu->arch.hcrx_el2 = host_vcpu->arch.hcrx_el2;
arch/arm64/kvm/hyp/nvhe/pkvm.c
263
if (!hyp_vm || hyp_vm->kvm.arch.pkvm.is_dying)
arch/arm64/kvm/hyp/nvhe/pkvm.c
341
unsigned long host_arch_flags = READ_ONCE(host_kvm->arch.flags);
arch/arm64/kvm/hyp/nvhe/pkvm.c
345
hyp_vm->kvm.arch.ctr_el0 = host_kvm->arch.ctr_el0;
arch/arm64/kvm/hyp/nvhe/pkvm.c
348
hyp_vm->kvm.arch.vgic.vgic_model = host_kvm->arch.vgic.vgic_model;
arch/arm64/kvm/hyp/nvhe/pkvm.c
35
vcpu->arch.hcr_el2 = HCR_GUEST_FLAGS;
arch/arm64/kvm/hyp/nvhe/pkvm.c
352
hyp_vm->kvm.arch.flags = host_arch_flags;
arch/arm64/kvm/hyp/nvhe/pkvm.c
353
hyp_vm->kvm.arch.flags &= ~BIT_ULL(KVM_ARCH_FLAG_ID_REGS_INITIALIZED);
arch/arm64/kvm/hyp/nvhe/pkvm.c
355
bitmap_copy(kvm->arch.vcpu_features,
arch/arm64/kvm/hyp/nvhe/pkvm.c
356
host_kvm->arch.vcpu_features,
arch/arm64/kvm/hyp/nvhe/pkvm.c
360
hyp_vm->kvm.arch.midr_el1 = host_kvm->arch.midr_el1;
arch/arm64/kvm/hyp/nvhe/pkvm.c
366
kvm->arch.flags |= host_arch_flags & BIT(KVM_ARCH_FLAG_MTE_ENABLED);
arch/arm64/kvm/hyp/nvhe/pkvm.c
38
vcpu->arch.hcr_el2 |= HCR_E2H;
arch/arm64/kvm/hyp/nvhe/pkvm.c
383
kvm->arch.flags |= host_arch_flags & BIT(KVM_ARCH_FLAG_GUEST_HAS_SVE);
arch/arm64/kvm/hyp/nvhe/pkvm.c
386
bitmap_and(kvm->arch.vcpu_features, host_kvm->arch.vcpu_features,
arch/arm64/kvm/hyp/nvhe/pkvm.c
403
sve_state = hyp_vcpu->vcpu.arch.sve_state;
arch/arm64/kvm/hyp/nvhe/pkvm.c
42
vcpu->arch.hcr_el2 |= HCR_TEA;
arch/arm64/kvm/hyp/nvhe/pkvm.c
427
struct kvm_s2_mmu *mmu = &hyp_vm->kvm.arch.mmu;
arch/arm64/kvm/hyp/nvhe/pkvm.c
430
hyp_vm->kvm.arch.pkvm.handle = handle;
arch/arm64/kvm/hyp/nvhe/pkvm.c
434
hyp_vm->kvm.arch.pkvm.is_protected = READ_ONCE(host_kvm->arch.pkvm.is_protected);
arch/arm64/kvm/hyp/nvhe/pkvm.c
435
hyp_vm->kvm.arch.pkvm.is_created = true;
arch/arm64/kvm/hyp/nvhe/pkvm.c
436
hyp_vm->kvm.arch.flags = 0;
arch/arm64/kvm/hyp/nvhe/pkvm.c
44
vcpu->arch.hcr_el2 |= HCR_TERR;
arch/arm64/kvm/hyp/nvhe/pkvm.c
442
mmu->vtcr = host_mmu.arch.mmu.vtcr;
arch/arm64/kvm/hyp/nvhe/pkvm.c
443
mmu->arch = &hyp_vm->kvm.arch;
arch/arm64/kvm/hyp/nvhe/pkvm.c
461
sve_max_vl = min(READ_ONCE(host_vcpu->arch.sve_max_vl), kvm_host_sve_max_vl);
arch/arm64/kvm/hyp/nvhe/pkvm.c
463
sve_state = kern_hyp_va(READ_ONCE(host_vcpu->arch.sve_state));
arch/arm64/kvm/hyp/nvhe/pkvm.c
474
vcpu->arch.sve_state = sve_state;
arch/arm64/kvm/hyp/nvhe/pkvm.c
475
vcpu->arch.sve_max_vl = sve_max_vl;
arch/arm64/kvm/hyp/nvhe/pkvm.c
479
clear_bit(KVM_ARM_VCPU_SVE, vcpu->kvm->arch.vcpu_features);
arch/arm64/kvm/hyp/nvhe/pkvm.c
48
vcpu->arch.hcr_el2 |= HCR_FWB;
arch/arm64/kvm/hyp/nvhe/pkvm.c
489
if (!test_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &host_kvm->arch.flags))
arch/arm64/kvm/hyp/nvhe/pkvm.c
492
if (test_and_set_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &kvm->arch.flags))
arch/arm64/kvm/hyp/nvhe/pkvm.c
495
memcpy(kvm->arch.id_regs, host_kvm->arch.id_regs, sizeof(kvm->arch.id_regs));
arch/arm64/kvm/hyp/nvhe/pkvm.c
527
hyp_vcpu->vcpu.arch.hw_mmu = &hyp_vm->kvm.arch.mmu;
arch/arm64/kvm/hyp/nvhe/pkvm.c
528
hyp_vcpu->vcpu.arch.cflags = READ_ONCE(host_vcpu->arch.cflags);
arch/arm64/kvm/hyp/nvhe/pkvm.c
529
hyp_vcpu->vcpu.arch.mp_state.mp_state = KVM_MP_STATE_STOPPED;
arch/arm64/kvm/hyp/nvhe/pkvm.c
53
vcpu->arch.hcr_el2 |= HCR_TID4;
arch/arm64/kvm/hyp/nvhe/pkvm.c
55
vcpu->arch.hcr_el2 |= HCR_TID2;
arch/arm64/kvm/hyp/nvhe/pkvm.c
58
vcpu->arch.hcr_el2 |= (HCR_API | HCR_APK);
arch/arm64/kvm/hyp/nvhe/pkvm.c
61
vcpu->arch.hcr_el2 |= HCR_ATA;
arch/arm64/kvm/hyp/nvhe/pkvm.c
67
u64 val = vcpu->arch.hcr_el2;
arch/arm64/kvm/hyp/nvhe/pkvm.c
733
.arch = {
arch/arm64/kvm/hyp/nvhe/pkvm.c
735
.arch = &selftest_vm.kvm.arch,
arch/arm64/kvm/hyp/nvhe/pkvm.c
744
.arch = {
arch/arm64/kvm/hyp/nvhe/pkvm.c
745
.hw_mmu = &selftest_vm.kvm.arch.mmu,
arch/arm64/kvm/hyp/nvhe/pkvm.c
756
selftest_vm.kvm.arch.mmu.vtcr = host_mmu.arch.mmu.vtcr;
arch/arm64/kvm/hyp/nvhe/pkvm.c
766
selftest_vm.kvm.arch.pkvm.handle = __pkvm_reserve_vm();
arch/arm64/kvm/hyp/nvhe/pkvm.c
767
insert_vm_table_entry(selftest_vm.kvm.arch.pkvm.handle, &selftest_vm);
arch/arm64/kvm/hyp/nvhe/pkvm.c
774
remove_vm_table_entry(selftest_vm.kvm.arch.pkvm.handle);
arch/arm64/kvm/hyp/nvhe/pkvm.c
813
handle = READ_ONCE(host_kvm->arch.pkvm.handle);
arch/arm64/kvm/hyp/nvhe/pkvm.c
820
pgd_size = kvm_pgtable_stage2_pgd_size(host_mmu.arch.mmu.vtcr);
arch/arm64/kvm/hyp/nvhe/pkvm.c
927
if (hyp_vm->kvm.arch.pkvm.is_dying)
arch/arm64/kvm/hyp/nvhe/pkvm.c
95
vcpu->arch.hcr_el2 = val;
arch/arm64/kvm/hyp/nvhe/pkvm.c
954
if (!hyp_vm || hyp_vm->kvm.arch.pkvm.is_dying) {
arch/arm64/kvm/hyp/nvhe/pkvm.c
959
hyp_vm->kvm.arch.pkvm.is_dying = true;
arch/arm64/kvm/hyp/nvhe/pkvm.c
977
if (!hyp_vm || !hyp_vm->kvm.arch.pkvm.is_dying) {
arch/arm64/kvm/hyp/nvhe/pkvm.c
985
__kvm_tlb_flush_vmid(&hyp_vm->kvm.arch.mmu);
arch/arm64/kvm/hyp/nvhe/pkvm.c
990
mc = &host_kvm->arch.pkvm.teardown_mc;
arch/arm64/kvm/hyp/nvhe/pkvm.c
991
stage2_mc = &host_kvm->arch.pkvm.stage2_teardown_mc;
arch/arm64/kvm/hyp/nvhe/switch.c
117
__vgic_v5_save_state(&vcpu->arch.vgic_cpu.vgic_v5);
arch/arm64/kvm/hyp/nvhe/switch.c
118
__vgic_v5_save_ppi_state(&vcpu->arch.vgic_cpu.vgic_v5);
arch/arm64/kvm/hyp/nvhe/switch.c
123
__vgic_v3_save_state(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/hyp/nvhe/switch.c
124
__vgic_v3_deactivate_traps(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/hyp/nvhe/switch.c
132
__vgic_v5_restore_state(&vcpu->arch.vgic_cpu.vgic_v5);
arch/arm64/kvm/hyp/nvhe/switch.c
133
__vgic_v5_restore_ppi_state(&vcpu->arch.vgic_cpu.vgic_v5);
arch/arm64/kvm/hyp/nvhe/switch.c
138
__vgic_v3_activate_traps(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/hyp/nvhe/switch.c
139
__vgic_v3_restore_state(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/hyp/nvhe/switch.c
149
struct kvm_pmu_events *pmu = &vcpu->arch.pmu.events;
arch/arm64/kvm/hyp/nvhe/switch.c
165
struct kvm_pmu_events *pmu = &vcpu->arch.pmu.events;
arch/arm64/kvm/hyp/nvhe/switch.c
279
guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/nvhe/switch.c
317
mmu = kern_hyp_va(vcpu->arch.hw_mmu);
arch/arm64/kvm/hyp/nvhe/switch.c
318
__load_stage2(mmu, kern_hyp_va(mmu->arch));
arch/arm64/kvm/hyp/nvhe/switch.c
55
___activate_traps(vcpu, vcpu->arch.hcr_el2);
arch/arm64/kvm/hyp/nvhe/switch.c
58
write_sysreg(vcpu->arch.mdcr_el2, mdcr_el2);
arch/arm64/kvm/hyp/nvhe/switch.c
66
struct kvm_cpu_context *ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/nvhe/sys_regs.c
290
if (WARN_ON_ONCE(!test_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &kvm->arch.flags)))
arch/arm64/kvm/hyp/nvhe/sys_regs.c
294
return kvm->arch.id_regs[IDREG_IDX(reg)];
arch/arm64/kvm/hyp/nvhe/sys_regs.c
532
struct kvm_arch *ka = &kvm->arch;
arch/arm64/kvm/hyp/nvhe/sys_regs.c
537
if (test_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &kvm->arch.flags))
arch/arm64/kvm/hyp/nvhe/sys_regs.c
547
set_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &kvm->arch.flags);
arch/arm64/kvm/hyp/nvhe/timer-sr.c
52
!kern_hyp_va(vcpu->kvm)->arch.timer_data.poffset)
arch/arm64/kvm/hyp/nvhe/tlb.c
113
__load_stage2(mmu, kern_hyp_va(mmu->arch));
arch/arm64/kvm/hyp/nvhe/tlb.c
131
__load_stage2(mmu, kern_hyp_va(mmu->arch));
arch/arm64/kvm/hyp/nvhe/tlb.c
23
struct kvm_s2_mmu *host_s2_mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/tlb.c
62
if (mmu == vcpu->arch.hw_mmu || WARN_ON(mmu != host_s2_mmu))
arch/arm64/kvm/hyp/nvhe/tlb.c
65
cxt->mmu = vcpu->arch.hw_mmu;
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
40
struct vgic_dist *vgic = &kvm->arch.vgic;
arch/arm64/kvm/hyp/vgic-v3-sr.c
1150
if (kern_hyp_va(vcpu->kvm)->arch.vgic.vgic_model != KVM_DEV_TYPE_ARM_VGIC_V3)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1249
if (vcpu->arch.vgic_cpu.vgic_v3.vgic_hcr & ICH_HCR_EL2_TDIR) {
arch/arm64/kvm/hyp/vgic-v3-sr.c
559
unsigned int used_lrs = vcpu->arch.vgic_cpu.vgic_v3.used_lrs;
arch/arm64/kvm/hyp/vgic-v3-sr.c
598
unsigned int used_lrs = vcpu->arch.vgic_cpu.vgic_v3.used_lrs;
arch/arm64/kvm/hyp/vhe/switch.c
222
__load_stage2(vcpu->arch.hw_mmu, vcpu->arch.hw_mmu->arch);
arch/arm64/kvm/hyp/vhe/switch.c
431
atomic_read(&vcpu->kvm->arch.vncr_map_count))
arch/arm64/kvm/hyp/vhe/switch.c
515
vcpu->arch.fault.esr_el2 = FIELD_PREP(ESR_ELx_EC_MASK, ESR_ELx_EC_SYS64) |
arch/arm64/kvm/hyp/vhe/switch.c
54
u64 guest_hcr, hcr = vcpu->arch.hcr_el2;
arch/arm64/kvm/hyp/vhe/switch.c
579
guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/vhe/switch.c
83
write_sysreg_s(vcpu->arch.ctxt.vncr_array, SYS_VNCR_EL2);
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
126
if (ctxt_has_tcrx(&vcpu->arch.ctxt)) {
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
129
if (ctxt_has_s1pie(&vcpu->arch.ctxt)) {
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
134
if (ctxt_has_s1poe(&vcpu->arch.ctxt))
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
146
if (ctxt_has_sctlr2(&vcpu->arch.ctxt))
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
214
struct kvm_cpu_context *guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
276
struct kvm_cpu_context *guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
54
if (ctxt_has_tcrx(&vcpu->arch.ctxt)) {
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
57
if (ctxt_has_s1pie(&vcpu->arch.ctxt)) {
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
62
if (ctxt_has_s1poe(&vcpu->arch.ctxt))
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
81
if (ctxt_has_sctlr2(&vcpu->arch.ctxt))
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
93
write_sysreg(ctxt_midr_el1(&vcpu->arch.ctxt), vpidr_el2);
arch/arm64/kvm/hyp/vhe/tlb.c
28
if (vcpu && mmu != vcpu->arch.hw_mmu)
arch/arm64/kvm/hyp/vhe/tlb.c
29
cxt->mmu = vcpu->arch.hw_mmu;
arch/arm64/kvm/hyp/vhe/tlb.c
63
__load_stage2(mmu, mmu->arch);
arch/arm64/kvm/hyp/vhe/tlb.c
81
__load_stage2(cxt->mmu, cxt->mmu->arch);
arch/arm64/kvm/hypercalls.c
100
struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
arch/arm64/kvm/hypercalls.c
147
r = mtree_insert_range(&kvm->arch.smccc_filter,
arch/arm64/kvm/hypercalls.c
154
r = mtree_insert_range(&kvm->arch.smccc_filter,
arch/arm64/kvm/hypercalls.c
163
mtree_destroy(&kvm->arch.smccc_filter);
arch/arm64/kvm/hypercalls.c
169
return !mtree_empty(&kvm->arch.smccc_filter);
arch/arm64/kvm/hypercalls.c
191
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/hypercalls.c
204
r = mtree_insert_range(&kvm->arch.smccc_filter, start, end,
arch/arm64/kvm/hypercalls.c
207
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/hypercalls.c
225
val = mt_find(&kvm->arch.smccc_filter, &idx, idx);
arch/arm64/kvm/hypercalls.c
267
struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
arch/arm64/kvm/hypercalls.c
401
struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
arch/arm64/kvm/hypercalls.c
407
mt_init(&kvm->arch.smccc_filter);
arch/arm64/kvm/hypercalls.c
412
mtree_destroy(&kvm->arch.smccc_filter);
arch/arm64/kvm/hypercalls.c
485
struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
arch/arm64/kvm/hypercalls.c
49
cycles = systime_snapshot.cycles - vcpu->kvm->arch.timer_data.voffset;
arch/arm64/kvm/hypercalls.c
52
cycles = systime_snapshot.cycles - vcpu->kvm->arch.timer_data.poffset;
arch/arm64/kvm/hypercalls.c
524
struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
arch/arm64/kvm/hypercalls.c
552
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/hypercalls.c
561
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/hypercalls.c
587
vcpu->kvm->arch.psci_version = val;
arch/arm64/kvm/hypercalls.c
596
vcpu->kvm->arch.psci_version = val;
arch/arm64/kvm/mmio.c
181
&vcpu->kvm->arch.flags)) {
arch/arm64/kvm/mmu.c
1024
kvm_free_stage2_pgd(&kvm->arch.mmu);
arch/arm64/kvm/mmu.c
1025
kvm_mmu_free_memory_cache(&kvm->arch.mmu.split_page_cache);
arch/arm64/kvm/mmu.c
1064
kvm_stage2_unmap_range(&kvm->arch.mmu, gpa, vm_end - vm_start, true);
arch/arm64/kvm/mmu.c
114
chunk_size = kvm->arch.mmu.split_page_chunk_size;
arch/arm64/kvm/mmu.c
116
cache = &kvm->arch.mmu.split_page_cache;
arch/arm64/kvm/mmu.c
1184
struct kvm_s2_mmu *mmu = &kvm->arch.mmu;
arch/arm64/kvm/mmu.c
1253
kvm_stage2_wp_range(&kvm->arch.mmu, start, end);
arch/arm64/kvm/mmu.c
130
chunk_size = kvm->arch.mmu.split_page_chunk_size;
arch/arm64/kvm/mmu.c
1308
kvm_stage2_wp_range(&kvm->arch.mmu, start, end);
arch/arm64/kvm/mmu.c
136
cache = &kvm->arch.mmu.split_page_cache;
arch/arm64/kvm/mmu.c
151
pgt = kvm->arch.mmu.pgt;
arch/arm64/kvm/mmu.c
1520
return &vcpu->arch.mmu_page_cache;
arch/arm64/kvm/mmu.c
1522
return &vcpu->arch.pkvm_memcache;
arch/arm64/kvm/mmu.c
1527
int min_pages = kvm_mmu_cache_min_pages(vcpu->arch.hw_mmu);
arch/arm64/kvm/mmu.c
1581
struct kvm_pgtable *pgt = s2fd->vcpu->arch.hw_mmu->pgt;
arch/arm64/kvm/mmu.c
1667
struct kvm_pgtable *pgt = vcpu->arch.hw_mmu->pgt;
arch/arm64/kvm/mmu.c
178
kvm_call_hyp_nvhe(__pkvm_tlb_flush_vmid, kvm->arch.pkvm.handle);
arch/arm64/kvm/mmu.c
180
kvm_call_hyp(__kvm_tlb_flush_vmid, &kvm->arch.mmu);
arch/arm64/kvm/mmu.c
191
kvm_call_hyp_nvhe(__pkvm_tlb_flush_vmid, kvm->arch.pkvm.handle);
arch/arm64/kvm/mmu.c
193
kvm_tlb_flush_vmid_range(&kvm->arch.mmu, addr, size);
arch/arm64/kvm/mmu.c
1987
pgt = s2fd->vcpu->arch.hw_mmu->pgt;
arch/arm64/kvm/mmu.c
2107
mmu = vcpu->arch.hw_mmu;
arch/arm64/kvm/mmu.c
2163
!test_bit(KVM_ARCH_FLAG_EXIT_SEA, &vcpu->kvm->arch.flags))
arch/arm64/kvm/mmu.c
2237
if (fault_ipa >= BIT_ULL(VTCR_EL2_IPA(vcpu->arch.hw_mmu->vtcr))) {
arch/arm64/kvm/mmu.c
2274
if (kvm_is_nested_s2_mmu(vcpu->kvm,vcpu->arch.hw_mmu) &&
arch/arm64/kvm/mmu.c
2275
vcpu->arch.hw_mmu->nested_stage2_enabled) {
arch/arm64/kvm/mmu.c
2350
VM_BUG_ON(ipa >= kvm_phys_size(vcpu->arch.hw_mmu));
arch/arm64/kvm/mmu.c
2391
if (!kvm->arch.mmu.pgt || kvm_vm_is_protected(kvm))
arch/arm64/kvm/mmu.c
2394
__unmap_stage2_range(&kvm->arch.mmu, range->start << PAGE_SHIFT,
arch/arm64/kvm/mmu.c
2406
if (!kvm->arch.mmu.pgt || kvm_vm_is_protected(kvm))
arch/arm64/kvm/mmu.c
2409
return KVM_PGT_FN(kvm_pgtable_stage2_test_clear_young)(kvm->arch.mmu.pgt,
arch/arm64/kvm/mmu.c
2422
if (!kvm->arch.mmu.pgt || kvm_vm_is_protected(kvm))
arch/arm64/kvm/mmu.c
2425
return KVM_PGT_FN(kvm_pgtable_stage2_test_clear_young)(kvm->arch.mmu.pgt,
arch/arm64/kvm/mmu.c
2567
kvm_mmu_free_memory_cache(&kvm->arch.mmu.split_page_cache);
arch/arm64/kvm/mmu.c
2600
if ((new->base_gfn + new->npages) > (kvm_phys_size(&kvm->arch.mmu) >> PAGE_SHIFT))
arch/arm64/kvm/mmu.c
2675
kvm_stage2_unmap_range(&kvm->arch.mmu, gpa, size, true);
arch/arm64/kvm/mmu.c
360
kvm_stage2_flush_range(&kvm->arch.mmu, addr, end);
arch/arm64/kvm/mmu.c
985
mmu->arch = &kvm->arch;
arch/arm64/kvm/nested.c
104
if (kvm->arch.nested_mmus != tmp)
arch/arm64/kvm/nested.c
105
for (int i = 0; i < kvm->arch.nested_mmus_size; i++)
arch/arm64/kvm/nested.c
106
kvm->arch.nested_mmus[i].pgt->mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
108
for (int i = kvm->arch.nested_mmus_size; !ret && i < num_mmus; i++)
arch/arm64/kvm/nested.c
109
ret = init_nested_s2_mmu(kvm, &kvm->arch.nested_mmus[i]);
arch/arm64/kvm/nested.c
112
for (int i = kvm->arch.nested_mmus_size; i < num_mmus; i++)
arch/arm64/kvm/nested.c
113
kvm_free_stage2_pgd(&kvm->arch.nested_mmus[i]);
arch/arm64/kvm/nested.c
1140
if (!kvm->arch.nested_mmus_size)
arch/arm64/kvm/nested.c
1143
for (i = 0; i < kvm->arch.nested_mmus_size; i++) {
arch/arm64/kvm/nested.c
1144
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
115
free_page((unsigned long)vcpu->arch.ctxt.vncr_array);
arch/arm64/kvm/nested.c
1150
kvm_invalidate_vncr_ipa(kvm, 0, BIT(kvm->arch.mmu.pgt->ia_bits));
arch/arm64/kvm/nested.c
1159
if (!kvm->arch.nested_mmus_size)
arch/arm64/kvm/nested.c
116
vcpu->arch.ctxt.vncr_array = NULL;
arch/arm64/kvm/nested.c
1162
for (i = 0; i < kvm->arch.nested_mmus_size; i++) {
arch/arm64/kvm/nested.c
1163
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
1169
kvm_invalidate_vncr_ipa(kvm, 0, BIT(kvm->arch.mmu.pgt->ia_bits));
arch/arm64/kvm/nested.c
1178
if (!kvm->arch.nested_mmus_size)
arch/arm64/kvm/nested.c
1181
for (i = 0; i < kvm->arch.nested_mmus_size; i++) {
arch/arm64/kvm/nested.c
1182
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
1193
for (i = 0; i < kvm->arch.nested_mmus_size; i++) {
arch/arm64/kvm/nested.c
1194
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
1199
kvfree(kvm->arch.nested_mmus);
arch/arm64/kvm/nested.c
1200
kvm->arch.nested_mmus = NULL;
arch/arm64/kvm/nested.c
1201
kvm->arch.nested_mmus_size = 0;
arch/arm64/kvm/nested.c
121
kvm->arch.nested_mmus_size = num_mmus;
arch/arm64/kvm/nested.c
1251
vcpu->arch.vncr_tlb = kzalloc_obj(*vcpu->arch.vncr_tlb,
arch/arm64/kvm/nested.c
1253
if (!vcpu->arch.vncr_tlb)
arch/arm64/kvm/nested.c
1274
vt = vcpu->arch.vncr_tlb;
arch/arm64/kvm/nested.c
1349
struct vncr_tlb *vt = vcpu->arch.vncr_tlb;
arch/arm64/kvm/nested.c
1362
struct vncr_tlb *vt = vcpu->arch.vncr_tlb;
arch/arm64/kvm/nested.c
1380
struct vncr_tlb *vt = vcpu->arch.vncr_tlb;
arch/arm64/kvm/nested.c
1451
struct vncr_tlb *vt = vcpu->arch.vncr_tlb;
arch/arm64/kvm/nested.c
1497
atomic_inc(&vcpu->kvm->arch.vncr_map_count);
arch/arm64/kvm/nested.c
1625
if (test_bit(KVM_ARM_VCPU_HAS_EL2_E2H0, kvm->arch.vcpu_features))
arch/arm64/kvm/nested.c
1657
if (test_bit(KVM_ARM_VCPU_HAS_EL2_E2H0, kvm->arch.vcpu_features)) {
arch/arm64/kvm/nested.c
1715
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/nested.c
1717
if (kvm->arch.sysreg_masks)
arch/arm64/kvm/nested.c
1720
kvm->arch.sysreg_masks = kzalloc_obj(*(kvm->arch.sysreg_masks),
arch/arm64/kvm/nested.c
1722
if (!kvm->arch.sysreg_masks)
arch/arm64/kvm/nested.c
1847
struct kvm_s2_mmu *mmu = vcpu->arch.hw_mmu;
arch/arm64/kvm/nested.c
1925
vcpu->arch.mdcr_el2 |= (guest_mdcr & NV_MDCR_GUEST_INCLUDE);
arch/arm64/kvm/nested.c
1935
vcpu->arch.mdcr_el2 |= MDCR_EL2_TDA;
arch/arm64/kvm/nested.c
48
kvm->arch.nested_mmus = NULL;
arch/arm64/kvm/nested.c
49
kvm->arch.nested_mmus_size = 0;
arch/arm64/kvm/nested.c
50
atomic_set(&kvm->arch.vncr_map_count, 0);
arch/arm64/kvm/nested.c
645
for (int i = 0; i < kvm->arch.nested_mmus_size; i++) {
arch/arm64/kvm/nested.c
646
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
687
for (int i = 0; i < kvm->arch.nested_mmus_size; i++) {
arch/arm64/kvm/nested.c
688
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
724
for (i = kvm->arch.nested_mmus_next;
arch/arm64/kvm/nested.c
725
i < (kvm->arch.nested_mmus_size + kvm->arch.nested_mmus_next);
arch/arm64/kvm/nested.c
727
s2_mmu = &kvm->arch.nested_mmus[i % kvm->arch.nested_mmus_size];
arch/arm64/kvm/nested.c
735
kvm->arch.nested_mmus_next = (i + 1) % kvm->arch.nested_mmus_size;
arch/arm64/kvm/nested.c
74
if (test_bit(KVM_ARM_VCPU_HAS_EL2_E2H0, kvm->arch.vcpu_features) &&
arch/arm64/kvm/nested.c
78
if (!vcpu->arch.ctxt.vncr_array)
arch/arm64/kvm/nested.c
787
if (!vcpu->arch.hw_mmu)
arch/arm64/kvm/nested.c
788
vcpu->arch.hw_mmu = &vcpu->kvm->arch.mmu;
arch/arm64/kvm/nested.c
79
vcpu->arch.ctxt.vncr_array = (u64 *)__get_free_page(GFP_KERNEL_ACCOUNT |
arch/arm64/kvm/nested.c
790
if (!vcpu->arch.hw_mmu) {
arch/arm64/kvm/nested.c
792
vcpu->arch.hw_mmu = get_s2_mmu_nested(vcpu);
arch/arm64/kvm/nested.c
804
BUG_ON(vcpu->arch.vncr_tlb->cpu != smp_processor_id());
arch/arm64/kvm/nested.c
807
clear_fixmap(vncr_fixmap(vcpu->arch.vncr_tlb->cpu));
arch/arm64/kvm/nested.c
808
vcpu->arch.vncr_tlb->cpu = -1;
arch/arm64/kvm/nested.c
810
atomic_dec(&vcpu->kvm->arch.vncr_map_count);
arch/arm64/kvm/nested.c
82
if (!vcpu->arch.ctxt.vncr_array)
arch/arm64/kvm/nested.c
821
if (kvm_is_nested_s2_mmu(vcpu->kvm, vcpu->arch.hw_mmu))
arch/arm64/kvm/nested.c
822
atomic_dec(&vcpu->arch.hw_mmu->refcnt);
arch/arm64/kvm/nested.c
824
vcpu->arch.hw_mmu = NULL;
arch/arm64/kvm/nested.c
860
vcpu_write_sys_reg(vcpu, vcpu->arch.fault.far_el2, FAR_EL2);
arch/arm64/kvm/nested.c
861
vcpu_write_sys_reg(vcpu, vcpu->arch.fault.hpfar_el2, HPFAR_EL2);
arch/arm64/kvm/nested.c
911
struct vncr_tlb *vt = vcpu->arch.vncr_tlb;
arch/arm64/kvm/nested.c
92
tmp = kvrealloc(kvm->arch.nested_mmus,
arch/arm64/kvm/nested.c
93
size_mul(sizeof(*kvm->arch.nested_mmus), num_mmus),
arch/arm64/kvm/nested.c
962
struct vncr_tlb *vt = vcpu->arch.vncr_tlb;
arch/arm64/kvm/nested.c
98
swap(kvm->arch.nested_mmus, tmp);
arch/arm64/kvm/pkvm.c
101
kvm->arch.pkvm.handle = 0;
arch/arm64/kvm/pkvm.c
102
kvm->arch.pkvm.is_created = false;
arch/arm64/kvm/pkvm.c
103
free_hyp_memcache(&kvm->arch.pkvm.teardown_mc);
arch/arm64/kvm/pkvm.c
104
free_hyp_memcache(&kvm->arch.pkvm.stage2_teardown_mc);
arch/arm64/kvm/pkvm.c
110
pkvm_handle_t handle = vcpu->kvm->arch.pkvm.handle;
arch/arm64/kvm/pkvm.c
114
vcpu->arch.pkvm_memcache.flags |= HYP_MEMCACHE_ACCOUNT_STAGE2;
arch/arm64/kvm/pkvm.c
148
pgd_sz = kvm_pgtable_stage2_pgd_size(kvm->arch.mmu.vtcr);
arch/arm64/kvm/pkvm.c
174
kvm->arch.pkvm.is_created = true;
arch/arm64/kvm/pkvm.c
175
kvm->arch.pkvm.stage2_teardown_mc.flags |= HYP_MEMCACHE_ACCOUNT_STAGE2;
arch/arm64/kvm/pkvm.c
188
return READ_ONCE(kvm->arch.pkvm.is_created);
arch/arm64/kvm/pkvm.c
200
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/pkvm.c
203
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/pkvm.c
213
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/pkvm.c
216
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/pkvm.c
223
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/pkvm.c
225
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/pkvm.c
237
if (kvm->arch.pkvm.handle)
arch/arm64/kvm/pkvm.c
245
kvm->arch.pkvm.handle = ret;
arch/arm64/kvm/pkvm.c
246
kvm->arch.pkvm.is_protected = protected;
arch/arm64/kvm/pkvm.c
340
pkvm_handle_t handle = kvm->arch.pkvm.handle;
arch/arm64/kvm/pkvm.c
366
pkvm_handle_t handle = kvm->arch.pkvm.handle;
arch/arm64/kvm/pkvm.c
386
pkvm_handle_t handle = kvm->arch.pkvm.handle;
arch/arm64/kvm/pkvm.c
391
if (pkvm_hyp_vm_is_created(kvm) && !kvm->arch.pkvm.is_dying) {
arch/arm64/kvm/pkvm.c
393
kvm->arch.pkvm.is_dying = true;
arch/arm64/kvm/pkvm.c
496
pkvm_handle_t handle = kvm->arch.pkvm.handle;
arch/arm64/kvm/pkvm.c
530
pkvm_handle_t handle = kvm->arch.pkvm.handle;
arch/arm64/kvm/pkvm.c
92
kvm->arch.pkvm.handle));
arch/arm64/kvm/pkvm.c
93
} else if (kvm->arch.pkvm.handle) {
arch/arm64/kvm/pkvm.c
98
kvm_call_hyp_nvhe(__pkvm_unreserve_vm, kvm->arch.pkvm.handle);
arch/arm64/kvm/pmu-emul.c
1005
if (vcpu->arch.pmu.irq_num != irq)
arch/arm64/kvm/pmu-emul.c
1008
if (vcpu->arch.pmu.irq_num == irq)
arch/arm64/kvm/pmu-emul.c
1022
struct arm_pmu *arm_pmu = kvm->arch.arm_pmu;
arch/arm64/kvm/pmu-emul.c
1040
kvm->arch.nr_pmu_counters = nr;
arch/arm64/kvm/pmu-emul.c
1043
if (test_bit(KVM_ARM_VCPU_HAS_EL2, kvm->arch.vcpu_features)) {
arch/arm64/kvm/pmu-emul.c
1050
val |= FIELD_PREP(MDCR_EL2_HPMN, kvm->arch.nr_pmu_counters);
arch/arm64/kvm/pmu-emul.c
1058
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/pmu-emul.c
1060
kvm->arch.arm_pmu = arm_pmu;
arch/arm64/kvm/pmu-emul.c
1094
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/pmu-emul.c
1101
(kvm->arch.pmu_filter && kvm->arch.arm_pmu != arm_pmu)) {
arch/arm64/kvm/pmu-emul.c
1107
cpumask_copy(kvm->arch.supported_cpus, &arm_pmu->supported_cpus);
arch/arm64/kvm/pmu-emul.c
1121
if (!kvm->arch.arm_pmu)
arch/arm64/kvm/pmu-emul.c
1135
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/pmu-emul.c
1140
if (vcpu->arch.pmu.created)
arch/arm64/kvm/pmu-emul.c
1165
vcpu->arch.pmu.irq_num = irq;
arch/arm64/kvm/pmu-emul.c
1194
if (!kvm->arch.pmu_filter) {
arch/arm64/kvm/pmu-emul.c
1195
kvm->arch.pmu_filter = bitmap_alloc(nr_events, GFP_KERNEL_ACCOUNT);
arch/arm64/kvm/pmu-emul.c
1196
if (!kvm->arch.pmu_filter)
arch/arm64/kvm/pmu-emul.c
1206
bitmap_zero(kvm->arch.pmu_filter, nr_events);
arch/arm64/kvm/pmu-emul.c
1208
bitmap_fill(kvm->arch.pmu_filter, nr_events);
arch/arm64/kvm/pmu-emul.c
1212
bitmap_set(kvm->arch.pmu_filter, filter.base_event, filter.nevents);
arch/arm64/kvm/pmu-emul.c
1214
bitmap_clear(kvm->arch.pmu_filter, filter.base_event, filter.nevents);
arch/arm64/kvm/pmu-emul.c
1259
irq = vcpu->arch.pmu.irq_num;
arch/arm64/kvm/pmu-emul.c
1313
u64 n = vcpu->kvm->arch.nr_pmu_counters;
arch/arm64/kvm/pmu-emul.c
255
struct kvm_pmu *pmu = &vcpu->arch.pmu;
arch/arm64/kvm/pmu-emul.c
272
irq_work_sync(&vcpu->arch.pmu.overflow_work);
arch/arm64/kvm/pmu-emul.c
283
n = vcpu->kvm->arch.nr_pmu_counters;
arch/arm64/kvm/pmu-emul.c
35
return container_of(pmc, struct kvm_vcpu, arch.pmu.pmc[pmc->idx]);
arch/arm64/kvm/pmu-emul.c
398
struct kvm_pmu *pmu = &vcpu->arch.pmu;
arch/arm64/kvm/pmu-emul.c
40
return &vcpu->arch.pmu.pmc[cnt_idx];
arch/arm64/kvm/pmu-emul.c
416
struct kvm_pmu *pmu = &vcpu->arch.pmu;
arch/arm64/kvm/pmu-emul.c
435
if (vcpu->arch.pmu.irq_level)
arch/arm64/kvm/pmu-emul.c
472
vcpu = container_of(work, struct kvm_vcpu, arch.pmu.overflow_work);
arch/arm64/kvm/pmu-emul.c
571
irq_work_queue(&vcpu->arch.pmu.overflow_work);
arch/arm64/kvm/pmu-emul.c
667
struct arm_pmu *pmu = kvm->arch.arm_pmu;
arch/arm64/kvm/pmu-emul.c
686
struct arm_pmu *arm_pmu = vcpu->kvm->arch.arm_pmu;
arch/arm64/kvm/pmu-emul.c
712
if (vcpu->kvm->arch.pmu_filter &&
arch/arm64/kvm/pmu-emul.c
713
!test_bit(eventsel, vcpu->kvm->arch.pmu_filter))
arch/arm64/kvm/pmu-emul.c
881
struct arm_pmu *cpu_pmu = vcpu->kvm->arch.arm_pmu;
arch/arm64/kvm/pmu-emul.c
882
unsigned long *bmap = vcpu->kvm->arch.pmu_filter;
arch/arm64/kvm/pmu-emul.c
926
if (!vcpu->arch.pmu.created)
arch/arm64/kvm/pmu-emul.c
935
int irq = vcpu->arch.pmu.irq_num;
arch/arm64/kvm/pmu-emul.c
970
vcpu->arch.pmu.irq_num = KVM_ARMV8_PMU_GICV5_IRQ;
arch/arm64/kvm/pmu-emul.c
973
ret = kvm_vgic_set_owner(vcpu, vcpu->arch.pmu.irq_num,
arch/arm64/kvm/pmu-emul.c
974
&vcpu->arch.pmu);
arch/arm64/kvm/pmu-emul.c
979
init_irq_work(&vcpu->arch.pmu.overflow_work,
arch/arm64/kvm/pmu-emul.c
982
vcpu->arch.pmu.created = true;
arch/arm64/kvm/psci.c
113
WRITE_ONCE(vcpu->arch.mp_state.mp_state, KVM_MP_STATE_RUNNABLE);
arch/arm64/kvm/psci.c
117
spin_unlock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/psci.c
179
spin_lock(&tmp->arch.mp_state_lock);
arch/arm64/kvm/psci.c
180
WRITE_ONCE(tmp->arch.mp_state.mp_state, KVM_MP_STATE_STOPPED);
arch/arm64/kvm/psci.c
181
spin_unlock(&tmp->arch.mp_state_lock);
arch/arm64/kvm/psci.c
359
if (test_bit(KVM_ARCH_FLAG_SYSTEM_SUSPEND_ENABLED, &kvm->arch.flags))
arch/arm64/kvm/psci.c
383
if (test_bit(KVM_ARCH_FLAG_SYSTEM_SUSPEND_ENABLED, &kvm->arch.flags)) {
arch/arm64/kvm/psci.c
81
spin_lock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/psci.c
91
reset_state = &vcpu->arch.reset_state;
arch/arm64/kvm/ptdump.c
287
mmu->arch->debugfs_nv_dentry, mmu,
arch/arm64/kvm/ptdump.c
301
&kvm->arch.mmu, &kvm_ptdump_guest_fops);
arch/arm64/kvm/ptdump.c
303
&kvm->arch.mmu, &kvm_pgtable_range_fops);
arch/arm64/kvm/ptdump.c
305
&kvm->arch.mmu, &kvm_pgtable_levels_fops);
arch/arm64/kvm/ptdump.c
307
kvm->arch.debugfs_nv_dentry = debugfs_create_dir("nested", kvm->debugfs_dentry);
arch/arm64/kvm/pvtime.c
102
vcpu->arch.steal.base = ipa;
arch/arm64/kvm/pvtime.c
117
ipa = vcpu->arch.steal.base;
arch/arm64/kvm/pvtime.c
16
u64 base = vcpu->arch.steal.base;
arch/arm64/kvm/pvtime.c
17
u64 last_steal = vcpu->arch.steal.last_steal;
arch/arm64/kvm/pvtime.c
28
vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay);
arch/arm64/kvm/pvtime.c
29
steal += vcpu->arch.steal.last_steal - last_steal;
arch/arm64/kvm/pvtime.c
43
if (vcpu->arch.steal.base != INVALID_GPA)
arch/arm64/kvm/pvtime.c
55
u64 base = vcpu->arch.steal.base;
arch/arm64/kvm/pvtime.c
64
vcpu->arch.steal.last_steal = current->sched_info.run_delay;
arch/arm64/kvm/pvtime.c
92
if (vcpu->arch.steal.base != INVALID_GPA)
arch/arm64/kvm/reset.c
102
vl = vcpu->arch.sve_max_vl;
arch/arm64/kvm/reset.c
124
vcpu->arch.sve_state = buf;
arch/arm64/kvm/reset.c
155
void *sve_state = vcpu->arch.sve_state;
arch/arm64/kvm/reset.c
161
free_page((unsigned long)vcpu->arch.ctxt.vncr_array);
arch/arm64/kvm/reset.c
162
kfree(vcpu->arch.vncr_tlb);
arch/arm64/kvm/reset.c
163
kfree(vcpu->arch.ccsidr);
arch/arm64/kvm/reset.c
169
memset(vcpu->arch.sve_state, 0, vcpu_sve_state_size(vcpu));
arch/arm64/kvm/reset.c
196
spin_lock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/reset.c
197
reset_state = vcpu->arch.reset_state;
arch/arm64/kvm/reset.c
198
vcpu->arch.reset_state.reset = false;
arch/arm64/kvm/reset.c
199
spin_unlock(&vcpu->arch.mp_state_lock);
arch/arm64/kvm/reset.c
222
memset(&vcpu->arch.ctxt.fp_regs, 0, sizeof(vcpu->arch.ctxt.fp_regs));
arch/arm64/kvm/reset.c
223
vcpu->arch.ctxt.spsr_abt = 0;
arch/arm64/kvm/reset.c
224
vcpu->arch.ctxt.spsr_und = 0;
arch/arm64/kvm/reset.c
225
vcpu->arch.ctxt.spsr_irq = 0;
arch/arm64/kvm/reset.c
226
vcpu->arch.ctxt.spsr_fiq = 0;
arch/arm64/kvm/reset.c
81
vcpu->arch.sve_max_vl = kvm_sve_max_vl;
arch/arm64/kvm/reset.c
88
set_bit(KVM_ARCH_FLAG_GUEST_HAS_SVE, &vcpu->kvm->arch.flags);
arch/arm64/kvm/sys_regs.c
1018
u8 n = vcpu->kvm->arch.nr_pmu_counters;
arch/arm64/kvm/sys_regs.c
1441
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
1451
kvm->arch.nr_pmu_counters = new_n;
arch/arm64/kvm/sys_regs.c
1453
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
1709
if (!test_bit(KVM_ARCH_FLAG_VM_COUNTER_OFFSET, &vcpu->kvm->arch.flags))
arch/arm64/kvm/sys_regs.c
1713
if (!test_bit(KVM_ARCH_FLAG_VM_COUNTER_OFFSET, &vcpu->kvm->arch.flags))
arch/arm64/kvm/sys_regs.c
2416
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
2418
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
2429
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
2441
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
2449
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
2465
u64 *p = __vm_id_reg(&kvm->arch, reg);
arch/arm64/kvm/sys_regs.c
2467
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
3025
if (hpmn > vcpu->kvm->arch.nr_pmu_counters) {
arch/arm64/kvm/sys_regs.c
3026
hpmn = vcpu->kvm->arch.nr_pmu_counters;
arch/arm64/kvm/sys_regs.c
3097
if (test_bit(KVM_ARCH_FLAG_WRITABLE_IMP_ID_REGS, &vcpu->kvm->arch.flags))
arch/arm64/kvm/sys_regs.c
3150
guard(mutex)(&kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
3156
if (!test_bit(KVM_ARCH_FLAG_WRITABLE_IMP_ID_REGS, &kvm->arch.flags))
arch/arm64/kvm/sys_regs.c
3188
__vcpu_assign_sys_reg(vcpu, r->reg, vcpu->kvm->arch.nr_pmu_counters);
arch/arm64/kvm/sys_regs.c
3189
return vcpu->kvm->arch.nr_pmu_counters;
arch/arm64/kvm/sys_regs.c
427
if (vcpu->arch.ccsidr)
arch/arm64/kvm/sys_regs.c
428
return vcpu->arch.ccsidr[csselr];
arch/arm64/kvm/sys_regs.c
461
u32 *ccsidr = vcpu->arch.ccsidr;
arch/arm64/kvm/sys_regs.c
479
vcpu->arch.ccsidr = ccsidr;
arch/arm64/kvm/sys_regs.c
5126
if (!test_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &kvm->arch.flags))
arch/arm64/kvm/sys_regs.c
5189
if (!kvm->arch.sysreg_masks)
arch/arm64/kvm/sys_regs.c
5247
if (test_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &kvm->arch.flags))
arch/arm64/kvm/sys_regs.c
5291
set_bit(KVM_ARCH_FLAG_ID_REGS_INITIALIZED, &kvm->arch.flags);
arch/arm64/kvm/sys_regs.c
5702
vcpu->arch.hcr_el2 |= HCR_E2H;
arch/arm64/kvm/sys_regs.c
5705
vcpu->arch.hcr_el2 |= HCR_TEA;
arch/arm64/kvm/sys_regs.c
5707
vcpu->arch.hcr_el2 |= HCR_TERR;
arch/arm64/kvm/sys_regs.c
5711
vcpu->arch.hcr_el2 |= HCR_FWB;
arch/arm64/kvm/sys_regs.c
5716
vcpu->arch.hcr_el2 |= HCR_TID4;
arch/arm64/kvm/sys_regs.c
5718
vcpu->arch.hcr_el2 |= HCR_TID2;
arch/arm64/kvm/sys_regs.c
5721
vcpu->arch.hcr_el2 &= ~HCR_RW;
arch/arm64/kvm/sys_regs.c
5724
vcpu->arch.hcr_el2 |= HCR_ATA;
arch/arm64/kvm/sys_regs.c
5726
vcpu->arch.hcr_el2 |= HCR_TID5;
arch/arm64/kvm/sys_regs.c
5734
vcpu->arch.hcr_el2 |= HCR_TTLBOS;
arch/arm64/kvm/sys_regs.c
5741
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
5746
if (test_bit(KVM_ARCH_FLAG_FGU_INITIALIZED, &kvm->arch.flags))
arch/arm64/kvm/sys_regs.c
5759
set_bit(KVM_ARCH_FLAG_FGU_INITIALIZED, &kvm->arch.flags);
arch/arm64/kvm/sys_regs.c
5761
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
5776
guard(mutex)(&kvm->arch.config_lock);
arch/arm64/kvm/sys_regs.c
663
p->regval = vcpu->arch.vgic_cpu.vgic_v3.vgic_sre;
arch/arm64/kvm/sys_regs.c
700
p->regval = FIELD_PREP(ICC_IDR0_EL1_PRI_BITS, vcpu->arch.vgic_cpu.num_pri_bits - 1) |
arch/arm64/kvm/sys_regs.c
701
FIELD_PREP(ICC_IDR0_EL1_ID_BITS, vcpu->arch.vgic_cpu.num_id_bits);
arch/arm64/kvm/sys_regs.c
725
unsigned long *mask = vcpu->kvm->arch.vgic.gicv5_vm.vgic_ppi_mask;
arch/arm64/kvm/sys_regs.c
726
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/sys_regs.c
892
struct kvm_guest_debug_arch *dbg = &vcpu->arch.vcpu_debug_state;
arch/arm64/kvm/vgic-sys-reg-v3.c
17
struct vgic_cpu *vgic_v3_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic-sys-reg-v3.c
210
struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic-sys-reg-v3.c
220
struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic-sys-reg-v3.c
293
struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic-sys-reg-v3.c
62
struct vgic_cpu *vgic_v3_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/trace.h
33
#define TRACE_INCLUDE_PATH ../../arch/arm64/kvm/vgic
arch/arm64/kvm/vgic/vgic-debug.c
243
print_dist_state(s, &kvm->arch.vgic, iter);
arch/arm64/kvm/vgic/vgic-debug.c
247
if (!kvm->arch.vgic.initialized)
arch/arm64/kvm/vgic/vgic-debug.c
35
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-debug.c
66
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-debug.c
87
iter->nr_spis = kvm->arch.vgic.nr_spis;
arch/arm64/kvm/vgic/vgic-init.c
109
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
145
kvm->arch.vgic.in_kernel = true;
arch/arm64/kvm/vgic/vgic-init.c
146
kvm->arch.vgic.vgic_model = type;
arch/arm64/kvm/vgic/vgic-init.c
147
kvm->arch.vgic.implementation_rev = KVM_VGIC_IMP_REV_LATEST;
arch/arm64/kvm/vgic/vgic-init.c
148
kvm->arch.vgic.vgic_dist_base = VGIC_ADDR_UNDEF;
arch/arm64/kvm/vgic/vgic-init.c
152
kvm->arch.vgic.vgic_cpu_base = VGIC_ADDR_UNDEF;
arch/arm64/kvm/vgic/vgic-init.c
155
INIT_LIST_HEAD(&kvm->arch.vgic.rd_regions);
arch/arm64/kvm/vgic/vgic-init.c
173
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-init.c
178
kvm->arch.vgic.vgic_model = 0;
arch/arm64/kvm/vgic/vgic-init.c
183
kvm->arch.vgic.nassgicap = system_supports_direct_sgis();
arch/arm64/kvm/vgic/vgic-init.c
195
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
209
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-init.c
260
guard(mutex)(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
267
if (vcpu->kvm->arch.vgic.mi_intid == 0)
arch/arm64/kvm/vgic/vgic-init.c
268
vcpu->kvm->arch.vgic.mi_intid = DEFAULT_MI_INTID;
arch/arm64/kvm/vgic/vgic-init.c
269
ret = kvm_vgic_set_owner(vcpu, vcpu->kvm->arch.vgic.mi_intid, vcpu);
arch/arm64/kvm/vgic/vgic-init.c
276
struct vgic_irq *irq = &vcpu->arch.vgic_cpu.private_irqs[i];
arch/arm64/kvm/vgic/vgic-init.c
308
struct vgic_irq *irq = &vcpu->arch.vgic_cpu.private_irqs[i];
arch/arm64/kvm/vgic/vgic-init.c
331
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-init.c
335
lockdep_assert_held(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
370
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
372
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
388
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-init.c
389
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-init.c
419
const struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-init.c
440
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-init.c
445
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
493
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-init.c
520
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-init.c
532
if (vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3) {
arch/arm64/kvm/vgic/vgic-init.c
54
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-init.c
571
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
580
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
582
if (kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3)
arch/arm64/kvm/vgic/vgic-init.c
606
if (kvm->arch.vgic.vgic_model != KVM_DEV_TYPE_ARM_VGIC_V2)
arch/arm64/kvm/vgic/vgic-init.c
609
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
611
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
631
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-init.c
641
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
665
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
673
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
679
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-init.c
691
u32 type = kvm->arch.vgic.vgic_model;
arch/arm64/kvm/vgic/vgic-irqfd.c
139
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-its.c
1047
lpi_nr >= max_lpis_propbaser(kvm->arch.vgic.propbaser))
arch/arm64/kvm/vgic/vgic-its.c
1297
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-its.c
1310
if (vcpu->arch.vgic_cpu.vgic_v3.its_vpe.its_vm)
arch/arm64/kvm/vgic/vgic-its.c
1311
its_invall_vpe(&vcpu->arch.vgic_cpu.vgic_v3.its_vpe);
arch/arm64/kvm/vgic/vgic-its.c
1352
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-its.c
1798
if (!(vcpu->arch.vgic_cpu.pendbaser & GICR_PENDBASER_PTZ))
arch/arm64/kvm/vgic/vgic-its.c
1853
mutex_lock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-its.c
1858
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-its.c
1881
dev->kvm->arch.vgic.msis_require_devid = true;
arch/arm64/kvm/vgic/vgic-its.c
1882
dev->kvm->arch.vgic.has_its = true;
arch/arm64/kvm/vgic/vgic-its.c
1890
dev->kvm->arch.vgic.propbaser = INITIAL_PROPBASER_VALUE;
arch/arm64/kvm/vgic/vgic-its.c
1896
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-its.c
1976
mutex_lock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-its.c
2005
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-its.c
2681
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-its.c
2700
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-its.c
2718
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-its.c
283
u64 propbase = GICR_PROPBASER_ADDRESS(kvm->arch.vgic.propbaser);
arch/arm64/kvm/vgic/vgic-its.c
331
map.vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe;
arch/arm64/kvm/vgic/vgic-its.c
391
gpa_t pendbase = GICR_PENDBASER_ADDRESS(vcpu->arch.vgic_cpu.pendbaser);
arch/arm64/kvm/vgic/vgic-its.c
392
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-its.c
79
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-kvm-device.c
191
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
199
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
238
mutex_lock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
245
if (dev->kvm->arch.vgic.nr_spis)
arch/arm64/kvm/vgic/vgic-kvm-device.c
248
dev->kvm->arch.vgic.nr_spis =
arch/arm64/kvm/vgic/vgic-kvm-device.c
251
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
258
mutex_lock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
260
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
277
mutex_lock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
279
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
303
r = put_user(dev->kvm->arch.vgic.nr_spis +
arch/arm64/kvm/vgic/vgic-kvm-device.c
31
if (addr & ~kvm_phys_mask(&kvm->arch.mmu) ||
arch/arm64/kvm/vgic/vgic-kvm-device.c
32
(addr + size) > kvm_phys_size(&kvm->arch.mmu))
arch/arm64/kvm/vgic/vgic-kvm-device.c
397
mutex_lock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
40
if (kvm->arch.vgic.vgic_model != type_needed)
arch/arm64/kvm/vgic/vgic-kvm-device.c
416
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
48
struct vgic_dist *vgic = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-kvm-device.c
51
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
576
mutex_lock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
614
mutex_unlock(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
642
guard(mutex)(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
649
dev->kvm->arch.vgic.mi_intid = val;
arch/arm64/kvm/vgic/vgic-kvm-device.c
669
guard(mutex)(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
670
return put_user(dev->kvm->arch.vgic.mi_intid, uaddr);
arch/arm64/kvm/vgic/vgic-kvm-device.c
726
struct vgic_v5_vm *gicv5_vm = &dev->kvm->arch.vgic.gicv5_vm;
arch/arm64/kvm/vgic/vgic-kvm-device.c
73
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
730
guard(mutex)(&dev->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-kvm-device.c
96
struct vgic_dist *vgic = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
114
if (vcpu->kvm->arch.vgic.v2_groups_user_writable)
arch/arm64/kvm/vgic/vgic-mmio-v2.c
28
struct vgic_dist *vgic = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
383
return vcpu->arch.vgic_cpu.vgic_v2.vgic_apr;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
385
struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
409
vcpu->arch.vgic_cpu.vgic_v2.vgic_apr = val;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
411
struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
56
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
76
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
98
vcpu->kvm->arch.vgic.v2_groups_user_writable = true;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
125
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
131
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
159
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
174
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
267
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
275
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
289
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
321
struct vgic_dist *vgic = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
322
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
43
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
482
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
491
struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
510
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
522
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
541
return !!atomic_read(&vcpu->arch.vgic_cpu.syncr_busy);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
547
atomic_inc(&vcpu->arch.vgic_cpu.syncr_busy);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
551
atomic_dec(&vcpu->arch.vgic_cpu.syncr_busy);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
71
return kvm->arch.vgic.nassgicap;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
781
struct vgic_dist *vgic = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
782
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
783
struct vgic_io_device *rd_dev = &vcpu->arch.vgic_cpu.rd_iodev;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
789
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
821
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
833
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
839
struct vgic_io_device *rd_dev = &vcpu->arch.vgic_cpu.rd_iodev;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
84
struct vgic_dist *vgic = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
889
struct vgic_dist *d = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
959
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
963
if (vcpu->arch.vgic_cpu.rdreg == rdreg)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
964
vcpu->arch.vgic_cpu.rdreg = NULL;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
975
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
977
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
989
mutex_lock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio-v3.c
992
mutex_unlock(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio.c
1111
struct vgic_io_device *io_device = &kvm->arch.vgic.dist_iodev;
arch/arm64/kvm/vgic/vgic-mmio.c
264
switch (vcpu->kvm->arch.vgic.vgic_model) {
arch/arm64/kvm/vgic/vgic-mmio.c
301
vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V2);
arch/arm64/kvm/vgic/vgic-mmio.c
473
if ((vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3 &&
arch/arm64/kvm/vgic/vgic-mmio.c
482
if ((vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3 &&
arch/arm64/kvm/vgic/vgic-mmio.c
518
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio.c
524
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio.c
564
u32 model = vcpu->kvm->arch.vgic.vgic_model;
arch/arm64/kvm/vgic/vgic-mmio.c
613
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio.c
619
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio.c
650
mutex_lock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio.c
656
mutex_unlock(&vcpu->kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-mmio.c
770
int nr_irqs = vcpu->kvm->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS;
arch/arm64/kvm/vgic/vgic-mmio.c
792
int nr_irqs = vcpu->kvm->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS;
arch/arm64/kvm/vgic/vgic-mmio.c
845
const struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio.c
867
const struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-mmio.c
950
int flags, nr_irqs = kvm->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS;
arch/arm64/kvm/vgic/vgic-v2.c
115
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-v2.c
153
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-v2.c
304
vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr] = val;
arch/arm64/kvm/vgic/vgic-v2.c
33
struct vgic_v2_cpu_if *cpuif = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v2.c
336
vcpu->arch.vgic_cpu.vgic_v2.vgic_lr[lr] = 0;
arch/arm64/kvm/vgic/vgic-v2.c
341
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v2.c
368
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v2.c
401
vcpu->arch.vgic_cpu.vgic_v2.vgic_vmcr = 0;
arch/arm64/kvm/vgic/vgic-v2.c
422
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v2.c
548
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v2.c
569
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v2.c
571
u64 used_lrs = vcpu->arch.vgic_cpu.vgic_v2.used_lrs;
arch/arm64/kvm/vgic/vgic-v2.c
593
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v2.c
609
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v2.c
619
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
arch/arm64/kvm/vgic/vgic-v3-nested.c
326
struct vgic_v3_cpu_if *host_if = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v3-nested.c
360
vcpu->arch.vgic_cpu.vgic_v3.used_lrs = cpu_if->used_lrs;
arch/arm64/kvm/vgic/vgic-v3-nested.c
381
vcpu->arch.vgic_cpu.vgic_v3.used_lrs = 0;
arch/arm64/kvm/vgic/vgic-v3-nested.c
395
vcpu->kvm->arch.vgic.mi_intid, state, vcpu);
arch/arm64/kvm/vgic/vgic-v3-nested.c
406
vcpu->kvm->arch.vgic.mi_intid, level, vcpu);
arch/arm64/kvm/vgic/vgic-v3.c
130
atomic_dec_if_positive(&vcpu->kvm->arch.vgic.active_spis);
arch/arm64/kvm/vgic/vgic-v3.c
148
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-v3.c
197
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-v3.c
199
u32 model = vcpu->kvm->arch.vgic.vgic_model;
arch/arm64/kvm/vgic/vgic-v3.c
219
val >= vcpu->kvm->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS)
arch/arm64/kvm/vgic/vgic-v3.c
27
struct vgic_v3_cpu_if *cpuif = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v3.c
300
u32 model = vcpu->kvm->arch.vgic.vgic_model;
arch/arm64/kvm/vgic/vgic-v3.c
368
u32 model = vcpu->kvm->arch.vgic.vgic_model;
arch/arm64/kvm/vgic/vgic-v3.c
371
vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[lr] = val;
arch/arm64/kvm/vgic/vgic-v3.c
401
vcpu->arch.vgic_cpu.vgic_v3.vgic_lr[lr] = 0;
arch/arm64/kvm/vgic/vgic-v3.c
406
struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v3.c
407
u32 model = vcpu->kvm->arch.vgic.vgic_model;
arch/arm64/kvm/vgic/vgic-v3.c
434
struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v3.c
435
u32 model = vcpu->kvm->arch.vgic.vgic_model;
arch/arm64/kvm/vgic/vgic-v3.c
468
struct vgic_v3_cpu_if *vgic_v3 = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v3.c
483
if (vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3) {
arch/arm64/kvm/vgic/vgic-v3.c
487
vcpu->arch.vgic_cpu.pendbaser = INITIAL_PENDBASER_VALUE;
arch/arm64/kvm/vgic/vgic-v3.c
492
vcpu->arch.vgic_cpu.num_id_bits = FIELD_GET(ICH_VTR_EL2_IDbits,
arch/arm64/kvm/vgic/vgic-v3.c
494
vcpu->arch.vgic_cpu.num_pri_bits = FIELD_GET(ICH_VTR_EL2_PRIbits,
arch/arm64/kvm/vgic/vgic-v3.c
500
struct vgic_v3_cpu_if *vgic_v3 = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v3.c
506
if (vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V2 ||
arch/arm64/kvm/vgic/vgic-v3.c
527
pendbase = GICR_PENDBASER_ADDRESS(vcpu->arch.vgic_cpu.pendbaser);
arch/arm64/kvm/vgic/vgic-v3.c
563
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v3.c
572
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v3.c
586
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v3.c
618
pendbase = GICR_PENDBASER_ADDRESS(vcpu->arch.vgic_cpu.pendbaser);
arch/arm64/kvm/vgic/vgic-v3.c
64
atomic_read(&vcpu->kvm->arch.vgic.active_spis))
arch/arm64/kvm/vgic/vgic-v3.c
670
struct vgic_dist *d = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v3.c
687
struct vgic_dist *d = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v3.c
735
struct list_head *rd_regions = &kvm->arch.vgic.rd_regions;
arch/arm64/kvm/vgic/vgic-v3.c
748
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v3.c
753
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic-v3.c
81
if (vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3) {
arch/arm64/kvm/vgic/vgic-v3.c
978
struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v3.c
997
struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3;
arch/arm64/kvm/vgic/vgic-v4.c
100
raw_spin_unlock(&vcpu->arch.vgic_cpu.vgic_v3.its_vpe.vpe_lock);
arch/arm64/kvm/vgic/vgic-v4.c
117
struct its_vpe *vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe;
arch/arm64/kvm/vgic/vgic-v4.c
192
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v4.c
196
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-v4.c
218
struct its_vpe *vpe = &irq->target_vcpu->arch.vgic_cpu.vgic_v3.its_vpe;
arch/arm64/kvm/vgic/vgic-v4.c
244
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic-v4.c
249
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-v4.c
267
dist->its_vm.vpes[i] = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe;
arch/arm64/kvm/vgic/vgic-v4.c
320
struct its_vm *its_vm = &kvm->arch.vgic.its_vm;
arch/arm64/kvm/vgic/vgic-v4.c
323
lockdep_assert_held(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-v4.c
360
struct its_vpe *vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe;
arch/arm64/kvm/vgic/vgic-v4.c
370
struct its_vpe *vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe;
arch/arm64/kvm/vgic/vgic-v4.c
389
err = its_make_vpe_resident(vpe, false, vcpu->kvm->arch.vgic.enabled);
arch/arm64/kvm/vgic/vgic-v4.c
406
struct its_vpe *vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe;
arch/arm64/kvm/vgic/vgic-v4.c
476
.vm = &kvm->arch.vgic.its_vm,
arch/arm64/kvm/vgic/vgic-v4.c
477
.vpe = &irq->target_vcpu->arch.vgic_cpu.vgic_v3.its_vpe,
arch/arm64/kvm/vgic/vgic-v4.c
520
xa_for_each(&kvm->arch.vgic.lpi_xa, idx, irq) {
arch/arm64/kvm/vgic/vgic-v4.c
548
atomic_dec(&irq->target_vcpu->arch.vgic_cpu.vgic_v3.its_vpe.vlpi_count);
arch/arm64/kvm/vgic/vgic-v4.c
98
raw_spin_lock(&vcpu->arch.vgic_cpu.vgic_v3.its_vpe.vpe_lock);
arch/arm64/kvm/vgic/vgic-v4.c
99
vcpu->arch.vgic_cpu.vgic_v3.its_vpe.pending_last = true;
arch/arm64/kvm/vgic/vgic-v5.c
122
vcpu->arch.vgic_cpu.num_id_bits = ICC_IDR0_EL1_ID_BITS_16BITS;
arch/arm64/kvm/vgic/vgic-v5.c
128
vcpu->arch.vgic_cpu.num_pri_bits = 5;
arch/arm64/kvm/vgic/vgic-v5.c
147
bitmap_zero(kvm->arch.vgic.gicv5_vm.userspace_ppis,
arch/arm64/kvm/vgic/vgic-v5.c
150
kvm->arch.vgic.gicv5_vm.userspace_ppis,
arch/arm64/kvm/vgic/vgic-v5.c
152
bitmap_and(kvm->arch.vgic.gicv5_vm.userspace_ppis,
arch/arm64/kvm/vgic/vgic-v5.c
153
kvm->arch.vgic.gicv5_vm.userspace_ppis,
arch/arm64/kvm/vgic/vgic-v5.c
175
guard(mutex)(&kvm->arch.config_lock);
arch/arm64/kvm/vgic/vgic-v5.c
182
if (test_bit(GICV5_ARCH_PPI_SW_PPI, kvm->arch.vgic.gicv5_vm.vgic_ppi_mask))
arch/arm64/kvm/vgic/vgic-v5.c
188
bitmap_zero(kvm->arch.vgic.gicv5_vm.vgic_ppi_mask, VGIC_V5_NR_PRIVATE_IRQS);
arch/arm64/kvm/vgic/vgic-v5.c
189
bitmap_zero(kvm->arch.vgic.gicv5_vm.vgic_ppi_hmr, VGIC_V5_NR_PRIVATE_IRQS);
arch/arm64/kvm/vgic/vgic-v5.c
200
__assign_bit(i, kvm->arch.vgic.gicv5_vm.vgic_ppi_mask, 1);
arch/arm64/kvm/vgic/vgic-v5.c
201
__assign_bit(i, kvm->arch.vgic.gicv5_vm.vgic_ppi_hmr,
arch/arm64/kvm/vgic/vgic-v5.c
214
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
284
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
309
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
319
for_each_set_bit(i, vcpu->kvm->arch.vgic.gicv5_vm.vgic_ppi_mask, VGIC_V5_NR_PRIVATE_IRQS) {
arch/arm64/kvm/vgic/vgic-v5.c
361
for_each_set_bit(i, vcpu->kvm->arch.vgic.gicv5_vm.vgic_ppi_mask, VGIC_V5_NR_PRIVATE_IRQS) {
arch/arm64/kvm/vgic/vgic-v5.c
387
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
394
for_each_set_bit(i, vcpu->kvm->arch.vgic.gicv5_vm.vgic_ppi_mask,
arch/arm64/kvm/vgic/vgic-v5.c
432
for_each_set_bit(i, vcpu->kvm->arch.vgic.gicv5_vm.vgic_ppi_mask,
arch/arm64/kvm/vgic/vgic-v5.c
461
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
479
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
501
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
510
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
521
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic-v5.c
530
struct vgic_v5_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v5;
arch/arm64/kvm/vgic/vgic.c
1000
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.c
1031
vcpu->arch.vgic_cpu.vgic_v2.used_lrs = count;
arch/arm64/kvm/vgic/vgic.c
1034
vcpu->arch.vgic_cpu.vgic_v3.used_lrs = count;
arch/arm64/kvm/vgic/vgic.c
1057
__vgic_v3_save_state(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/vgic/vgic.c
1110
__vgic_v3_restore_state(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/vgic/vgic.c
1120
scoped_guard(raw_spinlock, &vcpu->arch.vgic_cpu.ap_list_lock)
arch/arm64/kvm/vgic/vgic.c
1171
const struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic.c
1175
__vgic_v3_activate_traps(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/vgic/vgic.c
1199
const struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic.c
1203
__vgic_v3_deactivate_traps(&vcpu->arch.vgic_cpu.vgic_v3);
arch/arm64/kvm/vgic/vgic.c
121
return &vcpu->arch.vgic_cpu.private_irqs[int_num];
arch/arm64/kvm/vgic/vgic.c
1227
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.c
1236
if (!vcpu->kvm->arch.vgic.enabled)
arch/arm64/kvm/vgic/vgic.c
1239
if (vcpu->arch.vgic_cpu.vgic_v3.its_vpe.pending_last)
arch/arm64/kvm/vgic/vgic.c
127
return &vcpu->arch.vgic_cpu.private_irqs[intid];
arch/arm64/kvm/vgic/vgic.c
159
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic.c
181
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic.c
197
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.c
279
!irq->target_vcpu->kvm->arch.vgic.enabled))
arch/arm64/kvm/vgic/vgic.c
364
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.c
404
kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3);
arch/arm64/kvm/vgic/vgic.c
464
raw_spin_lock_irqsave(&vcpu->arch.vgic_cpu.ap_list_lock, flags);
arch/arm64/kvm/vgic/vgic.c
481
raw_spin_unlock_irqrestore(&vcpu->arch.vgic_cpu.ap_list_lock,
arch/arm64/kvm/vgic/vgic.c
494
list_add_tail(&irq->ap_list, &vcpu->arch.vgic_cpu.ap_list_head);
arch/arm64/kvm/vgic/vgic.c
500
atomic_fetch_inc(&vcpu->kvm->arch.vgic.active_spis) == 0);
arch/arm64/kvm/vgic/vgic.c
503
raw_spin_unlock_irqrestore(&vcpu->arch.vgic_cpu.ap_list_lock, flags);
arch/arm64/kvm/vgic/vgic.c
68
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic.c
753
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.c
815
raw_spin_lock(&vcpuA->arch.vgic_cpu.ap_list_lock);
arch/arm64/kvm/vgic/vgic.c
816
raw_spin_lock_nested(&vcpuB->arch.vgic_cpu.ap_list_lock,
arch/arm64/kvm/vgic/vgic.c
830
struct vgic_cpu *new_cpu = &target_vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.c
839
raw_spin_unlock(&vcpuB->arch.vgic_cpu.ap_list_lock);
arch/arm64/kvm/vgic/vgic.c
840
raw_spin_unlock(&vcpuA->arch.vgic_cpu.ap_list_lock);
arch/arm64/kvm/vgic/vgic.c
895
struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.c
95
intid < (kvm->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS)) {
arch/arm64/kvm/vgic/vgic.c
96
intid = array_index_nospec(intid, kvm->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS);
arch/arm64/kvm/vgic/vgic.c
97
return &kvm->arch.vgic.spis[intid - VGIC_NR_PRIVATE_IRQS];
arch/arm64/kvm/vgic/vgic.h
123
return vcpu->kvm->arch.vgic.implementation_rev;
arch/arm64/kvm/vgic/vgic.h
157
struct vgic_dist *dist = &kvm->arch.vgic;
arch/arm64/kvm/vgic/vgic.h
383
struct vgic_cpu *cpu_if = &vcpu->arch.vgic_cpu;
arch/arm64/kvm/vgic/vgic.h
425
struct vgic_dist *d = &kvm->arch.vgic;
arch/loongarch/include/asm/kvm_host.h
275
static inline bool kvm_guest_has_msgint(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
277
return arch->cpucfg[1] & CPUCFG1_MSGINT;
arch/loongarch/include/asm/kvm_host.h
280
static inline bool kvm_guest_has_fpu(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
282
return arch->cpucfg[2] & CPUCFG2_FP;
arch/loongarch/include/asm/kvm_host.h
285
static inline bool kvm_guest_has_lsx(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
287
return arch->cpucfg[2] & CPUCFG2_LSX;
arch/loongarch/include/asm/kvm_host.h
290
static inline bool kvm_guest_has_lasx(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
292
return arch->cpucfg[2] & CPUCFG2_LASX;
arch/loongarch/include/asm/kvm_host.h
295
static inline bool kvm_guest_has_lbt(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
297
return arch->cpucfg[2] & (CPUCFG2_X86BT | CPUCFG2_ARMBT | CPUCFG2_MIPSBT);
arch/loongarch/include/asm/kvm_host.h
300
static inline bool kvm_guest_has_pmu(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
302
return arch->cpucfg[6] & CPUCFG6_PMP;
arch/loongarch/include/asm/kvm_host.h
305
static inline int kvm_get_pmu_num(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
307
return (arch->cpucfg[6] & CPUCFG6_PMNUM) >> CPUCFG6_PMNUM_SHIFT;
arch/loongarch/include/asm/kvm_host.h
311
static inline bool kvm_vm_support(struct kvm_arch *arch, int feature)
arch/loongarch/include/asm/kvm_host.h
313
return !!(arch->kvm_features & BIT_ULL(feature));
arch/loongarch/include/asm/kvm_host.h
330
static inline void update_pc(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
332
arch->pc += 4;
arch/loongarch/include/asm/kvm_host.h
342
static inline bool kvm_is_ifetch_fault(struct kvm_vcpu_arch *arch)
arch/loongarch/include/asm/kvm_host.h
344
return arch->pc == arch->badv;
arch/loongarch/include/asm/kvm_vcpu.h
104
clear_bit(irq, &vcpu->arch.irq_pending);
arch/loongarch/include/asm/kvm_vcpu.h
105
set_bit(irq, &vcpu->arch.irq_clear);
arch/loongarch/include/asm/kvm_vcpu.h
112
if (!vcpu->arch.exception_pending) {
arch/loongarch/include/asm/kvm_vcpu.h
113
set_bit(code, &vcpu->arch.exception_pending);
arch/loongarch/include/asm/kvm_vcpu.h
114
vcpu->arch.esubcode = subcode;
arch/loongarch/include/asm/kvm_vcpu.h
122
return vcpu->arch.gprs[num];
arch/loongarch/include/asm/kvm_vcpu.h
127
vcpu->arch.gprs[num] = val;
arch/loongarch/include/asm/kvm_vcpu.h
137
return vcpu->kvm->arch.pv_features & BIT(feature);
arch/loongarch/include/asm/kvm_vcpu.h
98
set_bit(irq, &vcpu->arch.irq_pending);
arch/loongarch/include/asm/kvm_vcpu.h
99
clear_bit(irq, &vcpu->arch.irq_clear);
arch/loongarch/kernel/asm-offsets.c
295
OFFSET(KVM_VCPU_ARCH, kvm_vcpu, arch);
arch/loongarch/kernel/asm-offsets.c
315
OFFSET(KVM_GPGD, kvm, arch.pgd);
arch/loongarch/kernel/ftrace_dyn.c
43
struct plt_entry *plt = mod->arch.ftrace_trampolines;
arch/loongarch/kernel/machine_kexec.c
249
struct kimage_arch *internal = &image->arch;
arch/loongarch/kernel/machine_kexec.c
48
kimage->arch.efi_boot = fw_arg0;
arch/loongarch/kernel/machine_kexec.c
49
kimage->arch.systable_ptr = fw_arg2;
arch/loongarch/kernel/machine_kexec.c
56
memcpy((void *)KEXEC_CMDLINE_ADDR, (void *)kimage->arch.cmdline_ptr,
arch/loongarch/kernel/machine_kexec.c
57
strlen((char *)kimage->arch.cmdline_ptr) + 1);
arch/loongarch/kernel/machine_kexec.c
58
kimage->arch.cmdline_ptr = (unsigned long)KEXEC_CMDLINE_ADDR;
arch/loongarch/kernel/machine_kexec.c
64
kimage->arch.cmdline_ptr = (unsigned long)cmdline_ptr;
arch/loongarch/kernel/machine_kexec.c
69
if (!kimage->arch.cmdline_ptr) {
arch/loongarch/kernel/machine_kexec_file.c
231
image->arch.cmdline_ptr = (unsigned long)cmdline;
arch/loongarch/kernel/module-sections.c
116
mod->arch.got.shndx = i;
arch/loongarch/kernel/module-sections.c
118
mod->arch.plt.shndx = i;
arch/loongarch/kernel/module-sections.c
120
mod->arch.plt_idx.shndx = i;
arch/loongarch/kernel/module-sections.c
125
if (!mod->arch.got.shndx) {
arch/loongarch/kernel/module-sections.c
129
if (!mod->arch.plt.shndx) {
arch/loongarch/kernel/module-sections.c
133
if (!mod->arch.plt_idx.shndx) {
arch/loongarch/kernel/module-sections.c
15
struct mod_section *got_sec = &mod->arch.got;
arch/loongarch/kernel/module-sections.c
154
got_sec = sechdrs + mod->arch.got.shndx;
arch/loongarch/kernel/module-sections.c
159
mod->arch.got.num_entries = 0;
arch/loongarch/kernel/module-sections.c
160
mod->arch.got.max_entries = num_gots;
arch/loongarch/kernel/module-sections.c
162
plt_sec = sechdrs + mod->arch.plt.shndx;
arch/loongarch/kernel/module-sections.c
167
mod->arch.plt.num_entries = 0;
arch/loongarch/kernel/module-sections.c
168
mod->arch.plt.max_entries = num_plts;
arch/loongarch/kernel/module-sections.c
170
plt_idx_sec = sechdrs + mod->arch.plt_idx.shndx;
arch/loongarch/kernel/module-sections.c
175
mod->arch.plt_idx.num_entries = 0;
arch/loongarch/kernel/module-sections.c
176
mod->arch.plt_idx.max_entries = num_plts;
arch/loongarch/kernel/module-sections.c
42
struct mod_section *plt_sec = &mod->arch.plt;
arch/loongarch/kernel/module-sections.c
43
struct mod_section *plt_idx_sec = &mod->arch.plt_idx;
arch/loongarch/kernel/module.c
602
mod->arch.ftrace_trampolines = ftrace_plts;
arch/loongarch/kernel/unwind_orc.c
265
mod->arch.orc_unwind_ip = orc_ip;
arch/loongarch/kernel/unwind_orc.c
266
mod->arch.orc_unwind = orc;
arch/loongarch/kernel/unwind_orc.c
267
mod->arch.num_orcs = num_entries;
arch/loongarch/kernel/unwind_orc.c
94
if (!mod || !mod->arch.orc_unwind || !mod->arch.orc_unwind_ip)
arch/loongarch/kernel/unwind_orc.c
97
return __orc_find(mod->arch.orc_unwind_ip, mod->arch.orc_unwind, mod->arch.num_orcs, ip);
arch/loongarch/kvm/exit.c
100
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/exit.c
108
pr_warn_once("Unsupported csrxchg 0x%x with pc %lx\n", csrid, vcpu->arch.pc);
arch/loongarch/kvm/exit.c
128
if (csrid >= LOONGARCH_CSR_PERFCTRL0 && csrid <= vcpu->arch.max_pmu_csrid) {
arch/loongarch/kvm/exit.c
129
if (kvm_guest_has_pmu(&vcpu->arch)) {
arch/loongarch/kvm/exit.c
130
vcpu->arch.pc -= 4;
arch/loongarch/kvm/exit.c
140
vcpu->arch.gprs[rd] = val;
arch/loongarch/kvm/exit.c
143
val = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
145
vcpu->arch.gprs[rd] = val;
arch/loongarch/kvm/exit.c
148
val = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
149
csr_mask = vcpu->arch.gprs[rj];
arch/loongarch/kvm/exit.c
151
vcpu->arch.gprs[rd] = val;
arch/loongarch/kvm/exit.c
169
addr = vcpu->arch.gprs[rj];
arch/loongarch/kvm/exit.c
172
val = &vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
221
vcpu->arch.io_gpr = rd; /* Set register id for iocsr read completion */
arch/loongarch/kvm/exit.c
240
unsigned long *gpr = &vcpu->arch.gprs[vcpu->arch.io_gpr];
arch/loongarch/kvm/exit.c
257
run->iocsr_io.len, vcpu->arch.badv);
arch/loongarch/kvm/exit.c
284
inst.word = vcpu->arch.badi;
arch/loongarch/kvm/exit.c
285
curr_pc = vcpu->arch.pc;
arch/loongarch/kvm/exit.c
286
update_pc(&vcpu->arch);
arch/loongarch/kvm/exit.c
334
vcpu->arch.pc = curr_pc;
arch/loongarch/kvm/exit.c
35
index = vcpu->arch.gprs[rj];
arch/loongarch/kvm/exit.c
375
run->mmio.phys_addr = vcpu->arch.badv;
arch/loongarch/kvm/exit.c
46
vcpu->arch.gprs[rd] = vcpu->arch.cpucfg[index];
arch/loongarch/kvm/exit.c
471
vcpu->arch.io_gpr = rd; /* Set for kvm_complete_mmio_read() use */
arch/loongarch/kvm/exit.c
479
ret = kvm_io_bus_read(vcpu, KVM_MMIO_BUS, vcpu->arch.badv,
arch/loongarch/kvm/exit.c
484
update_pc(&vcpu->arch);
arch/loongarch/kvm/exit.c
495
inst.word, vcpu->arch.pc, vcpu->arch.badv);
arch/loongarch/kvm/exit.c
50
vcpu->arch.gprs[rd] = *(unsigned int *)KVM_SIGNATURE;
arch/loongarch/kvm/exit.c
505
unsigned long *gpr = &vcpu->arch.gprs[vcpu->arch.io_gpr];
arch/loongarch/kvm/exit.c
508
update_pc(&vcpu->arch);
arch/loongarch/kvm/exit.c
53
ret = vcpu->kvm->arch.pv_features & LOONGARCH_PV_FEAT_MASK;
arch/loongarch/kvm/exit.c
533
run->mmio.len, vcpu->arch.badv);
arch/loongarch/kvm/exit.c
54
vcpu->arch.gprs[rd] = ret;
arch/loongarch/kvm/exit.c
556
curr_pc = vcpu->arch.pc;
arch/loongarch/kvm/exit.c
557
update_pc(&vcpu->arch);
arch/loongarch/kvm/exit.c
560
run->mmio.phys_addr = vcpu->arch.badv;
arch/loongarch/kvm/exit.c
57
vcpu->arch.gprs[rd] = 0;
arch/loongarch/kvm/exit.c
570
*(unsigned int *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
574
*(unsigned long *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
584
rd_val = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
615
*(unsigned char *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
619
*(unsigned short *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
623
*(unsigned int *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
627
*(unsigned long *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
647
ret = kvm_io_bus_write(vcpu, KVM_MMIO_BUS, vcpu->arch.badv, run->mmio.len, data);
arch/loongarch/kvm/exit.c
658
vcpu->arch.pc = curr_pc;
arch/loongarch/kvm/exit.c
660
inst.word, vcpu->arch.pc, vcpu->arch.badv);
arch/loongarch/kvm/exit.c
673
unsigned long badv = vcpu->arch.badv;
arch/loongarch/kvm/exit.c
676
if (unlikely(badv >= vcpu->kvm->arch.gpa_size)) {
arch/loongarch/kvm/exit.c
68
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/exit.c
684
inst.word = vcpu->arch.badi;
arch/loongarch/kvm/exit.c
689
if (kvm_is_ifetch_fault(&vcpu->arch)) {
arch/loongarch/kvm/exit.c
723
update_pc(&vcpu->arch);
arch/loongarch/kvm/exit.c
741
if (!kvm_guest_has_fpu(&vcpu->arch)) {
arch/loongarch/kvm/exit.c
751
if (WARN_ON(vcpu->arch.aux_inuse & KVM_LARCH_FPU)) {
arch/loongarch/kvm/exit.c
757
vcpu->arch.aux_ldtype = KVM_LARCH_FPU;
arch/loongarch/kvm/exit.c
77
pr_warn_once("Unsupported csrrd 0x%x with pc %lx\n", csrid, vcpu->arch.pc);
arch/loongarch/kvm/exit.c
774
vcpu->arch.st.guest_addr = data;
arch/loongarch/kvm/exit.c
778
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/loongarch/kvm/exit.c
796
if (!kvm_guest_has_lsx(&vcpu->arch))
arch/loongarch/kvm/exit.c
799
vcpu->arch.aux_ldtype = KVM_LARCH_LSX;
arch/loongarch/kvm/exit.c
816
if (!kvm_guest_has_lasx(&vcpu->arch))
arch/loongarch/kvm/exit.c
819
vcpu->arch.aux_ldtype = KVM_LARCH_LASX;
arch/loongarch/kvm/exit.c
828
if (!kvm_guest_has_lbt(&vcpu->arch))
arch/loongarch/kvm/exit.c
831
vcpu->arch.aux_ldtype = KVM_LARCH_LBT;
arch/loongarch/kvm/exit.c
85
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/exit.c
891
inst.word = vcpu->arch.badi;
arch/loongarch/kvm/exit.c
91
pr_warn_once("Unsupported csrwr 0x%x with pc %lx\n", csrid, vcpu->arch.pc);
arch/loongarch/kvm/exit.c
937
update_pc(&vcpu->arch);
arch/loongarch/kvm/exit.c
951
inst = vcpu->arch.badi;
arch/loongarch/kvm/exit.c
952
badv = vcpu->arch.badv;
arch/loongarch/kvm/exit.c
954
ecode, vcpu->arch.pc, inst, badv, read_gcsr_estat());
arch/loongarch/kvm/intc/dmsintc.c
147
if (kvm->arch.dmsintc) {
arch/loongarch/kvm/intc/dmsintc.c
15
struct dmsintc_state *ds = &vcpu->arch.dmsintc_state;
arch/loongarch/kvm/intc/dmsintc.c
157
kvm->arch.dmsintc = s;
arch/loongarch/kvm/intc/dmsintc.c
165
if (!dev || !dev->kvm || !dev->kvm->arch.dmsintc)
arch/loongarch/kvm/intc/dmsintc.c
168
kfree(dev->kvm->arch.dmsintc);
arch/loongarch/kvm/intc/dmsintc.c
51
struct dmsintc_state *ds = &vcpu->arch.dmsintc_state;
arch/loongarch/kvm/intc/dmsintc.c
74
cpu = (addr >> AVEC_CPU_SHIFT) & kvm->arch.dmsintc->cpu_mask;
arch/loongarch/kvm/intc/dmsintc.c
90
struct loongarch_dmsintc *s = dev->kvm->arch.dmsintc;
arch/loongarch/kvm/intc/eiointc.c
165
struct loongarch_eiointc *eiointc = vcpu->kvm->arch.eiointc;
arch/loongarch/kvm/intc/eiointc.c
297
struct loongarch_eiointc *eiointc = vcpu->kvm->arch.eiointc;
arch/loongarch/kvm/intc/eiointc.c
345
struct loongarch_eiointc *eiointc = vcpu->kvm->arch.eiointc;
arch/loongarch/kvm/intc/eiointc.c
375
struct loongarch_eiointc *eiointc = vcpu->kvm->arch.eiointc;
arch/loongarch/kvm/intc/eiointc.c
417
struct loongarch_eiointc *s = dev->kvm->arch.eiointc;
arch/loongarch/kvm/intc/eiointc.c
468
s = dev->kvm->arch.eiointc;
arch/loongarch/kvm/intc/eiointc.c
528
s = dev->kvm->arch.eiointc;
arch/loongarch/kvm/intc/eiointc.c
622
if (kvm->arch.eiointc)
arch/loongarch/kvm/intc/eiointc.c
655
kvm->arch.eiointc = s;
arch/loongarch/kvm/intc/eiointc.c
665
if (!dev || !dev->kvm || !dev->kvm->arch.eiointc)
arch/loongarch/kvm/intc/eiointc.c
669
eiointc = kvm->arch.eiointc;
arch/loongarch/kvm/intc/ipi.c
101
spin_unlock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
15
spin_lock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
16
status = vcpu->arch.ipi_state.status;
arch/loongarch/kvm/intc/ipi.c
17
vcpu->arch.ipi_state.status |= data;
arch/loongarch/kvm/intc/ipi.c
18
spin_unlock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
194
spin_lock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
195
res = vcpu->arch.ipi_state.status;
arch/loongarch/kvm/intc/ipi.c
196
spin_unlock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
199
spin_lock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
200
res = vcpu->arch.ipi_state.en;
arch/loongarch/kvm/intc/ipi.c
201
spin_unlock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
237
spin_lock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
238
vcpu->arch.ipi_state.en = data;
arch/loongarch/kvm/intc/ipi.c
239
spin_unlock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
315
p = &vcpu->arch.ipi_state.status;
arch/loongarch/kvm/intc/ipi.c
318
p = &vcpu->arch.ipi_state.en;
arch/loongarch/kvm/intc/ipi.c
321
p = &vcpu->arch.ipi_state.set;
arch/loongarch/kvm/intc/ipi.c
324
p = &vcpu->arch.ipi_state.clear;
arch/loongarch/kvm/intc/ipi.c
327
p = &vcpu->arch.ipi_state.buf[0];
arch/loongarch/kvm/intc/ipi.c
331
p = &vcpu->arch.ipi_state.buf[1];
arch/loongarch/kvm/intc/ipi.c
335
p = &vcpu->arch.ipi_state.buf[2];
arch/loongarch/kvm/intc/ipi.c
339
p = &vcpu->arch.ipi_state.buf[3];
arch/loongarch/kvm/intc/ipi.c
407
if (kvm->arch.ipi) {
arch/loongarch/kvm/intc/ipi.c
432
kvm->arch.ipi = s;
arch/loongarch/kvm/intc/ipi.c
445
if (!dev || !dev->kvm || !dev->kvm->arch.ipi)
arch/loongarch/kvm/intc/ipi.c
449
ipi = kvm->arch.ipi;
arch/loongarch/kvm/intc/ipi.c
45
spin_lock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
46
vcpu->arch.ipi_state.status &= ~data;
arch/loongarch/kvm/intc/ipi.c
47
status = vcpu->arch.ipi_state.status;
arch/loongarch/kvm/intc/ipi.c
48
spin_unlock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
59
spin_lock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
60
data = *(ulong *)((void *)vcpu->arch.ipi_state.buf + (offset - 0x20));
arch/loongarch/kvm/intc/ipi.c
61
spin_unlock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
82
spin_lock(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/intc/ipi.c
83
pbuf = (void *)vcpu->arch.ipi_state.buf + (offset - 0x20);
arch/loongarch/kvm/intc/pch_pic.c
151
struct loongarch_pch_pic *s = vcpu->kvm->arch.pch_pic;
arch/loongarch/kvm/intc/pch_pic.c
25
eiointc_set_irq(s->kvm->arch.eiointc, irq, level);
arch/loongarch/kvm/intc/pch_pic.c
250
struct loongarch_pch_pic *s = vcpu->kvm->arch.pch_pic;
arch/loongarch/kvm/intc/pch_pic.c
279
struct loongarch_pch_pic *s = dev->kvm->arch.pch_pic;
arch/loongarch/kvm/intc/pch_pic.c
304
s = dev->kvm->arch.pch_pic;
arch/loongarch/kvm/intc/pch_pic.c
31
eiointc_set_irq(s->kvm->arch.eiointc, irq, level);
arch/loongarch/kvm/intc/pch_pic.c
392
if (!dev->kvm->arch.pch_pic) {
arch/loongarch/kvm/intc/pch_pic.c
439
if (kvm->arch.pch_pic)
arch/loongarch/kvm/intc/pch_pic.c
468
kvm->arch.pch_pic = s;
arch/loongarch/kvm/intc/pch_pic.c
478
if (!dev || !dev->kvm || !dev->kvm->arch.pch_pic)
arch/loongarch/kvm/intc/pch_pic.c
482
s = kvm->arch.pch_pic;
arch/loongarch/kvm/intc/pch_pic.c
75
if (cpu_has_msgint && kvm->arch.dmsintc &&
arch/loongarch/kvm/intc/pch_pic.c
76
msg_addr >= kvm->arch.dmsintc->msg_addr_base &&
arch/loongarch/kvm/intc/pch_pic.c
77
msg_addr < (kvm->arch.dmsintc->msg_addr_base + kvm->arch.dmsintc->msg_addr_size)) {
arch/loongarch/kvm/intc/pch_pic.c
81
eiointc_set_irq(kvm->arch.eiointc, e->msi.data, level);
arch/loongarch/kvm/interrupt.c
108
return test_bit(INT_TI, &vcpu->arch.irq_pending);
arch/loongarch/kvm/interrupt.c
127
val = vcpu->arch.pc;
arch/loongarch/kvm/interrupt.c
129
val = vcpu->arch.badv;
arch/loongarch/kvm/interrupt.c
134
kvm_write_hw_gcsr(LOONGARCH_CSR_BADI, vcpu->arch.badi);
arch/loongarch/kvm/interrupt.c
146
kvm_write_hw_gcsr(LOONGARCH_CSR_ERA, vcpu->arch.pc);
arch/loongarch/kvm/interrupt.c
165
vcpu->arch.pc = val + code * vec_size;
arch/loongarch/kvm/interrupt.c
171
unsigned long *pending = &vcpu->arch.exception_pending;
arch/loongarch/kvm/interrupt.c
175
_kvm_deliver_exception(vcpu, code, vcpu->arch.esubcode);
arch/loongarch/kvm/interrupt.c
177
vcpu->arch.esubcode = 0;
arch/loongarch/kvm/interrupt.c
32
clear_bit(priority, &vcpu->arch.irq_pending);
arch/loongarch/kvm/interrupt.c
36
if (kvm_guest_has_msgint(&vcpu->arch) && (priority == INT_AVEC)) {
arch/loongarch/kvm/interrupt.c
65
clear_bit(priority, &vcpu->arch.irq_clear);
arch/loongarch/kvm/interrupt.c
69
if (kvm_guest_has_msgint(&vcpu->arch) && (priority == INT_AVEC)) {
arch/loongarch/kvm/interrupt.c
96
unsigned long *pending = &vcpu->arch.irq_pending;
arch/loongarch/kvm/interrupt.c
97
unsigned long *pending_clr = &vcpu->arch.irq_clear;
arch/loongarch/kvm/irqfd.c
14
pch_pic_set_irq(kvm->arch.pch_pic, e->irqchip.pin, level);
arch/loongarch/kvm/irqfd.c
77
pch_pic_set_irq(kvm->arch.pch_pic, e->irqchip.pin, level);
arch/loongarch/kvm/main.c
210
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/main.c
224
vcpu->arch.vpid = vpid;
arch/loongarch/kvm/main.c
239
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/main.c
249
ver = vcpu->arch.vpid & ~vpid_mask;
arch/loongarch/kvm/main.c
253
trace_kvm_vpid_change(vcpu, vcpu->arch.vpid);
arch/loongarch/kvm/main.c
276
vpid = (vcpu->arch.vpid & vpid_mask) << CSR_GSTAT_GID_SHIFT;
arch/loongarch/kvm/main.c
283
kvm->arch.vmcs = vmcs;
arch/loongarch/kvm/mmu.c
157
child = kvm->arch.pgd;
arch/loongarch/kvm/mmu.c
18
return slot->arch.flags & KVM_MEM_HUGEPAGE_CAPABLE;
arch/loongarch/kvm/mmu.c
23
return slot->arch.flags & KVM_MEM_HUGEPAGE_INCAPABLE;
arch/loongarch/kvm/mmu.c
28
ctx->level = kvm->arch.root_level;
arch/loongarch/kvm/mmu.c
30
ctx->invalid_ptes = kvm->arch.invalid_ptes;
arch/loongarch/kvm/mmu.c
305
ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
309
ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
31
ctx->pte_shifts = kvm->arch.pte_shifts;
arch/loongarch/kvm/mmu.c
348
return kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, end_gfn << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
376
kvm_ptw_top(kvm->arch.pgd, start << PAGE_SHIFT, end << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
392
if ((new->base_gfn + new->npages) > (kvm->arch.gpa_size >> PAGE_SHIFT))
arch/loongarch/kvm/mmu.c
395
new->arch.flags = 0;
arch/loongarch/kvm/mmu.c
401
new->arch.flags |= KVM_MEM_HUGEPAGE_CAPABLE;
arch/loongarch/kvm/mmu.c
430
new->arch.flags |= KVM_MEM_HUGEPAGE_INCAPABLE;
arch/loongarch/kvm/mmu.c
435
new->arch.flags |= KVM_MEM_HUGEPAGE_INCAPABLE;
arch/loongarch/kvm/mmu.c
489
kvm_flush_range(kvm, 0, kvm->arch.gpa_size >> PAGE_SHIFT, 0);
arch/loongarch/kvm/mmu.c
510
return kvm_ptw_top(kvm->arch.pgd, range->start << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
522
return kvm_ptw_top(kvm->arch.pgd, range->start << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
735
memcache = &vcpu->arch.mmu_page_cache;
arch/loongarch/kvm/mmu.c
782
struct kvm_mmu_memory_cache *memcache = &vcpu->arch.mmu_page_cache;
arch/loongarch/kvm/mmu.c
932
vcpu->arch.flush_gpa = gpa;
arch/loongarch/kvm/timer.c
116
expire = vcpu->arch.expire;
arch/loongarch/kvm/timer.c
144
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/timer.c
162
vcpu->arch.expire = expire;
arch/loongarch/kvm/timer.c
170
hrtimer_start(&vcpu->arch.swtimer, expire, HRTIMER_MODE_ABS_PINNED_HARD);
arch/loongarch/kvm/timer.c
180
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/timer.c
19
return div_u64(delta * vcpu->arch.timer_mhz, MNSEC_PER_SEC);
arch/loongarch/kvm/timer.c
24
return div_u64(tick * MNSEC_PER_SEC, vcpu->arch.timer_mhz);
arch/loongarch/kvm/timer.c
32
vcpu = container_of(timer, struct kvm_vcpu, arch.swtimer);
arch/loongarch/kvm/timer.c
44
vcpu->arch.timer_mhz = timer_hz >> 20;
arch/loongarch/kvm/timer.c
47
kvm_write_sw_gcsr(vcpu->arch.csr, LOONGARCH_CSR_TVAL, 0);
arch/loongarch/kvm/timer.c
58
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/timer.c
80
hrtimer_cancel(&vcpu->arch.swtimer);
arch/loongarch/kvm/trace.h
151
__entry->pc = vcpu->arch.pc;
arch/loongarch/kvm/trace.h
216
#define TRACE_INCLUDE_PATH ../../arch/loongarch/kvm
arch/loongarch/kvm/trace.h
28
__entry->pc = vcpu->arch.pc;
arch/loongarch/kvm/trace.h
70
__entry->pc = vcpu->arch.pc;
arch/loongarch/kvm/vcpu.c
1018
for (i = 0; i < ARRAY_SIZE(vcpu->arch.gprs); i++)
arch/loongarch/kvm/vcpu.c
1019
regs->gpr[i] = vcpu->arch.gprs[i];
arch/loongarch/kvm/vcpu.c
1021
regs->pc = vcpu->arch.pc;
arch/loongarch/kvm/vcpu.c
1030
for (i = 1; i < ARRAY_SIZE(vcpu->arch.gprs); i++)
arch/loongarch/kvm/vcpu.c
1031
vcpu->arch.gprs[i] = regs->gpr[i];
arch/loongarch/kvm/vcpu.c
1033
vcpu->arch.gprs[0] = 0; /* zero is special, and cannot be set. */
arch/loongarch/kvm/vcpu.c
1034
vcpu->arch.pc = regs->pc;
arch/loongarch/kvm/vcpu.c
104
if (!kvm_guest_has_pmu(&vcpu->arch))
arch/loongarch/kvm/vcpu.c
1105
val = vcpu->kvm->arch.pv_features & LOONGARCH_PV_FEAT_MASK;
arch/loongarch/kvm/vcpu.c
111
val |= (kvm_get_pmu_num(&vcpu->arch) + 1) << CSR_GCFG_GPERF_SHIFT;
arch/loongarch/kvm/vcpu.c
1126
gpa = vcpu->arch.st.guest_addr;
arch/loongarch/kvm/vcpu.c
1169
if ((kvm->arch.pv_features & LOONGARCH_PV_FEAT_UPDATED)
arch/loongarch/kvm/vcpu.c
1170
&& ((kvm->arch.pv_features & valid) != val))
arch/loongarch/kvm/vcpu.c
1172
kvm->arch.pv_features = val | LOONGARCH_PV_FEAT_UPDATED;
arch/loongarch/kvm/vcpu.c
1197
vcpu->arch.st.guest_addr = gpa;
arch/loongarch/kvm/vcpu.c
1208
vcpu->arch.st.guest_addr = gpa;
arch/loongarch/kvm/vcpu.c
1209
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/loongarch/kvm/vcpu.c
122
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
124
if (!(vcpu->arch.aux_inuse & KVM_LARCH_PMU))
arch/loongarch/kvm/vcpu.c
1265
vcpu->arch.aux_inuse &= ~KVM_LARCH_HWCSR_USABLE;
arch/loongarch/kvm/vcpu.c
1312
fpu->fcc = vcpu->arch.fpu.fcc;
arch/loongarch/kvm/vcpu.c
1313
fpu->fcsr = vcpu->arch.fpu.fcsr;
arch/loongarch/kvm/vcpu.c
1315
memcpy(&fpu->fpr[i], &vcpu->arch.fpu.fpr[i], FPU_REG_WIDTH / 64);
arch/loongarch/kvm/vcpu.c
1324
vcpu->arch.fpu.fcc = fpu->fcc;
arch/loongarch/kvm/vcpu.c
1325
vcpu->arch.fpu.fcsr = fpu->fcsr;
arch/loongarch/kvm/vcpu.c
1327
memcpy(&vcpu->arch.fpu.fpr[i], &fpu->fpr[i], FPU_REG_WIDTH / 64);
arch/loongarch/kvm/vcpu.c
1335
if (!(vcpu->arch.aux_inuse & KVM_LARCH_LBT)) {
arch/loongarch/kvm/vcpu.c
1337
_restore_lbt(&vcpu->arch.lbt);
arch/loongarch/kvm/vcpu.c
1338
vcpu->arch.aux_inuse |= KVM_LARCH_LBT;
arch/loongarch/kvm/vcpu.c
1347
if (vcpu->arch.aux_inuse & KVM_LARCH_LBT) {
arch/loongarch/kvm/vcpu.c
1348
_save_lbt(&vcpu->arch.lbt);
arch/loongarch/kvm/vcpu.c
1350
vcpu->arch.aux_inuse &= ~KVM_LARCH_LBT;
arch/loongarch/kvm/vcpu.c
1367
if (vcpu->arch.aux_inuse & KVM_LARCH_FPU) {
arch/loongarch/kvm/vcpu.c
1368
if (vcpu->arch.aux_inuse & KVM_LARCH_LBT)
arch/loongarch/kvm/vcpu.c
1386
kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr);
arch/loongarch/kvm/vcpu.c
1389
kvm_restore_fpu(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1390
vcpu->arch.aux_inuse |= KVM_LARCH_FPU;
arch/loongarch/kvm/vcpu.c
1399
kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr);
arch/loongarch/kvm/vcpu.c
1401
switch (vcpu->arch.aux_inuse & KVM_LARCH_FPU) {
arch/loongarch/kvm/vcpu.c
1407
_restore_lsx_upper(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1413
kvm_restore_lsx(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1418
vcpu->arch.aux_inuse |= KVM_LARCH_LSX | KVM_LARCH_FPU;
arch/loongarch/kvm/vcpu.c
1428
kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr);
arch/loongarch/kvm/vcpu.c
1430
switch (vcpu->arch.aux_inuse & (KVM_LARCH_FPU | KVM_LARCH_LSX)) {
arch/loongarch/kvm/vcpu.c
1434
_restore_lasx_upper(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1438
_restore_lsx_upper(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1439
_restore_lasx_upper(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1443
kvm_restore_lasx(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1448
vcpu->arch.aux_inuse |= KVM_LARCH_LASX | KVM_LARCH_LSX | KVM_LARCH_FPU;
arch/loongarch/kvm/vcpu.c
145
vcpu->arch.aux_inuse &= ~KVM_LARCH_PMU;
arch/loongarch/kvm/vcpu.c
1460
if (vcpu->arch.aux_inuse & KVM_LARCH_LASX) {
arch/loongarch/kvm/vcpu.c
1461
kvm_save_lasx(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1462
vcpu->arch.aux_inuse &= ~(KVM_LARCH_LSX | KVM_LARCH_FPU | KVM_LARCH_LASX);
arch/loongarch/kvm/vcpu.c
1467
} else if (vcpu->arch.aux_inuse & KVM_LARCH_LSX) {
arch/loongarch/kvm/vcpu.c
1468
kvm_save_lsx(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1469
vcpu->arch.aux_inuse &= ~(KVM_LARCH_LSX | KVM_LARCH_FPU);
arch/loongarch/kvm/vcpu.c
1474
} else if (vcpu->arch.aux_inuse & KVM_LARCH_FPU) {
arch/loongarch/kvm/vcpu.c
1475
kvm_save_fpu(&vcpu->arch.fpu);
arch/loongarch/kvm/vcpu.c
1476
vcpu->arch.aux_inuse &= ~KVM_LARCH_FPU;
arch/loongarch/kvm/vcpu.c
1535
vcpu->arch.vpid = 0;
arch/loongarch/kvm/vcpu.c
1536
vcpu->arch.flush_gpa = INVALID_GPA;
arch/loongarch/kvm/vcpu.c
1538
hrtimer_setup(&vcpu->arch.swtimer, kvm_swtimer_wakeup, CLOCK_MONOTONIC,
arch/loongarch/kvm/vcpu.c
1542
vcpu->arch.kvm_pgd = __pa(vcpu->kvm->arch.pgd);
arch/loongarch/kvm/vcpu.c
1548
vcpu->arch.host_pgd = (unsigned long)vcpu->kvm->mm->pgd;
arch/loongarch/kvm/vcpu.c
1550
vcpu->arch.handle_exit = kvm_handle_exit;
arch/loongarch/kvm/vcpu.c
1551
vcpu->arch.guest_eentry = (unsigned long)kvm_loongarch_ops->exc_entry;
arch/loongarch/kvm/vcpu.c
1552
vcpu->arch.csr = kzalloc_obj(struct loongarch_csrs);
arch/loongarch/kvm/vcpu.c
1553
if (!vcpu->arch.csr)
arch/loongarch/kvm/vcpu.c
1560
vcpu->arch.host_ecfg = (read_csr_ecfg() & CSR_ECFG_VS);
arch/loongarch/kvm/vcpu.c
1563
vcpu->arch.last_sched_cpu = -1;
arch/loongarch/kvm/vcpu.c
1566
spin_lock_init(&vcpu->arch.ipi_state.lock);
arch/loongarch/kvm/vcpu.c
1575
csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
1597
hrtimer_cancel(&vcpu->arch.swtimer);
arch/loongarch/kvm/vcpu.c
1598
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
arch/loongarch/kvm/vcpu.c
1600
kfree(vcpu->arch.csr);
arch/loongarch/kvm/vcpu.c
1607
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/vcpu.c
161
ghc = &vcpu->arch.st.cache;
arch/loongarch/kvm/vcpu.c
1617
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
162
gpa = vcpu->arch.st.guest_addr;
arch/loongarch/kvm/vcpu.c
1623
migrated = (vcpu->arch.last_sched_cpu != cpu);
arch/loongarch/kvm/vcpu.c
1629
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/vcpu.c
1632
vcpu->arch.aux_inuse &= ~KVM_LARCH_HWCSR_USABLE;
arch/loongarch/kvm/vcpu.c
1633
vcpu->arch.host_eentry = csr_read64(LOONGARCH_CSR_EENTRY);
arch/loongarch/kvm/vcpu.c
1641
if (vcpu->arch.aux_inuse & KVM_LARCH_HWCSR_USABLE)
arch/loongarch/kvm/vcpu.c
1644
write_csr_gcntc((ulong)vcpu->kvm->arch.time_offset);
arch/loongarch/kvm/vcpu.c
1692
if (kvm_guest_has_msgint(&vcpu->arch)) {
arch/loongarch/kvm/vcpu.c
1712
vcpu->arch.aux_inuse |= KVM_LARCH_HWCSR_USABLE;
arch/loongarch/kvm/vcpu.c
1729
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
1739
if (vcpu->arch.aux_inuse & KVM_LARCH_SWCSR_LATEST)
arch/loongarch/kvm/vcpu.c
178
vcpu->arch.st.preempted = 0;
arch/loongarch/kvm/vcpu.c
1790
if (kvm_guest_has_msgint(&vcpu->arch)) {
arch/loongarch/kvm/vcpu.c
1799
vcpu->arch.aux_inuse |= KVM_LARCH_SWCSR_LATEST;
arch/loongarch/kvm/vcpu.c
1816
gpa = vcpu->arch.st.guest_addr;
arch/loongarch/kvm/vcpu.c
1821
if (vcpu->arch.st.preempted)
arch/loongarch/kvm/vcpu.c
1829
ghc = &vcpu->arch.st.cache;
arch/loongarch/kvm/vcpu.c
1840
vcpu->arch.st.preempted = KVM_VCPU_PREEMPTED;
arch/loongarch/kvm/vcpu.c
1862
vcpu->arch.last_sched_cpu = cpu;
arch/loongarch/kvm/vcpu.c
190
steal += current->sched_info.run_delay - vcpu->arch.st.last_steal;
arch/loongarch/kvm/vcpu.c
191
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/loongarch/kvm/vcpu.c
213
vcpu->arch.vpid = 0; /* Drop vpid for this vCPU */
arch/loongarch/kvm/vcpu.c
233
vcpu->arch.aux_inuse |= KVM_LARCH_PMU;
arch/loongarch/kvm/vcpu.c
237
if (vcpu->arch.flush_gpa != INVALID_GPA) {
arch/loongarch/kvm/vcpu.c
238
kvm_flush_tlb_gpa(vcpu, vcpu->arch.flush_gpa);
arch/loongarch/kvm/vcpu.c
239
vcpu->arch.flush_gpa = INVALID_GPA;
arch/loongarch/kvm/vcpu.c
243
switch (vcpu->arch.aux_ldtype) {
arch/loongarch/kvm/vcpu.c
260
vcpu->arch.aux_ldtype = 0;
arch/loongarch/kvm/vcpu.c
324
vcpu->arch.aux_inuse &= ~KVM_LARCH_SWCSR_LATEST;
arch/loongarch/kvm/vcpu.c
327
if (vcpu->arch.aux_inuse & KVM_LARCH_PMU) {
arch/loongarch/kvm/vcpu.c
347
unsigned long estat = vcpu->arch.host_estat;
arch/loongarch/kvm/vcpu.c
387
return !!(vcpu->arch.irq_pending) &&
arch/loongarch/kvm/vcpu.c
388
vcpu->arch.mp_state.mp_state == KVM_MP_STATE_RUNNABLE;
arch/loongarch/kvm/vcpu.c
410
return vcpu->arch.pc;
arch/loongarch/kvm/vcpu.c
420
return (vcpu && !(vcpu->arch.aux_inuse & KVM_LARCH_PMU));
arch/loongarch/kvm/vcpu.c
45
context = this_cpu_ptr(vcpu->kvm->arch.vmcs);
arch/loongarch/kvm/vcpu.c
458
kvm_debug("\tPC = 0x%08lx\n", vcpu->arch.pc);
arch/loongarch/kvm/vcpu.c
459
kvm_debug("\tExceptions: %08lx\n", vcpu->arch.irq_pending);
arch/loongarch/kvm/vcpu.c
463
vcpu->arch.gprs[i], vcpu->arch.gprs[i + 1],
arch/loongarch/kvm/vcpu.c
464
vcpu->arch.gprs[i + 2], vcpu->arch.gprs[i + 3]);
arch/loongarch/kvm/vcpu.c
479
*mp_state = vcpu->arch.mp_state;
arch/loongarch/kvm/vcpu.c
491
vcpu->arch.mp_state = *mp_state;
arch/loongarch/kvm/vcpu.c
518
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
523
map = vcpu->kvm->arch.phyid_map;
arch/loongarch/kvm/vcpu.c
526
spin_lock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
530
spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
538
spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
545
spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
553
spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
560
spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
569
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
571
map = vcpu->kvm->arch.phyid_map;
arch/loongarch/kvm/vcpu.c
577
spin_lock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
583
spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vcpu.c
596
map = kvm->arch.phyid_map;
arch/loongarch/kvm/vcpu.c
60
context = this_cpu_ptr(vcpu->kvm->arch.vmcs);
arch/loongarch/kvm/vcpu.c
606
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
619
vcpu->arch.aux_inuse &= ~KVM_LARCH_SWCSR_LATEST;
arch/loongarch/kvm/vcpu.c
641
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
74
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
829
*v = vcpu->arch.cpucfg[id];
arch/loongarch/kvm/vcpu.c
834
if (!kvm_guest_has_lbt(&vcpu->arch))
arch/loongarch/kvm/vcpu.c
839
*v = vcpu->arch.lbt.scr0;
arch/loongarch/kvm/vcpu.c
842
*v = vcpu->arch.lbt.scr1;
arch/loongarch/kvm/vcpu.c
845
*v = vcpu->arch.lbt.scr2;
arch/loongarch/kvm/vcpu.c
848
*v = vcpu->arch.lbt.scr3;
arch/loongarch/kvm/vcpu.c
851
*v = vcpu->arch.lbt.eflags;
arch/loongarch/kvm/vcpu.c
854
*v = vcpu->arch.fpu.ftop;
arch/loongarch/kvm/vcpu.c
864
*v = get_cycles() + vcpu->kvm->arch.time_offset;
arch/loongarch/kvm/vcpu.c
88
struct loongarch_csrs *csr = vcpu->arch.csr;
arch/loongarch/kvm/vcpu.c
918
vcpu->arch.cpucfg[id] = (u32)v;
arch/loongarch/kvm/vcpu.c
920
vcpu->arch.max_pmu_csrid =
arch/loongarch/kvm/vcpu.c
921
LOONGARCH_CSR_PERFCTRL0 + 2 * kvm_get_pmu_num(&vcpu->arch) + 1;
arch/loongarch/kvm/vcpu.c
924
if (!kvm_guest_has_lbt(&vcpu->arch))
arch/loongarch/kvm/vcpu.c
929
vcpu->arch.lbt.scr0 = v;
arch/loongarch/kvm/vcpu.c
932
vcpu->arch.lbt.scr1 = v;
arch/loongarch/kvm/vcpu.c
935
vcpu->arch.lbt.scr2 = v;
arch/loongarch/kvm/vcpu.c
938
vcpu->arch.lbt.scr3 = v;
arch/loongarch/kvm/vcpu.c
941
vcpu->arch.lbt.eflags = v;
arch/loongarch/kvm/vcpu.c
944
vcpu->arch.fpu.ftop = v;
arch/loongarch/kvm/vcpu.c
959
vcpu->kvm->arch.time_offset = (signed long)(v - get_cycles());
arch/loongarch/kvm/vcpu.c
962
vcpu->arch.st.guest_addr = 0;
arch/loongarch/kvm/vcpu.c
963
memset(&vcpu->arch.irq_pending, 0, sizeof(vcpu->arch.irq_pending));
arch/loongarch/kvm/vcpu.c
964
memset(&vcpu->arch.irq_clear, 0, sizeof(vcpu->arch.irq_clear));
arch/loongarch/kvm/vcpu.c
970
kvm_write_sw_gcsr(vcpu->arch.csr, LOONGARCH_CSR_GINTC, 0);
arch/loongarch/kvm/vcpu.c
971
kvm_write_sw_gcsr(vcpu->arch.csr, LOONGARCH_CSR_ESTAT, 0);
arch/loongarch/kvm/vm.c
106
free_page((unsigned long)kvm->arch.pgd);
arch/loongarch/kvm/vm.c
107
kvm->arch.pgd = NULL;
arch/loongarch/kvm/vm.c
108
kvfree(kvm->arch.phyid_map);
arch/loongarch/kvm/vm.c
109
kvm->arch.phyid_map = NULL;
arch/loongarch/kvm/vm.c
161
if (kvm_vm_support(&kvm->arch, attr->attr))
arch/loongarch/kvm/vm.c
211
return (kvm->arch.ipi && kvm->arch.eiointc && kvm->arch.pch_pic);
arch/loongarch/kvm/vm.c
33
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_LSX);
arch/loongarch/kvm/vm.c
35
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_LASX);
arch/loongarch/kvm/vm.c
37
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_X86BT);
arch/loongarch/kvm/vm.c
39
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_ARMBT);
arch/loongarch/kvm/vm.c
41
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_MIPSBT);
arch/loongarch/kvm/vm.c
43
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_PTW);
arch/loongarch/kvm/vm.c
45
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_MSGINT);
arch/loongarch/kvm/vm.c
49
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_PMU);
arch/loongarch/kvm/vm.c
52
kvm->arch.pv_features |= BIT(KVM_FEATURE_IPI);
arch/loongarch/kvm/vm.c
53
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_PV_IPI);
arch/loongarch/kvm/vm.c
55
kvm->arch.pv_features |= BIT(KVM_FEATURE_PREEMPT);
arch/loongarch/kvm/vm.c
56
kvm->arch.pv_features |= BIT(KVM_FEATURE_STEAL_TIME);
arch/loongarch/kvm/vm.c
57
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_PV_PREEMPT);
arch/loongarch/kvm/vm.c
58
kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_PV_STEALTIME);
arch/loongarch/kvm/vm.c
67
kvm->arch.pgd = kvm_pgd_alloc();
arch/loongarch/kvm/vm.c
68
if (!kvm->arch.pgd)
arch/loongarch/kvm/vm.c
71
kvm->arch.phyid_map = kvzalloc_obj(struct kvm_phyid_map,
arch/loongarch/kvm/vm.c
73
if (!kvm->arch.phyid_map) {
arch/loongarch/kvm/vm.c
74
free_page((unsigned long)kvm->arch.pgd);
arch/loongarch/kvm/vm.c
75
kvm->arch.pgd = NULL;
arch/loongarch/kvm/vm.c
78
spin_lock_init(&kvm->arch.phyid_map_lock);
arch/loongarch/kvm/vm.c
87
kvm->arch.gpa_size = BIT(cpu_vabits);
arch/loongarch/kvm/vm.c
88
kvm->arch.root_level = CONFIG_PGTABLE_LEVELS - 1;
arch/loongarch/kvm/vm.c
89
kvm->arch.invalid_ptes[0] = 0;
arch/loongarch/kvm/vm.c
90
kvm->arch.invalid_ptes[1] = (unsigned long)invalid_pte_table;
arch/loongarch/kvm/vm.c
92
kvm->arch.invalid_ptes[2] = (unsigned long)invalid_pmd_table;
arch/loongarch/kvm/vm.c
95
kvm->arch.invalid_ptes[3] = (unsigned long)invalid_pud_table;
arch/loongarch/kvm/vm.c
97
for (i = 0; i <= kvm->arch.root_level; i++)
arch/loongarch/kvm/vm.c
98
kvm->arch.pte_shifts[i] = PAGE_SHIFT + i * (PAGE_SHIFT - 3);
arch/m68k/kernel/module.c
106
module_fixup(mod, mod->arch.fixup_start, mod->arch.fixup_end);
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
28
.set arch=octeon
arch/mips/include/asm/stackframe.h
434
.set arch=r4000
arch/mips/include/asm/syscall.h
172
int arch = AUDIT_ARCH_MIPS;
arch/mips/include/asm/syscall.h
175
arch |= __AUDIT_ARCH_64BIT;
arch/mips/include/asm/syscall.h
178
arch |= __AUDIT_ARCH_CONVENTION_MIPS64_N32;
arch/mips/include/asm/syscall.h
182
arch |= __AUDIT_ARCH_LE;
arch/mips/include/asm/syscall.h
184
return arch;
arch/mips/kernel/module.c
113
if (me->arch.r_mips_hi16_list != NULL) {
arch/mips/kernel/module.c
114
l = me->arch.r_mips_hi16_list;
arch/mips/kernel/module.c
149
me->arch.r_mips_hi16_list = NULL;
arch/mips/kernel/module.c
163
me->arch.r_mips_hi16_list = NULL;
arch/mips/kernel/module.c
323
me->arch.r_mips_hi16_list = NULL;
arch/mips/kernel/module.c
367
if (me->arch.r_mips_hi16_list) {
arch/mips/kernel/module.c
368
free_relocation_chain(me->arch.r_mips_hi16_list);
arch/mips/kernel/module.c
369
me->arch.r_mips_hi16_list = NULL;
arch/mips/kernel/module.c
424
INIT_LIST_HEAD(&me->arch.dbe_list);
arch/mips/kernel/module.c
428
me->arch.dbe_start = (void *)s->sh_addr;
arch/mips/kernel/module.c
429
me->arch.dbe_end = (void *)s->sh_addr + s->sh_size;
arch/mips/kernel/module.c
431
list_add(&me->arch.dbe_list, &dbe_list);
arch/mips/kernel/module.c
440
list_del(&mod->arch.dbe_list);
arch/mips/kernel/module.c
81
n->next = me->arch.r_mips_hi16_list;
arch/mips/kernel/module.c
82
me->arch.r_mips_hi16_list = n;
arch/mips/kvm/emulate.c
1000
*(u64 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1003
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1004
vcpu->arch.gprs[rt], *(u64 *)data);
arch/mips/kvm/emulate.c
1010
*(u32 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1013
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1014
vcpu->arch.gprs[rt], *(u32 *)data);
arch/mips/kvm/emulate.c
1019
*(u16 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
102
arch->gprs[31] = epc + 8;
arch/mips/kvm/emulate.c
1022
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1023
vcpu->arch.gprs[rt], *(u16 *)data);
arch/mips/kvm/emulate.c
1028
*(u8 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
103
if ((long)arch->gprs[insn.i_format.rs] < 0)
arch/mips/kvm/emulate.c
1031
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1032
vcpu->arch.gprs[rt], *(u8 *)data);
arch/mips/kvm/emulate.c
1037
vcpu->arch.host_cp0_badvaddr) & (~0x3);
arch/mips/kvm/emulate.c
1039
imme = vcpu->arch.host_cp0_badvaddr & 0x3;
arch/mips/kvm/emulate.c
1043
(vcpu->arch.gprs[rt] >> 24);
arch/mips/kvm/emulate.c
1047
(vcpu->arch.gprs[rt] >> 16);
arch/mips/kvm/emulate.c
1051
(vcpu->arch.gprs[rt] >> 8);
arch/mips/kvm/emulate.c
1054
*(u32 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1061
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1062
vcpu->arch.gprs[rt], *(u32 *)data);
arch/mips/kvm/emulate.c
1067
vcpu->arch.host_cp0_badvaddr) & (~0x3);
arch/mips/kvm/emulate.c
1069
imme = vcpu->arch.host_cp0_badvaddr & 0x3;
arch/mips/kvm/emulate.c
1072
*(u32 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1076
(vcpu->arch.gprs[rt] << 8);
arch/mips/kvm/emulate.c
1080
(vcpu->arch.gprs[rt] << 16);
arch/mips/kvm/emulate.c
1084
(vcpu->arch.gprs[rt] << 24);
arch/mips/kvm/emulate.c
1091
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1092
vcpu->arch.gprs[rt], *(u32 *)data);
arch/mips/kvm/emulate.c
1098
vcpu->arch.host_cp0_badvaddr) & (~0x7);
arch/mips/kvm/emulate.c
1101
imme = vcpu->arch.host_cp0_badvaddr & 0x7;
arch/mips/kvm/emulate.c
1105
((vcpu->arch.gprs[rt] >> 56) & 0xff);
arch/mips/kvm/emulate.c
1109
((vcpu->arch.gprs[rt] >> 48) & 0xffff);
arch/mips/kvm/emulate.c
1113
((vcpu->arch.gprs[rt] >> 40) & 0xffffff);
arch/mips/kvm/emulate.c
1117
((vcpu->arch.gprs[rt] >> 32) & 0xffffffff);
arch/mips/kvm/emulate.c
112
arch->gprs[31] = epc + 8;
arch/mips/kvm/emulate.c
1121
((vcpu->arch.gprs[rt] >> 24) & 0xffffffffff);
arch/mips/kvm/emulate.c
1125
((vcpu->arch.gprs[rt] >> 16) & 0xffffffffffff);
arch/mips/kvm/emulate.c
1129
((vcpu->arch.gprs[rt] >> 8) & 0xffffffffffffff);
arch/mips/kvm/emulate.c
113
if ((long)arch->gprs[insn.i_format.rs] >= 0)
arch/mips/kvm/emulate.c
1132
*(u64 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1139
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1140
vcpu->arch.gprs[rt], *(u64 *)data);
arch/mips/kvm/emulate.c
1145
vcpu->arch.host_cp0_badvaddr) & (~0x7);
arch/mips/kvm/emulate.c
1148
imme = vcpu->arch.host_cp0_badvaddr & 0x7;
arch/mips/kvm/emulate.c
1151
*(u64 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1155
(vcpu->arch.gprs[rt] << 8);
arch/mips/kvm/emulate.c
1159
(vcpu->arch.gprs[rt] << 16);
arch/mips/kvm/emulate.c
1163
(vcpu->arch.gprs[rt] << 24);
arch/mips/kvm/emulate.c
1167
(vcpu->arch.gprs[rt] << 32);
arch/mips/kvm/emulate.c
1171
(vcpu->arch.gprs[rt] << 40);
arch/mips/kvm/emulate.c
1175
(vcpu->arch.gprs[rt] << 48);
arch/mips/kvm/emulate.c
1179
(vcpu->arch.gprs[rt] << 56);
arch/mips/kvm/emulate.c
1186
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1187
vcpu->arch.gprs[rt], *(u64 *)data);
arch/mips/kvm/emulate.c
1205
*(u8 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1208
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1209
vcpu->arch.gprs[rt], *(u8 *)data);
arch/mips/kvm/emulate.c
1213
*(u16 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1216
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1217
vcpu->arch.gprs[rt], *(u16 *)data);
arch/mips/kvm/emulate.c
1221
*(u32 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1224
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1225
vcpu->arch.gprs[rt], *(u32 *)data);
arch/mips/kvm/emulate.c
1229
*(u64 *)data = vcpu->arch.gprs[rt];
arch/mips/kvm/emulate.c
1232
vcpu->arch.pc, vcpu->arch.host_cp0_badvaddr,
arch/mips/kvm/emulate.c
1233
vcpu->arch.gprs[rt], *(u64 *)data);
arch/mips/kvm/emulate.c
1264
vcpu->arch.pc = curr_pc;
arch/mips/kvm/emulate.c
1286
curr_pc = vcpu->arch.pc;
arch/mips/kvm/emulate.c
1290
vcpu->arch.io_pc = vcpu->arch.pc;
arch/mips/kvm/emulate.c
1291
vcpu->arch.pc = curr_pc;
arch/mips/kvm/emulate.c
1293
vcpu->arch.io_gpr = rt;
arch/mips/kvm/emulate.c
1296
vcpu->arch.host_cp0_badvaddr);
arch/mips/kvm/emulate.c
1331
vcpu->arch.host_cp0_badvaddr) & (~0x3);
arch/mips/kvm/emulate.c
1334
imme = vcpu->arch.host_cp0_badvaddr & 0x3;
arch/mips/kvm/emulate.c
1355
vcpu->arch.host_cp0_badvaddr) & (~0x3);
arch/mips/kvm/emulate.c
1358
imme = vcpu->arch.host_cp0_badvaddr & 0x3;
arch/mips/kvm/emulate.c
1380
vcpu->arch.host_cp0_badvaddr) & (~0x7);
arch/mips/kvm/emulate.c
1383
imme = vcpu->arch.host_cp0_badvaddr & 0x7;
arch/mips/kvm/emulate.c
141
arch->gprs[31] = instpc + 8;
arch/mips/kvm/emulate.c
1416
vcpu->arch.host_cp0_badvaddr) & (~0x7);
arch/mips/kvm/emulate.c
1419
imme = vcpu->arch.host_cp0_badvaddr & 0x7;
arch/mips/kvm/emulate.c
1512
unsigned long *gpr = &vcpu->arch.gprs[vcpu->arch.io_gpr];
arch/mips/kvm/emulate.c
1522
vcpu->arch.pc = vcpu->arch.io_pc;
arch/mips/kvm/emulate.c
1528
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffffffffff) |
arch/mips/kvm/emulate.c
1532
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffffffff) |
arch/mips/kvm/emulate.c
1536
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffffff) |
arch/mips/kvm/emulate.c
154
if (arch->gprs[insn.i_format.rs] ==
arch/mips/kvm/emulate.c
1540
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffff) |
arch/mips/kvm/emulate.c
1544
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffff) |
arch/mips/kvm/emulate.c
1548
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffff) |
arch/mips/kvm/emulate.c
155
arch->gprs[insn.i_format.rt])
arch/mips/kvm/emulate.c
1552
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xff) |
arch/mips/kvm/emulate.c
1560
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xff00000000000000) |
arch/mips/kvm/emulate.c
1564
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffff000000000000) |
arch/mips/kvm/emulate.c
1568
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffff0000000000) |
arch/mips/kvm/emulate.c
1572
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffff00000000) |
arch/mips/kvm/emulate.c
1576
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffffff000000) |
arch/mips/kvm/emulate.c
1580
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffffffff0000) |
arch/mips/kvm/emulate.c
1584
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffffffffffff00) |
arch/mips/kvm/emulate.c
1601
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffff) |
arch/mips/kvm/emulate.c
1605
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffff) |
arch/mips/kvm/emulate.c
1609
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xff) |
arch/mips/kvm/emulate.c
1617
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xff000000) |
arch/mips/kvm/emulate.c
1621
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffff0000) |
arch/mips/kvm/emulate.c
1625
*gpr = (vcpu->arch.gprs[vcpu->arch.io_gpr] & 0xffffff00) |
arch/mips/kvm/emulate.c
164
if (arch->gprs[insn.i_format.rs] !=
arch/mips/kvm/emulate.c
165
arch->gprs[insn.i_format.rt])
arch/mips/kvm/emulate.c
178
if ((long)arch->gprs[insn.i_format.rs] <= 0)
arch/mips/kvm/emulate.c
191
if ((long)arch->gprs[insn.i_format.rs] > 0)
arch/mips/kvm/emulate.c
248
err = kvm_compute_return_epc(vcpu, vcpu->arch.pc,
arch/mips/kvm/emulate.c
249
&vcpu->arch.pc);
arch/mips/kvm/emulate.c
253
vcpu->arch.pc += 4;
arch/mips/kvm/emulate.c
256
kvm_debug("update_pc(): New PC: %#lx\n", vcpu->arch.pc);
arch/mips/kvm/emulate.c
275
*out = vcpu->arch.host_cp0_badinstr;
arch/mips/kvm/emulate.c
297
*out = vcpu->arch.host_cp0_badinstrp;
arch/mips/kvm/emulate.c
315
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
317
return (vcpu->arch.count_ctl & KVM_REG_MIPS_COUNT_CTL_DC) ||
arch/mips/kvm/emulate.c
334
delta = now_ns + vcpu->arch.count_dyn_bias;
arch/mips/kvm/emulate.c
336
if (delta >= vcpu->arch.count_period) {
arch/mips/kvm/emulate.c
338
periods = div64_s64(now_ns, vcpu->arch.count_period);
arch/mips/kvm/emulate.c
339
vcpu->arch.count_dyn_bias = -periods * vcpu->arch.count_period;
arch/mips/kvm/emulate.c
341
delta = now_ns + vcpu->arch.count_dyn_bias;
arch/mips/kvm/emulate.c
354
return div_u64(delta * vcpu->arch.count_hz, NSEC_PER_SEC);
arch/mips/kvm/emulate.c
369
if (unlikely(vcpu->arch.count_ctl & KVM_REG_MIPS_COUNT_CTL_DC))
arch/mips/kvm/emulate.c
370
return vcpu->arch.count_resume;
arch/mips/kvm/emulate.c
387
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
393
count = vcpu->arch.count_bias + kvm_mips_ktime_to_count(vcpu, now);
arch/mips/kvm/emulate.c
409
expires = hrtimer_get_expires(&vcpu->arch.comparecount_timer);
arch/mips/kvm/emulate.c
410
threshold = ktime_add_ns(now, vcpu->arch.count_period / 4);
arch/mips/kvm/emulate.c
416
running = hrtimer_cancel(&vcpu->arch.comparecount_timer);
arch/mips/kvm/emulate.c
427
vcpu->arch.count_period);
arch/mips/kvm/emulate.c
428
hrtimer_start(&vcpu->arch.comparecount_timer, expires,
arch/mips/kvm/emulate.c
447
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
45
struct kvm_vcpu_arch *arch = &vcpu->arch;
arch/mips/kvm/emulate.c
477
hrtimer_cancel(&vcpu->arch.comparecount_timer);
arch/mips/kvm/emulate.c
505
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
513
delta = div_u64(delta * NSEC_PER_SEC, vcpu->arch.count_hz);
arch/mips/kvm/emulate.c
517
hrtimer_cancel(&vcpu->arch.comparecount_timer);
arch/mips/kvm/emulate.c
518
hrtimer_start(&vcpu->arch.comparecount_timer, expire, HRTIMER_MODE_ABS);
arch/mips/kvm/emulate.c
552
before_count = vcpu->arch.count_bias +
arch/mips/kvm/emulate.c
565
vcpu->arch.count_bias += drift;
arch/mips/kvm/emulate.c
572
now_count = vcpu->arch.count_bias + kvm_mips_ktime_to_count(vcpu, now);
arch/mips/kvm/emulate.c
581
vcpu->arch.count_bias += drift;
arch/mips/kvm/emulate.c
588
delta = div_u64(delta * NSEC_PER_SEC, vcpu->arch.count_hz);
arch/mips/kvm/emulate.c
606
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
611
vcpu->arch.count_bias = count - kvm_mips_ktime_to_count(vcpu, now);
arch/mips/kvm/emulate.c
631
vcpu->arch.count_hz = count_hz;
arch/mips/kvm/emulate.c
632
vcpu->arch.count_period = div_u64((u64)NSEC_PER_SEC << 32, count_hz);
arch/mips/kvm/emulate.c
633
vcpu->arch.count_dyn_bias = 0;
arch/mips/kvm/emulate.c
65
arch->gprs[insn.r_format.rd] = epc + 8;
arch/mips/kvm/emulate.c
652
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
661
if (vcpu->arch.count_hz == count_hz)
arch/mips/kvm/emulate.c
674
vcpu->arch.count_hz = count_hz;
arch/mips/kvm/emulate.c
675
vcpu->arch.count_period = div_u64((u64)NSEC_PER_SEC << 32, count_hz);
arch/mips/kvm/emulate.c
676
vcpu->arch.count_dyn_bias = 0;
arch/mips/kvm/emulate.c
679
vcpu->arch.count_bias = count - kvm_mips_ktime_to_count(vcpu, now);
arch/mips/kvm/emulate.c
68
nextpc = arch->gprs[insn.r_format.rs];
arch/mips/kvm/emulate.c
699
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
782
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
787
hrtimer_cancel(&vcpu->arch.comparecount_timer);
arch/mips/kvm/emulate.c
809
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
812
if (!(vcpu->arch.count_ctl & KVM_REG_MIPS_COUNT_CTL_DC))
arch/mips/kvm/emulate.c
829
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
84
if ((long)arch->gprs[insn.i_format.rs] < 0)
arch/mips/kvm/emulate.c
855
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/emulate.c
856
s64 changed = count_ctl ^ vcpu->arch.count_ctl;
arch/mips/kvm/emulate.c
866
vcpu->arch.count_ctl = count_ctl;
arch/mips/kvm/emulate.c
874
vcpu->arch.count_resume = ktime_get();
arch/mips/kvm/emulate.c
877
vcpu->arch.count_resume = kvm_mips_count_disable(vcpu);
arch/mips/kvm/emulate.c
887
vcpu->arch.count_hz);
arch/mips/kvm/emulate.c
888
expire = ktime_add_ns(vcpu->arch.count_resume, delta);
arch/mips/kvm/emulate.c
925
vcpu->arch.count_resume = ns_to_ktime(count_resume);
arch/mips/kvm/emulate.c
93
if ((long)arch->gprs[insn.i_format.rs] >= 0)
arch/mips/kvm/emulate.c
940
hrtimer_add_expires_ns(&vcpu->arch.comparecount_timer,
arch/mips/kvm/emulate.c
941
vcpu->arch.count_period);
arch/mips/kvm/emulate.c
947
kvm_debug("[%#lx] !!!WAIT!!! (%#lx)\n", vcpu->arch.pc,
arch/mips/kvm/emulate.c
948
vcpu->arch.pending_exceptions);
arch/mips/kvm/emulate.c
952
if (!vcpu->arch.pending_exceptions) {
arch/mips/kvm/emulate.c
954
vcpu->arch.wait = 1;
arch/mips/kvm/emulate.c
984
curr_pc = vcpu->arch.pc;
arch/mips/kvm/emulate.c
992
vcpu->arch.host_cp0_badvaddr);
arch/mips/kvm/entry.c
184
UASM_i_ADDIU(&p, GPR_K1, GPR_A0, offsetof(struct kvm_vcpu, arch));
arch/mips/kvm/entry.c
265
(int)offsetof(struct kvm_vcpu, arch), GPR_K1);
arch/mips/kvm/entry.c
266
UASM_i_LW(&p, GPR_A0, offsetof(struct kvm, arch.gpa_mm.pgd), GPR_S0);
arch/mips/kvm/entry.c
309
offsetof(struct kvm, arch.gpa_mm.context.asid));
arch/mips/kvm/entry.c
400
UASM_i_SW(&p, GPR_K0, offsetof(struct kvm_vcpu, arch.gprs[GPR_K0]), GPR_K1);
arch/mips/kvm/entry.c
448
UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu, arch.gprs[GPR_K0]), GPR_K1);
arch/mips/kvm/entry.c
484
UASM_i_ADDIU(&p, GPR_K1, GPR_K1, offsetof(struct kvm_vcpu, arch));
arch/mips/kvm/entry.c
759
UASM_i_ADDIU(&p, GPR_K1, GPR_K1, offsetof(struct kvm_vcpu, arch));
arch/mips/kvm/hypcall.c
22
kvm_debug("[%#lx] HYPCALL %#03x\n", vcpu->arch.pc, code);
arch/mips/kvm/hypcall.c
45
num = vcpu->arch.gprs[2]; /* v0 */
arch/mips/kvm/hypcall.c
46
args[0] = vcpu->arch.gprs[4]; /* a0 */
arch/mips/kvm/hypcall.c
47
args[1] = vcpu->arch.gprs[5]; /* a1 */
arch/mips/kvm/hypcall.c
48
args[2] = vcpu->arch.gprs[6]; /* a2 */
arch/mips/kvm/hypcall.c
49
args[3] = vcpu->arch.gprs[7]; /* a3 */
arch/mips/kvm/hypcall.c
52
args, &vcpu->arch.gprs[2] /* v0 */);
arch/mips/kvm/interrupt.c
26
unsigned long *pending = &vcpu->arch.pending_exceptions;
arch/mips/kvm/interrupt.c
27
unsigned long *pending_clr = &vcpu->arch.pending_exceptions_clr;
arch/mips/kvm/interrupt.c
39
return test_bit(MIPS_EXC_INT_TIMER, &vcpu->arch.pending_exceptions);
arch/mips/kvm/loongson_ipi.c
199
s = &kvm->arch.ipi;
arch/mips/kvm/mips.c
1079
kvm_read_c0_guest_cause(&vcpu->arch.cop0) & C_TI;
arch/mips/kvm/mips.c
1091
kvm_debug("\tpc = 0x%08lx\n", vcpu->arch.pc);
arch/mips/kvm/mips.c
1092
kvm_debug("\texceptions: %08lx\n", vcpu->arch.pending_exceptions);
arch/mips/kvm/mips.c
1096
vcpu->arch.gprs[i],
arch/mips/kvm/mips.c
1097
vcpu->arch.gprs[i + 1],
arch/mips/kvm/mips.c
1098
vcpu->arch.gprs[i + 2], vcpu->arch.gprs[i + 3]);
arch/mips/kvm/mips.c
1100
kvm_debug("\thi: 0x%08lx\n", vcpu->arch.hi);
arch/mips/kvm/mips.c
1101
kvm_debug("\tlo: 0x%08lx\n", vcpu->arch.lo);
arch/mips/kvm/mips.c
1103
cop0 = &vcpu->arch.cop0;
arch/mips/kvm/mips.c
1119
for (i = 1; i < ARRAY_SIZE(vcpu->arch.gprs); i++)
arch/mips/kvm/mips.c
1120
vcpu->arch.gprs[i] = regs->gpr[i];
arch/mips/kvm/mips.c
1121
vcpu->arch.gprs[0] = 0; /* zero is special, and cannot be set. */
arch/mips/kvm/mips.c
1122
vcpu->arch.hi = regs->hi;
arch/mips/kvm/mips.c
1123
vcpu->arch.lo = regs->lo;
arch/mips/kvm/mips.c
1124
vcpu->arch.pc = regs->pc;
arch/mips/kvm/mips.c
1136
for (i = 0; i < ARRAY_SIZE(vcpu->arch.gprs); i++)
arch/mips/kvm/mips.c
1137
regs->gpr[i] = vcpu->arch.gprs[i];
arch/mips/kvm/mips.c
1139
regs->hi = vcpu->arch.hi;
arch/mips/kvm/mips.c
1140
regs->lo = vcpu->arch.lo;
arch/mips/kvm/mips.c
1141
regs->pc = vcpu->arch.pc;
arch/mips/kvm/mips.c
115
return !!(vcpu->arch.pending_exceptions);
arch/mips/kvm/mips.c
1170
u32 cause = vcpu->arch.host_cp0_cause;
arch/mips/kvm/mips.c
1172
u32 __user *opc = (u32 __user *) vcpu->arch.pc;
arch/mips/kvm/mips.c
1173
unsigned long badvaddr = vcpu->arch.host_cp0_badvaddr;
arch/mips/kvm/mips.c
1225
cause, kvm_read_c0_guest_status(&vcpu->arch.cop0), opc,
arch/mips/kvm/mips.c
1297
kvm_read_c0_guest_status(&vcpu->arch.cop0));
arch/mips/kvm/mips.c
1345
if (kvm_mips_guest_has_fpu(&vcpu->arch) &&
arch/mips/kvm/mips.c
1347
__kvm_restore_fcsr(&vcpu->arch);
arch/mips/kvm/mips.c
1349
if (kvm_mips_guest_has_msa(&vcpu->arch) &&
arch/mips/kvm/mips.c
1351
__kvm_restore_msacsr(&vcpu->arch);
arch/mips/kvm/mips.c
1370
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/mips.c
1384
vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA)
arch/mips/kvm/mips.c
1399
if (!(vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU)) {
arch/mips/kvm/mips.c
1400
__kvm_restore_fpu(&vcpu->arch);
arch/mips/kvm/mips.c
1401
vcpu->arch.aux_inuse |= KVM_MIPS_AUX_FPU;
arch/mips/kvm/mips.c
1414
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/mips.c
1423
if (kvm_mips_guest_has_fpu(&vcpu->arch)) {
arch/mips/kvm/mips.c
1431
(vcpu->arch.aux_inuse & (KVM_MIPS_AUX_FPU |
arch/mips/kvm/mips.c
1446
switch (vcpu->arch.aux_inuse & (KVM_MIPS_AUX_FPU | KVM_MIPS_AUX_MSA)) {
arch/mips/kvm/mips.c
1451
__kvm_restore_msa_upper(&vcpu->arch);
arch/mips/kvm/mips.c
1452
vcpu->arch.aux_inuse |= KVM_MIPS_AUX_MSA;
arch/mips/kvm/mips.c
1457
__kvm_restore_msa(&vcpu->arch);
arch/mips/kvm/mips.c
1458
vcpu->arch.aux_inuse |= KVM_MIPS_AUX_MSA;
arch/mips/kvm/mips.c
1459
if (kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
1460
vcpu->arch.aux_inuse |= KVM_MIPS_AUX_FPU;
arch/mips/kvm/mips.c
1477
if (cpu_has_msa && vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA) {
arch/mips/kvm/mips.c
1480
vcpu->arch.aux_inuse &= ~KVM_MIPS_AUX_MSA;
arch/mips/kvm/mips.c
1482
if (vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU) {
arch/mips/kvm/mips.c
1485
vcpu->arch.aux_inuse &= ~KVM_MIPS_AUX_FPU;
arch/mips/kvm/mips.c
1501
if (cpu_has_msa && vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA) {
arch/mips/kvm/mips.c
1502
__kvm_save_msa(&vcpu->arch);
arch/mips/kvm/mips.c
1507
if (vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU) {
arch/mips/kvm/mips.c
151
kvm->arch.gpa_mm.pgd = kvm_pgd_alloc();
arch/mips/kvm/mips.c
1511
vcpu->arch.aux_inuse &= ~(KVM_MIPS_AUX_FPU | KVM_MIPS_AUX_MSA);
arch/mips/kvm/mips.c
1512
} else if (vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU) {
arch/mips/kvm/mips.c
1513
__kvm_save_fpu(&vcpu->arch);
arch/mips/kvm/mips.c
1514
vcpu->arch.aux_inuse &= ~KVM_MIPS_AUX_FPU;
arch/mips/kvm/mips.c
152
if (!kvm->arch.gpa_mm.pgd)
arch/mips/kvm/mips.c
166
pgd_free(NULL, kvm->arch.gpa_mm.pgd);
arch/mips/kvm/mips.c
263
vcpu = container_of(timer, struct kvm_vcpu, arch.comparecount_timer);
arch/mips/kvm/mips.c
267
vcpu->arch.wait = 0;
arch/mips/kvm/mips.c
291
hrtimer_setup(&vcpu->arch.comparecount_timer, kvm_mips_comparecount_wakeup, CLOCK_MONOTONIC,
arch/mips/kvm/mips.c
325
vcpu->arch.guest_ebase = gebase;
arch/mips/kvm/mips.c
352
vcpu->arch.vcpu_run = p;
arch/mips/kvm/mips.c
359
dump_handler("kvm_vcpu_run", vcpu->arch.vcpu_run, p);
arch/mips/kvm/mips.c
362
dump_handler("kvm_exit", gebase + 0x2000, vcpu->arch.vcpu_run);
arch/mips/kvm/mips.c
369
vcpu->arch.last_sched_cpu = -1;
arch/mips/kvm/mips.c
370
vcpu->arch.last_exec_cpu = -1;
arch/mips/kvm/mips.c
388
hrtimer_cancel(&vcpu->arch.comparecount_timer);
arch/mips/kvm/mips.c
393
kfree(vcpu->arch.guest_ebase);
arch/mips/kvm/mips.c
508
dvcpu->arch.wait = 0;
arch/mips/kvm/mips.c
583
if (kvm_mips_guest_can_have_fpu(&vcpu->arch)) {
arch/mips/kvm/mips.c
589
if (kvm_mips_guest_can_have_msa(&vcpu->arch))
arch/mips/kvm/mips.c
606
if (kvm_mips_guest_can_have_fpu(&vcpu->arch)) {
arch/mips/kvm/mips.c
629
if (kvm_mips_guest_can_have_msa(&vcpu->arch)) {
arch/mips/kvm/mips.c
649
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/mips.c
650
struct mips_fpu_struct *fpu = &vcpu->arch.fpu;
arch/mips/kvm/mips.c
659
v = (long)vcpu->arch.gprs[reg->id - KVM_REG_MIPS_R0];
arch/mips/kvm/mips.c
663
v = (long)vcpu->arch.hi;
arch/mips/kvm/mips.c
666
v = (long)vcpu->arch.lo;
arch/mips/kvm/mips.c
670
v = (long)vcpu->arch.pc;
arch/mips/kvm/mips.c
675
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
685
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
694
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
699
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
706
if (!kvm_mips_guest_has_msa(&vcpu->arch))
arch/mips/kvm/mips.c
723
if (!kvm_mips_guest_has_msa(&vcpu->arch))
arch/mips/kvm/mips.c
728
if (!kvm_mips_guest_has_msa(&vcpu->arch))
arch/mips/kvm/mips.c
761
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/mips.c
762
struct mips_fpu_struct *fpu = &vcpu->arch.fpu;
arch/mips/kvm/mips.c
793
vcpu->arch.gprs[reg->id - KVM_REG_MIPS_R0] = v;
arch/mips/kvm/mips.c
797
vcpu->arch.hi = v;
arch/mips/kvm/mips.c
800
vcpu->arch.lo = v;
arch/mips/kvm/mips.c
804
vcpu->arch.pc = v;
arch/mips/kvm/mips.c
809
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
819
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
828
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
833
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/mips.c
840
if (!kvm_mips_guest_has_msa(&vcpu->arch))
arch/mips/kvm/mips.c
854
if (!kvm_mips_guest_has_msa(&vcpu->arch))
arch/mips/kvm/mips.c
859
if (!kvm_mips_guest_has_msa(&vcpu->arch))
arch/mips/kvm/mips.c
885
vcpu->arch.fpu_enabled = true;
arch/mips/kvm/mips.c
888
vcpu->arch.msa_enabled = true;
arch/mips/kvm/mmu.c
146
return kvm_mips_walk_pgd(kvm->arch.gpa_mm.pgd, cache, addr);
arch/mips/kvm/mmu.c
272
return kvm_mips_flush_gpa_pgd(kvm->arch.gpa_mm.pgd,
arch/mips/kvm/mmu.c
30
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
arch/mips/kvm/mmu.c
398
return kvm_mips_mkclean_pgd(kvm->arch.gpa_mm.pgd,
arch/mips/kvm/mmu.c
436
return kvm_mips_mkold_pgd(kvm->arch.gpa_mm.pgd,
arch/mips/kvm/mmu.c
552
struct kvm_mmu_memory_cache *memcache = &vcpu->arch.mmu_page_cache;
arch/mips/kvm/mmu.c
670
if (hrtimer_cancel(&vcpu->arch.comparecount_timer))
arch/mips/kvm/mmu.c
671
hrtimer_restart(&vcpu->arch.comparecount_timer);
arch/mips/kvm/mmu.c
684
if (vcpu->arch.last_sched_cpu != cpu) {
arch/mips/kvm/mmu.c
686
vcpu->arch.last_sched_cpu, cpu, vcpu->vcpu_id);
arch/mips/kvm/mmu.c
710
vcpu->arch.last_sched_cpu = cpu;
arch/mips/kvm/stats.c
57
if (vcpu->arch.cop0.stat[i][j])
arch/mips/kvm/stats.c
59
vcpu->arch.cop0.stat[i][j]);
arch/mips/kvm/tlb.c
42
struct mm_struct *gpa_mm = &vcpu->kvm->arch.gpa_mm;
arch/mips/kvm/trace.h
124
__entry->pc = vcpu->arch.pc;
arch/mips/kvm/trace.h
263
__entry->pc = vcpu->arch.pc;
arch/mips/kvm/trace.h
287
__entry->pc = vcpu->arch.pc;
arch/mips/kvm/trace.h
325
__entry->epc = kvm_read_c0_guest_epc(&vcpu->arch.cop0);
arch/mips/kvm/trace.h
326
__entry->pc = vcpu->arch.pc;
arch/mips/kvm/trace.h
327
__entry->badvaddr = kvm_read_c0_guest_badvaddr(&vcpu->arch.cop0);
arch/mips/kvm/trace.h
328
__entry->status = kvm_read_c0_guest_status(&vcpu->arch.cop0);
arch/mips/kvm/trace.h
329
__entry->cause = kvm_read_c0_guest_cause(&vcpu->arch.cop0);
arch/mips/kvm/trace.h
38
__entry->pc = vcpu->arch.pc;
arch/mips/kvm/vz.c
1011
ARRAY_SIZE(vcpu->arch.maar));
arch/mips/kvm/vz.c
1012
vcpu->arch.maar[kvm_read_sw_gc0_maari(cop0)] =
arch/mips/kvm/vz.c
1068
vcpu->arch.pc = curr_pc;
arch/mips/kvm/vz.c
1081
struct kvm_vcpu_arch *arch = &vcpu->arch;
arch/mips/kvm/vz.c
1088
curr_pc = vcpu->arch.pc;
arch/mips/kvm/vz.c
1102
va = arch->gprs[base] + offset;
arch/mips/kvm/vz.c
1105
cache, op, base, arch->gprs[base], offset);
arch/mips/kvm/vz.c
1134
curr_pc, vcpu->arch.gprs[31], cache, op, base, arch->gprs[base],
arch/mips/kvm/vz.c
1137
vcpu->arch.pc = curr_pc;
arch/mips/kvm/vz.c
115
if (kvm_mips_guest_has_msa(&vcpu->arch))
arch/mips/kvm/vz.c
1156
curr_pc = vcpu->arch.pc;
arch/mips/kvm/vz.c
1166
hostcfg = read_cpucfg(vcpu->arch.gprs[rs]);
arch/mips/kvm/vz.c
1168
switch (vcpu->arch.gprs[rs]) {
arch/mips/kvm/vz.c
1170
vcpu->arch.gprs[rd] = 0x14c000;
arch/mips/kvm/vz.c
1176
vcpu->arch.gprs[rd] = hostcfg;
arch/mips/kvm/vz.c
1181
vcpu->arch.gprs[rd] = hostcfg;
arch/mips/kvm/vz.c
1184
vcpu->arch.gprs[rd] = hostcfg;
arch/mips/kvm/vz.c
1188
vcpu->arch.gprs[rd] = 0;
arch/mips/kvm/vz.c
1195
inst.loongson3_lscsr_format.fr, vcpu->arch.gprs[rs], curr_pc);
arch/mips/kvm/vz.c
1205
vcpu->arch.pc = curr_pc;
arch/mips/kvm/vz.c
1216
struct kvm_vcpu_arch *arch = &vcpu->arch;
arch/mips/kvm/vz.c
122
if (kvm_mips_guest_has_fpu(&vcpu->arch)) {
arch/mips/kvm/vz.c
1263
arch->gprs[rt] =
arch/mips/kvm/vz.c
1273
KVM_TRACE_HWR(rd, sel), arch->gprs[rt]);
arch/mips/kvm/vz.c
1298
struct kvm_vcpu_arch *arch = &vcpu->arch;
arch/mips/kvm/vz.c
1318
unsigned int val = arch->gprs[rt];
arch/mips/kvm/vz.c
1326
if (!kvm_mips_guest_has_fpu(&vcpu->arch))
arch/mips/kvm/vz.c
1356
vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA)
arch/mips/kvm/vz.c
1395
vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU)
arch/mips/kvm/vz.c
1450
curr_pc = vcpu->arch.pc;
arch/mips/kvm/vz.c
1457
vcpu->arch.pc = curr_pc;
arch/mips/kvm/vz.c
1484
u32 *opc = (u32 *) vcpu->arch.pc;
arch/mips/kvm/vz.c
1485
u32 cause = vcpu->arch.host_cp0_cause;
arch/mips/kvm/vz.c
1487
u32 gexccode = (vcpu->arch.host_cp0_guestctl0 &
arch/mips/kvm/vz.c
1555
u32 cause = vcpu->arch.host_cp0_cause;
arch/mips/kvm/vz.c
1565
if (WARN_ON(!kvm_mips_guest_has_fpu(&vcpu->arch) ||
arch/mips/kvm/vz.c
1566
vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU)) {
arch/mips/kvm/vz.c
158
if (kvm_mips_guest_can_have_fpu(&vcpu->arch))
arch/mips/kvm/vz.c
1610
if (!kvm_mips_guest_has_msa(&vcpu->arch) ||
arch/mips/kvm/vz.c
1613
vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA) {
arch/mips/kvm/vz.c
1626
u32 *opc = (u32 *) vcpu->arch.pc;
arch/mips/kvm/vz.c
1627
u32 cause = vcpu->arch.host_cp0_cause;
arch/mips/kvm/vz.c
1628
ulong badvaddr = vcpu->arch.host_cp0_badvaddr;
arch/mips/kvm/vz.c
1635
if (kvm_is_ifetch_fault(&vcpu->arch)) {
arch/mips/kvm/vz.c
1673
u32 *opc = (u32 *) vcpu->arch.pc;
arch/mips/kvm/vz.c
1674
u32 cause = vcpu->arch.host_cp0_cause;
arch/mips/kvm/vz.c
1675
ulong badvaddr = vcpu->arch.host_cp0_badvaddr;
arch/mips/kvm/vz.c
1684
vcpu->arch.host_cp0_badvaddr = badvaddr;
arch/mips/kvm/vz.c
175
if (kvm_mips_guest_can_have_msa(&vcpu->arch))
arch/mips/kvm/vz.c
1800
ret += 1 + ARRAY_SIZE(vcpu->arch.maar);
arch/mips/kvm/vz.c
1853
for (i = 0; i < ARRAY_SIZE(vcpu->arch.maar); ++i) {
arch/mips/kvm/vz.c
1914
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
205
set_bit(priority, &vcpu->arch.pending_exceptions);
arch/mips/kvm/vz.c
206
clear_bit(priority, &vcpu->arch.pending_exceptions_clr);
arch/mips/kvm/vz.c
2077
if (idx >= ARRAY_SIZE(vcpu->arch.maar))
arch/mips/kvm/vz.c
2079
*v = vcpu->arch.maar[idx];
arch/mips/kvm/vz.c
2084
*v = kvm_read_sw_gc0_maari(&vcpu->arch.cop0);
arch/mips/kvm/vz.c
211
clear_bit(priority, &vcpu->arch.pending_exceptions);
arch/mips/kvm/vz.c
212
set_bit(priority, &vcpu->arch.pending_exceptions_clr);
arch/mips/kvm/vz.c
2120
*v = vcpu->arch.count_ctl;
arch/mips/kvm/vz.c
2123
*v = ktime_to_ns(vcpu->arch.count_resume);
arch/mips/kvm/vz.c
2126
*v = vcpu->arch.count_hz;
arch/mips/kvm/vz.c
2138
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
2354
if (idx >= ARRAY_SIZE(vcpu->arch.maar))
arch/mips/kvm/vz.c
2356
vcpu->arch.maar[idx] = mips_process_maar(dmtc_op, v);
arch/mips/kvm/vz.c
2446
vcpu->arch.vzguestid[i] = 0;
arch/mips/kvm/vz.c
2469
if (wired > vcpu->arch.wired_tlb_limit) {
arch/mips/kvm/vz.c
2470
tlbs = krealloc(vcpu->arch.wired_tlb, wired *
arch/mips/kvm/vz.c
2471
sizeof(*vcpu->arch.wired_tlb), GFP_ATOMIC);
arch/mips/kvm/vz.c
2474
wired = vcpu->arch.wired_tlb_limit;
arch/mips/kvm/vz.c
2476
vcpu->arch.wired_tlb = tlbs;
arch/mips/kvm/vz.c
2477
vcpu->arch.wired_tlb_limit = wired;
arch/mips/kvm/vz.c
2483
kvm_vz_save_guesttlb(vcpu->arch.wired_tlb, 0, wired);
arch/mips/kvm/vz.c
2485
for (i = wired; i < vcpu->arch.wired_tlb_used; ++i) {
arch/mips/kvm/vz.c
2486
vcpu->arch.wired_tlb[i].tlb_hi = UNIQUE_GUEST_ENTRYHI(i);
arch/mips/kvm/vz.c
2487
vcpu->arch.wired_tlb[i].tlb_lo[0] = 0;
arch/mips/kvm/vz.c
2488
vcpu->arch.wired_tlb[i].tlb_lo[1] = 0;
arch/mips/kvm/vz.c
2489
vcpu->arch.wired_tlb[i].tlb_mask = 0;
arch/mips/kvm/vz.c
2491
vcpu->arch.wired_tlb_used = wired;
arch/mips/kvm/vz.c
2497
if (vcpu->arch.wired_tlb)
arch/mips/kvm/vz.c
2498
kvm_vz_load_guesttlb(vcpu->arch.wired_tlb, 0,
arch/mips/kvm/vz.c
2499
vcpu->arch.wired_tlb_used);
arch/mips/kvm/vz.c
2505
struct mm_struct *gpa_mm = &kvm->arch.gpa_mm;
arch/mips/kvm/vz.c
2512
migrated = (vcpu->arch.last_exec_cpu != cpu);
arch/mips/kvm/vz.c
2513
vcpu->arch.last_exec_cpu = cpu;
arch/mips/kvm/vz.c
2530
(vcpu->arch.vzguestid[cpu] ^ guestid_cache(cpu)) &
arch/mips/kvm/vz.c
2533
vcpu->arch.vzguestid[cpu] = guestid_cache(cpu);
arch/mips/kvm/vz.c
2535
vcpu->arch.vzguestid[cpu]);
arch/mips/kvm/vz.c
2539
change_c0_guestctl1(GUESTID_MASK, vcpu->arch.vzguestid[cpu]);
arch/mips/kvm/vz.c
2556
if (cpumask_test_and_clear_cpu(cpu, &kvm->arch.asid_flush_mask))
arch/mips/kvm/vz.c
2565
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
2572
migrated = (vcpu->arch.last_sched_cpu != cpu);
arch/mips/kvm/vz.c
2707
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
282
clear_bit(priority, &vcpu->arch.pending_exceptions);
arch/mips/kvm/vz.c
3055
vcpu->arch.vzguestid[i] = 0;
arch/mips/kvm/vz.c
3079
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
3213
vcpu->arch.pc = CKSEG1ADDR(0x1fc00000);
arch/mips/kvm/vz.c
3230
cpumask_setall(&kvm->arch.asid_flush_mask);
arch/mips/kvm/vz.c
324
clear_bit(priority, &vcpu->arch.pending_exceptions_clr);
arch/mips/kvm/vz.c
3263
r = vcpu->arch.vcpu_run(vcpu);
arch/mips/kvm/vz.c
346
if (mips_hpt_frequency != vcpu->arch.count_hz)
arch/mips/kvm/vz.c
425
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
520
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
665
opc = (u32 *)vcpu->arch.pc;
arch/mips/kvm/vz.c
666
if (vcpu->arch.host_cp0_cause & CAUSEF_BD)
arch/mips/kvm/vz.c
799
unsigned int gexccode = (vcpu->arch.host_cp0_guestctl0 &
arch/mips/kvm/vz.c
819
u32 *opc = (u32 *) vcpu->arch.pc;
arch/mips/kvm/vz.c
820
u32 cause = vcpu->arch.host_cp0_cause;
arch/mips/kvm/vz.c
822
unsigned long badvaddr = vcpu->arch.host_cp0_badvaddr;
arch/mips/kvm/vz.c
866
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
870
kvm_write_sw_gc0_maari(cop0, ARRAY_SIZE(vcpu->arch.maar) - 1);
arch/mips/kvm/vz.c
871
else if (val < ARRAY_SIZE(vcpu->arch.maar))
arch/mips/kvm/vz.c
879
struct mips_coproc *cop0 = &vcpu->arch.cop0;
arch/mips/kvm/vz.c
889
curr_pc = vcpu->arch.pc;
arch/mips/kvm/vz.c
932
ARRAY_SIZE(vcpu->arch.maar));
arch/mips/kvm/vz.c
933
val = vcpu->arch.maar[
arch/mips/kvm/vz.c
966
vcpu->arch.gprs[rt] = val;
arch/mips/kvm/vz.c
979
val = vcpu->arch.gprs[rt];
arch/mips/kvm/vz.c
987
kvm_mips_write_count(vcpu, vcpu->arch.gprs[rt]);
arch/mips/kvm/vz.c
991
vcpu->arch.gprs[rt],
arch/parisc/include/asm/syscall.h
79
int arch = AUDIT_ARCH_PARISC;
arch/parisc/include/asm/syscall.h
82
arch = AUDIT_ARCH_PARISC64;
arch/parisc/include/asm/syscall.h
84
return arch;
arch/parisc/kernel/audit.c
32
int audit_classify_arch(int arch)
arch/parisc/kernel/audit.c
35
if (arch == AUDIT_ARCH_PARISC)
arch/parisc/kernel/kexec.c
100
*(unsigned long *)(virt + kexec_initrd_end_offset) = arch->initrd_end;
arch/parisc/kernel/kexec.c
82
struct kimage_arch *arch = &image->arch;
arch/parisc/kernel/kexec.c
98
*(unsigned long *)(virt + kexec_cmdline_offset) = arch->cmdline;
arch/parisc/kernel/kexec.c
99
*(unsigned long *)(virt + kexec_initrd_start_offset) = arch->initrd_start;
arch/parisc/kernel/kexec_file.c
55
image->arch.initrd_start = kbuf.mem;
arch/parisc/kernel/kexec_file.c
56
image->arch.initrd_end = kbuf.mem + initrd_len;
arch/parisc/kernel/kexec_file.c
72
image->arch.cmdline = kbuf.mem;
arch/parisc/kernel/module.c
252
kfree(mod->arch.section);
arch/parisc/kernel/module.c
253
mod->arch.section = NULL;
arch/parisc/kernel/module.c
262
return (mod->arch.section[section].stub_entries + 1)
arch/parisc/kernel/module.c
276
len = hdr->e_shnum * sizeof(me->arch.section[0]);
arch/parisc/kernel/module.c
277
me->arch.section = kzalloc(len, GFP_KERNEL);
arch/parisc/kernel/module.c
278
if (!me->arch.section)
arch/parisc/kernel/module.c
288
me->arch.unwind_section = i;
arch/parisc/kernel/module.c
313
WARN_ON(me->arch.section[s].stub_entries);
arch/parisc/kernel/module.c
316
me->arch.section[s].stub_entries += count;
arch/parisc/kernel/module.c
322
me->arch.got_offset = mod_mem->size;
arch/parisc/kernel/module.c
326
me->arch.fdesc_offset = mod_mem->size;
arch/parisc/kernel/module.c
329
me->arch.got_max = gots;
arch/parisc/kernel/module.c
330
me->arch.fdesc_max = fdescs;
arch/parisc/kernel/module.c
345
got = me->mem[MOD_TEXT].base + me->arch.got_offset;
arch/parisc/kernel/module.c
350
BUG_ON(++me->arch.got_count > me->arch.got_max);
arch/parisc/kernel/module.c
363
Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
arch/parisc/kernel/module.c
377
BUG_ON(++me->arch.fdesc_count > me->arch.fdesc_max);
arch/parisc/kernel/module.c
381
fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
arch/parisc/kernel/module.c
399
if (!me->arch.section[targetsec].stub_offset) {
arch/parisc/kernel/module.c
400
loc0 -= (me->arch.section[targetsec].stub_entries + 1) *
arch/parisc/kernel/module.c
404
me->arch.section[targetsec].stub_offset = loc0;
arch/parisc/kernel/module.c
408
stub = (void *) me->arch.section[targetsec].stub_offset;
arch/parisc/kernel/module.c
409
me->arch.section[targetsec].stub_offset += sizeof(struct stub_entry);
arch/parisc/kernel/module.c
412
BUG_ON(0 == me->arch.section[targetsec].stub_entries--);
arch/parisc/kernel/module.c
808
if (!me->arch.unwind_section)
arch/parisc/kernel/module.c
811
table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
arch/parisc/kernel/module.c
812
end = table + sechdrs[me->arch.unwind_section].sh_size;
arch/parisc/kernel/module.c
813
gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
arch/parisc/kernel/module.c
816
me->arch.unwind_section, table, end, gp);
arch/parisc/kernel/module.c
817
me->arch.unwind = unwind_table_add(me->name, 0, gp, table, end);
arch/parisc/kernel/module.c
823
if (me->arch.unwind)
arch/parisc/kernel/module.c
824
unwind_table_remove(me->arch.unwind);
arch/parisc/kernel/module.c
851
me->arch.got_count, me->arch.got_max,
arch/parisc/kernel/module.c
852
me->arch.fdesc_count, me->arch.fdesc_max);
arch/parisc/kernel/module.c
876
if(me->arch.got_count > MAX_GOTS) {
arch/parisc/kernel/module.c
878
me->name, me->arch.got_count, MAX_GOTS);
arch/parisc/kernel/module.c
882
kfree(me->arch.section);
arch/parisc/kernel/module.c
883
me->arch.section = NULL;
arch/parisc/kernel/module.c
952
mod->arch.fdesc_offset;
arch/parisc/kernel/module.c
954
mod->arch.fdesc_count * sizeof(Elf64_Fdesc);
arch/powerpc/boot/addnote.c
132
nnote = 12 + ROUNDUP(strlen(arch) + 1) + sizeof(descr);
arch/powerpc/boot/addnote.c
189
PUT_32(ns, strlen(arch) + 1);
arch/powerpc/boot/addnote.c
192
strcpy((char *) &buf[ns + 12], arch);
arch/powerpc/boot/addnote.c
193
ns += 12 + strlen(arch) + 1;
arch/powerpc/boot/addnote.c
21
static const char arch[] = "PowerPC";
arch/powerpc/include/asm/kvm_book3s.h
389
return vcpu->arch.book3s;
arch/powerpc/include/asm/kvm_book3s.h
403
vcpu->arch.regs.gpr[num] = val;
arch/powerpc/include/asm/kvm_book3s.h
410
return vcpu->arch.regs.gpr[num];
arch/powerpc/include/asm/kvm_book3s.h
415
vcpu->arch.regs.ccr = val;
arch/powerpc/include/asm/kvm_book3s.h
422
return vcpu->arch.regs.ccr;
arch/powerpc/include/asm/kvm_book3s.h
427
vcpu->arch.regs.xer = val;
arch/powerpc/include/asm/kvm_book3s.h
434
return vcpu->arch.regs.xer;
arch/powerpc/include/asm/kvm_book3s.h
439
vcpu->arch.regs.ctr = val;
arch/powerpc/include/asm/kvm_book3s.h
446
return vcpu->arch.regs.ctr;
arch/powerpc/include/asm/kvm_book3s.h
451
vcpu->arch.regs.link = val;
arch/powerpc/include/asm/kvm_book3s.h
458
return vcpu->arch.regs.link;
arch/powerpc/include/asm/kvm_book3s.h
463
vcpu->arch.regs.nip = val;
arch/powerpc/include/asm/kvm_book3s.h
470
return vcpu->arch.regs.nip;
arch/powerpc/include/asm/kvm_book3s.h
481
return vcpu->arch.fault_dar;
arch/powerpc/include/asm/kvm_book3s.h
487
return vcpu->arch.fp.fpr[i][TS_FPROFFSET];
arch/powerpc/include/asm/kvm_book3s.h
492
vcpu->arch.fp.fpr[i][TS_FPROFFSET] = val;
arch/powerpc/include/asm/kvm_book3s.h
499
return vcpu->arch.fp.fpscr;
arch/powerpc/include/asm/kvm_book3s.h
504
vcpu->arch.fp.fpscr = val;
arch/powerpc/include/asm/kvm_book3s.h
512
return vcpu->arch.fp.fpr[i][j];
arch/powerpc/include/asm/kvm_book3s.h
518
vcpu->arch.fp.fpr[i][j] = val;
arch/powerpc/include/asm/kvm_book3s.h
526
*v = vcpu->arch.vr.vr[i];
arch/powerpc/include/asm/kvm_book3s.h
532
vcpu->arch.vr.vr[i] = *val;
arch/powerpc/include/asm/kvm_book3s.h
539
return vcpu->arch.vr.vscr.u[3];
arch/powerpc/include/asm/kvm_book3s.h
544
vcpu->arch.vr.vscr.u[3] = val;
arch/powerpc/include/asm/kvm_book3s.h
553
vcpu->arch.reg = val; \
arch/powerpc/include/asm/kvm_book3s.h
561
return vcpu->arch.reg; \
arch/powerpc/include/asm/kvm_book3s.h
580
vcpu->arch.vcore->reg = val; \
arch/powerpc/include/asm/kvm_book3s.h
588
return vcpu->arch.vcore->reg; \
arch/powerpc/include/asm/kvm_book3s.h
604
return vcpu->arch.vcore->tb_offset;
arch/powerpc/include/asm/kvm_book3s.h
610
return vcpu->arch.dec_expires;
arch/powerpc/include/asm/kvm_book3s.h
615
vcpu->arch.dec_expires = val;
arch/powerpc/include/asm/kvm_book3s.h
687
int stride = kvm->arch.emul_smt_mode;
arch/powerpc/include/asm/kvm_book3s_32.h
14
return vcpu->arch.shadow_vcpu;
arch/powerpc/include/asm/kvm_book3s_64.h
128
return kvm->arch.radix;
arch/powerpc/include/asm/kvm_book3s_64.h
135
if (vcpu->arch.nested)
arch/powerpc/include/asm/kvm_book3s_64.h
136
radix = vcpu->arch.nested->radix;
arch/powerpc/include/asm/kvm_book3s_64.h
528
if (atomic_read(&kvm->arch.hpte_mod_interest))
arch/powerpc/include/asm/kvm_book3s_64.h
592
vcpu->arch.regs.ccr = vcpu->arch.cr_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
593
vcpu->arch.regs.xer = vcpu->arch.xer_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
594
vcpu->arch.regs.link = vcpu->arch.lr_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
595
vcpu->arch.regs.ctr = vcpu->arch.ctr_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
596
vcpu->arch.amr = vcpu->arch.amr_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
597
vcpu->arch.ppr = vcpu->arch.ppr_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
598
vcpu->arch.dscr = vcpu->arch.dscr_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
599
vcpu->arch.tar = vcpu->arch.tar_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
600
memcpy(vcpu->arch.regs.gpr, vcpu->arch.gpr_tm,
arch/powerpc/include/asm/kvm_book3s_64.h
601
sizeof(vcpu->arch.regs.gpr));
arch/powerpc/include/asm/kvm_book3s_64.h
602
vcpu->arch.fp = vcpu->arch.fp_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
603
vcpu->arch.vr = vcpu->arch.vr_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
604
vcpu->arch.vrsave = vcpu->arch.vrsave_tm;
arch/powerpc/include/asm/kvm_book3s_64.h
609
vcpu->arch.cr_tm = vcpu->arch.regs.ccr;
arch/powerpc/include/asm/kvm_book3s_64.h
610
vcpu->arch.xer_tm = vcpu->arch.regs.xer;
arch/powerpc/include/asm/kvm_book3s_64.h
611
vcpu->arch.lr_tm = vcpu->arch.regs.link;
arch/powerpc/include/asm/kvm_book3s_64.h
612
vcpu->arch.ctr_tm = vcpu->arch.regs.ctr;
arch/powerpc/include/asm/kvm_book3s_64.h
613
vcpu->arch.amr_tm = vcpu->arch.amr;
arch/powerpc/include/asm/kvm_book3s_64.h
614
vcpu->arch.ppr_tm = vcpu->arch.ppr;
arch/powerpc/include/asm/kvm_book3s_64.h
615
vcpu->arch.dscr_tm = vcpu->arch.dscr;
arch/powerpc/include/asm/kvm_book3s_64.h
616
vcpu->arch.tar_tm = vcpu->arch.tar;
arch/powerpc/include/asm/kvm_book3s_64.h
617
memcpy(vcpu->arch.gpr_tm, vcpu->arch.regs.gpr,
arch/powerpc/include/asm/kvm_book3s_64.h
618
sizeof(vcpu->arch.regs.gpr));
arch/powerpc/include/asm/kvm_book3s_64.h
619
vcpu->arch.fp_tm = vcpu->arch.fp;
arch/powerpc/include/asm/kvm_book3s_64.h
620
vcpu->arch.vr_tm = vcpu->arch.vr;
arch/powerpc/include/asm/kvm_book3s_64.h
621
vcpu->arch.vrsave_tm = vcpu->arch.vrsave;
arch/powerpc/include/asm/kvm_book3s_64.h
645
pte = __find_linux_pte(kvm->arch.pgtable, ea, NULL, hshift);
arch/powerpc/include/asm/kvm_book3s_64.h
656
pte = __find_linux_pte(kvm->arch.pgtable, ea, NULL, hshift);
arch/powerpc/include/asm/kvm_booke.h
105
return vcpu->arch.fault_dear;
arch/powerpc/include/asm/kvm_booke.h
28
vcpu->arch.regs.gpr[num] = val;
arch/powerpc/include/asm/kvm_booke.h
33
return vcpu->arch.regs.gpr[num];
arch/powerpc/include/asm/kvm_booke.h
38
vcpu->arch.regs.ccr = val;
arch/powerpc/include/asm/kvm_booke.h
43
return vcpu->arch.regs.ccr;
arch/powerpc/include/asm/kvm_booke.h
48
vcpu->arch.regs.xer = val;
arch/powerpc/include/asm/kvm_booke.h
53
return vcpu->arch.regs.xer;
arch/powerpc/include/asm/kvm_booke.h
64
vcpu->arch.regs.ctr = val;
arch/powerpc/include/asm/kvm_booke.h
69
return vcpu->arch.regs.ctr;
arch/powerpc/include/asm/kvm_booke.h
74
vcpu->arch.regs.link = val;
arch/powerpc/include/asm/kvm_booke.h
79
return vcpu->arch.regs.link;
arch/powerpc/include/asm/kvm_booke.h
84
vcpu->arch.regs.nip = val;
arch/powerpc/include/asm/kvm_booke.h
89
return vcpu->arch.regs.nip;
arch/powerpc/include/asm/kvm_booke.h
94
vcpu->arch.fp.fpr[i][TS_FPROFFSET] = val;
arch/powerpc/include/asm/kvm_booke.h
99
return vcpu->arch.fp.fpr[i][TS_FPROFFSET];
arch/powerpc/include/asm/kvm_host.h
881
#define VCPU_FPR(vcpu, i) (vcpu)->arch.fp.fpr[i][TS_FPROFFSET]
arch/powerpc/include/asm/kvm_host.h
882
#define VCPU_VSX_FPR(vcpu, i, j) ((vcpu)->arch.fp.fpr[i][j])
arch/powerpc/include/asm/kvm_host.h
883
#define VCPU_VSX_VR(vcpu, i) ((vcpu)->arch.vr.vr[i])
arch/powerpc/include/asm/kvm_ppc.h
1033
vcpu->arch.shared->msr = cpu_to_be64(val);
arch/powerpc/include/asm/kvm_ppc.h
1035
vcpu->arch.shared->msr = cpu_to_le64(val);
arch/powerpc/include/asm/kvm_ppc.h
1048
return be32_to_cpu(vcpu->arch.shared->sr[nr]);
arch/powerpc/include/asm/kvm_ppc.h
1050
return le32_to_cpu(vcpu->arch.shared->sr[nr]);
arch/powerpc/include/asm/kvm_ppc.h
1056
vcpu->arch.shared->sr[nr] = cpu_to_be32(val);
arch/powerpc/include/asm/kvm_ppc.h
1058
vcpu->arch.shared->sr[nr] = cpu_to_le32(val);
arch/powerpc/include/asm/kvm_ppc.h
335
if (vcpu->arch.last_inst == KVM_INST_FETCH_FAILED)
arch/powerpc/include/asm/kvm_ppc.h
336
ret = kvmppc_load_last_inst(vcpu, type, &vcpu->arch.last_inst);
arch/powerpc/include/asm/kvm_ppc.h
346
if ((vcpu->arch.last_inst >> 32) != 0) {
arch/powerpc/include/asm/kvm_ppc.h
347
u32 prefix = vcpu->arch.last_inst >> 32;
arch/powerpc/include/asm/kvm_ppc.h
348
u32 suffix = vcpu->arch.last_inst;
arch/powerpc/include/asm/kvm_ppc.h
359
swab32(vcpu->arch.last_inst) :
arch/powerpc/include/asm/kvm_ppc.h
360
vcpu->arch.last_inst;
arch/powerpc/include/asm/kvm_ppc.h
367
return kvm->arch.kvm_ops == kvmppc_hv_ops;
arch/powerpc/include/asm/kvm_ppc.h
566
vcpu->kvm->arch.kvm_ops->fast_vcpu_kick(vcpu);
arch/powerpc/include/asm/kvm_ppc.h
658
return vcpu->arch.irq_type == KVMPPC_IRQ_XICS;
arch/powerpc/include/asm/kvm_ppc.h
665
return kvm->arch.pimap;
arch/powerpc/include/asm/kvm_ppc.h
746
return vcpu->arch.irq_type == KVMPPC_IRQ_XIVE;
arch/powerpc/include/asm/kvm_ppc.h
880
return vcpu->arch.epr;
arch/powerpc/include/asm/kvm_ppc.h
891
vcpu->arch.epr = epr;
arch/powerpc/include/asm/kvm_ppc.h
956
return vcpu->arch.shared_big_endian;
arch/powerpc/include/asm/kvm_ppc.h
983
return be##size##_to_cpu((__be##size __force)vcpu->arch.shared->reg); \
arch/powerpc/include/asm/kvm_ppc.h
985
return le##size##_to_cpu((__le##size __force)vcpu->arch.shared->reg); \
arch/powerpc/include/asm/kvm_ppc.h
992
vcpu->arch.shared->reg = (u##size __force)cpu_to_be##size(val); \
arch/powerpc/include/asm/kvm_ppc.h
994
vcpu->arch.shared->reg = (u##size __force)cpu_to_le##size(val); \
arch/powerpc/kernel/asm-offsets.c
355
OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack);
arch/powerpc/kernel/asm-offsets.c
356
OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid);
arch/powerpc/kernel/asm-offsets.c
357
OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid);
arch/powerpc/kernel/asm-offsets.c
358
OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr);
arch/powerpc/kernel/asm-offsets.c
359
OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave);
arch/powerpc/kernel/asm-offsets.c
360
OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr);
arch/powerpc/kernel/asm-offsets.c
362
OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr);
arch/powerpc/kernel/asm-offsets.c
364
OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer);
arch/powerpc/kernel/asm-offsets.c
365
OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr);
arch/powerpc/kernel/asm-offsets.c
366
OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link);
arch/powerpc/kernel/asm-offsets.c
368
OFFSET(VCPU_TAR, kvm_vcpu, arch.tar);
arch/powerpc/kernel/asm-offsets.c
370
OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr);
arch/powerpc/kernel/asm-offsets.c
371
OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip);
arch/powerpc/kernel/asm-offsets.c
373
OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr);
arch/powerpc/kernel/asm-offsets.c
374
OFFSET(VCPU_SRR0, kvm_vcpu, arch.shregs.srr0);
arch/powerpc/kernel/asm-offsets.c
375
OFFSET(VCPU_SRR1, kvm_vcpu, arch.shregs.srr1);
arch/powerpc/kernel/asm-offsets.c
376
OFFSET(VCPU_SPRG0, kvm_vcpu, arch.shregs.sprg0);
arch/powerpc/kernel/asm-offsets.c
377
OFFSET(VCPU_SPRG1, kvm_vcpu, arch.shregs.sprg1);
arch/powerpc/kernel/asm-offsets.c
378
OFFSET(VCPU_SPRG2, kvm_vcpu, arch.shregs.sprg2);
arch/powerpc/kernel/asm-offsets.c
379
OFFSET(VCPU_SPRG3, kvm_vcpu, arch.shregs.sprg3);
arch/powerpc/kernel/asm-offsets.c
382
OFFSET(VCPU_TB_RMENTRY, kvm_vcpu, arch.rm_entry);
arch/powerpc/kernel/asm-offsets.c
383
OFFSET(VCPU_TB_RMINTR, kvm_vcpu, arch.rm_intr);
arch/powerpc/kernel/asm-offsets.c
384
OFFSET(VCPU_TB_RMEXIT, kvm_vcpu, arch.rm_exit);
arch/powerpc/kernel/asm-offsets.c
385
OFFSET(VCPU_TB_GUEST, kvm_vcpu, arch.guest_time);
arch/powerpc/kernel/asm-offsets.c
386
OFFSET(VCPU_TB_CEDE, kvm_vcpu, arch.cede_time);
arch/powerpc/kernel/asm-offsets.c
387
OFFSET(VCPU_CUR_ACTIVITY, kvm_vcpu, arch.cur_activity);
arch/powerpc/kernel/asm-offsets.c
388
OFFSET(VCPU_ACTIVITY_START, kvm_vcpu, arch.cur_tb_start);
arch/powerpc/kernel/asm-offsets.c
399
OFFSET(VCPU_SHADOW_PID, kvm_vcpu, arch.shadow_pid);
arch/powerpc/kernel/asm-offsets.c
400
OFFSET(VCPU_SHADOW_PID1, kvm_vcpu, arch.shadow_pid1);
arch/powerpc/kernel/asm-offsets.c
401
OFFSET(VCPU_SHARED, kvm_vcpu, arch.shared);
arch/powerpc/kernel/asm-offsets.c
403
OFFSET(VCPU_SHADOW_MSR, kvm_vcpu, arch.shadow_msr);
arch/powerpc/kernel/asm-offsets.c
405
OFFSET(VCPU_SHAREDBE, kvm_vcpu, arch.shared_big_endian);
arch/powerpc/kernel/asm-offsets.c
416
OFFSET(KVM_LPID, kvm, arch.lpid);
arch/powerpc/kernel/asm-offsets.c
420
OFFSET(KVM_SDR1, kvm, arch.sdr1);
arch/powerpc/kernel/asm-offsets.c
421
OFFSET(KVM_HOST_LPID, kvm, arch.host_lpid);
arch/powerpc/kernel/asm-offsets.c
422
OFFSET(KVM_HOST_LPCR, kvm, arch.host_lpcr);
arch/powerpc/kernel/asm-offsets.c
423
OFFSET(KVM_HOST_SDR1, kvm, arch.host_sdr1);
arch/powerpc/kernel/asm-offsets.c
424
OFFSET(KVM_ENABLED_HCALLS, kvm, arch.enabled_hcalls);
arch/powerpc/kernel/asm-offsets.c
425
OFFSET(KVM_VRMA_SLB_V, kvm, arch.vrma_slb_v);
arch/powerpc/kernel/asm-offsets.c
426
OFFSET(KVM_SECURE_GUEST, kvm, arch.secure_guest);
arch/powerpc/kernel/asm-offsets.c
427
OFFSET(VCPU_DSISR, kvm_vcpu, arch.shregs.dsisr);
arch/powerpc/kernel/asm-offsets.c
428
OFFSET(VCPU_DAR, kvm_vcpu, arch.shregs.dar);
arch/powerpc/kernel/asm-offsets.c
429
OFFSET(VCPU_VPA, kvm_vcpu, arch.vpa.pinned_addr);
arch/powerpc/kernel/asm-offsets.c
430
OFFSET(VCPU_VPA_DIRTY, kvm_vcpu, arch.vpa.dirty);
arch/powerpc/kernel/asm-offsets.c
431
OFFSET(VCPU_HEIR, kvm_vcpu, arch.emul_inst);
arch/powerpc/kernel/asm-offsets.c
433
OFFSET(VCPU_THREAD_CPU, kvm_vcpu, arch.thread_cpu);
arch/powerpc/kernel/asm-offsets.c
436
OFFSET(VCPU_PURR, kvm_vcpu, arch.purr);
arch/powerpc/kernel/asm-offsets.c
437
OFFSET(VCPU_SPURR, kvm_vcpu, arch.spurr);
arch/powerpc/kernel/asm-offsets.c
438
OFFSET(VCPU_IC, kvm_vcpu, arch.ic);
arch/powerpc/kernel/asm-offsets.c
439
OFFSET(VCPU_DSCR, kvm_vcpu, arch.dscr);
arch/powerpc/kernel/asm-offsets.c
440
OFFSET(VCPU_AMR, kvm_vcpu, arch.amr);
arch/powerpc/kernel/asm-offsets.c
441
OFFSET(VCPU_UAMOR, kvm_vcpu, arch.uamor);
arch/powerpc/kernel/asm-offsets.c
442
OFFSET(VCPU_IAMR, kvm_vcpu, arch.iamr);
arch/powerpc/kernel/asm-offsets.c
443
OFFSET(VCPU_CTRL, kvm_vcpu, arch.ctrl);
arch/powerpc/kernel/asm-offsets.c
444
OFFSET(VCPU_DABR, kvm_vcpu, arch.dabr);
arch/powerpc/kernel/asm-offsets.c
445
OFFSET(VCPU_DABRX, kvm_vcpu, arch.dabrx);
arch/powerpc/kernel/asm-offsets.c
446
OFFSET(VCPU_DAWR0, kvm_vcpu, arch.dawr0);
arch/powerpc/kernel/asm-offsets.c
447
OFFSET(VCPU_DAWRX0, kvm_vcpu, arch.dawrx0);
arch/powerpc/kernel/asm-offsets.c
448
OFFSET(VCPU_CIABR, kvm_vcpu, arch.ciabr);
arch/powerpc/kernel/asm-offsets.c
449
OFFSET(VCPU_HFLAGS, kvm_vcpu, arch.hflags);
arch/powerpc/kernel/asm-offsets.c
450
OFFSET(VCPU_DEC_EXPIRES, kvm_vcpu, arch.dec_expires);
arch/powerpc/kernel/asm-offsets.c
451
OFFSET(VCPU_PENDING_EXC, kvm_vcpu, arch.pending_exceptions);
arch/powerpc/kernel/asm-offsets.c
452
OFFSET(VCPU_CEDED, kvm_vcpu, arch.ceded);
arch/powerpc/kernel/asm-offsets.c
453
OFFSET(VCPU_PRODDED, kvm_vcpu, arch.prodded);
arch/powerpc/kernel/asm-offsets.c
454
OFFSET(VCPU_MMCR, kvm_vcpu, arch.mmcr);
arch/powerpc/kernel/asm-offsets.c
455
OFFSET(VCPU_MMCRA, kvm_vcpu, arch.mmcra);
arch/powerpc/kernel/asm-offsets.c
456
OFFSET(VCPU_MMCRS, kvm_vcpu, arch.mmcrs);
arch/powerpc/kernel/asm-offsets.c
457
OFFSET(VCPU_PMC, kvm_vcpu, arch.pmc);
arch/powerpc/kernel/asm-offsets.c
458
OFFSET(VCPU_SIAR, kvm_vcpu, arch.siar);
arch/powerpc/kernel/asm-offsets.c
459
OFFSET(VCPU_SDAR, kvm_vcpu, arch.sdar);
arch/powerpc/kernel/asm-offsets.c
460
OFFSET(VCPU_SIER, kvm_vcpu, arch.sier);
arch/powerpc/kernel/asm-offsets.c
461
OFFSET(VCPU_SLB, kvm_vcpu, arch.slb);
arch/powerpc/kernel/asm-offsets.c
462
OFFSET(VCPU_SLB_MAX, kvm_vcpu, arch.slb_max);
arch/powerpc/kernel/asm-offsets.c
463
OFFSET(VCPU_SLB_NR, kvm_vcpu, arch.slb_nr);
arch/powerpc/kernel/asm-offsets.c
464
OFFSET(VCPU_FAULT_DSISR, kvm_vcpu, arch.fault_dsisr);
arch/powerpc/kernel/asm-offsets.c
465
OFFSET(VCPU_FAULT_DAR, kvm_vcpu, arch.fault_dar);
arch/powerpc/kernel/asm-offsets.c
466
OFFSET(VCPU_INTR_MSR, kvm_vcpu, arch.intr_msr);
arch/powerpc/kernel/asm-offsets.c
467
OFFSET(VCPU_LAST_INST, kvm_vcpu, arch.last_inst);
arch/powerpc/kernel/asm-offsets.c
468
OFFSET(VCPU_TRAP, kvm_vcpu, arch.trap);
arch/powerpc/kernel/asm-offsets.c
469
OFFSET(VCPU_CFAR, kvm_vcpu, arch.cfar);
arch/powerpc/kernel/asm-offsets.c
470
OFFSET(VCPU_PPR, kvm_vcpu, arch.ppr);
arch/powerpc/kernel/asm-offsets.c
471
OFFSET(VCPU_FSCR, kvm_vcpu, arch.fscr);
arch/powerpc/kernel/asm-offsets.c
472
OFFSET(VCPU_PSPB, kvm_vcpu, arch.pspb);
arch/powerpc/kernel/asm-offsets.c
473
OFFSET(VCPU_EBBHR, kvm_vcpu, arch.ebbhr);
arch/powerpc/kernel/asm-offsets.c
474
OFFSET(VCPU_EBBRR, kvm_vcpu, arch.ebbrr);
arch/powerpc/kernel/asm-offsets.c
475
OFFSET(VCPU_BESCR, kvm_vcpu, arch.bescr);
arch/powerpc/kernel/asm-offsets.c
476
OFFSET(VCPU_CSIGR, kvm_vcpu, arch.csigr);
arch/powerpc/kernel/asm-offsets.c
477
OFFSET(VCPU_TACR, kvm_vcpu, arch.tacr);
arch/powerpc/kernel/asm-offsets.c
478
OFFSET(VCPU_TCSCR, kvm_vcpu, arch.tcscr);
arch/powerpc/kernel/asm-offsets.c
479
OFFSET(VCPU_ACOP, kvm_vcpu, arch.acop);
arch/powerpc/kernel/asm-offsets.c
480
OFFSET(VCPU_WORT, kvm_vcpu, arch.wort);
arch/powerpc/kernel/asm-offsets.c
481
OFFSET(VCPU_HFSCR, kvm_vcpu, arch.hfscr);
arch/powerpc/kernel/asm-offsets.c
496
OFFSET(VCPU_TFHAR, kvm_vcpu, arch.tfhar);
arch/powerpc/kernel/asm-offsets.c
497
OFFSET(VCPU_TFIAR, kvm_vcpu, arch.tfiar);
arch/powerpc/kernel/asm-offsets.c
498
OFFSET(VCPU_TEXASR, kvm_vcpu, arch.texasr);
arch/powerpc/kernel/asm-offsets.c
499
OFFSET(VCPU_ORIG_TEXASR, kvm_vcpu, arch.orig_texasr);
arch/powerpc/kernel/asm-offsets.c
500
OFFSET(VCPU_GPR_TM, kvm_vcpu, arch.gpr_tm);
arch/powerpc/kernel/asm-offsets.c
501
OFFSET(VCPU_FPRS_TM, kvm_vcpu, arch.fp_tm.fpr);
arch/powerpc/kernel/asm-offsets.c
502
OFFSET(VCPU_VRS_TM, kvm_vcpu, arch.vr_tm.vr);
arch/powerpc/kernel/asm-offsets.c
503
OFFSET(VCPU_VRSAVE_TM, kvm_vcpu, arch.vrsave_tm);
arch/powerpc/kernel/asm-offsets.c
504
OFFSET(VCPU_CR_TM, kvm_vcpu, arch.cr_tm);
arch/powerpc/kernel/asm-offsets.c
505
OFFSET(VCPU_XER_TM, kvm_vcpu, arch.xer_tm);
arch/powerpc/kernel/asm-offsets.c
506
OFFSET(VCPU_LR_TM, kvm_vcpu, arch.lr_tm);
arch/powerpc/kernel/asm-offsets.c
507
OFFSET(VCPU_CTR_TM, kvm_vcpu, arch.ctr_tm);
arch/powerpc/kernel/asm-offsets.c
508
OFFSET(VCPU_AMR_TM, kvm_vcpu, arch.amr_tm);
arch/powerpc/kernel/asm-offsets.c
509
OFFSET(VCPU_PPR_TM, kvm_vcpu, arch.ppr_tm);
arch/powerpc/kernel/asm-offsets.c
510
OFFSET(VCPU_DSCR_TM, kvm_vcpu, arch.dscr_tm);
arch/powerpc/kernel/asm-offsets.c
511
OFFSET(VCPU_TAR_TM, kvm_vcpu, arch.tar_tm);
arch/powerpc/kernel/asm-offsets.c
611
OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr);
arch/powerpc/kernel/asm-offsets.c
612
OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer);
arch/powerpc/kernel/asm-offsets.c
613
OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link);
arch/powerpc/kernel/asm-offsets.c
614
OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr);
arch/powerpc/kernel/asm-offsets.c
615
OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip);
arch/powerpc/kernel/asm-offsets.c
616
OFFSET(VCPU_SPRG9, kvm_vcpu, arch.sprg9);
arch/powerpc/kernel/asm-offsets.c
617
OFFSET(VCPU_LAST_INST, kvm_vcpu, arch.last_inst);
arch/powerpc/kernel/asm-offsets.c
618
OFFSET(VCPU_FAULT_DEAR, kvm_vcpu, arch.fault_dear);
arch/powerpc/kernel/asm-offsets.c
619
OFFSET(VCPU_FAULT_ESR, kvm_vcpu, arch.fault_esr);
arch/powerpc/kernel/asm-offsets.c
620
OFFSET(VCPU_CRIT_SAVE, kvm_vcpu, arch.crit_save);
arch/powerpc/kernel/asm-offsets.c
648
OFFSET(VCPU_EVR, kvm_vcpu, arch.evr[0]);
arch/powerpc/kernel/asm-offsets.c
649
OFFSET(VCPU_ACC, kvm_vcpu, arch.acc);
arch/powerpc/kernel/asm-offsets.c
650
OFFSET(VCPU_SPEFSCR, kvm_vcpu, arch.spefscr);
arch/powerpc/kernel/asm-offsets.c
651
OFFSET(VCPU_HOST_SPEFSCR, kvm_vcpu, arch.host_spefscr);
arch/powerpc/kernel/asm-offsets.c
655
OFFSET(VCPU_HOST_MAS4, kvm_vcpu, arch.host_mas4);
arch/powerpc/kernel/asm-offsets.c
656
OFFSET(VCPU_HOST_MAS6, kvm_vcpu, arch.host_mas6);
arch/powerpc/kernel/asm-offsets.c
660
OFFSET(VCPU_TIMING_EXIT_TBU, kvm_vcpu, arch.timing_exit.tv32.tbu);
arch/powerpc/kernel/asm-offsets.c
661
OFFSET(VCPU_TIMING_EXIT_TBL, kvm_vcpu, arch.timing_exit.tv32.tbl);
arch/powerpc/kernel/asm-offsets.c
662
OFFSET(VCPU_TIMING_LAST_ENTER_TBU, kvm_vcpu, arch.timing_last_enter.tv32.tbu);
arch/powerpc/kernel/asm-offsets.c
663
OFFSET(VCPU_TIMING_LAST_ENTER_TBL, kvm_vcpu, arch.timing_last_enter.tv32.tbl);
arch/powerpc/kernel/module.c
67
me->arch.start_opd = sect->sh_addr;
arch/powerpc/kernel/module.c
68
me->arch.end_opd = sect->sh_addr + sect->sh_size;
arch/powerpc/kernel/module_32.c
130
me->arch.init_plt_section = i;
arch/powerpc/kernel/module_32.c
132
me->arch.core_plt_section = i;
arch/powerpc/kernel/module_32.c
134
if (!me->arch.core_plt_section || !me->arch.init_plt_section) {
arch/powerpc/kernel/module_32.c
140
sechdrs[me->arch.core_plt_section].sh_size
arch/powerpc/kernel/module_32.c
142
sechdrs[me->arch.init_plt_section].sh_size
arch/powerpc/kernel/module_32.c
167
entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
arch/powerpc/kernel/module_32.c
169
entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
arch/powerpc/kernel/module_32.c
324
module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
arch/powerpc/kernel/module_32.c
327
if (!module->arch.tramp)
arch/powerpc/kernel/module_32.c
331
module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
arch/powerpc/kernel/module_32.c
334
if (!module->arch.tramp_regs)
arch/powerpc/kernel/module_64.c
1124
total_stubs = sechdrs[me->arch.stubs_section].sh_size / sizeof(*stub);
arch/powerpc/kernel/module_64.c
1125
num_stubs = roundup(me->arch.ool_stub_count * sizeof(struct ftrace_ool_stub),
arch/powerpc/kernel/module_64.c
1128
if (WARN_ON(me->arch.stub_count + num_stubs > total_stubs))
arch/powerpc/kernel/module_64.c
1131
stub = (void *)sechdrs[me->arch.stubs_section].sh_addr;
arch/powerpc/kernel/module_64.c
1132
me->arch.ool_stubs = (struct ftrace_ool_stub *)(stub + me->arch.stub_count);
arch/powerpc/kernel/module_64.c
1133
me->arch.stub_count += num_stubs;
arch/powerpc/kernel/module_64.c
1141
mod->arch.tramp = stub_for_addr(sechdrs,
arch/powerpc/kernel/module_64.c
1146
mod->arch.tramp_regs = stub_for_addr(sechdrs,
arch/powerpc/kernel/module_64.c
1150
if (!mod->arch.tramp_regs)
arch/powerpc/kernel/module_64.c
1154
if (!mod->arch.tramp)
arch/powerpc/kernel/module_64.c
1157
if (setup_ftrace_ool_stubs(sechdrs, mod->arch.tramp, mod))
arch/powerpc/kernel/module_64.c
252
me->arch.ool_stub_count = sechdrs[i].sh_size / sizeof(unsigned long);
arch/powerpc/kernel/module_64.c
253
me->arch.ool_stub_index = 0;
arch/powerpc/kernel/module_64.c
254
relocs += roundup(me->arch.ool_stub_count * sizeof(struct ftrace_ool_stub),
arch/powerpc/kernel/module_64.c
346
symindex, me->arch.pcpu_section);
arch/powerpc/kernel/module_64.c
440
me->arch.stubs_section = i;
arch/powerpc/kernel/module_64.c
443
me->arch.pcpu_section = i;
arch/powerpc/kernel/module_64.c
445
me->arch.got_section = i;
arch/powerpc/kernel/module_64.c
451
me->arch.toc_section = i;
arch/powerpc/kernel/module_64.c
469
if (!me->arch.stubs_section) {
arch/powerpc/kernel/module_64.c
475
if (!me->arch.got_section) {
arch/powerpc/kernel/module_64.c
481
sechdrs[me->arch.got_section].sh_size = get_got_size(hdr, sechdrs, me);
arch/powerpc/kernel/module_64.c
487
if (!me->arch.toc_section)
arch/powerpc/kernel/module_64.c
488
me->arch.toc_section = me->arch.stubs_section;
arch/powerpc/kernel/module_64.c
492
sechdrs[me->arch.stubs_section].sh_size = get_stubs_size(hdr, sechdrs, secstrings, me);
arch/powerpc/kernel/module_64.c
612
return (sechdrs[me->arch.toc_section].sh_addr & ~0xfful) + 0x8000;
arch/powerpc/kernel/module_64.c
703
num_stubs = sechdrs[me->arch.stubs_section].sh_size / sizeof(*stubs);
arch/powerpc/kernel/module_64.c
706
stubs = (void *)sechdrs[me->arch.stubs_section].sh_addr;
arch/powerpc/kernel/module_64.c
707
for (i = 0; i < me->arch.stub_count; i++) {
arch/powerpc/kernel/module_64.c
718
me->arch.stub_count++;
arch/powerpc/kernel/module_64.c
735
num_got = sechdrs[me->arch.got_section].sh_size / sizeof(*got);
arch/powerpc/kernel/module_64.c
738
got = (void *)sechdrs[me->arch.got_section].sh_addr;
arch/powerpc/kernel/module_64.c
81
if (ptr < (void *)mod->arch.start_opd ||
arch/powerpc/kernel/module_64.c
810
if (!me->arch.toc_fixed) {
arch/powerpc/kernel/module_64.c
816
me->arch.toc_fixed = true;
arch/powerpc/kernel/module_64.c
82
ptr >= (void *)mod->arch.end_opd)
arch/powerpc/kernel/module_64.c
968
if (sym->st_shndx != me->arch.pcpu_section) {
arch/powerpc/kernel/trace/ftrace.c
141
return (addr == (unsigned long)ftrace_caller ? mod->arch.tramp : mod->arch.tramp_regs);
arch/powerpc/kernel/trace/ftrace.c
153
return rec->arch.ool_stub;
arch/powerpc/kernel/trace/ftrace.c
217
WARN_ON(rec->arch.ool_stub);
arch/powerpc/kernel/trace/ftrace.c
242
ool_stub = mod->arch.ool_stubs;
arch/powerpc/kernel/trace/ftrace.c
243
ool_stub_index = &mod->arch.ool_stub_index;
arch/powerpc/kernel/trace/ftrace.c
244
ool_stub_count = mod->arch.ool_stub_count;
arch/powerpc/kernel/trace/ftrace.c
263
rec->arch.ool_stub = (unsigned long)&ool_stub->insn[0];
arch/powerpc/kernel/trace/ftrace.c
278
mod->arch.tramp, 1);
arch/powerpc/kernel/trace/ftrace_64_pg.c
431
if (!mod->arch.tramp ||
arch/powerpc/kernel/trace/ftrace_64_pg.c
432
(IS_ENABLED(CONFIG_DYNAMIC_FTRACE_WITH_REGS) && !mod->arch.tramp_regs)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
438
tramp = mod->arch.tramp_regs;
arch/powerpc/kernel/trace/ftrace_64_pg.c
440
tramp = mod->arch.tramp;
arch/powerpc/kernel/trace/ftrace_64_pg.c
554
if (!mod->arch.tramp || !mod->arch.tramp_regs) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
605
tramp = mod->arch.tramp_regs;
arch/powerpc/kernel/trace/ftrace_64_pg.c
607
tramp = mod->arch.tramp;
arch/powerpc/kexec/crash.c
429
image->arch.backup_start = phdr->p_offset;
arch/powerpc/kexec/crash.c
431
phdr->p_offset = image->arch.backup_start;
arch/powerpc/kexec/crash.c
434
image->arch.backup_start);
arch/powerpc/kexec/elf_64.c
141
image->arch.fdt = fdt;
arch/powerpc/kexec/file_load_64.c
372
image->arch.backup_buf = buf;
arch/powerpc/kexec/file_load_64.c
373
image->arch.backup_start = kbuf->mem;
arch/powerpc/kexec/file_load_64.c
530
&image->arch.backup_start,
arch/powerpc/kexec/file_load_64.c
531
sizeof(image->arch.backup_start),
arch/powerpc/kexec/file_load_64.c
58
emem = image->arch.exclude_ranges;
arch/powerpc/kexec/file_load_64.c
762
ret = fdt_add_mem_rsv(fdt, image->arch.backup_start,
arch/powerpc/kexec/file_load_64.c
824
ret = get_exclude_memory_ranges(&(image->arch.exclude_ranges));
arch/powerpc/kexec/file_load_64.c
842
kfree(image->arch.exclude_ranges);
arch/powerpc/kexec/file_load_64.c
843
image->arch.exclude_ranges = NULL;
arch/powerpc/kexec/file_load_64.c
845
vfree(image->arch.backup_buf);
arch/powerpc/kexec/file_load_64.c
846
image->arch.backup_buf = NULL;
arch/powerpc/kexec/file_load_64.c
852
kvfree(image->arch.fdt);
arch/powerpc/kexec/file_load_64.c
853
image->arch.fdt = NULL;
arch/powerpc/kvm/book3s.c
1018
return kvm->arch.kvm_ops->hcall_implemented(hcall);
arch/powerpc/kvm/book3s.c
137
vcpu->kvm->arch.kvm_ops->inject_interrupt(vcpu, vec, flags);
arch/powerpc/kvm/book3s.c
170
unsigned long old_pending = vcpu->arch.pending_exceptions;
arch/powerpc/kvm/book3s.c
173
&vcpu->arch.pending_exceptions);
arch/powerpc/kvm/book3s.c
175
kvmppc_update_int_pending(vcpu, vcpu->arch.pending_exceptions,
arch/powerpc/kvm/book3s.c
184
&vcpu->arch.pending_exceptions);
arch/powerpc/kvm/book3s.c
237
return test_bit(BOOK3S_IRQPRIO_DECREMENTER, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/book3s.c
271
vcpu->arch.external_oneshot = 1;
arch/powerpc/kvm/book3s.c
384
if (vcpu->arch.external_oneshot) {
arch/powerpc/kvm/book3s.c
385
vcpu->arch.external_oneshot = 0;
arch/powerpc/kvm/book3s.c
396
unsigned long *pending = &vcpu->arch.pending_exceptions;
arch/powerpc/kvm/book3s.c
397
unsigned long old_pending = vcpu->arch.pending_exceptions;
arch/powerpc/kvm/book3s.c
401
if (vcpu->arch.pending_exceptions)
arch/powerpc/kvm/book3s.c
402
printk(KERN_EMERG "KVM: Check pending: %lx\n", vcpu->arch.pending_exceptions);
arch/powerpc/kvm/book3s.c
408
clear_bit(priority, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/book3s.c
427
ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM;
arch/powerpc/kvm/book3s.c
436
ulong shared_page = ((ulong)vcpu->arch.shared) & PAGE_MASK;
arch/powerpc/kvm/book3s.c
460
r = vcpu->arch.mmu.xlate(vcpu, eaddr, pte, data, iswrite);
arch/powerpc/kvm/book3s.c
472
if ((vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) &&
arch/powerpc/kvm/book3s.c
535
ret = vcpu->kvm->arch.kvm_ops->get_sregs(vcpu, sregs);
arch/powerpc/kvm/book3s.c
547
ret = vcpu->kvm->arch.kvm_ops->set_sregs(vcpu, sregs);
arch/powerpc/kvm/book3s.c
624
r = vcpu->kvm->arch.kvm_ops->get_one_reg(vcpu, id, val);
arch/powerpc/kvm/book3s.c
657
if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) {
arch/powerpc/kvm/book3s.c
669
if (!vcpu->arch.xive_vcpu) {
arch/powerpc/kvm/book3s.c
680
*val = get_reg_val(id, vcpu->arch.fscr);
arch/powerpc/kvm/book3s.c
712
r = vcpu->kvm->arch.kvm_ops->set_one_reg(vcpu, id, val);
arch/powerpc/kvm/book3s.c
727
vcpu->arch.fp.fpscr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s.c
742
if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) {
arch/powerpc/kvm/book3s.c
754
if (!vcpu->arch.xive_vcpu) {
arch/powerpc/kvm/book3s.c
793
vcpu->kvm->arch.kvm_ops->vcpu_load(vcpu, cpu);
arch/powerpc/kvm/book3s.c
798
vcpu->kvm->arch.kvm_ops->vcpu_put(vcpu);
arch/powerpc/kvm/book3s.c
803
vcpu->kvm->arch.kvm_ops->set_msr(vcpu, msr);
arch/powerpc/kvm/book3s.c
809
return vcpu->kvm->arch.kvm_ops->vcpu_run(vcpu);
arch/powerpc/kvm/book3s.c
835
return vcpu->kvm->arch.kvm_ops->vcpu_create(vcpu);
arch/powerpc/kvm/book3s.c
840
vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu);
arch/powerpc/kvm/book3s.c
845
return vcpu->kvm->arch.kvm_ops->check_requests(vcpu);
arch/powerpc/kvm/book3s.c
855
return kvm->arch.kvm_ops->get_dirty_log(kvm, log);
arch/powerpc/kvm/book3s.c
860
kvm->arch.kvm_ops->free_memslot(slot);
arch/powerpc/kvm/book3s.c
865
kvm->arch.kvm_ops->flush_memslot(kvm, memslot);
arch/powerpc/kvm/book3s.c
873
return kvm->arch.kvm_ops->prepare_memory_region(kvm, old, new, change);
arch/powerpc/kvm/book3s.c
881
kvm->arch.kvm_ops->commit_memory_region(kvm, old, new, change);
arch/powerpc/kvm/book3s.c
886
return kvm->arch.kvm_ops->unmap_gfn_range(kvm, range);
arch/powerpc/kvm/book3s.c
891
return kvm->arch.kvm_ops->age_gfn(kvm, range);
arch/powerpc/kvm/book3s.c
896
return kvm->arch.kvm_ops->test_age_gfn(kvm, range);
arch/powerpc/kvm/book3s.c
903
INIT_LIST_HEAD_RCU(&kvm->arch.spapr_tce_tables);
arch/powerpc/kvm/book3s.c
904
INIT_LIST_HEAD(&kvm->arch.rtas_tokens);
arch/powerpc/kvm/book3s.c
905
mutex_init(&kvm->arch.rtas_token_lock);
arch/powerpc/kvm/book3s.c
908
return kvm->arch.kvm_ops->init_vm(kvm);
arch/powerpc/kvm/book3s.c
913
kvm->arch.kvm_ops->destroy_vm(kvm);
arch/powerpc/kvm/book3s.c
917
WARN_ON(!list_empty(&kvm->arch.spapr_tce_tables));
arch/powerpc/kvm/book3s.c
925
kfree(kvm->arch.xive_devices.native);
arch/powerpc/kvm/book3s.c
926
kvm->arch.xive_devices.native = NULL;
arch/powerpc/kvm/book3s.c
927
kfree(kvm->arch.xive_devices.xics_on_xive);
arch/powerpc/kvm/book3s.c
928
kvm->arch.xive_devices.xics_on_xive = NULL;
arch/powerpc/kvm/book3s.c
929
kfree(kvm->arch.xics_device);
arch/powerpc/kvm/book3s.c
930
kvm->arch.xics_device = NULL;
arch/powerpc/kvm/book3s_32_mmu.c
295
ulong mp_ea = vcpu->arch.magic_page_ea;
arch/powerpc/kvm/book3s_32_mmu.c
305
pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff);
arch/powerpc/kvm/book3s_32_mmu.c
399
struct kvmppc_mmu *mmu = &vcpu->arch.mmu;
arch/powerpc/kvm/book3s_32_mmu.c
43
return vcpu->arch.regs.nip == DEBUG_MMU_PTE_IP;
arch/powerpc/kvm/book3s_32_mmu_host.c
157
vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid);
arch/powerpc/kvm/book3s_32_mmu_host.c
310
if (vcpu->arch.mmu.esid_to_vsid(vcpu, esid, &gvsid)) {
arch/powerpc/kvm/book3s_64_mmu.c
145
if (vcpu->arch.papr_enabled)
arch/powerpc/kvm/book3s_64_mmu.c
207
ulong mp_ea = vcpu->arch.magic_page_ea;
arch/powerpc/kvm/book3s_64_mmu.c
215
gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff);
arch/powerpc/kvm/book3s_64_mmu.c
244
mutex_lock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_64_mmu.c
270
(vcpu->arch.hflags & BOOK3S_HFLAG_MULTI_PGSIZE)) {
arch/powerpc/kvm/book3s_64_mmu.c
300
if (unlikely(vcpu->arch.disable_kernel_nx) &&
arch/powerpc/kvm/book3s_64_mmu.c
348
mutex_unlock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_64_mmu.c
35
for (i = 0; i < vcpu->arch.slb_nr; i++) {
arch/powerpc/kvm/book3s_64_mmu.c
355
mutex_unlock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_64_mmu.c
375
if (slb_nr > vcpu->arch.slb_nr)
arch/powerpc/kvm/book3s_64_mmu.c
378
slbe = &vcpu->arch.slb[slb_nr];
arch/powerpc/kvm/book3s_64_mmu.c
38
if (!vcpu->arch.slb[i].valid)
arch/powerpc/kvm/book3s_64_mmu.c
392
if (vcpu->arch.hflags & BOOK3S_HFLAG_MULTI_PGSIZE) {
arch/powerpc/kvm/book3s_64_mmu.c
41
if (vcpu->arch.slb[i].tb)
arch/powerpc/kvm/book3s_64_mmu.c
429
if (slb_nr > vcpu->arch.slb_nr)
arch/powerpc/kvm/book3s_64_mmu.c
432
slbe = &vcpu->arch.slb[slb_nr];
arch/powerpc/kvm/book3s_64_mmu.c
44
if (vcpu->arch.slb[i].esid == cmp_esid)
arch/powerpc/kvm/book3s_64_mmu.c
441
if (slb_nr > vcpu->arch.slb_nr)
arch/powerpc/kvm/book3s_64_mmu.c
444
slbe = &vcpu->arch.slb[slb_nr];
arch/powerpc/kvm/book3s_64_mmu.c
45
return &vcpu->arch.slb[i];
arch/powerpc/kvm/book3s_64_mmu.c
477
for (i = 1; i < vcpu->arch.slb_nr; i++) {
arch/powerpc/kvm/book3s_64_mmu.c
478
vcpu->arch.slb[i].valid = false;
arch/powerpc/kvm/book3s_64_mmu.c
479
vcpu->arch.slb[i].orige = 0;
arch/powerpc/kvm/book3s_64_mmu.c
480
vcpu->arch.slb[i].origv = 0;
arch/powerpc/kvm/book3s_64_mmu.c
50
for (i = 0; i < vcpu->arch.slb_nr; i++) {
arch/powerpc/kvm/book3s_64_mmu.c
51
if (vcpu->arch.slb[i].vsid)
arch/powerpc/kvm/book3s_64_mmu.c
53
vcpu->arch.slb[i].valid ? 'v' : ' ',
arch/powerpc/kvm/book3s_64_mmu.c
54
vcpu->arch.slb[i].large ? 'l' : ' ',
arch/powerpc/kvm/book3s_64_mmu.c
544
if (vcpu->arch.hflags & BOOK3S_HFLAG_NEW_TLBIE) {
arch/powerpc/kvm/book3s_64_mmu.c
55
vcpu->arch.slb[i].tb ? 't' : ' ',
arch/powerpc/kvm/book3s_64_mmu.c
56
vcpu->arch.slb[i].esid,
arch/powerpc/kvm/book3s_64_mmu.c
565
ulong mp_ea = vcpu->arch.magic_page_ea;
arch/powerpc/kvm/book3s_64_mmu.c
57
vcpu->arch.slb[i].vsid);
arch/powerpc/kvm/book3s_64_mmu.c
578
ulong mp_ea = vcpu->arch.magic_page_ea;
arch/powerpc/kvm/book3s_64_mmu.c
653
struct kvmppc_mmu *mmu = &vcpu->arch.mmu;
arch/powerpc/kvm/book3s_64_mmu.c
669
vcpu->arch.hflags |= BOOK3S_HFLAG_SLB;
arch/powerpc/kvm/book3s_64_mmu_host.c
108
vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid);
arch/powerpc/kvm/book3s_64_mmu_host.c
219
vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid);
arch/powerpc/kvm/book3s_64_mmu_host.c
325
if (vcpu->arch.mmu.esid_to_vsid(vcpu, esid, &gvsid)) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1007
struct revmap_entry *rev = kvm->arch.hpt.rev;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1024
hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4));
arch/powerpc/kvm/book3s_64_mmu_hv.c
1111
rmapp = memslot->arch.rmap;
arch/powerpc/kvm/book3s_64_mmu_hv.c
120
atomic64_set(&kvm->arch.mmio_update, 0);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1201
struct kvm_hpt_info *old = &kvm->arch.hpt;
arch/powerpc/kvm/book3s_64_mmu_hv.c
121
kvm->arch.hpt = *info;
arch/powerpc/kvm/book3s_64_mmu_hv.c
122
kvm->arch.sdr1 = __pa(info->virt) | (info->order - 18);
arch/powerpc/kvm/book3s_64_mmu_hv.c
125
info->virt, (long)info->order, kvm->arch.lpid);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1254
rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
arch/powerpc/kvm/book3s_64_mmu_hv.c
133
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
134
if (kvm->arch.mmu_ready) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
135
kvm->arch.mmu_ready = 0;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1358
for (i = 0; i < kvmppc_hpt_npte(&kvm->arch.hpt); i++) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
138
if (atomic_read(&kvm->arch.vcpus_running)) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1380
hpt_tmp = kvm->arch.hpt;
arch/powerpc/kvm/book3s_64_mmu_hv.c
139
kvm->arch.mmu_ready = 1;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1396
if (WARN_ON(!mutex_is_locked(&kvm->arch.mmu_setup_lock)))
arch/powerpc/kvm/book3s_64_mmu_hv.c
1408
if (kvm->arch.resize_hpt == resize)
arch/powerpc/kvm/book3s_64_mmu_hv.c
1409
kvm->arch.resize_hpt = NULL;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1423
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1426
if (kvm->arch.resize_hpt == resize) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1430
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1443
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1451
if (kvm->arch.resize_hpt != resize)
arch/powerpc/kvm/book3s_64_mmu_hv.c
1454
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1471
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1473
resize = kvm->arch.resize_hpt;
arch/powerpc/kvm/book3s_64_mmu_hv.c
149
if (kvm->arch.hpt.order == order) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1507
kvm->arch.resize_hpt = resize;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1514
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
153
memset((void *)kvm->arch.hpt.virt, 0, 1ul << order);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1537
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1539
resize = kvm->arch.resize_hpt;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1543
if (WARN_ON(!kvm->arch.mmu_ready))
arch/powerpc/kvm/book3s_64_mmu_hv.c
1547
kvm->arch.mmu_ready = 0;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1570
kvm->arch.mmu_ready = 1;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1574
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
162
if (kvm->arch.hpt.virt) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
163
kvmppc_free_hpt(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1720
hptp = (__be64 *)(kvm->arch.hpt.virt + (i * HPTE_SIZE));
arch/powerpc/kvm/book3s_64_mmu_hv.c
1721
revp = kvm->arch.hpt.rev + i;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1736
while (i < kvmppc_hpt_npte(&kvm->arch.hpt) &&
arch/powerpc/kvm/book3s_64_mmu_hv.c
1746
while (i < kvmppc_hpt_npte(&kvm->arch.hpt) &&
arch/powerpc/kvm/book3s_64_mmu_hv.c
175
cpumask_setall(&kvm->arch.need_tlb_flush);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1762
while (i < kvmppc_hpt_npte(&kvm->arch.hpt) &&
arch/powerpc/kvm/book3s_64_mmu_hv.c
177
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1783
if (i >= kvmppc_hpt_npte(&kvm->arch.hpt)) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1817
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1818
mmu_ready = kvm->arch.mmu_ready;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1820
kvm->arch.mmu_ready = 0; /* temporarily */
arch/powerpc/kvm/book3s_64_mmu_hv.c
1823
if (atomic_read(&kvm->arch.vcpus_running)) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1824
kvm->arch.mmu_ready = 1;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1825
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1845
if (i >= kvmppc_hpt_npte(&kvm->arch.hpt) ||
arch/powerpc/kvm/book3s_64_mmu_hv.c
1846
i + hdr.n_valid + hdr.n_invalid > kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_64_mmu_hv.c
1849
hptp = (__be64 *)(kvm->arch.hpt.virt + (i * HPTE_SIZE));
arch/powerpc/kvm/book3s_64_mmu_hv.c
1883
kvm->arch.vrma_slb_v = senc | SLB_VSID_B_1T |
arch/powerpc/kvm/book3s_64_mmu_hv.c
1910
kvm->arch.mmu_ready = mmu_ready;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1911
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1924
atomic_dec(&ctx->kvm->arch.hpte_mod_interest);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1965
atomic_inc(&kvm->arch.hpte_mod_interest);
arch/powerpc/kvm/book3s_64_mmu_hv.c
2046
hptp = (__be64 *)(kvm->arch.hpt.virt + (i * HPTE_SIZE));
arch/powerpc/kvm/book3s_64_mmu_hv.c
2047
for (; len != 0 && i < kvmppc_hpt_npte(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_64_mmu_hv.c
2058
gr = kvm->arch.hpt.rev[i].guest_rpte;
arch/powerpc/kvm/book3s_64_mmu_hv.c
2114
struct kvmppc_mmu *mmu = &vcpu->arch.mmu;
arch/powerpc/kvm/book3s_64_mmu_hv.c
2116
vcpu->arch.slb_nr = 32; /* POWER7/POWER8 */
arch/powerpc/kvm/book3s_64_mmu_hv.c
2120
vcpu->arch.hflags |= BOOK3S_HFLAG_SLB;
arch/powerpc/kvm/book3s_64_mmu_hv.c
226
if (npages > kvmppc_hpt_mask(&kvm->arch.hpt) + 1)
arch/powerpc/kvm/book3s_64_mmu_hv.c
227
npages = kvmppc_hpt_mask(&kvm->arch.hpt) + 1;
arch/powerpc/kvm/book3s_64_mmu_hv.c
238
& kvmppc_hpt_mask(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_64_mmu_hv.c
317
for (i = 0; i < vcpu->arch.slb_nr; i++) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
318
if (!(vcpu->arch.slb[i].orige & SLB_ESID_V))
arch/powerpc/kvm/book3s_64_mmu_hv.c
321
if (vcpu->arch.slb[i].origv & SLB_VSID_B_1T)
arch/powerpc/kvm/book3s_64_mmu_hv.c
326
if (((vcpu->arch.slb[i].orige ^ eaddr) & mask) == 0)
arch/powerpc/kvm/book3s_64_mmu_hv.c
327
return &vcpu->arch.slb[i];
arch/powerpc/kvm/book3s_64_mmu_hv.c
364
slb_v = vcpu->kvm->arch.vrma_slb_v;
arch/powerpc/kvm/book3s_64_mmu_hv.c
375
hptep = (__be64 *)(kvm->arch.hpt.virt + (index << 4));
arch/powerpc/kvm/book3s_64_mmu_hv.c
379
gr = kvm->arch.hpt.rev[index].guest_rpte;
arch/powerpc/kvm/book3s_64_mmu_hv.c
399
int amrfield = hpte_get_skey_perm(gr, vcpu->arch.amr);
arch/powerpc/kvm/book3s_64_mmu_hv.c
501
vcpu->arch.paddr_accessed = gpa;
arch/powerpc/kvm/book3s_64_mmu_hv.c
502
vcpu->arch.vaddr_accessed = ea;
arch/powerpc/kvm/book3s_64_mmu_hv.c
537
if (ea != vcpu->arch.pgfault_addr)
arch/powerpc/kvm/book3s_64_mmu_hv.c
540
if (vcpu->arch.pgfault_cache) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
541
mmio_update = atomic64_read(&kvm->arch.mmio_update);
arch/powerpc/kvm/book3s_64_mmu_hv.c
542
if (mmio_update == vcpu->arch.pgfault_cache->mmio_update) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
543
r = vcpu->arch.pgfault_cache->rpte;
arch/powerpc/kvm/book3s_64_mmu_hv.c
544
psize = kvmppc_actual_pgsz(vcpu->arch.pgfault_hpte[0],
arch/powerpc/kvm/book3s_64_mmu_hv.c
553
index = vcpu->arch.pgfault_index;
arch/powerpc/kvm/book3s_64_mmu_hv.c
554
hptep = (__be64 *)(kvm->arch.hpt.virt + (index << 4));
arch/powerpc/kvm/book3s_64_mmu_hv.c
555
rev = &kvm->arch.hpt.rev[index];
arch/powerpc/kvm/book3s_64_mmu_hv.c
569
if (hpte[0] != vcpu->arch.pgfault_hpte[0] ||
arch/powerpc/kvm/book3s_64_mmu_hv.c
570
hpte[1] != vcpu->arch.pgfault_hpte[1])
arch/powerpc/kvm/book3s_64_mmu_hv.c
679
if (!kvm->arch.mmu_ready)
arch/powerpc/kvm/book3s_64_mmu_hv.c
689
rmap = &memslot->arch.rmap[gfn_base - memslot->base_gfn];
arch/powerpc/kvm/book3s_64_mmu_hv.c
754
memset(memslot->arch.rmap, 0,
arch/powerpc/kvm/book3s_64_mmu_hv.c
755
memslot->npages * sizeof(*memslot->arch.rmap));
arch/powerpc/kvm/book3s_64_mmu_hv.c
766
__be64 *hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4));
arch/powerpc/kvm/book3s_64_mmu_hv.c
767
struct revmap_entry *rev = kvm->arch.hpt.rev;
arch/powerpc/kvm/book3s_64_mmu_hv.c
811
rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
arch/powerpc/kvm/book3s_64_mmu_hv.c
825
hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4));
arch/powerpc/kvm/book3s_64_mmu_hv.c
863
rmapp = memslot->arch.rmap;
arch/powerpc/kvm/book3s_64_mmu_hv.c
885
struct revmap_entry *rev = kvm->arch.hpt.rev;
arch/powerpc/kvm/book3s_64_mmu_hv.c
891
rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
arch/powerpc/kvm/book3s_64_mmu_hv.c
905
hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4));
arch/powerpc/kvm/book3s_64_mmu_hv.c
956
struct revmap_entry *rev = kvm->arch.hpt.rev;
arch/powerpc/kvm/book3s_64_mmu_hv.c
962
rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
arch/powerpc/kvm/book3s_64_mmu_hv.c
973
hp = (unsigned long *)(kvm->arch.hpt.virt + (i << 4));
arch/powerpc/kvm/book3s_64_mmu_hv.c
998
return atomic_read(&kvm->arch.vcpus_running) != 0;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1018
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
1019
uv_page_inval(kvm->arch.lpid, gpa, PAGE_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_radix.c
102
int lpid = vcpu->kvm->arch.lpid;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1026
kvm->arch.lpid);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1039
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE)
arch/powerpc/kvm/book3s_64_mmu_radix.c
1048
rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
arch/powerpc/kvm/book3s_64_mmu_radix.c
1067
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE)
arch/powerpc/kvm/book3s_64_mmu_radix.c
1087
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE)
arch/powerpc/kvm/book3s_64_mmu_radix.c
110
if (vcpu->arch.nested)
arch/powerpc/kvm/book3s_64_mmu_radix.c
111
lpid = vcpu->arch.nested->shadow_lpid;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1121
kvmppc_radix_tlbie_page(kvm, gpa, shift, kvm->arch.lpid);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1123
rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
arch/powerpc/kvm/book3s_64_mmu_radix.c
1165
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_START)
arch/powerpc/kvm/book3s_64_mmu_radix.c
1168
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE)
arch/powerpc/kvm/book3s_64_mmu_radix.c
1177
kvm->arch.lpid);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1227
kvm->arch.pgtable = pgd_alloc(kvm->mm);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1228
if (!kvm->arch.pgtable)
arch/powerpc/kvm/book3s_64_mmu_radix.c
1342
pgt = kvm->arch.pgtable;
arch/powerpc/kvm/book3s_64_mmu_radix.c
289
vcpu->kvm->arch.process_table, pid, &pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
307
if (vcpu->arch.iamr & (1ul << 62))
arch/powerpc/kvm/book3s_64_mmu_radix.c
438
if (lpid != kvm->arch.lpid)
arch/powerpc/kvm/book3s_64_mmu_radix.c
566
if (kvm->arch.pgtable) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
567
kvmppc_free_pgtable_radix(kvm, kvm->arch.pgtable,
arch/powerpc/kvm/book3s_64_mmu_radix.c
568
kvm->arch.lpid);
arch/powerpc/kvm/book3s_64_mmu_radix.c
569
pgd_free(kvm->mm, kvm->arch.pgtable);
arch/powerpc/kvm/book3s_64_mmu_radix.c
570
kvm->arch.pgtable = NULL;
arch/powerpc/kvm/book3s_64_mmu_radix.c
902
ret = kvmppc_create_pte(kvm, kvm->arch.pgtable, pte, gpa, level,
arch/powerpc/kvm/book3s_64_mmu_radix.c
903
mmu_seq, kvm->arch.lpid, NULL, NULL);
arch/powerpc/kvm/book3s_64_mmu_radix.c
950
gpa = vcpu->arch.fault_gpa & ~0xfffUL;
arch/powerpc/kvm/book3s_64_mmu_radix.c
956
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE)
arch/powerpc/kvm/book3s_64_mmu_radix.c
992
gpa, kvm->arch.lpid))
arch/powerpc/kvm/book3s_64_vio.c
124
list_for_each_entry_rcu(stt, &kvm->arch.spapr_tce_tables, list) {
arch/powerpc/kvm/book3s_64_vio.c
324
list_for_each_entry(siter, &kvm->arch.spapr_tce_tables, list) {
arch/powerpc/kvm/book3s_64_vio.c
337
list_add_rcu(&stt->list, &kvm->arch.spapr_tce_tables);
arch/powerpc/kvm/book3s_64_vio.c
41
list_for_each_entry_lockless(stt, &kvm->arch.spapr_tce_tables, list)
arch/powerpc/kvm/book3s_64_vio.c
90
list_for_each_entry_rcu(stt, &kvm->arch.spapr_tce_tables, list) {
arch/powerpc/kvm/book3s_emulate.c
100
vcpu->arch.vrsave_tm = vcpu->arch.vrsave;
arch/powerpc/kvm/book3s_emulate.c
1045
return vcpu->arch.fault_dar;
arch/powerpc/kvm/book3s_emulate.c
105
memcpy(&vcpu->arch.regs.gpr[0], &vcpu->arch.gpr_tm[0],
arch/powerpc/kvm/book3s_emulate.c
106
sizeof(vcpu->arch.regs.gpr));
arch/powerpc/kvm/book3s_emulate.c
107
memcpy(&vcpu->arch.fp, &vcpu->arch.fp_tm,
arch/powerpc/kvm/book3s_emulate.c
109
memcpy(&vcpu->arch.vr, &vcpu->arch.vr_tm,
arch/powerpc/kvm/book3s_emulate.c
111
vcpu->arch.ppr = vcpu->arch.ppr_tm;
arch/powerpc/kvm/book3s_emulate.c
112
vcpu->arch.dscr = vcpu->arch.dscr_tm;
arch/powerpc/kvm/book3s_emulate.c
113
vcpu->arch.amr = vcpu->arch.amr_tm;
arch/powerpc/kvm/book3s_emulate.c
114
vcpu->arch.regs.ctr = vcpu->arch.ctr_tm;
arch/powerpc/kvm/book3s_emulate.c
115
vcpu->arch.tar = vcpu->arch.tar_tm;
arch/powerpc/kvm/book3s_emulate.c
116
vcpu->arch.regs.link = vcpu->arch.lr_tm;
arch/powerpc/kvm/book3s_emulate.c
117
vcpu->arch.regs.ccr = vcpu->arch.cr_tm;
arch/powerpc/kvm/book3s_emulate.c
118
vcpu->arch.regs.xer = vcpu->arch.xer_tm;
arch/powerpc/kvm/book3s_emulate.c
119
vcpu->arch.vrsave = vcpu->arch.vrsave_tm;
arch/powerpc/kvm/book3s_emulate.c
129
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & ~(CR0_MASK << CR0_SHIFT)) |
arch/powerpc/kvm/book3s_emulate.c
151
vcpu->arch.texasr = texasr;
arch/powerpc/kvm/book3s_emulate.c
152
vcpu->arch.tfiar = kvmppc_get_pc(vcpu);
arch/powerpc/kvm/book3s_emulate.c
154
mtspr(SPRN_TFIAR, vcpu->arch.tfiar);
arch/powerpc/kvm/book3s_emulate.c
164
if (vcpu->arch.shadow_fscr & FSCR_TAR)
arch/powerpc/kvm/book3s_emulate.c
165
mtspr(SPRN_TAR, vcpu->arch.tar);
arch/powerpc/kvm/book3s_emulate.c
208
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & ~(CR0_MASK << CR0_SHIFT)) |
arch/powerpc/kvm/book3s_emulate.c
212
vcpu->arch.texasr = mfspr(SPRN_TEXASR);
arch/powerpc/kvm/book3s_emulate.c
219
vcpu->arch.texasr &= ~(TEXASR_PR | TEXASR_HV);
arch/powerpc/kvm/book3s_emulate.c
221
vcpu->arch.texasr |= TEXASR_PR;
arch/powerpc/kvm/book3s_emulate.c
224
vcpu->arch.texasr |= TEXASR_HV;
arch/powerpc/kvm/book3s_emulate.c
226
vcpu->arch.tfiar = kvmppc_get_pc(vcpu);
arch/powerpc/kvm/book3s_emulate.c
317
if (vcpu->arch.mmu.mfsrin) {
arch/powerpc/kvm/book3s_emulate.c
319
sr = vcpu->arch.mmu.mfsrin(vcpu, srnum);
arch/powerpc/kvm/book3s_emulate.c
329
if (vcpu->arch.mmu.mfsrin) {
arch/powerpc/kvm/book3s_emulate.c
331
sr = vcpu->arch.mmu.mfsrin(vcpu, srnum);
arch/powerpc/kvm/book3s_emulate.c
337
vcpu->arch.mmu.mtsrin(vcpu,
arch/powerpc/kvm/book3s_emulate.c
342
vcpu->arch.mmu.mtsrin(vcpu,
arch/powerpc/kvm/book3s_emulate.c
351
vcpu->arch.mmu.tlbie(vcpu, addr, large);
arch/powerpc/kvm/book3s_emulate.c
362
!vcpu->arch.papr_enabled) {
arch/powerpc/kvm/book3s_emulate.c
377
vcpu->arch.hcall_needed = 1;
arch/powerpc/kvm/book3s_emulate.c
385
if (!vcpu->arch.mmu.slbmte)
arch/powerpc/kvm/book3s_emulate.c
388
vcpu->arch.mmu.slbmte(vcpu,
arch/powerpc/kvm/book3s_emulate.c
393
if (!vcpu->arch.mmu.slbie)
arch/powerpc/kvm/book3s_emulate.c
396
vcpu->arch.mmu.slbie(vcpu,
arch/powerpc/kvm/book3s_emulate.c
400
if (!vcpu->arch.mmu.slbia)
arch/powerpc/kvm/book3s_emulate.c
403
vcpu->arch.mmu.slbia(vcpu);
arch/powerpc/kvm/book3s_emulate.c
406
if (!(inst & 1) || !vcpu->arch.mmu.slbfee) {
arch/powerpc/kvm/book3s_emulate.c
413
if (!vcpu->arch.mmu.slbfee(vcpu, b, &t))
arch/powerpc/kvm/book3s_emulate.c
417
cr |= (vcpu->arch.regs.xer & 0x80000000) >>
arch/powerpc/kvm/book3s_emulate.c
423
if (!vcpu->arch.mmu.slbmfee) {
arch/powerpc/kvm/book3s_emulate.c
429
t = vcpu->arch.mmu.slbmfee(vcpu, rb_val);
arch/powerpc/kvm/book3s_emulate.c
434
if (!vcpu->arch.mmu.slbmfev) {
arch/powerpc/kvm/book3s_emulate.c
440
t = vcpu->arch.mmu.slbmfev(vcpu, rb_val);
arch/powerpc/kvm/book3s_emulate.c
468
vcpu->arch.fault_dar = vaddr;
arch/powerpc/kvm/book3s_emulate.c
477
vcpu->arch.fault_dsisr = dsisr;
arch/powerpc/kvm/book3s_emulate.c
499
vcpu->arch.regs.ccr = (CR0_TBEGIN_FAILURE |
arch/powerpc/kvm/book3s_emulate.c
500
(vcpu->arch.regs.ccr & ~(CR0_MASK << CR0_SHIFT)));
arch/powerpc/kvm/book3s_emulate.c
502
vcpu->arch.texasr = (TEXASR_FS | TEXASR_EXACT |
arch/powerpc/kvm/book3s_emulate.c
507
vcpu->arch.texasr |= TEXASR_ROT;
arch/powerpc/kvm/book3s_emulate.c
510
vcpu->arch.texasr |= TEXASR_HV;
arch/powerpc/kvm/book3s_emulate.c
512
vcpu->arch.tfhar = kvmppc_get_pc(vcpu) + 4;
arch/powerpc/kvm/book3s_emulate.c
513
vcpu->arch.tfiar = kvmppc_get_pc(vcpu);
arch/powerpc/kvm/book3s_emulate.c
723
switch (vcpu->arch.pvr) {
arch/powerpc/kvm/book3s_emulate.c
73
if (vcpu->arch.papr_enabled && (level > PRIV_SUPER))
arch/powerpc/kvm/book3s_emulate.c
733
if (vcpu->arch.hflags & BOOK3S_HFLAG_NATIVE_PS) {
arch/powerpc/kvm/book3s_emulate.c
736
vcpu->arch.hflags |= BOOK3S_HFLAG_PAIRED_SINGLE;
arch/powerpc/kvm/book3s_emulate.c
739
vcpu->arch.hflags &= ~BOOK3S_HFLAG_PAIRED_SINGLE;
arch/powerpc/kvm/book3s_emulate.c
751
if (vcpu->arch.mmu.is_dcbz32(vcpu) &&
arch/powerpc/kvm/book3s_emulate.c
753
vcpu->arch.hflags |= BOOK3S_HFLAG_DCBZ32;
arch/powerpc/kvm/book3s_emulate.c
770
vcpu->arch.bescr = spr_val;
arch/powerpc/kvm/book3s_emulate.c
773
vcpu->arch.ebbhr = spr_val;
arch/powerpc/kvm/book3s_emulate.c
776
vcpu->arch.ebbrr = spr_val;
arch/powerpc/kvm/book3s_emulate.c
86
memcpy(&vcpu->arch.gpr_tm[0], &vcpu->arch.regs.gpr[0],
arch/powerpc/kvm/book3s_emulate.c
87
sizeof(vcpu->arch.gpr_tm));
arch/powerpc/kvm/book3s_emulate.c
88
memcpy(&vcpu->arch.fp_tm, &vcpu->arch.fp,
arch/powerpc/kvm/book3s_emulate.c
90
memcpy(&vcpu->arch.vr_tm, &vcpu->arch.vr,
arch/powerpc/kvm/book3s_emulate.c
92
vcpu->arch.ppr_tm = vcpu->arch.ppr;
arch/powerpc/kvm/book3s_emulate.c
922
*spr_val = vcpu->arch.purr;
arch/powerpc/kvm/book3s_emulate.c
928
*spr_val = vcpu->arch.spurr;
arch/powerpc/kvm/book3s_emulate.c
93
vcpu->arch.dscr_tm = vcpu->arch.dscr;
arch/powerpc/kvm/book3s_emulate.c
934
*spr_val = vcpu->arch.ic;
arch/powerpc/kvm/book3s_emulate.c
94
vcpu->arch.amr_tm = vcpu->arch.amr;
arch/powerpc/kvm/book3s_emulate.c
948
*spr_val = vcpu->arch.fscr;
arch/powerpc/kvm/book3s_emulate.c
95
vcpu->arch.ctr_tm = vcpu->arch.regs.ctr;
arch/powerpc/kvm/book3s_emulate.c
951
*spr_val = vcpu->arch.bescr;
arch/powerpc/kvm/book3s_emulate.c
954
*spr_val = vcpu->arch.ebbhr;
arch/powerpc/kvm/book3s_emulate.c
957
*spr_val = vcpu->arch.ebbrr;
arch/powerpc/kvm/book3s_emulate.c
96
vcpu->arch.tar_tm = vcpu->arch.tar;
arch/powerpc/kvm/book3s_emulate.c
97
vcpu->arch.lr_tm = vcpu->arch.regs.link;
arch/powerpc/kvm/book3s_emulate.c
98
vcpu->arch.cr_tm = vcpu->arch.regs.ccr;
arch/powerpc/kvm/book3s_emulate.c
99
vcpu->arch.xer_tm = vcpu->arch.regs.xer;
arch/powerpc/kvm/book3s_hv.c
1044
struct kvmppc_vcore *vcore = target->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
1059
if (target->arch.state == KVMPPC_VCPU_RUNNABLE &&
arch/powerpc/kvm/book3s_hv.c
1074
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
1075
lppaca = (struct lppaca *)vcpu->arch.vpa.pinned_addr;
arch/powerpc/kvm/book3s_hv.c
1078
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
1103
do_h_rpt_invalidate_prt(pid, vcpu->arch.nested->shadow_lpid,
arch/powerpc/kvm/book3s_hv.c
1139
do_h_rpt_invalidate_prt(id, vcpu->kvm->arch.lpid,
arch/powerpc/kvm/book3s_hv.c
1154
!test_bit(req/4, vcpu->kvm->arch.enabled_hcalls))
arch/powerpc/kvm/book3s_hv.c
1213
tvcpu->arch.prodded = 1;
arch/powerpc/kvm/book3s_hv.c
1215
if (tvcpu->arch.ceded)
arch/powerpc/kvm/book3s_hv.c
1238
if (list_empty(&kvm->arch.rtas_tokens))
arch/powerpc/kvm/book3s_hv.c
1352
vcpu->arch.hcall_needed = 0;
arch/powerpc/kvm/book3s_hv.c
1356
vcpu->arch.hcall_needed = 0;
arch/powerpc/kvm/book3s_hv.c
1417
vcpu->arch.hcall_needed = 0;
arch/powerpc/kvm/book3s_hv.c
1431
vcpu->arch.ceded = 1;
arch/powerpc/kvm/book3s_hv.c
1433
if (vcpu->arch.prodded) {
arch/powerpc/kvm/book3s_hv.c
1434
vcpu->arch.prodded = 0;
arch/powerpc/kvm/book3s_hv.c
1436
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv.c
1488
vcpu->run->debug.arch.address = kvmppc_get_pc(vcpu);
arch/powerpc/kvm/book3s_hv.c
1507
nthreads = vcpu->kvm->arch.emul_smt_mode;
arch/powerpc/kvm/book3s_hv.c
1548
thr = vcpu->vcpu_id & (kvm->arch.emul_smt_mode - 1);
arch/powerpc/kvm/book3s_hv.c
1555
if (arg >= kvm->arch.emul_smt_mode)
arch/powerpc/kvm/book3s_hv.c
1560
if (!tvcpu->arch.doorbell_request) {
arch/powerpc/kvm/book3s_hv.c
1561
tvcpu->arch.doorbell_request = 1;
arch/powerpc/kvm/book3s_hv.c
1569
vcpu->arch.vcore->dpdes = 0;
arch/powerpc/kvm/book3s_hv.c
1570
vcpu->arch.doorbell_request = 0;
arch/powerpc/kvm/book3s_hv.c
1601
if (!(vcpu->arch.hfscr_permitted & HFSCR_PM))
arch/powerpc/kvm/book3s_hv.c
1611
if (!(vcpu->arch.hfscr_permitted & HFSCR_EBB))
arch/powerpc/kvm/book3s_hv.c
1621
if (!(vcpu->arch.hfscr_permitted & HFSCR_TM))
arch/powerpc/kvm/book3s_hv.c
1648
vcpu->arch.trap, kvmppc_get_pc(vcpu),
arch/powerpc/kvm/book3s_hv.c
1649
vcpu->arch.shregs.msr);
arch/powerpc/kvm/book3s_hv.c
1652
run->hw.hardware_exit_reason = vcpu->arch.trap;
arch/powerpc/kvm/book3s_hv.c
1657
switch (vcpu->arch.trap) {
arch/powerpc/kvm/book3s_hv.c
1661
vcpu->arch.trap = BOOK3S_INTERRUPT_HV_DECREMENTER;
arch/powerpc/kvm/book3s_hv.c
1687
machine_check_print_event_info(&vcpu->arch.mce_evt,false, true);
arch/powerpc/kvm/book3s_hv.c
1695
if (!vcpu->kvm->arch.fwnmi_enabled) {
arch/powerpc/kvm/book3s_hv.c
1705
run->hw.hardware_exit_reason = vcpu->arch.trap;
arch/powerpc/kvm/book3s_hv.c
1709
if (vcpu->arch.mce_evt.disposition == MCE_DISPOSITION_RECOVERED)
arch/powerpc/kvm/book3s_hv.c
1771
vcpu->arch.hcall_needed = 1;
arch/powerpc/kvm/book3s_hv.c
1790
unlikely(vcpu->arch.fault_dsisr == HDSISR_CANARY)) {
arch/powerpc/kvm/book3s_hv.c
1806
if (!(vcpu->arch.fault_dsisr & (DSISR_NOHPTE | DSISR_PROTFAULT))) {
arch/powerpc/kvm/book3s_hv.c
1809
vcpu->arch.fault_dar, vcpu->arch.fault_dsisr);
arch/powerpc/kvm/book3s_hv.c
1815
vsid = vcpu->kvm->arch.vrma_slb_v;
arch/powerpc/kvm/book3s_hv.c
1817
vsid = vcpu->arch.fault_gpa;
arch/powerpc/kvm/book3s_hv.c
1819
err = kvmppc_hpte_hv_fault(vcpu, vcpu->arch.fault_dar,
arch/powerpc/kvm/book3s_hv.c
1820
vsid, vcpu->arch.fault_dsisr, true);
arch/powerpc/kvm/book3s_hv.c
1828
vcpu->arch.fault_dar, err);
arch/powerpc/kvm/book3s_hv.c
1837
vcpu->arch.fault_dar = kvmppc_get_pc(vcpu);
arch/powerpc/kvm/book3s_hv.c
1838
vcpu->arch.fault_dsisr = __kvmppc_get_msr_hv(vcpu) &
arch/powerpc/kvm/book3s_hv.c
1848
vcpu->arch.fault_dsisr |= DSISR_ISSTORE;
arch/powerpc/kvm/book3s_hv.c
1853
if (!(vcpu->arch.fault_dsisr & SRR1_ISI_NOPT)) {
arch/powerpc/kvm/book3s_hv.c
1855
vcpu->arch.fault_dsisr |
arch/powerpc/kvm/book3s_hv.c
1862
vsid = vcpu->kvm->arch.vrma_slb_v;
arch/powerpc/kvm/book3s_hv.c
1864
vsid = vcpu->arch.fault_gpa;
arch/powerpc/kvm/book3s_hv.c
1866
err = kvmppc_hpte_hv_fault(vcpu, vcpu->arch.fault_dar,
arch/powerpc/kvm/book3s_hv.c
1867
vsid, vcpu->arch.fault_dsisr, false);
arch/powerpc/kvm/book3s_hv.c
1888
if (vcpu->arch.emul_inst != KVM_INST_FETCH_FAILED)
arch/powerpc/kvm/book3s_hv.c
1889
vcpu->arch.last_inst = kvmppc_need_byteswap(vcpu) ?
arch/powerpc/kvm/book3s_hv.c
1890
swab32(vcpu->arch.emul_inst) :
arch/powerpc/kvm/book3s_hv.c
1891
vcpu->arch.emul_inst;
arch/powerpc/kvm/book3s_hv.c
1958
vcpu->arch.trap, kvmppc_get_pc(vcpu),
arch/powerpc/kvm/book3s_hv.c
1960
run->hw.hardware_exit_reason = vcpu->arch.trap;
arch/powerpc/kvm/book3s_hv.c
1986
vcpu->arch.trap, kvmppc_get_pc(vcpu),
arch/powerpc/kvm/book3s_hv.c
1991
switch (vcpu->arch.trap) {
arch/powerpc/kvm/book3s_hv.c
2008
vcpu->arch.trap = BOOK3S_INTERRUPT_HV_DECREMENTER;
arch/powerpc/kvm/book3s_hv.c
2026
machine_check_print_event_info(&vcpu->arch.mce_evt, false, true);
arch/powerpc/kvm/book3s_hv.c
2041
vcpu->arch.fault_dar = kvmppc_get_pc(vcpu);
arch/powerpc/kvm/book3s_hv.c
2042
vcpu->arch.fault_dsisr = kvmppc_get_msr(vcpu) &
arch/powerpc/kvm/book3s_hv.c
2045
vcpu->arch.fault_dsisr |= DSISR_ISSTORE;
arch/powerpc/kvm/book3s_hv.c
2070
vcpu->arch.trap = 0;
arch/powerpc/kvm/book3s_hv.c
2106
sregs->pvr = vcpu->arch.pvr;
arch/powerpc/kvm/book3s_hv.c
2107
for (i = 0; i < vcpu->arch.slb_max; i++) {
arch/powerpc/kvm/book3s_hv.c
2108
sregs->u.s.ppc64.slb[i].slbe = vcpu->arch.slb[i].orige;
arch/powerpc/kvm/book3s_hv.c
2109
sregs->u.s.ppc64.slb[i].slbv = vcpu->arch.slb[i].origv;
arch/powerpc/kvm/book3s_hv.c
2121
if (sregs->pvr != vcpu->arch.pvr)
arch/powerpc/kvm/book3s_hv.c
2125
for (i = 0; i < vcpu->arch.slb_nr; i++) {
arch/powerpc/kvm/book3s_hv.c
2127
vcpu->arch.slb[j].orige = sregs->u.s.ppc64.slb[i].slbe;
arch/powerpc/kvm/book3s_hv.c
2128
vcpu->arch.slb[j].origv = sregs->u.s.ppc64.slb[i].slbv;
arch/powerpc/kvm/book3s_hv.c
2132
vcpu->arch.slb_max = j;
arch/powerpc/kvm/book3s_hv.c
2184
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
2214
if (vcpu->arch.vcore != vc)
arch/powerpc/kvm/book3s_hv.c
2217
vcpu->arch.intr_msr |= MSR_LE;
arch/powerpc/kvm/book3s_hv.c
2219
vcpu->arch.intr_msr &= ~MSR_LE;
arch/powerpc/kvm/book3s_hv.c
2243
*val = get_reg_val(id, vcpu->arch.dabr);
arch/powerpc/kvm/book3s_hv.c
2246
*val = get_reg_val(id, vcpu->arch.dabrx);
arch/powerpc/kvm/book3s_hv.c
2274
*val = get_reg_val(id, vcpu->arch.mmcrs);
arch/powerpc/kvm/book3s_hv.c
2285
*val = get_reg_val(id, vcpu->arch.spmc[i]);
arch/powerpc/kvm/book3s_hv.c
2316
*val = get_reg_val(id, vcpu->arch.doorbell_request);
arch/powerpc/kvm/book3s_hv.c
2318
*val = get_reg_val(id, vcpu->arch.vcore->dpdes);
arch/powerpc/kvm/book3s_hv.c
2348
*val = get_reg_val(id, vcpu->arch.csigr);
arch/powerpc/kvm/book3s_hv.c
2351
*val = get_reg_val(id, vcpu->arch.tacr);
arch/powerpc/kvm/book3s_hv.c
2354
*val = get_reg_val(id, vcpu->arch.tcscr);
arch/powerpc/kvm/book3s_hv.c
2360
*val = get_reg_val(id, vcpu->arch.acop);
arch/powerpc/kvm/book3s_hv.c
2366
*val = get_reg_val(id, vcpu->arch.tid);
arch/powerpc/kvm/book3s_hv.c
2369
*val = get_reg_val(id, vcpu->arch.psscr);
arch/powerpc/kvm/book3s_hv.c
2372
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
2373
*val = get_reg_val(id, vcpu->arch.vpa.next_gpa);
arch/powerpc/kvm/book3s_hv.c
2374
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
2377
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
2378
val->vpaval.addr = vcpu->arch.slb_shadow.next_gpa;
arch/powerpc/kvm/book3s_hv.c
2379
val->vpaval.length = vcpu->arch.slb_shadow.len;
arch/powerpc/kvm/book3s_hv.c
2380
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
2383
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
2384
val->vpaval.addr = vcpu->arch.dtl.next_gpa;
arch/powerpc/kvm/book3s_hv.c
2385
val->vpaval.length = vcpu->arch.dtl.len;
arch/powerpc/kvm/book3s_hv.c
2386
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
2400
*val = get_reg_val(id, vcpu->arch.tfhar);
arch/powerpc/kvm/book3s_hv.c
2403
*val = get_reg_val(id, vcpu->arch.tfiar);
arch/powerpc/kvm/book3s_hv.c
2406
*val = get_reg_val(id, vcpu->arch.texasr);
arch/powerpc/kvm/book3s_hv.c
2410
*val = get_reg_val(id, vcpu->arch.gpr_tm[i]);
arch/powerpc/kvm/book3s_hv.c
2418
val->vsxval[j] = vcpu->arch.fp_tm.fpr[i][j];
arch/powerpc/kvm/book3s_hv.c
242
cpu = READ_ONCE(vcpu->arch.thread_cpu);
arch/powerpc/kvm/book3s_hv.c
2421
val->vval = vcpu->arch.vr_tm.vr[i-32];
arch/powerpc/kvm/book3s_hv.c
2428
*val = get_reg_val(id, vcpu->arch.cr_tm);
arch/powerpc/kvm/book3s_hv.c
2431
*val = get_reg_val(id, vcpu->arch.xer_tm);
arch/powerpc/kvm/book3s_hv.c
2434
*val = get_reg_val(id, vcpu->arch.lr_tm);
arch/powerpc/kvm/book3s_hv.c
2437
*val = get_reg_val(id, vcpu->arch.ctr_tm);
arch/powerpc/kvm/book3s_hv.c
2440
*val = get_reg_val(id, vcpu->arch.fp_tm.fpscr);
arch/powerpc/kvm/book3s_hv.c
2443
*val = get_reg_val(id, vcpu->arch.amr_tm);
arch/powerpc/kvm/book3s_hv.c
2446
*val = get_reg_val(id, vcpu->arch.ppr_tm);
arch/powerpc/kvm/book3s_hv.c
2449
*val = get_reg_val(id, vcpu->arch.vrsave_tm);
arch/powerpc/kvm/book3s_hv.c
2453
*val = get_reg_val(id, vcpu->arch.vr_tm.vscr.u[3]);
arch/powerpc/kvm/book3s_hv.c
2458
*val = get_reg_val(id, vcpu->arch.dscr_tm);
arch/powerpc/kvm/book3s_hv.c
2461
*val = get_reg_val(id, vcpu->arch.tar_tm);
arch/powerpc/kvm/book3s_hv.c
2471
*val = get_reg_val(id, vcpu->arch.online);
arch/powerpc/kvm/book3s_hv.c
2474
*val = get_reg_val(id, vcpu->kvm->arch.l1_ptcr);
arch/powerpc/kvm/book3s_hv.c
2501
vcpu->arch.dabr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2504
vcpu->arch.dabrx = set_reg_val(id, *val) & ~DABRX_HYP;
arch/powerpc/kvm/book3s_hv.c
2532
vcpu->arch.mmcrs = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2543
vcpu->arch.spmc[i] = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2568
vcpu->arch.doorbell_request = set_reg_val(id, *val) & 1;
arch/powerpc/kvm/book3s_hv.c
2570
vcpu->arch.vcore->dpdes = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2603
vcpu->arch.csigr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2606
vcpu->arch.tacr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2609
vcpu->arch.tcscr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2615
vcpu->arch.acop = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2621
vcpu->arch.tid = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2624
vcpu->arch.psscr = set_reg_val(id, *val) & PSSCR_GUEST_VIS;
arch/powerpc/kvm/book3s_hv.c
2629
if (!addr && (vcpu->arch.slb_shadow.next_gpa ||
arch/powerpc/kvm/book3s_hv.c
2630
vcpu->arch.dtl.next_gpa))
arch/powerpc/kvm/book3s_hv.c
2632
r = set_vpa(vcpu, &vcpu->arch.vpa, addr, sizeof(struct lppaca));
arch/powerpc/kvm/book3s_hv.c
2638
if (addr && !vcpu->arch.vpa.next_gpa)
arch/powerpc/kvm/book3s_hv.c
2640
r = set_vpa(vcpu, &vcpu->arch.slb_shadow, addr, len);
arch/powerpc/kvm/book3s_hv.c
2647
!vcpu->arch.vpa.next_gpa))
arch/powerpc/kvm/book3s_hv.c
2650
r = set_vpa(vcpu, &vcpu->arch.dtl, addr, len);
arch/powerpc/kvm/book3s_hv.c
2683
vcpu->arch.tfhar = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2686
vcpu->arch.tfiar = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2689
vcpu->arch.texasr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2693
vcpu->arch.gpr_tm[i] = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2701
vcpu->arch.fp_tm.fpr[i][j] = val->vsxval[j];
arch/powerpc/kvm/book3s_hv.c
2704
vcpu->arch.vr_tm.vr[i-32] = val->vval;
arch/powerpc/kvm/book3s_hv.c
2710
vcpu->arch.cr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2713
vcpu->arch.xer_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2716
vcpu->arch.lr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2719
vcpu->arch.ctr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2722
vcpu->arch.fp_tm.fpscr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2725
vcpu->arch.amr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2728
vcpu->arch.ppr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2731
vcpu->arch.vrsave_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2735
vcpu->arch.vr.vscr.u[3] = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2740
vcpu->arch.dscr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2743
vcpu->arch.tar_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2754
if (i && !vcpu->arch.online)
arch/powerpc/kvm/book3s_hv.c
2755
atomic_inc(&vcpu->arch.vcore->online_count);
arch/powerpc/kvm/book3s_hv.c
2756
else if (!i && vcpu->arch.online)
arch/powerpc/kvm/book3s_hv.c
2757
atomic_dec(&vcpu->arch.vcore->online_count);
arch/powerpc/kvm/book3s_hv.c
2758
vcpu->arch.online = i;
arch/powerpc/kvm/book3s_hv.c
2761
vcpu->kvm->arch.l1_ptcr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_hv.c
2801
vcore->lpcr = kvm->arch.lpcr;
arch/powerpc/kvm/book3s_hv.c
2815
{"vcpu_entry", offsetof(struct kvm_vcpu, arch.vcpu_entry)},
arch/powerpc/kvm/book3s_hv.c
2816
{"guest_entry", offsetof(struct kvm_vcpu, arch.guest_entry)},
arch/powerpc/kvm/book3s_hv.c
2817
{"in_guest", offsetof(struct kvm_vcpu, arch.in_guest)},
arch/powerpc/kvm/book3s_hv.c
2818
{"guest_exit", offsetof(struct kvm_vcpu, arch.guest_exit)},
arch/powerpc/kvm/book3s_hv.c
2819
{"vcpu_exit", offsetof(struct kvm_vcpu, arch.vcpu_exit)},
arch/powerpc/kvm/book3s_hv.c
2820
{"hypercall", offsetof(struct kvm_vcpu, arch.hcall)},
arch/powerpc/kvm/book3s_hv.c
2821
{"page_fault", offsetof(struct kvm_vcpu, arch.pg_fault)},
arch/powerpc/kvm/book3s_hv.c
2823
{"rm_entry", offsetof(struct kvm_vcpu, arch.rm_entry)},
arch/powerpc/kvm/book3s_hv.c
2824
{"rm_intr", offsetof(struct kvm_vcpu, arch.rm_intr)},
arch/powerpc/kvm/book3s_hv.c
2825
{"rm_exit", offsetof(struct kvm_vcpu, arch.rm_exit)},
arch/powerpc/kvm/book3s_hv.c
2826
{"guest", offsetof(struct kvm_vcpu, arch.guest_time)},
arch/powerpc/kvm/book3s_hv.c
2827
{"cede", offsetof(struct kvm_vcpu, arch.cede_time)},
arch/powerpc/kvm/book3s_hv.c
2971
vcpu->arch.shared = &vcpu->arch.shregs;
arch/powerpc/kvm/book3s_hv.c
2978
vcpu->arch.shared_big_endian = true;
arch/powerpc/kvm/book3s_hv.c
2980
vcpu->arch.shared_big_endian = false;
arch/powerpc/kvm/book3s_hv.c
2985
err = kvmhv_nestedv2_vcpu_create(vcpu, &vcpu->arch.nestedv2_io);
arch/powerpc/kvm/book3s_hv.c
2999
spin_lock_init(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
3000
spin_lock_init(&vcpu->arch.tbacct_lock);
arch/powerpc/kvm/book3s_hv.c
3001
vcpu->arch.busy_preempt = TB_NIL;
arch/powerpc/kvm/book3s_hv.c
3003
vcpu->arch.intr_msr = MSR_SF | MSR_ME;
arch/powerpc/kvm/book3s_hv.c
3028
vcpu->arch.hfscr |= HFSCR_TM;
arch/powerpc/kvm/book3s_hv.c
3030
vcpu->arch.hfscr_permitted = kvmppc_get_hfscr_hv(vcpu);
arch/powerpc/kvm/book3s_hv.c
3039
vcpu->arch.state = KVMPPC_VCPU_NOTREADY;
arch/powerpc/kvm/book3s_hv.c
3041
init_waitqueue_head(&vcpu->arch.cpu_run);
arch/powerpc/kvm/book3s_hv.c
3047
if (id >= (KVM_MAX_VCPUS * kvm->arch.emul_smt_mode)) {
arch/powerpc/kvm/book3s_hv.c
3051
BUG_ON(kvm->arch.smt_mode != 1);
arch/powerpc/kvm/book3s_hv.c
3055
core = id / kvm->arch.smt_mode;
arch/powerpc/kvm/book3s_hv.c
3058
vcore = kvm->arch.vcores[core];
arch/powerpc/kvm/book3s_hv.c
3069
id & ~(kvm->arch.smt_mode - 1));
arch/powerpc/kvm/book3s_hv.c
3070
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
3071
kvm->arch.vcores[core] = vcore;
arch/powerpc/kvm/book3s_hv.c
3072
kvm->arch.online_vcores++;
arch/powerpc/kvm/book3s_hv.c
3073
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
3084
vcpu->arch.vcore = vcore;
arch/powerpc/kvm/book3s_hv.c
3085
vcpu->arch.ptid = vcpu->vcpu_id - vcore->first_vcpuid;
arch/powerpc/kvm/book3s_hv.c
3086
vcpu->arch.thread_cpu = -1;
arch/powerpc/kvm/book3s_hv.c
3087
vcpu->arch.prev_cpu = -1;
arch/powerpc/kvm/book3s_hv.c
3089
vcpu->arch.cpu_type = KVM_CPU_3S_64;
arch/powerpc/kvm/book3s_hv.c
3122
if (!kvm->arch.online_vcores) {
arch/powerpc/kvm/book3s_hv.c
3123
kvm->arch.smt_mode = smt_mode;
arch/powerpc/kvm/book3s_hv.c
3124
kvm->arch.emul_smt_mode = esmt;
arch/powerpc/kvm/book3s_hv.c
3141
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
3142
unpin_vpa(vcpu->kvm, &vcpu->arch.dtl);
arch/powerpc/kvm/book3s_hv.c
3143
unpin_vpa(vcpu->kvm, &vcpu->arch.slb_shadow);
arch/powerpc/kvm/book3s_hv.c
3144
unpin_vpa(vcpu->kvm, &vcpu->arch.vpa);
arch/powerpc/kvm/book3s_hv.c
3145
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
3147
kvmhv_nestedv2_vcpu_free(vcpu, &vcpu->arch.nestedv2_io);
arch/powerpc/kvm/book3s_hv.c
3168
hrtimer_start(&vcpu->arch.dec_timer, dec_nsec, HRTIMER_MODE_REL);
arch/powerpc/kvm/book3s_hv.c
3169
vcpu->arch.timer_running = 1;
arch/powerpc/kvm/book3s_hv.c
3179
if (vcpu->arch.state != KVMPPC_VCPU_RUNNABLE)
arch/powerpc/kvm/book3s_hv.c
3181
spin_lock_irq(&vcpu->arch.tbacct_lock);
arch/powerpc/kvm/book3s_hv.c
3183
vcpu->arch.busy_stolen += vcore_stolen_time(vc, now) -
arch/powerpc/kvm/book3s_hv.c
3184
vcpu->arch.stolen_logged;
arch/powerpc/kvm/book3s_hv.c
3185
vcpu->arch.busy_preempt = now;
arch/powerpc/kvm/book3s_hv.c
3186
vcpu->arch.state = KVMPPC_VCPU_BUSY_IN_HOST;
arch/powerpc/kvm/book3s_hv.c
3187
spin_unlock_irq(&vcpu->arch.tbacct_lock);
arch/powerpc/kvm/book3s_hv.c
3189
WRITE_ONCE(vc->runnable_threads[vcpu->arch.ptid], NULL);
arch/powerpc/kvm/book3s_hv.c
319
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
324
if (vcpu->arch.busy_preempt != TB_NIL) {
arch/powerpc/kvm/book3s_hv.c
3241
struct kvm_nested_guest *nested = vcpu->arch.nested;
arch/powerpc/kvm/book3s_hv.c
3248
need_tlb_flush = &kvm->arch.need_tlb_flush;
arch/powerpc/kvm/book3s_hv.c
325
WARN_ON_ONCE(vcpu->arch.state != KVMPPC_VCPU_BUSY_IN_HOST);
arch/powerpc/kvm/book3s_hv.c
326
vc->stolen_tb += mftb() - vcpu->arch.busy_preempt;
arch/powerpc/kvm/book3s_hv.c
327
vcpu->arch.busy_preempt = TB_NIL;
arch/powerpc/kvm/book3s_hv.c
3286
if (kvm->arch.lpcr & LPCR_GTSE)
arch/powerpc/kvm/book3s_hv.c
3294
struct kvm_nested_guest *nested = vcpu->arch.nested;
arch/powerpc/kvm/book3s_hv.c
3302
prev_cpu = nested->prev_cpu[vcpu->arch.nested_vcpu_id];
arch/powerpc/kvm/book3s_hv.c
3304
prev_cpu = vcpu->arch.prev_cpu;
arch/powerpc/kvm/book3s_hv.c
3327
nested->prev_cpu[vcpu->arch.nested_vcpu_id] = pcpu;
arch/powerpc/kvm/book3s_hv.c
3329
vcpu->arch.prev_cpu = pcpu;
arch/powerpc/kvm/book3s_hv.c
3340
if (vcpu->arch.timer_running) {
arch/powerpc/kvm/book3s_hv.c
3341
hrtimer_try_to_cancel(&vcpu->arch.dec_timer);
arch/powerpc/kvm/book3s_hv.c
3342
vcpu->arch.timer_running = 0;
arch/powerpc/kvm/book3s_hv.c
3344
cpu += vcpu->arch.ptid;
arch/powerpc/kvm/book3s_hv.c
3346
vcpu->arch.thread_cpu = cpu;
arch/powerpc/kvm/book3s_hv.c
343
spin_lock_irqsave(&vcpu->arch.tbacct_lock, flags);
arch/powerpc/kvm/book3s_hv.c
344
if (vcpu->arch.state == KVMPPC_VCPU_BUSY_IN_HOST &&
arch/powerpc/kvm/book3s_hv.c
345
vcpu->arch.busy_preempt != TB_NIL) {
arch/powerpc/kvm/book3s_hv.c
346
vcpu->arch.busy_stolen += now - vcpu->arch.busy_preempt;
arch/powerpc/kvm/book3s_hv.c
347
vcpu->arch.busy_preempt = TB_NIL;
arch/powerpc/kvm/book3s_hv.c
349
spin_unlock_irqrestore(&vcpu->arch.tbacct_lock, flags);
arch/powerpc/kvm/book3s_hv.c
354
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
3586
if (signal_pending(vcpu->arch.run_task))
arch/powerpc/kvm/book3s_hv.c
3587
vcpu->arch.ret = -EINTR;
arch/powerpc/kvm/book3s_hv.c
3588
else if (vcpu->arch.vpa.update_pending ||
arch/powerpc/kvm/book3s_hv.c
3589
vcpu->arch.slb_shadow.update_pending ||
arch/powerpc/kvm/book3s_hv.c
3590
vcpu->arch.dtl.update_pending)
arch/powerpc/kvm/book3s_hv.c
3591
vcpu->arch.ret = RESUME_GUEST;
arch/powerpc/kvm/book3s_hv.c
3595
wake_up(&vcpu->arch.cpu_run);
arch/powerpc/kvm/book3s_hv.c
3609
if (!pvc->n_runnable || !pvc->kvm->arch.mmu_ready) {
arch/powerpc/kvm/book3s_hv.c
363
WARN_ON_ONCE(vcpu->arch.state == KVMPPC_VCPU_RUNNABLE);
arch/powerpc/kvm/book3s_hv.c
3638
if (!vc->kvm->arch.mmu_ready)
arch/powerpc/kvm/book3s_hv.c
3641
if (signal_pending(vcpu->arch.run_task))
arch/powerpc/kvm/book3s_hv.c
3673
if (vcpu->arch.trap)
arch/powerpc/kvm/book3s_hv.c
3675
vcpu->arch.run_task);
arch/powerpc/kvm/book3s_hv.c
3677
vcpu->arch.ret = ret;
arch/powerpc/kvm/book3s_hv.c
3678
vcpu->arch.trap = 0;
arch/powerpc/kvm/book3s_hv.c
3681
if (is_kvmppc_resume_guest(vcpu->arch.ret)) {
arch/powerpc/kvm/book3s_hv.c
3682
if (vcpu->arch.pending_exceptions)
arch/powerpc/kvm/book3s_hv.c
3684
if (vcpu->arch.ceded)
arch/powerpc/kvm/book3s_hv.c
369
vcpu->arch.state == KVMPPC_VCPU_BUSY_IN_HOST)
arch/powerpc/kvm/book3s_hv.c
3690
wake_up(&vcpu->arch.cpu_run);
arch/powerpc/kvm/book3s_hv.c
370
vcpu->arch.busy_preempt = mftb();
arch/powerpc/kvm/book3s_hv.c
3706
wake_up(&vcpu->arch.cpu_run);
arch/powerpc/kvm/book3s_hv.c
379
spin_lock_irqsave(&vcpu->arch.tbacct_lock, flags);
arch/powerpc/kvm/book3s_hv.c
380
if (vcpu->arch.state == KVMPPC_VCPU_BUSY_IN_HOST)
arch/powerpc/kvm/book3s_hv.c
3804
if (vc->runner->arch.state != KVMPPC_VCPU_RUNNABLE)
arch/powerpc/kvm/book3s_hv.c
381
vcpu->arch.busy_preempt = now;
arch/powerpc/kvm/book3s_hv.c
382
spin_unlock_irqrestore(&vcpu->arch.tbacct_lock, flags);
arch/powerpc/kvm/book3s_hv.c
3828
vcpu->arch.ret = -EBUSY;
arch/powerpc/kvm/book3s_hv.c
3830
wake_up(&vcpu->arch.cpu_run);
arch/powerpc/kvm/book3s_hv.c
387
vcpu->arch.pvr = pvr;
arch/powerpc/kvm/book3s_hv.c
3971
if (!vcpu->arch.ptid)
arch/powerpc/kvm/book3s_hv.c
3973
active |= 1 << (thr + vcpu->arch.ptid);
arch/powerpc/kvm/book3s_hv.c
4102
struct lppaca *lp = vcpu->arch.vpa.pinned_addr;
arch/powerpc/kvm/book3s_hv.c
4106
vcpu->arch.vpa.dirty = 1;
arch/powerpc/kvm/book3s_hv.c
4169
vcpu->arch.l1_to_l2_cs = l1_to_l2_ns;
arch/powerpc/kvm/book3s_hv.c
417
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
4170
vcpu->arch.l2_to_l1_cs = l2_to_l1_ns;
arch/powerpc/kvm/book3s_hv.c
4171
vcpu->arch.l2_runtime_agg = l2_runtime_ns;
arch/powerpc/kvm/book3s_hv.c
4195
struct kvm_vcpu_arch *arch;
arch/powerpc/kvm/book3s_hv.c
4199
arch = &vcpu->arch;
arch/powerpc/kvm/book3s_hv.c
4200
return arch->l1_to_l2_cs;
arch/powerpc/kvm/book3s_hv.c
4210
struct kvm_vcpu_arch *arch;
arch/powerpc/kvm/book3s_hv.c
4214
arch = &vcpu->arch;
arch/powerpc/kvm/book3s_hv.c
4215
return arch->l2_to_l1_cs;
arch/powerpc/kvm/book3s_hv.c
4225
struct kvm_vcpu_arch *arch;
arch/powerpc/kvm/book3s_hv.c
4229
arch = &vcpu->arch;
arch/powerpc/kvm/book3s_hv.c
4230
return arch->l2_runtime_agg;
arch/powerpc/kvm/book3s_hv.c
4256
if (vcpu->arch.doorbell_request) {
arch/powerpc/kvm/book3s_hv.c
4257
vcpu->arch.doorbell_request = 0;
arch/powerpc/kvm/book3s_hv.c
4261
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv.c
4274
accumulate_time(vcpu, &vcpu->arch.in_guest);
arch/powerpc/kvm/book3s_hv.c
4275
rc = plpar_guest_run_vcpu(0, vcpu->kvm->arch.lpid, vcpu->vcpu_id,
arch/powerpc/kvm/book3s_hv.c
4288
accumulate_time(vcpu, &vcpu->arch.guest_exit);
arch/powerpc/kvm/book3s_hv.c
4337
if (vcpu->arch.psscr != host_psscr)
arch/powerpc/kvm/book3s_hv.c
4338
mtspr(SPRN_PSSCR_PR, vcpu->arch.psscr);
arch/powerpc/kvm/book3s_hv.c
4343
vcpu->arch.regs.msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv.c
4345
if (vcpu->arch.nested) {
arch/powerpc/kvm/book3s_hv.c
4346
hvregs.lpid = vcpu->arch.nested->shadow_lpid;
arch/powerpc/kvm/book3s_hv.c
4347
hvregs.vcpu_token = vcpu->arch.nested_vcpu_id;
arch/powerpc/kvm/book3s_hv.c
4349
hvregs.lpid = vcpu->kvm->arch.lpid;
arch/powerpc/kvm/book3s_hv.c
4360
if (vcpu->arch.doorbell_request)
arch/powerpc/kvm/book3s_hv.c
4361
vcpu->arch.doorbell_request = 0;
arch/powerpc/kvm/book3s_hv.c
4379
mtspr(SPRN_DAR, vcpu->arch.shregs.dar);
arch/powerpc/kvm/book3s_hv.c
4380
mtspr(SPRN_DSISR, vcpu->arch.shregs.dsisr);
arch/powerpc/kvm/book3s_hv.c
4382
accumulate_time(vcpu, &vcpu->arch.in_guest);
arch/powerpc/kvm/book3s_hv.c
4384
__pa(&vcpu->arch.regs));
arch/powerpc/kvm/book3s_hv.c
4385
accumulate_time(vcpu, &vcpu->arch.guest_exit);
arch/powerpc/kvm/book3s_hv.c
4388
vcpu->arch.shregs.msr = vcpu->arch.regs.msr;
arch/powerpc/kvm/book3s_hv.c
4389
vcpu->arch.shregs.dar = mfspr(SPRN_DAR);
arch/powerpc/kvm/book3s_hv.c
4390
vcpu->arch.shregs.dsisr = mfspr(SPRN_DSISR);
arch/powerpc/kvm/book3s_hv.c
4391
vcpu->arch.psscr = mfspr(SPRN_PSSCR_PR);
arch/powerpc/kvm/book3s_hv.c
4399
vcpu->arch.dec_expires = dec + (*tb + kvmppc_get_tb_offset(vcpu));
arch/powerpc/kvm/book3s_hv.c
4404
if (vcpu->arch.psscr != host_psscr)
arch/powerpc/kvm/book3s_hv.c
4417
struct kvm_nested_guest *nested = vcpu->arch.nested;
arch/powerpc/kvm/book3s_hv.c
4429
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv.c
4476
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv.c
4508
vcpu->arch.slb_max = 0;
arch/powerpc/kvm/book3s_hv.c
4525
prepare_to_wait(&vcpu->arch.cpu_run, &wait, wait_state);
arch/powerpc/kvm/book3s_hv.c
4526
if (vcpu->arch.state == KVMPPC_VCPU_RUNNABLE) {
arch/powerpc/kvm/book3s_hv.c
4531
finish_wait(&vcpu->arch.cpu_run, &wait);
arch/powerpc/kvm/book3s_hv.c
4557
return vcpu->arch.irq_pending || vcpu->arch.xive_saved_state.pipr <
arch/powerpc/kvm/book3s_hv.c
4558
vcpu->arch.xive_saved_state.cppr;
arch/powerpc/kvm/book3s_hv.c
4569
if (vcpu->arch.pending_exceptions || vcpu->arch.prodded ||
arch/powerpc/kvm/book3s_hv.c
4578
if (!vcpu->arch.ceded || kvmppc_vcpu_woken(vcpu))
arch/powerpc/kvm/book3s_hv.c
4723
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
4724
if (!kvm->arch.mmu_ready) {
arch/powerpc/kvm/book3s_hv.c
4730
kvm->arch.mmu_ready = 1;
arch/powerpc/kvm/book3s_hv.c
4733
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
4747
vcpu->arch.ret = RESUME_GUEST;
arch/powerpc/kvm/book3s_hv.c
4748
vcpu->arch.trap = 0;
arch/powerpc/kvm/book3s_hv.c
4754
vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
4756
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv.c
4757
vcpu->arch.run_task = current;
arch/powerpc/kvm/book3s_hv.c
4758
vcpu->arch.stolen_logged = vcore_stolen_time(vc, mftb());
arch/powerpc/kvm/book3s_hv.c
4759
vcpu->arch.state = KVMPPC_VCPU_RUNNABLE;
arch/powerpc/kvm/book3s_hv.c
4760
vcpu->arch.busy_preempt = TB_NIL;
arch/powerpc/kvm/book3s_hv.c
4761
WRITE_ONCE(vc->runnable_threads[vcpu->arch.ptid], vcpu);
arch/powerpc/kvm/book3s_hv.c
4782
while (vcpu->arch.state == KVMPPC_VCPU_RUNNABLE &&
arch/powerpc/kvm/book3s_hv.c
4785
if (!vcpu->kvm->arch.mmu_ready) {
arch/powerpc/kvm/book3s_hv.c
4793
vcpu->arch.ret = r;
arch/powerpc/kvm/book3s_hv.c
4807
if (signal_pending(v->arch.run_task)) {
arch/powerpc/kvm/book3s_hv.c
4811
v->arch.ret = -EINTR;
arch/powerpc/kvm/book3s_hv.c
4812
wake_up(&v->arch.cpu_run);
arch/powerpc/kvm/book3s_hv.c
4815
if (!vc->n_runnable || vcpu->arch.state != KVMPPC_VCPU_RUNNABLE)
arch/powerpc/kvm/book3s_hv.c
4820
n_ceded += v->arch.ceded;
arch/powerpc/kvm/book3s_hv.c
4822
v->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv.c
4839
while (vcpu->arch.state == KVMPPC_VCPU_RUNNABLE &&
arch/powerpc/kvm/book3s_hv.c
4848
if (vcpu->arch.state == KVMPPC_VCPU_RUNNABLE) {
arch/powerpc/kvm/book3s_hv.c
4852
vcpu->arch.ret = -EINTR;
arch/powerpc/kvm/book3s_hv.c
4859
wake_up(&v->arch.cpu_run);
arch/powerpc/kvm/book3s_hv.c
4864
return vcpu->arch.ret;
arch/powerpc/kvm/book3s_hv.c
4876
struct kvm_nested_guest *nested = vcpu->arch.nested;
arch/powerpc/kvm/book3s_hv.c
4883
vcpu->arch.ret = RESUME_GUEST;
arch/powerpc/kvm/book3s_hv.c
4884
vcpu->arch.trap = 0;
arch/powerpc/kvm/book3s_hv.c
4886
vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
4887
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv.c
4888
vcpu->arch.run_task = current;
arch/powerpc/kvm/book3s_hv.c
4889
vcpu->arch.last_inst = KVM_INST_FETCH_FAILED;
arch/powerpc/kvm/book3s_hv.c
4892
if (unlikely(!kvm->arch.mmu_ready)) {
arch/powerpc/kvm/book3s_hv.c
4897
vcpu->arch.ret = r;
arch/powerpc/kvm/book3s_hv.c
491
vcpu->arch.regs.nip, vcpu->arch.shregs.msr, vcpu->arch.trap);
arch/powerpc/kvm/book3s_hv.c
4915
vcpu->arch.state = KVMPPC_VCPU_RUNNABLE;
arch/powerpc/kvm/book3s_hv.c
4919
if (need_resched() || !kvm->arch.mmu_ready)
arch/powerpc/kvm/book3s_hv.c
4923
vcpu->arch.thread_cpu = pcpu;
arch/powerpc/kvm/book3s_hv.c
4940
&vcpu->arch.pending_exceptions) ||
arch/powerpc/kvm/book3s_hv.c
4965
} else if (vcpu->arch.pending_exceptions ||
arch/powerpc/kvm/book3s_hv.c
4967
vcpu->arch.ret = RESUME_HOST;
arch/powerpc/kvm/book3s_hv.c
497
vcpu->arch.regs.ctr, vcpu->arch.regs.link);
arch/powerpc/kvm/book3s_hv.c
4971
if (vcpu->arch.timer_running) {
arch/powerpc/kvm/book3s_hv.c
4972
hrtimer_try_to_cancel(&vcpu->arch.dec_timer);
arch/powerpc/kvm/book3s_hv.c
4973
vcpu->arch.timer_running = 0;
arch/powerpc/kvm/book3s_hv.c
499
vcpu->arch.shregs.srr0, vcpu->arch.shregs.srr1);
arch/powerpc/kvm/book3s_hv.c
4990
vcpu->arch.trap = trap;
arch/powerpc/kvm/book3s_hv.c
5000
vcpu->arch.thread_cpu = -1;
arch/powerpc/kvm/book3s_hv.c
5001
vcpu->arch.state = KVMPPC_VCPU_BUSY_IN_HOST;
arch/powerpc/kvm/book3s_hv.c
501
vcpu->arch.shregs.sprg0, vcpu->arch.shregs.sprg1);
arch/powerpc/kvm/book3s_hv.c
503
vcpu->arch.shregs.sprg2, vcpu->arch.shregs.sprg3);
arch/powerpc/kvm/book3s_hv.c
5040
vcpu->arch.ret = r;
arch/powerpc/kvm/book3s_hv.c
505
vcpu->arch.regs.ccr, vcpu->arch.regs.xer, vcpu->arch.shregs.dsisr);
arch/powerpc/kvm/book3s_hv.c
5051
vcpu->arch.ret = -EINTR;
arch/powerpc/kvm/book3s_hv.c
506
pr_err("dar = %.16llx\n", vcpu->arch.shregs.dar);
arch/powerpc/kvm/book3s_hv.c
5064
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv.c
5069
return vcpu->arch.ret;
arch/powerpc/kvm/book3s_hv.c
5074
vcpu->arch.ret = -EINTR;
arch/powerpc/kvm/book3s_hv.c
5077
vcpu->arch.thread_cpu = -1;
arch/powerpc/kvm/book3s_hv.c
5078
vcpu->arch.state = KVMPPC_VCPU_BUSY_IN_HOST;
arch/powerpc/kvm/book3s_hv.c
508
vcpu->arch.fault_dar, vcpu->arch.fault_dsisr);
arch/powerpc/kvm/book3s_hv.c
509
pr_err("SLB (%d entries):\n", vcpu->arch.slb_max);
arch/powerpc/kvm/book3s_hv.c
5092
start_timing(vcpu, &vcpu->arch.vcpu_entry);
arch/powerpc/kvm/book3s_hv.c
5094
if (!vcpu->arch.sane) {
arch/powerpc/kvm/book3s_hv.c
510
for (r = 0; r < vcpu->arch.slb_max; ++r)
arch/powerpc/kvm/book3s_hv.c
512
vcpu->arch.slb[r].orige, vcpu->arch.slb[r].origv);
arch/powerpc/kvm/book3s_hv.c
5124
if (!vcpu->arch.online) {
arch/powerpc/kvm/book3s_hv.c
5125
atomic_inc(&vcpu->arch.vcore->online_count);
arch/powerpc/kvm/book3s_hv.c
5126
vcpu->arch.online = 1;
arch/powerpc/kvm/book3s_hv.c
5132
atomic_inc(&kvm->arch.vcpus_running);
arch/powerpc/kvm/book3s_hv.c
514
vcpu->arch.vcore->lpcr, vcpu->kvm->arch.sdr1,
arch/powerpc/kvm/book3s_hv.c
515
vcpu->arch.last_inst);
arch/powerpc/kvm/book3s_hv.c
5154
vcpu->arch.waitp = &vcpu->arch.vcore->wait;
arch/powerpc/kvm/book3s_hv.c
5155
vcpu->arch.pgdir = kvm->mm->pgd;
arch/powerpc/kvm/book3s_hv.c
5156
vcpu->arch.state = KVMPPC_VCPU_BUSY_IN_HOST;
arch/powerpc/kvm/book3s_hv.c
5159
accumulate_time(vcpu, &vcpu->arch.guest_entry);
arch/powerpc/kvm/book3s_hv.c
5162
vcpu->arch.vcore->lpcr);
arch/powerpc/kvm/book3s_hv.c
5167
accumulate_time(vcpu, &vcpu->arch.hcall);
arch/powerpc/kvm/book3s_hv.c
5184
accumulate_time(vcpu, &vcpu->arch.pg_fault);
arch/powerpc/kvm/book3s_hv.c
5187
vcpu->arch.fault_dar, vcpu->arch.fault_dsisr);
arch/powerpc/kvm/book3s_hv.c
5196
accumulate_time(vcpu, &vcpu->arch.vcpu_exit);
arch/powerpc/kvm/book3s_hv.c
5198
vcpu->arch.state = KVMPPC_VCPU_NOTREADY;
arch/powerpc/kvm/book3s_hv.c
5199
atomic_dec(&kvm->arch.vcpus_running);
arch/powerpc/kvm/book3s_hv.c
5311
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
5312
kvmppc_harvest_vpa_dirty(&vcpu->arch.vpa, memslot, buf);
arch/powerpc/kvm/book3s_hv.c
5313
kvmppc_harvest_vpa_dirty(&vcpu->arch.dtl, memslot, buf);
arch/powerpc/kvm/book3s_hv.c
5314
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
5329
vfree(slot->arch.rmap);
arch/powerpc/kvm/book3s_hv.c
5330
slot->arch.rmap = NULL;
arch/powerpc/kvm/book3s_hv.c
5339
unsigned long size = array_size(new->npages, sizeof(*new->arch.rmap));
arch/powerpc/kvm/book3s_hv.c
5344
new->arch.rmap = vzalloc(size);
arch/powerpc/kvm/book3s_hv.c
5345
if (!new->arch.rmap)
arch/powerpc/kvm/book3s_hv.c
5348
new->arch.rmap = old->arch.rmap;
arch/powerpc/kvm/book3s_hv.c
535
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
5366
atomic64_inc(&kvm->arch.mmio_update);
arch/powerpc/kvm/book3s_hv.c
5387
if (!kvm->arch.secure_guest)
arch/powerpc/kvm/book3s_hv.c
541
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
5417
if ((kvm->arch.lpcr & mask) == lpcr)
arch/powerpc/kvm/book3s_hv.c
5420
kvm->arch.lpcr = (kvm->arch.lpcr & ~mask) | lpcr;
arch/powerpc/kvm/book3s_hv.c
5423
struct kvmppc_vcore *vc = kvm->arch.vcores[i];
arch/powerpc/kvm/book3s_hv.c
5431
if (++cores_done >= kvm->arch.online_vcores)
arch/powerpc/kvm/book3s_hv.c
5450
dw0 = ((kvm->arch.vrma_slb_v & SLB_VSID_L) >> 1) |
arch/powerpc/kvm/book3s_hv.c
5451
((kvm->arch.vrma_slb_v & SLB_VSID_LP) << 1);
arch/powerpc/kvm/book3s_hv.c
5453
dw0 |= kvm->arch.sdr1;
arch/powerpc/kvm/book3s_hv.c
5456
dw1 = kvm->arch.process_table;
arch/powerpc/kvm/book3s_hv.c
5459
__pa(kvm->arch.pgtable) | RADIX_PGD_INDEX_SIZE;
arch/powerpc/kvm/book3s_hv.c
5460
dw1 = PATB_GR | kvm->arch.process_table;
arch/powerpc/kvm/book3s_hv.c
5462
kvmhv_set_ptbl_entry(kvm->arch.lpid, dw0, dw1);
arch/powerpc/kvm/book3s_hv.c
5481
if (!kvm->arch.hpt.virt) {
arch/powerpc/kvm/book3s_hv.c
5530
kvm->arch.vrma_slb_v = senc | SLB_VSID_B_1T |
arch/powerpc/kvm/book3s_hv.c
5566
kvm->arch.process_table = 0;
arch/powerpc/kvm/book3s_hv.c
5569
kvm->arch.radix = 0;
arch/powerpc/kvm/book3s_hv.c
5597
kvm->arch.radix = 1;
arch/powerpc/kvm/book3s_hv.c
5599
kvmppc_free_hpt(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_hv.c
5606
(kvm->arch.host_lpcr & LPCR_HAIL))
arch/powerpc/kvm/book3s_hv.c
5702
mutex_init(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv.c
5703
INIT_LIST_HEAD(&kvm->arch.uvmem_pfns);
arch/powerpc/kvm/book3s_hv.c
5704
mutex_init(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
5712
kvm->arch.lpid = lpid;
arch/powerpc/kvm/book3s_hv.c
5741
kvm->arch.lpid = guest_id;
arch/powerpc/kvm/book3s_hv.c
5753
cpumask_setall(&kvm->arch.need_tlb_flush);
arch/powerpc/kvm/book3s_hv.c
5756
memcpy(kvm->arch.enabled_hcalls, default_enabled_hcalls,
arch/powerpc/kvm/book3s_hv.c
5757
sizeof(kvm->arch.enabled_hcalls));
arch/powerpc/kvm/book3s_hv.c
5760
kvm->arch.host_sdr1 = mfspr(SPRN_SDR1);
arch/powerpc/kvm/book3s_hv.c
5764
kvm->arch.host_lpid = mfspr(SPRN_LPID);
arch/powerpc/kvm/book3s_hv.c
5765
kvm->arch.host_lpcr = lpcr = mfspr(SPRN_LPCR);
arch/powerpc/kvm/book3s_hv.c
5776
kvm->arch.vrma_slb_v = SLB_VSID_B_1T |
arch/powerpc/kvm/book3s_hv.c
5804
kvm->arch.radix = 1;
arch/powerpc/kvm/book3s_hv.c
5805
kvm->arch.mmu_ready = 1;
arch/powerpc/kvm/book3s_hv.c
5810
(kvm->arch.host_lpcr & LPCR_HAIL))
arch/powerpc/kvm/book3s_hv.c
5815
plpar_guest_delete(0, kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv.c
5817
kvmppc_free_lpid(kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv.c
5824
kvm->arch.lpcr = lpcr;
arch/powerpc/kvm/book3s_hv.c
5827
kvm->arch.resize_hpt = NULL;
arch/powerpc/kvm/book3s_hv.c
5837
kvm->arch.tlb_sets = 1;
arch/powerpc/kvm/book3s_hv.c
5839
kvm->arch.tlb_sets = POWER9_TLB_SETS_RADIX; /* 128 */
arch/powerpc/kvm/book3s_hv.c
5841
kvm->arch.tlb_sets = POWER9_TLB_SETS_HASH; /* 256 */
arch/powerpc/kvm/book3s_hv.c
5843
kvm->arch.tlb_sets = POWER8_TLB_SETS; /* 512 */
arch/powerpc/kvm/book3s_hv.c
5845
kvm->arch.tlb_sets = POWER7_TLB_SETS; /* 128 */
arch/powerpc/kvm/book3s_hv.c
5862
kvm->arch.smt_mode = threads_per_subcore;
arch/powerpc/kvm/book3s_hv.c
5864
kvm->arch.smt_mode = 1;
arch/powerpc/kvm/book3s_hv.c
5865
kvm->arch.emul_smt_mode = 1;
arch/powerpc/kvm/book3s_hv.c
5883
kfree(kvm->arch.vcores[i]);
arch/powerpc/kvm/book3s_hv.c
5884
kvm->arch.online_vcores = 0;
arch/powerpc/kvm/book3s_hv.c
5898
kvmppc_free_hpt(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_hv.c
5904
kvm->arch.process_table = 0;
arch/powerpc/kvm/book3s_hv.c
5905
if (kvm->arch.secure_guest)
arch/powerpc/kvm/book3s_hv.c
5906
uv_svm_terminate(kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv.c
5908
kvmhv_set_ptbl_entry(kvm->arch.lpid, 0, 0);
arch/powerpc/kvm/book3s_hv.c
5912
kvmhv_flush_lpid(kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv.c
5913
plpar_guest_delete(0, kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv.c
5915
kvmppc_free_lpid(kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv.c
5957
kfree(kvm->arch.pimap);
arch/powerpc/kvm/book3s_hv.c
5983
pimap = kvm->arch.pimap;
arch/powerpc/kvm/book3s_hv.c
5991
kvm->arch.pimap = pimap;
arch/powerpc/kvm/book3s_hv.c
604
spin_lock(&tvcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
6075
if (!kvm->arch.pimap)
arch/powerpc/kvm/book3s_hv.c
6078
pimap = kvm->arch.pimap;
arch/powerpc/kvm/book3s_hv.c
617
vpap = &tvcpu->arch.vpa;
arch/powerpc/kvm/book3s_hv.c
628
if (!vpa_is_registered(&tvcpu->arch.vpa))
arch/powerpc/kvm/book3s_hv.c
6288
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
6290
if (kvm->arch.mmu_ready) {
arch/powerpc/kvm/book3s_hv.c
6291
kvm->arch.mmu_ready = 0;
arch/powerpc/kvm/book3s_hv.c
6294
if (atomic_read(&kvm->arch.vcpus_running)) {
arch/powerpc/kvm/book3s_hv.c
6295
kvm->arch.mmu_ready = 1;
arch/powerpc/kvm/book3s_hv.c
6308
kvm->arch.process_table = cfg->process_table;
arch/powerpc/kvm/book3s_hv.c
631
vpap = &tvcpu->arch.dtl;
arch/powerpc/kvm/book3s_hv.c
6316
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
6333
kvm->arch.nested_enable = true;
arch/powerpc/kvm/book3s_hv.c
6350
if (rc && vcpu->arch.nested)
arch/powerpc/kvm/book3s_hv.c
6369
if (rc && vcpu->arch.nested)
arch/powerpc/kvm/book3s_hv.c
638
if (!vpa_is_registered(&tvcpu->arch.vpa))
arch/powerpc/kvm/book3s_hv.c
6395
kvm->arch.svm_enabled = 1;
arch/powerpc/kvm/book3s_hv.c
641
vpap = &tvcpu->arch.slb_shadow;
arch/powerpc/kvm/book3s_hv.c
6415
if (!(kvm->arch.secure_guest & KVMPPC_SECURE_INIT_START))
arch/powerpc/kvm/book3s_hv.c
6418
mutex_lock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
6419
mmu_was_ready = kvm->arch.mmu_ready;
arch/powerpc/kvm/book3s_hv.c
6420
if (kvm->arch.mmu_ready) {
arch/powerpc/kvm/book3s_hv.c
6421
kvm->arch.mmu_ready = 0;
arch/powerpc/kvm/book3s_hv.c
6424
if (atomic_read(&kvm->arch.vcpus_running)) {
arch/powerpc/kvm/book3s_hv.c
6425
kvm->arch.mmu_ready = 1;
arch/powerpc/kvm/book3s_hv.c
6442
uv_unregister_mem_slot(kvm->arch.lpid, memslot->id);
arch/powerpc/kvm/book3s_hv.c
6447
ret = uv_svm_terminate(kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv.c
6464
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
6465
unpin_vpa_reset(kvm, &vcpu->arch.dtl);
arch/powerpc/kvm/book3s_hv.c
6466
unpin_vpa_reset(kvm, &vcpu->arch.slb_shadow);
arch/powerpc/kvm/book3s_hv.c
6467
unpin_vpa_reset(kvm, &vcpu->arch.vpa);
arch/powerpc/kvm/book3s_hv.c
6468
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
6472
kvm->arch.secure_guest = 0;
arch/powerpc/kvm/book3s_hv.c
6473
kvm->arch.mmu_ready = mmu_was_ready;
arch/powerpc/kvm/book3s_hv.c
6475
mutex_unlock(&kvm->arch.mmu_setup_lock);
arch/powerpc/kvm/book3s_hv.c
648
if (vpa_is_registered(&tvcpu->arch.dtl) ||
arch/powerpc/kvm/book3s_hv.c
6486
kvm->arch.dawr1_enabled = true;
arch/powerpc/kvm/book3s_hv.c
649
vpa_is_registered(&tvcpu->arch.slb_shadow))
arch/powerpc/kvm/book3s_hv.c
652
vpap = &tvcpu->arch.vpa;
arch/powerpc/kvm/book3s_hv.c
657
vpap = &tvcpu->arch.dtl;
arch/powerpc/kvm/book3s_hv.c
662
vpap = &tvcpu->arch.slb_shadow;
arch/powerpc/kvm/book3s_hv.c
673
spin_unlock(&tvcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
696
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
701
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
733
if (!(vcpu->arch.vpa.update_pending ||
arch/powerpc/kvm/book3s_hv.c
734
vcpu->arch.slb_shadow.update_pending ||
arch/powerpc/kvm/book3s_hv.c
735
vcpu->arch.dtl.update_pending))
arch/powerpc/kvm/book3s_hv.c
738
spin_lock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
739
if (vcpu->arch.vpa.update_pending) {
arch/powerpc/kvm/book3s_hv.c
740
kvmppc_update_vpa(vcpu, &vcpu->arch.vpa, &old_vpa);
arch/powerpc/kvm/book3s_hv.c
747
if (vcpu->arch.vpa.pinned_addr) {
arch/powerpc/kvm/book3s_hv.c
748
init_vpa(vcpu, vcpu->arch.vpa.pinned_addr);
arch/powerpc/kvm/book3s_hv.c
750
kvmhv_nestedv2_set_vpa(vcpu, __pa(vcpu->arch.vpa.pinned_addr));
arch/powerpc/kvm/book3s_hv.c
753
if (vcpu->arch.dtl.update_pending) {
arch/powerpc/kvm/book3s_hv.c
754
kvmppc_update_vpa(vcpu, &vcpu->arch.dtl, &old_vpa);
arch/powerpc/kvm/book3s_hv.c
758
vcpu->arch.dtl_ptr = vcpu->arch.dtl.pinned_addr;
arch/powerpc/kvm/book3s_hv.c
759
vcpu->arch.dtl_index = 0;
arch/powerpc/kvm/book3s_hv.c
761
if (vcpu->arch.slb_shadow.update_pending) {
arch/powerpc/kvm/book3s_hv.c
762
kvmppc_update_vpa(vcpu, &vcpu->arch.slb_shadow, &old_vpa);
arch/powerpc/kvm/book3s_hv.c
768
spin_unlock(&vcpu->arch.vpa_update_lock);
arch/powerpc/kvm/book3s_hv.c
798
dt = vcpu->arch.dtl_ptr;
arch/powerpc/kvm/book3s_hv.c
805
dt->processor_id = cpu_to_be16(pcpu + vcpu->arch.ptid);
arch/powerpc/kvm/book3s_hv.c
812
dt->srr1 = cpu_to_be64(vcpu->arch.shregs.msr);
arch/powerpc/kvm/book3s_hv.c
815
if (dt == vcpu->arch.dtl.pinned_end)
arch/powerpc/kvm/book3s_hv.c
816
dt = vcpu->arch.dtl.pinned_addr;
arch/powerpc/kvm/book3s_hv.c
817
vcpu->arch.dtl_ptr = dt;
arch/powerpc/kvm/book3s_hv.c
820
vpa->dtl_idx = cpu_to_be64(++vcpu->arch.dtl_index);
arch/powerpc/kvm/book3s_hv.c
834
vpa = vcpu->arch.vpa.pinned_addr;
arch/powerpc/kvm/book3s_hv.c
841
stolen = core_stolen - vcpu->arch.stolen_logged;
arch/powerpc/kvm/book3s_hv.c
842
vcpu->arch.stolen_logged = core_stolen;
arch/powerpc/kvm/book3s_hv.c
843
spin_lock_irqsave(&vcpu->arch.tbacct_lock, flags);
arch/powerpc/kvm/book3s_hv.c
844
stolen += vcpu->arch.busy_stolen;
arch/powerpc/kvm/book3s_hv.c
845
vcpu->arch.busy_stolen = 0;
arch/powerpc/kvm/book3s_hv.c
846
spin_unlock_irqrestore(&vcpu->arch.tbacct_lock, flags);
arch/powerpc/kvm/book3s_hv.c
852
vcpu->arch.vpa.dirty = true;
arch/powerpc/kvm/book3s_hv.c
863
vpa = vcpu->arch.vpa.pinned_addr;
arch/powerpc/kvm/book3s_hv.c
868
stolen_delta = stolen - vcpu->arch.stolen_logged;
arch/powerpc/kvm/book3s_hv.c
869
vcpu->arch.stolen_logged = stolen;
arch/powerpc/kvm/book3s_hv.c
875
vcpu->arch.vpa.dirty = true;
arch/powerpc/kvm/book3s_hv.c
884
if (vcpu->arch.doorbell_request)
arch/powerpc/kvm/book3s_hv.c
894
vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv.c
945
if (!vcpu->kvm->arch.dawr1_enabled)
arch/powerpc/kvm/book3s_hv.h
32
return kvm->arch.nested_enable && kvm_is_radix(kvm);
arch/powerpc/kvm/book3s_hv.h
58
vcpu->arch.shregs.msr = val;
arch/powerpc/kvm/book3s_hv.h
65
return vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv.h
71
vcpu->arch.reg = val; \
arch/powerpc/kvm/book3s_hv.h
79
return vcpu->arch.reg; \
arch/powerpc/kvm/book3s_hv.h
89
vcpu->arch.reg[i] = val; \
arch/powerpc/kvm/book3s_hv.h
97
return vcpu->arch.reg[i]; \
arch/powerpc/kvm/book3s_hv_builtin.c
500
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv_builtin.c
501
if (vcpu->arch.timer_running) {
arch/powerpc/kvm/book3s_hv_builtin.c
502
hrtimer_try_to_cancel(&vcpu->arch.dec_timer);
arch/powerpc/kvm/book3s_hv_builtin.c
503
vcpu->arch.timer_running = 0;
arch/powerpc/kvm/book3s_hv_builtin.c
529
new_msr = vcpu->arch.intr_msr;
arch/powerpc/kvm/book3s_hv_builtin.c
547
(vcpu->arch.vcore->lpcr & LPCR_AIL) == LPCR_AIL_3 &&
arch/powerpc/kvm/book3s_hv_builtin.c
578
ext = (vcpu->arch.pending_exceptions >> BOOK3S_IRQPRIO_EXTERNAL) & 1;
arch/powerpc/kvm/book3s_hv_builtin.c
584
if (vcpu->arch.shregs.msr & MSR_EE) {
arch/powerpc/kvm/book3s_hv_builtin.c
597
if (vcpu->arch.doorbell_request) {
arch/powerpc/kvm/book3s_hv_builtin.c
599
vcpu->arch.vcore->dpdes = 1;
arch/powerpc/kvm/book3s_hv_builtin.c
601
vcpu->arch.doorbell_request = 0;
arch/powerpc/kvm/book3s_hv_builtin.c
610
for (set = 0; set < kvm->arch.tlb_sets; ++set) {
arch/powerpc/kvm/book3s_hv_builtin.c
622
if (cpumask_test_cpu(pcpu, &kvm->arch.need_tlb_flush)) {
arch/powerpc/kvm/book3s_hv_builtin.c
626
cpumask_clear_cpu(pcpu, &kvm->arch.need_tlb_flush);
arch/powerpc/kvm/book3s_hv_nested.c
1053
unsigned long *rmap = &memslot->arch.rmap[gfn];
arch/powerpc/kvm/book3s_hv_nested.c
106
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_nested.c
1063
unsigned long rmap, *rmapp = &free->arch.rmap[page];
arch/powerpc/kvm/book3s_hv_nested.c
108
hr->dpdes = vcpu->arch.doorbell_request;
arch/powerpc/kvm/book3s_hv_nested.c
109
hr->purr = vcpu->arch.purr;
arch/powerpc/kvm/book3s_hv_nested.c
110
hr->spurr = vcpu->arch.spurr;
arch/powerpc/kvm/book3s_hv_nested.c
111
hr->ic = vcpu->arch.ic;
arch/powerpc/kvm/book3s_hv_nested.c
113
hr->srr0 = vcpu->arch.shregs.srr0;
arch/powerpc/kvm/book3s_hv_nested.c
114
hr->srr1 = vcpu->arch.shregs.srr1;
arch/powerpc/kvm/book3s_hv_nested.c
115
hr->sprg[0] = vcpu->arch.shregs.sprg0;
arch/powerpc/kvm/book3s_hv_nested.c
116
hr->sprg[1] = vcpu->arch.shregs.sprg1;
arch/powerpc/kvm/book3s_hv_nested.c
117
hr->sprg[2] = vcpu->arch.shregs.sprg2;
arch/powerpc/kvm/book3s_hv_nested.c
118
hr->sprg[3] = vcpu->arch.shregs.sprg3;
arch/powerpc/kvm/book3s_hv_nested.c
119
hr->pidr = vcpu->arch.pid;
arch/powerpc/kvm/book3s_hv_nested.c
120
hr->cfar = vcpu->arch.cfar;
arch/powerpc/kvm/book3s_hv_nested.c
1206
idr_for_each_entry(&kvm->arch.kvm_nested_guest_idr, gp, lpid) {
arch/powerpc/kvm/book3s_hv_nested.c
121
hr->ppr = vcpu->arch.ppr;
arch/powerpc/kvm/book3s_hv_nested.c
122
switch (vcpu->arch.trap) {
arch/powerpc/kvm/book3s_hv_nested.c
124
hr->hdar = vcpu->arch.fault_dar;
arch/powerpc/kvm/book3s_hv_nested.c
125
hr->hdsisr = vcpu->arch.fault_dsisr;
arch/powerpc/kvm/book3s_hv_nested.c
126
hr->asdr = vcpu->arch.fault_gpa;
arch/powerpc/kvm/book3s_hv_nested.c
129
hr->asdr = vcpu->arch.fault_gpa;
arch/powerpc/kvm/book3s_hv_nested.c
133
(HFSCR_INTR_CAUSE & vcpu->arch.hfscr));
arch/powerpc/kvm/book3s_hv_nested.c
136
hr->heir = vcpu->arch.emul_inst;
arch/powerpc/kvm/book3s_hv_nested.c
1418
vcpu->arch.fault_gpa = fault_addr;
arch/powerpc/kvm/book3s_hv_nested.c
143
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_nested.c
1432
} else if (vcpu->arch.trap == BOOK3S_INTERRUPT_H_INST_STORAGE) {
arch/powerpc/kvm/book3s_hv_nested.c
1450
vcpu->arch.fault_dsisr = flags;
arch/powerpc/kvm/book3s_hv_nested.c
1451
if (vcpu->arch.trap == BOOK3S_INTERRUPT_H_INST_STORAGE) {
arch/powerpc/kvm/book3s_hv_nested.c
1452
vcpu->arch.shregs.msr &= SRR1_MSR_BITS;
arch/powerpc/kvm/book3s_hv_nested.c
1453
vcpu->arch.shregs.msr |= flags;
arch/powerpc/kvm/book3s_hv_nested.c
146
vcpu->arch.doorbell_request = hr->dpdes;
arch/powerpc/kvm/book3s_hv_nested.c
147
vcpu->arch.hfscr = hr->hfscr;
arch/powerpc/kvm/book3s_hv_nested.c
1479
gpte.raddr, kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv_nested.c
148
vcpu->arch.dawr0 = hr->dawr0;
arch/powerpc/kvm/book3s_hv_nested.c
149
vcpu->arch.dawrx0 = hr->dawrx0;
arch/powerpc/kvm/book3s_hv_nested.c
150
vcpu->arch.ciabr = hr->ciabr;
arch/powerpc/kvm/book3s_hv_nested.c
151
vcpu->arch.purr = hr->purr;
arch/powerpc/kvm/book3s_hv_nested.c
152
vcpu->arch.spurr = hr->spurr;
arch/powerpc/kvm/book3s_hv_nested.c
153
vcpu->arch.ic = hr->ic;
arch/powerpc/kvm/book3s_hv_nested.c
1532
unsigned long dsisr = vcpu->arch.fault_dsisr;
arch/powerpc/kvm/book3s_hv_nested.c
1533
unsigned long ea = vcpu->arch.fault_dar;
arch/powerpc/kvm/book3s_hv_nested.c
1548
n_gpa = vcpu->arch.fault_gpa & ~0xF000000000000FFFULL;
arch/powerpc/kvm/book3s_hv_nested.c
155
vcpu->arch.shregs.srr0 = hr->srr0;
arch/powerpc/kvm/book3s_hv_nested.c
156
vcpu->arch.shregs.srr1 = hr->srr1;
arch/powerpc/kvm/book3s_hv_nested.c
157
vcpu->arch.shregs.sprg0 = hr->sprg[0];
arch/powerpc/kvm/book3s_hv_nested.c
158
vcpu->arch.shregs.sprg1 = hr->sprg[1];
arch/powerpc/kvm/book3s_hv_nested.c
159
vcpu->arch.shregs.sprg2 = hr->sprg[2];
arch/powerpc/kvm/book3s_hv_nested.c
160
vcpu->arch.shregs.sprg3 = hr->sprg[3];
arch/powerpc/kvm/book3s_hv_nested.c
161
vcpu->arch.pid = hr->pidr;
arch/powerpc/kvm/book3s_hv_nested.c
162
vcpu->arch.cfar = hr->cfar;
arch/powerpc/kvm/book3s_hv_nested.c
163
vcpu->arch.ppr = hr->ppr;
arch/powerpc/kvm/book3s_hv_nested.c
164
vcpu->arch.dawr1 = hr->dawr1;
arch/powerpc/kvm/book3s_hv_nested.c
165
vcpu->arch.dawrx1 = hr->dawrx1;
arch/powerpc/kvm/book3s_hv_nested.c
1679
rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
arch/powerpc/kvm/book3s_hv_nested.c
1695
struct kvm_nested_guest *gp = vcpu->arch.nested;
arch/powerpc/kvm/book3s_hv_nested.c
1709
if (!idr_get_next(&kvm->arch.kvm_nested_guest_idr, &ret))
arch/powerpc/kvm/book3s_hv_nested.c
171
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_nested.c
179
vcpu->arch.doorbell_request = vcpu->arch.doorbell_request | hr->dpdes;
arch/powerpc/kvm/book3s_hv_nested.c
180
vcpu->arch.hfscr = hr->hfscr;
arch/powerpc/kvm/book3s_hv_nested.c
181
vcpu->arch.purr = hr->purr;
arch/powerpc/kvm/book3s_hv_nested.c
182
vcpu->arch.spurr = hr->spurr;
arch/powerpc/kvm/book3s_hv_nested.c
183
vcpu->arch.ic = hr->ic;
arch/powerpc/kvm/book3s_hv_nested.c
185
vcpu->arch.fault_dar = hr->hdar;
arch/powerpc/kvm/book3s_hv_nested.c
186
vcpu->arch.fault_dsisr = hr->hdsisr;
arch/powerpc/kvm/book3s_hv_nested.c
187
vcpu->arch.fault_gpa = hr->asdr;
arch/powerpc/kvm/book3s_hv_nested.c
188
vcpu->arch.emul_inst = hr->heir;
arch/powerpc/kvm/book3s_hv_nested.c
189
vcpu->arch.shregs.srr0 = hr->srr0;
arch/powerpc/kvm/book3s_hv_nested.c
190
vcpu->arch.shregs.srr1 = hr->srr1;
arch/powerpc/kvm/book3s_hv_nested.c
191
vcpu->arch.shregs.sprg0 = hr->sprg[0];
arch/powerpc/kvm/book3s_hv_nested.c
192
vcpu->arch.shregs.sprg1 = hr->sprg[1];
arch/powerpc/kvm/book3s_hv_nested.c
193
vcpu->arch.shregs.sprg2 = hr->sprg[2];
arch/powerpc/kvm/book3s_hv_nested.c
194
vcpu->arch.shregs.sprg3 = hr->sprg[3];
arch/powerpc/kvm/book3s_hv_nested.c
195
vcpu->arch.pid = hr->pidr;
arch/powerpc/kvm/book3s_hv_nested.c
196
vcpu->arch.cfar = hr->cfar;
arch/powerpc/kvm/book3s_hv_nested.c
197
vcpu->arch.ppr = hr->ppr;
arch/powerpc/kvm/book3s_hv_nested.c
203
vcpu->arch.trap = 0;
arch/powerpc/kvm/book3s_hv_nested.c
211
if (((vcpu->arch.io_gpr & KVM_MMIO_REG_EXT_MASK) == KVM_MMIO_REG_GPR)
arch/powerpc/kvm/book3s_hv_nested.c
213
vcpu->arch.nested_io_gpr = (gpa_t) regs_ptr +
arch/powerpc/kvm/book3s_hv_nested.c
215
gpr[vcpu->arch.io_gpr]);
arch/powerpc/kvm/book3s_hv_nested.c
216
vcpu->arch.io_gpr = KVM_MMIO_REG_NESTED_GPR;
arch/powerpc/kvm/book3s_hv_nested.c
263
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_nested.c
284
vcpu->arch.hfscr = l2_hv->hfscr & (HFSCR_INTR_CAUSE | vcpu->arch.hfscr_permitted);
arch/powerpc/kvm/book3s_hv_nested.c
287
vcpu->arch.dawrx0 = l2_hv->dawrx0 & ~DAWRX_HYP;
arch/powerpc/kvm/book3s_hv_nested.c
288
vcpu->arch.dawrx1 = l2_hv->dawrx1 & ~DAWRX_HYP;
arch/powerpc/kvm/book3s_hv_nested.c
292
vcpu->arch.ciabr = l2_hv->ciabr & ~CIABR_PRIV;
arch/powerpc/kvm/book3s_hv_nested.c
301
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_nested.c
306
if (vcpu->kvm->arch.l1_ptcr == 0)
arch/powerpc/kvm/book3s_hv_nested.c
309
if (MSR_TM_TRANSACTIONAL(vcpu->arch.shregs.msr))
arch/powerpc/kvm/book3s_hv_nested.c
32
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_nested.c
339
if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) {
arch/powerpc/kvm/book3s_hv_nested.c
345
if (WARN_ON_ONCE(vcpu->arch.shregs.msr & MSR_TS_MASK))
arch/powerpc/kvm/book3s_hv_nested.c
35
hr->dpdes = vcpu->arch.doorbell_request;
arch/powerpc/kvm/book3s_hv_nested.c
36
hr->hfscr = vcpu->arch.hfscr;
arch/powerpc/kvm/book3s_hv_nested.c
360
vcpu->arch.regs.msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_nested.c
361
saved_l1_regs = vcpu->arch.regs;
arch/powerpc/kvm/book3s_hv_nested.c
367
vcpu->arch.dec_expires += l2_hv.tb_offset;
arch/powerpc/kvm/book3s_hv_nested.c
370
vcpu->arch.nested = l2;
arch/powerpc/kvm/book3s_hv_nested.c
371
vcpu->arch.nested_vcpu_id = l2_hv.vcpu_token;
arch/powerpc/kvm/book3s_hv_nested.c
372
vcpu->arch.nested_hfscr = l2_hv.hfscr;
arch/powerpc/kvm/book3s_hv_nested.c
373
vcpu->arch.regs = l2_regs;
arch/powerpc/kvm/book3s_hv_nested.c
376
vcpu->arch.shregs.msr = (vcpu->arch.regs.msr | MSR_ME) & ~MSR_HV;
arch/powerpc/kvm/book3s_hv_nested.c
38
hr->dawr0 = vcpu->arch.dawr0;
arch/powerpc/kvm/book3s_hv_nested.c
381
vcpu->arch.ret = RESUME_GUEST;
arch/powerpc/kvm/book3s_hv_nested.c
382
vcpu->arch.trap = 0;
arch/powerpc/kvm/book3s_hv_nested.c
388
l2_regs = vcpu->arch.regs;
arch/powerpc/kvm/book3s_hv_nested.c
389
l2_regs.msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_nested.c
39
hr->dawrx0 = vcpu->arch.dawrx0;
arch/powerpc/kvm/book3s_hv_nested.c
390
delta_purr = vcpu->arch.purr - l2_hv.purr;
arch/powerpc/kvm/book3s_hv_nested.c
391
delta_spurr = vcpu->arch.spurr - l2_hv.spurr;
arch/powerpc/kvm/book3s_hv_nested.c
392
delta_ic = vcpu->arch.ic - l2_hv.ic;
arch/powerpc/kvm/book3s_hv_nested.c
397
vcpu->arch.nested = NULL;
arch/powerpc/kvm/book3s_hv_nested.c
398
vcpu->arch.regs = saved_l1_regs;
arch/powerpc/kvm/book3s_hv_nested.c
399
vcpu->arch.shregs.msr = saved_l1_regs.msr & ~MSR_TS_MASK;
arch/powerpc/kvm/book3s_hv_nested.c
40
hr->ciabr = vcpu->arch.ciabr;
arch/powerpc/kvm/book3s_hv_nested.c
402
vcpu->arch.shregs.msr |= MSR_TS_S;
arch/powerpc/kvm/book3s_hv_nested.c
405
vcpu->arch.dec_expires -= l2_hv.tb_offset;
arch/powerpc/kvm/book3s_hv_nested.c
407
vcpu->arch.purr += delta_purr;
arch/powerpc/kvm/book3s_hv_nested.c
408
vcpu->arch.spurr += delta_spurr;
arch/powerpc/kvm/book3s_hv_nested.c
409
vcpu->arch.ic += delta_ic;
arch/powerpc/kvm/book3s_hv_nested.c
41
hr->purr = vcpu->arch.purr;
arch/powerpc/kvm/book3s_hv_nested.c
42
hr->spurr = vcpu->arch.spurr;
arch/powerpc/kvm/book3s_hv_nested.c
43
hr->ic = vcpu->arch.ic;
arch/powerpc/kvm/book3s_hv_nested.c
434
return vcpu->arch.trap;
arch/powerpc/kvm/book3s_hv_nested.c
45
hr->srr0 = vcpu->arch.shregs.srr0;
arch/powerpc/kvm/book3s_hv_nested.c
46
hr->srr1 = vcpu->arch.shregs.srr1;
arch/powerpc/kvm/book3s_hv_nested.c
47
hr->sprg[0] = vcpu->arch.shregs.sprg0;
arch/powerpc/kvm/book3s_hv_nested.c
48
hr->sprg[1] = vcpu->arch.shregs.sprg1;
arch/powerpc/kvm/book3s_hv_nested.c
49
hr->sprg[2] = vcpu->arch.shregs.sprg2;
arch/powerpc/kvm/book3s_hv_nested.c
50
hr->sprg[3] = vcpu->arch.shregs.sprg3;
arch/powerpc/kvm/book3s_hv_nested.c
51
hr->pidr = vcpu->arch.pid;
arch/powerpc/kvm/book3s_hv_nested.c
52
hr->cfar = vcpu->arch.cfar;
arch/powerpc/kvm/book3s_hv_nested.c
53
hr->ppr = vcpu->arch.ppr;
arch/powerpc/kvm/book3s_hv_nested.c
54
hr->dawr1 = vcpu->arch.dawr1;
arch/powerpc/kvm/book3s_hv_nested.c
55
hr->dawrx1 = vcpu->arch.dawrx1;
arch/powerpc/kvm/book3s_hv_nested.c
582
kvm->arch.l1_ptcr = ptcr;
arch/powerpc/kvm/book3s_hv_nested.c
678
ptbl_addr = (kvm->arch.l1_ptcr & PRTB_MASK) + (gp->l1_lpid << 4);
arch/powerpc/kvm/book3s_hv_nested.c
679
if (gp->l1_lpid < (1ul << ((kvm->arch.l1_ptcr & PRTS_MASK) + 12 - 4))) {
arch/powerpc/kvm/book3s_hv_nested.c
697
idr_init(&kvm->arch.kvm_nested_guest_idr);
arch/powerpc/kvm/book3s_hv_nested.c
702
return idr_find(&kvm->arch.kvm_nested_guest_idr, lpid);
arch/powerpc/kvm/book3s_hv_nested.c
707
if (idr_alloc(&kvm->arch.kvm_nested_guest_idr,
arch/powerpc/kvm/book3s_hv_nested.c
715
if (idr_replace(&kvm->arch.kvm_nested_guest_idr, gp, lpid))
arch/powerpc/kvm/book3s_hv_nested.c
721
idr_remove(&kvm->arch.kvm_nested_guest_idr, lpid);
arch/powerpc/kvm/book3s_hv_nested.c
809
idr_for_each_entry(&kvm->arch.kvm_nested_guest_idr, gp, lpid) {
arch/powerpc/kvm/book3s_hv_nested.c
816
idr_destroy(&kvm->arch.kvm_nested_guest_idr);
arch/powerpc/kvm/book3s_hv_nested.c
849
if (l1_lpid >= (1ul << ((kvm->arch.l1_ptcr & PRTS_MASK) + 12 - 4)))
arch/powerpc/kvm/book3s_hv_nestedv2.c
1041
rc = plpar_guest_create_vcpu(0, vcpu->kvm->arch.lpid, vcpu->vcpu_id);
arch/powerpc/kvm/book3s_hv_nestedv2.c
161
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.dscr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
164
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.mmcra);
arch/powerpc/kvm/book3s_hv_nestedv2.c
167
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.hfscr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
170
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.purr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
173
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.spurr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
176
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.amr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
179
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.uamor);
arch/powerpc/kvm/book3s_hv_nestedv2.c
182
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.siar);
arch/powerpc/kvm/book3s_hv_nestedv2.c
185
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.sdar);
arch/powerpc/kvm/book3s_hv_nestedv2.c
188
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.iamr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
191
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.dawr0);
arch/powerpc/kvm/book3s_hv_nestedv2.c
194
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.dawr1);
arch/powerpc/kvm/book3s_hv_nestedv2.c
197
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.dawrx0);
arch/powerpc/kvm/book3s_hv_nestedv2.c
200
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.dawrx1);
arch/powerpc/kvm/book3s_hv_nestedv2.c
203
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.dexcr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
206
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.hashkeyr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
209
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.hashpkeyr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
212
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.ciabr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
215
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.wort);
arch/powerpc/kvm/book3s_hv_nestedv2.c
218
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.ppr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
221
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.pspb);
arch/powerpc/kvm/book3s_hv_nestedv2.c
224
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.tar);
arch/powerpc/kvm/book3s_hv_nestedv2.c
227
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.fscr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
230
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.ebbhr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
233
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.ebbrr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
236
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.bescr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
239
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.ic);
arch/powerpc/kvm/book3s_hv_nestedv2.c
242
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.ctrl);
arch/powerpc/kvm/book3s_hv_nestedv2.c
245
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.pid);
arch/powerpc/kvm/book3s_hv_nestedv2.c
254
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.vrsave);
arch/powerpc/kvm/book3s_hv_nestedv2.c
258
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.mmcr[i]);
arch/powerpc/kvm/book3s_hv_nestedv2.c
262
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.sier[i]);
arch/powerpc/kvm/book3s_hv_nestedv2.c
266
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.pmc[i]);
arch/powerpc/kvm/book3s_hv_nestedv2.c
271
vcpu->arch.regs.gpr[i]);
arch/powerpc/kvm/book3s_hv_nestedv2.c
274
rc = kvmppc_gse_put_u32(gsb, iden, vcpu->arch.regs.ccr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
277
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.regs.xer);
arch/powerpc/kvm/book3s_hv_nestedv2.c
280
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.regs.ctr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
284
vcpu->arch.regs.link);
arch/powerpc/kvm/book3s_hv_nestedv2.c
287
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.regs.nip);
arch/powerpc/kvm/book3s_hv_nestedv2.c
291
vcpu->arch.shregs.srr0);
arch/powerpc/kvm/book3s_hv_nestedv2.c
295
vcpu->arch.shregs.srr1);
arch/powerpc/kvm/book3s_hv_nestedv2.c
299
vcpu->arch.shregs.sprg0);
arch/powerpc/kvm/book3s_hv_nestedv2.c
303
vcpu->arch.shregs.sprg1);
arch/powerpc/kvm/book3s_hv_nestedv2.c
307
vcpu->arch.shregs.sprg2);
arch/powerpc/kvm/book3s_hv_nestedv2.c
311
vcpu->arch.shregs.sprg3);
arch/powerpc/kvm/book3s_hv_nestedv2.c
315
vcpu->arch.shregs.dar);
arch/powerpc/kvm/book3s_hv_nestedv2.c
319
vcpu->arch.shregs.dsisr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
323
vcpu->arch.shregs.msr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
327
vcpu->arch.vcore->vtb);
arch/powerpc/kvm/book3s_hv_nestedv2.c
331
vcpu->arch.vcore->dpdes);
arch/powerpc/kvm/book3s_hv_nestedv2.c
335
vcpu->arch.vcore->lpcr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
339
vcpu->arch.vcore->tb_offset);
arch/powerpc/kvm/book3s_hv_nestedv2.c
342
rc = kvmppc_gse_put_u64(gsb, iden, vcpu->arch.fp.fpscr);
arch/powerpc/kvm/book3s_hv_nestedv2.c
346
memcpy(&v, &vcpu->arch.fp.fpr[i],
arch/powerpc/kvm/book3s_hv_nestedv2.c
347
sizeof(vcpu->arch.fp.fpr[i]));
arch/powerpc/kvm/book3s_hv_nestedv2.c
353
vcpu->arch.vr.vscr.u[3]);
arch/powerpc/kvm/book3s_hv_nestedv2.c
358
&vcpu->arch.vr.vr[i]);
arch/powerpc/kvm/book3s_hv_nestedv2.c
364
dw = vcpu->arch.dec_expires -
arch/powerpc/kvm/book3s_hv_nestedv2.c
365
vcpu->arch.vcore->tb_offset;
arch/powerpc/kvm/book3s_hv_nestedv2.c
378
if (!vcpu->arch.vcore->arch_compat) {
arch/powerpc/kvm/book3s_hv_nestedv2.c
386
arch_compat = vcpu->arch.vcore->arch_compat;
arch/powerpc/kvm/book3s_hv_nestedv2.c
417
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv_nestedv2.c
424
vcpu->arch.dscr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
427
vcpu->arch.mmcra = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
430
vcpu->arch.hfscr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
433
vcpu->arch.purr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
436
vcpu->arch.spurr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
439
vcpu->arch.amr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
442
vcpu->arch.uamor = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
445
vcpu->arch.siar = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
448
vcpu->arch.sdar = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
451
vcpu->arch.iamr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
454
vcpu->arch.dawr0 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
457
vcpu->arch.dawr1 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
460
vcpu->arch.dawrx0 = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
463
vcpu->arch.dawrx1 = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
466
vcpu->arch.dexcr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
469
vcpu->arch.hashkeyr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
472
vcpu->arch.hashpkeyr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
475
vcpu->arch.ciabr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
478
vcpu->arch.wort = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
481
vcpu->arch.ppr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
484
vcpu->arch.pspb = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
487
vcpu->arch.tar = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
490
vcpu->arch.fscr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
493
vcpu->arch.ebbhr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
496
vcpu->arch.ebbrr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
499
vcpu->arch.bescr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
502
vcpu->arch.ic = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
505
vcpu->arch.ctrl = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
508
vcpu->arch.pid = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
513
vcpu->arch.vrsave = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
517
vcpu->arch.mmcr[i] = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
521
vcpu->arch.sier[i] = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
525
vcpu->arch.pmc[i] = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
529
vcpu->arch.regs.gpr[i] = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
532
vcpu->arch.regs.ccr = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
535
vcpu->arch.regs.xer = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
538
vcpu->arch.regs.ctr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
541
vcpu->arch.regs.link = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
544
vcpu->arch.regs.nip = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
547
vcpu->arch.shregs.srr0 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
550
vcpu->arch.shregs.srr1 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
553
vcpu->arch.shregs.sprg0 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
556
vcpu->arch.shregs.sprg1 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
559
vcpu->arch.shregs.sprg2 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
562
vcpu->arch.shregs.sprg3 = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
565
vcpu->arch.shregs.dar = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
568
vcpu->arch.shregs.dsisr = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
571
vcpu->arch.shregs.msr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
574
vcpu->arch.vcore->vtb = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
577
vcpu->arch.vcore->dpdes = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
580
vcpu->arch.vcore->lpcr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
583
vcpu->arch.vcore->tb_offset = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
586
vcpu->arch.fp.fpscr = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
591
memcpy(&vcpu->arch.fp.fpr[i], &v,
arch/powerpc/kvm/book3s_hv_nestedv2.c
592
sizeof(vcpu->arch.fp.fpr[i]));
arch/powerpc/kvm/book3s_hv_nestedv2.c
596
vcpu->arch.vr.vscr.u[3] = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
600
kvmppc_gse_get_vector128(gse, &vcpu->arch.vr.vr[i]);
arch/powerpc/kvm/book3s_hv_nestedv2.c
604
vcpu->arch.fault_dar = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
607
vcpu->arch.fault_dsisr = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
610
vcpu->arch.fault_gpa = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
613
vcpu->arch.emul_inst = kvmppc_gse_get_u64(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
619
vcpu->arch.dec_expires =
arch/powerpc/kvm/book3s_hv_nestedv2.c
620
dw + vcpu->arch.vcore->tb_offset;
arch/powerpc/kvm/book3s_hv_nestedv2.c
624
vcpu->arch.vcore->arch_compat = kvmppc_gse_get_u32(gse);
arch/powerpc/kvm/book3s_hv_nestedv2.c
651
guest_id = vcpu->kvm->arch.lpid;
arch/powerpc/kvm/book3s_hv_nestedv2.c
765
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv_nestedv2.c
795
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv_nestedv2.c
829
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv_nestedv2.c
918
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv_nestedv2.c
948
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv_nestedv2.c
951
vcpu->arch.fault_dar = 0;
arch/powerpc/kvm/book3s_hv_nestedv2.c
952
vcpu->arch.fault_dsisr = 0;
arch/powerpc/kvm/book3s_hv_nestedv2.c
953
vcpu->arch.fault_gpa = 0;
arch/powerpc/kvm/book3s_hv_nestedv2.c
954
vcpu->arch.emul_inst = KVM_INST_FETCH_FAILED;
arch/powerpc/kvm/book3s_hv_nestedv2.c
979
io = &vcpu->arch.nestedv2_io;
arch/powerpc/kvm/book3s_hv_p9_entry.c
100
mtspr(SPRN_TFHAR, vcpu->arch.tfhar);
arch/powerpc/kvm/book3s_hv_p9_entry.c
101
mtspr(SPRN_TFIAR, vcpu->arch.tfiar);
arch/powerpc/kvm/book3s_hv_p9_entry.c
108
load_fp_state(&vcpu->arch.fp);
arch/powerpc/kvm/book3s_hv_p9_entry.c
110
load_vr_state(&vcpu->arch.vr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
121
store_fp_state(&vcpu->arch.fp);
arch/powerpc/kvm/book3s_hv_p9_entry.c
123
store_vr_state(&vcpu->arch.vr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
129
unsigned long guest_msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
132
} else if (vcpu->arch.hfscr & HFSCR_TM) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
133
vcpu->arch.texasr = mfspr(SPRN_TEXASR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
134
vcpu->arch.tfhar = mfspr(SPRN_TFHAR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
135
vcpu->arch.tfiar = mfspr(SPRN_TFIAR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
137
if (!vcpu->arch.nested) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
138
vcpu->arch.load_tm++; /* see load_ebb comment */
arch/powerpc/kvm/book3s_hv_p9_entry.c
139
if (!vcpu->arch.load_tm)
arch/powerpc/kvm/book3s_hv_p9_entry.c
14
mtspr(SPRN_TAR, vcpu->arch.tar);
arch/powerpc/kvm/book3s_hv_p9_entry.c
140
vcpu->arch.hfscr &= ~HFSCR_TM;
arch/powerpc/kvm/book3s_hv_p9_entry.c
169
current->thread.tidr != vcpu->arch.tid)
arch/powerpc/kvm/book3s_hv_p9_entry.c
171
if (host_os_sprs->iamr != vcpu->arch.iamr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
173
if (vcpu->arch.uamor != 0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
175
if (host_os_sprs->amr != vcpu->arch.amr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
177
if (current->thread.fscr != vcpu->arch.fscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
179
if (current->thread.dscr != vcpu->arch.dscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
18
current->thread.vrsave != vcpu->arch.vrsave)
arch/powerpc/kvm/book3s_hv_p9_entry.c
181
if (vcpu->arch.pspb != 0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
185
if (!(vcpu->arch.ctrl & 1))
arch/powerpc/kvm/book3s_hv_p9_entry.c
19
mtspr(SPRN_VRSAVE, vcpu->arch.vrsave);
arch/powerpc/kvm/book3s_hv_p9_entry.c
190
vcpu->arch.vrsave != current->thread.vrsave)
arch/powerpc/kvm/book3s_hv_p9_entry.c
193
if (vcpu->arch.hfscr & HFSCR_EBB) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
194
if (vcpu->arch.bescr != current->thread.bescr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
196
if (vcpu->arch.ebbhr != current->thread.ebbhr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
198
if (vcpu->arch.ebbrr != current->thread.ebbrr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
201
if (!vcpu->arch.nested) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
208
vcpu->arch.load_ebb++;
arch/powerpc/kvm/book3s_hv_p9_entry.c
209
if (!vcpu->arch.load_ebb)
arch/powerpc/kvm/book3s_hv_p9_entry.c
210
vcpu->arch.hfscr &= ~HFSCR_EBB;
arch/powerpc/kvm/book3s_hv_p9_entry.c
214
if (vcpu->arch.tar != current->thread.tar)
arch/powerpc/kvm/book3s_hv_p9_entry.c
22
if (vcpu->arch.hfscr & HFSCR_EBB) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
222
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_p9_entry.c
229
curr = vcpu->arch.cur_activity;
arch/powerpc/kvm/book3s_hv_p9_entry.c
23
if (current->thread.ebbhr != vcpu->arch.ebbhr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
230
vcpu->arch.cur_activity = next;
arch/powerpc/kvm/book3s_hv_p9_entry.c
231
prev_tb = vcpu->arch.cur_tb_start;
arch/powerpc/kvm/book3s_hv_p9_entry.c
232
vcpu->arch.cur_tb_start = tb;
arch/powerpc/kvm/book3s_hv_p9_entry.c
24
mtspr(SPRN_EBBHR, vcpu->arch.ebbhr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
25
if (current->thread.ebbrr != vcpu->arch.ebbrr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
26
mtspr(SPRN_EBBRR, vcpu->arch.ebbrr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
27
if (current->thread.bescr != vcpu->arch.bescr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
28
mtspr(SPRN_BESCR, vcpu->arch.bescr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
303
struct kvm_nested_guest *nested = vcpu->arch.nested;
arch/powerpc/kvm/book3s_hv_p9_entry.c
307
lpid = nested ? nested->shadow_lpid : kvm->arch.lpid;
arch/powerpc/kvm/book3s_hv_p9_entry.c
32
current->thread.tidr != vcpu->arch.tid)
arch/powerpc/kvm/book3s_hv_p9_entry.c
33
mtspr(SPRN_TIDR, vcpu->arch.tid);
arch/powerpc/kvm/book3s_hv_p9_entry.c
332
lpid = kvm->arch.lpid;
arch/powerpc/kvm/book3s_hv_p9_entry.c
34
if (host_os_sprs->iamr != vcpu->arch.iamr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
346
for (i = 0; i < vcpu->arch.slb_max; i++)
arch/powerpc/kvm/book3s_hv_p9_entry.c
347
mtslb(vcpu->arch.slb[i].orige, vcpu->arch.slb[i].origv);
arch/powerpc/kvm/book3s_hv_p9_entry.c
35
mtspr(SPRN_IAMR, vcpu->arch.iamr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
355
u32 lpid = kvm->arch.host_lpid;
arch/powerpc/kvm/book3s_hv_p9_entry.c
356
u64 lpcr = kvm->arch.host_lpcr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
36
if (host_os_sprs->amr != vcpu->arch.amr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
37
mtspr(SPRN_AMR, vcpu->arch.amr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
38
if (vcpu->arch.uamor != 0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
39
mtspr(SPRN_UAMOR, vcpu->arch.uamor);
arch/powerpc/kvm/book3s_hv_p9_entry.c
40
if (current->thread.fscr != vcpu->arch.fscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
403
for (i = 0; i < vcpu->arch.slb_nr; i++) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
409
vcpu->arch.slb[nr].orige = slbee | i;
arch/powerpc/kvm/book3s_hv_p9_entry.c
41
mtspr(SPRN_FSCR, vcpu->arch.fscr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
410
vcpu->arch.slb[nr].origv = slbev;
arch/powerpc/kvm/book3s_hv_p9_entry.c
414
vcpu->arch.slb_max = nr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
42
if (current->thread.dscr != vcpu->arch.dscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
429
for (set = 1; set < kvm->arch.tlb_sets; ++set) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
43
mtspr(SPRN_DSCR, vcpu->arch.dscr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
44
if (vcpu->arch.pspb != 0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
440
for (set = 0; set < kvm->arch.tlb_sets; ++set) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
45
mtspr(SPRN_PSPB, vcpu->arch.pspb);
arch/powerpc/kvm/book3s_hv_p9_entry.c
463
need_tlb_flush = &kvm->arch.need_tlb_flush;
arch/powerpc/kvm/book3s_hv_p9_entry.c
509
(vcpu->arch.hfscr & HFSCR_TM))
arch/powerpc/kvm/book3s_hv_p9_entry.c
536
struct kvm_nested_guest *nested = vcpu->arch.nested;
arch/powerpc/kvm/book3s_hv_p9_entry.c
537
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_p9_entry.c
54
if (!(vcpu->arch.ctrl & 1))
arch/powerpc/kvm/book3s_hv_p9_entry.c
558
WARN_ON_ONCE(vcpu->arch.shregs.msr & MSR_HV);
arch/powerpc/kvm/book3s_hv_p9_entry.c
559
WARN_ON_ONCE(!(vcpu->arch.shregs.msr & MSR_ME));
arch/powerpc/kvm/book3s_hv_p9_entry.c
561
vcpu->arch.ceded = 0;
arch/powerpc/kvm/book3s_hv_p9_entry.c
60
vcpu->arch.tar = mfspr(SPRN_TAR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
608
mtspr(SPRN_PURR, vcpu->arch.purr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
609
mtspr(SPRN_SPURR, vcpu->arch.spurr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
613
if (vcpu->arch.doorbell_request) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
614
vcpu->arch.doorbell_request = 0;
arch/powerpc/kvm/book3s_hv_p9_entry.c
619
if (vcpu->arch.dawr0 != host_dawr0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
620
mtspr(SPRN_DAWR0, vcpu->arch.dawr0);
arch/powerpc/kvm/book3s_hv_p9_entry.c
621
if (vcpu->arch.dawrx0 != host_dawrx0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
622
mtspr(SPRN_DAWRX0, vcpu->arch.dawrx0);
arch/powerpc/kvm/book3s_hv_p9_entry.c
624
if (vcpu->arch.dawr1 != host_dawr1)
arch/powerpc/kvm/book3s_hv_p9_entry.c
625
mtspr(SPRN_DAWR1, vcpu->arch.dawr1);
arch/powerpc/kvm/book3s_hv_p9_entry.c
626
if (vcpu->arch.dawrx1 != host_dawrx1)
arch/powerpc/kvm/book3s_hv_p9_entry.c
627
mtspr(SPRN_DAWRX1, vcpu->arch.dawrx1);
arch/powerpc/kvm/book3s_hv_p9_entry.c
630
if (vcpu->arch.ciabr != host_ciabr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
631
mtspr(SPRN_CIABR, vcpu->arch.ciabr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
635
mtspr(SPRN_PSSCR, vcpu->arch.psscr | PSSCR_EC |
arch/powerpc/kvm/book3s_hv_p9_entry.c
638
if (vcpu->arch.psscr != host_psscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
639
mtspr(SPRN_PSSCR_PR, vcpu->arch.psscr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
64
vcpu->arch.vrsave = mfspr(SPRN_VRSAVE);
arch/powerpc/kvm/book3s_hv_p9_entry.c
642
mtspr(SPRN_HFSCR, vcpu->arch.hfscr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
644
mtspr(SPRN_HSRR0, vcpu->arch.regs.nip);
arch/powerpc/kvm/book3s_hv_p9_entry.c
645
mtspr(SPRN_HSRR1, (vcpu->arch.shregs.msr & ~MSR_HV) | MSR_ME);
arch/powerpc/kvm/book3s_hv_p9_entry.c
664
mtspr(SPRN_SPRG0, vcpu->arch.shregs.sprg0);
arch/powerpc/kvm/book3s_hv_p9_entry.c
665
mtspr(SPRN_SPRG1, vcpu->arch.shregs.sprg1);
arch/powerpc/kvm/book3s_hv_p9_entry.c
666
mtspr(SPRN_SPRG2, vcpu->arch.shregs.sprg2);
arch/powerpc/kvm/book3s_hv_p9_entry.c
667
mtspr(SPRN_SPRG3, vcpu->arch.shregs.sprg3);
arch/powerpc/kvm/book3s_hv_p9_entry.c
67
if (vcpu->arch.hfscr & HFSCR_EBB) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
68
vcpu->arch.ebbhr = mfspr(SPRN_EBBHR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
69
vcpu->arch.ebbrr = mfspr(SPRN_EBBRR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
70
vcpu->arch.bescr = mfspr(SPRN_BESCR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
711
mtspr(SPRN_DEC, vcpu->arch.dec_expires - *tb);
arch/powerpc/kvm/book3s_hv_p9_entry.c
716
mtspr(SPRN_DAR, vcpu->arch.shregs.dar);
arch/powerpc/kvm/book3s_hv_p9_entry.c
717
mtspr(SPRN_DSISR, vcpu->arch.shregs.dsisr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
718
mtspr(SPRN_SRR0, vcpu->arch.shregs.srr0);
arch/powerpc/kvm/book3s_hv_p9_entry.c
719
mtspr(SPRN_SRR1, vcpu->arch.shregs.srr1);
arch/powerpc/kvm/book3s_hv_p9_entry.c
722
accumulate_time(vcpu, &vcpu->arch.in_guest);
arch/powerpc/kvm/book3s_hv_p9_entry.c
726
accumulate_time(vcpu, &vcpu->arch.guest_exit);
arch/powerpc/kvm/book3s_hv_p9_entry.c
730
vcpu->arch.shregs.srr0 = mfspr(SPRN_SRR0);
arch/powerpc/kvm/book3s_hv_p9_entry.c
731
vcpu->arch.shregs.srr1 = mfspr(SPRN_SRR1);
arch/powerpc/kvm/book3s_hv_p9_entry.c
732
vcpu->arch.shregs.dar = mfspr(SPRN_DAR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
733
vcpu->arch.shregs.dsisr = mfspr(SPRN_DSISR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
74
vcpu->arch.tid = mfspr(SPRN_TIDR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
745
vcpu->arch.regs.gpr[1] = local_paca->kvm_hstate.scratch1;
arch/powerpc/kvm/book3s_hv_p9_entry.c
746
vcpu->arch.regs.gpr[3] = local_paca->kvm_hstate.scratch2;
arch/powerpc/kvm/book3s_hv_p9_entry.c
75
vcpu->arch.iamr = mfspr(SPRN_IAMR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
76
vcpu->arch.amr = mfspr(SPRN_AMR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
760
vcpu->arch.regs.gpr[9] = exsave[EX_R9/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
761
vcpu->arch.regs.gpr[10] = exsave[EX_R10/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
762
vcpu->arch.regs.gpr[11] = exsave[EX_R11/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
763
vcpu->arch.regs.gpr[12] = exsave[EX_R12/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
764
vcpu->arch.regs.gpr[13] = exsave[EX_R13/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
765
vcpu->arch.ppr = exsave[EX_PPR/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
766
vcpu->arch.cfar = exsave[EX_CFAR/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
767
vcpu->arch.regs.ctr = exsave[EX_CTR/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
769
vcpu->arch.last_inst = KVM_INST_FETCH_FAILED;
arch/powerpc/kvm/book3s_hv_p9_entry.c
77
vcpu->arch.uamor = mfspr(SPRN_UAMOR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
772
vcpu->arch.fault_dar = exsave[EX_DAR/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
773
vcpu->arch.fault_dsisr = exsave[EX_DSISR/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
78
vcpu->arch.fscr = mfspr(SPRN_FSCR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
780
vcpu->arch.emul_inst = mfspr(SPRN_HEIR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
783
vcpu->arch.fault_dar = exsave[EX_DAR/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
784
vcpu->arch.fault_dsisr = exsave[EX_DSISR/sizeof(u64)];
arch/powerpc/kvm/book3s_hv_p9_entry.c
785
vcpu->arch.fault_gpa = mfspr(SPRN_ASDR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
788
vcpu->arch.fault_gpa = mfspr(SPRN_ASDR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
79
vcpu->arch.dscr = mfspr(SPRN_DSCR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
791
vcpu->arch.hfscr = mfspr(SPRN_HFSCR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
80
vcpu->arch.pspb = mfspr(SPRN_PSPB);
arch/powerpc/kvm/book3s_hv_p9_entry.c
800
vcpu->arch.emul_inst = mfspr(SPRN_HEIR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
808
(vcpu->arch.shregs.msr & MSR_TS_S)) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
814
mtspr(SPRN_HSRR0, vcpu->arch.regs.nip);
arch/powerpc/kvm/book3s_hv_p9_entry.c
815
mtspr(SPRN_HSRR1, vcpu->arch.shregs.msr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
82
vcpu->arch.ctrl = mfspr(SPRN_CTRLF);
arch/powerpc/kvm/book3s_hv_p9_entry.c
825
local_paca->kvm_hstate.host_purr += purr - vcpu->arch.purr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
826
local_paca->kvm_hstate.host_spurr += spurr - vcpu->arch.spurr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
827
vcpu->arch.purr = purr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
828
vcpu->arch.spurr = spurr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
830
vcpu->arch.ic = mfspr(SPRN_IC);
arch/powerpc/kvm/book3s_hv_p9_entry.c
831
vcpu->arch.pid = mfspr(SPRN_PID);
arch/powerpc/kvm/book3s_hv_p9_entry.c
832
vcpu->arch.psscr = mfspr(SPRN_PSSCR_PR);
arch/powerpc/kvm/book3s_hv_p9_entry.c
834
vcpu->arch.shregs.sprg0 = mfspr(SPRN_SPRG0);
arch/powerpc/kvm/book3s_hv_p9_entry.c
835
vcpu->arch.shregs.sprg1 = mfspr(SPRN_SPRG1);
arch/powerpc/kvm/book3s_hv_p9_entry.c
836
vcpu->arch.shregs.sprg2 = mfspr(SPRN_SPRG2);
arch/powerpc/kvm/book3s_hv_p9_entry.c
837
vcpu->arch.shregs.sprg3 = mfspr(SPRN_SPRG3);
arch/powerpc/kvm/book3s_hv_p9_entry.c
841
vcpu->arch.doorbell_request = 1;
arch/powerpc/kvm/book3s_hv_p9_entry.c
849
vcpu->arch.dec_expires = dec + *tb;
arch/powerpc/kvm/book3s_hv_p9_entry.c
872
vcpu->arch.shregs.msr & MSR_TS_MASK)
arch/powerpc/kvm/book3s_hv_p9_entry.c
888
if (vcpu->arch.ciabr != host_ciabr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
892
if (vcpu->arch.dawr0 != host_dawr0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
894
if (vcpu->arch.dawrx0 != host_dawrx0)
arch/powerpc/kvm/book3s_hv_p9_entry.c
897
if (vcpu->arch.dawr1 != host_dawr1)
arch/powerpc/kvm/book3s_hv_p9_entry.c
899
if (vcpu->arch.dawrx1 != host_dawrx1)
arch/powerpc/kvm/book3s_hv_p9_entry.c
94
unsigned long guest_msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_p9_entry.c
98
} else if (vcpu->arch.hfscr & HFSCR_TM) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
99
mtspr(SPRN_TEXASR, vcpu->arch.texasr);
arch/powerpc/kvm/book3s_hv_p9_perf.c
101
mtspr(SPRN_MMCR3, vcpu->arch.mmcr[3]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
102
mtspr(SPRN_SIER2, vcpu->arch.sier[1]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
103
mtspr(SPRN_SIER3, vcpu->arch.sier[2]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
107
mtspr(SPRN_MMCRA, vcpu->arch.mmcra);
arch/powerpc/kvm/book3s_hv_p9_perf.c
108
mtspr(SPRN_MMCR0, vcpu->arch.mmcr[0]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
111
if (!vcpu->arch.nested &&
arch/powerpc/kvm/book3s_hv_p9_perf.c
112
(vcpu->arch.hfscr_permitted & HFSCR_PM))
arch/powerpc/kvm/book3s_hv_p9_perf.c
113
vcpu->arch.hfscr |= HFSCR_PM;
arch/powerpc/kvm/book3s_hv_p9_perf.c
124
lp = vcpu->arch.vpa.pinned_addr;
arch/powerpc/kvm/book3s_hv_p9_perf.c
137
vcpu->arch.mmcr[0] = mfspr(SPRN_MMCR0);
arch/powerpc/kvm/book3s_hv_p9_perf.c
138
vcpu->arch.mmcra = mfspr(SPRN_MMCRA);
arch/powerpc/kvm/book3s_hv_p9_perf.c
140
freeze_pmu(vcpu->arch.mmcr[0], vcpu->arch.mmcra);
arch/powerpc/kvm/book3s_hv_p9_perf.c
142
vcpu->arch.pmc[0] = mfspr(SPRN_PMC1);
arch/powerpc/kvm/book3s_hv_p9_perf.c
143
vcpu->arch.pmc[1] = mfspr(SPRN_PMC2);
arch/powerpc/kvm/book3s_hv_p9_perf.c
144
vcpu->arch.pmc[2] = mfspr(SPRN_PMC3);
arch/powerpc/kvm/book3s_hv_p9_perf.c
145
vcpu->arch.pmc[3] = mfspr(SPRN_PMC4);
arch/powerpc/kvm/book3s_hv_p9_perf.c
146
vcpu->arch.pmc[4] = mfspr(SPRN_PMC5);
arch/powerpc/kvm/book3s_hv_p9_perf.c
147
vcpu->arch.pmc[5] = mfspr(SPRN_PMC6);
arch/powerpc/kvm/book3s_hv_p9_perf.c
148
vcpu->arch.mmcr[1] = mfspr(SPRN_MMCR1);
arch/powerpc/kvm/book3s_hv_p9_perf.c
149
vcpu->arch.mmcr[2] = mfspr(SPRN_MMCR2);
arch/powerpc/kvm/book3s_hv_p9_perf.c
150
vcpu->arch.sdar = mfspr(SPRN_SDAR);
arch/powerpc/kvm/book3s_hv_p9_perf.c
151
vcpu->arch.siar = mfspr(SPRN_SIAR);
arch/powerpc/kvm/book3s_hv_p9_perf.c
152
vcpu->arch.sier[0] = mfspr(SPRN_SIER);
arch/powerpc/kvm/book3s_hv_p9_perf.c
155
vcpu->arch.mmcr[3] = mfspr(SPRN_MMCR3);
arch/powerpc/kvm/book3s_hv_p9_perf.c
156
vcpu->arch.sier[1] = mfspr(SPRN_SIER2);
arch/powerpc/kvm/book3s_hv_p9_perf.c
157
vcpu->arch.sier[2] = mfspr(SPRN_SIER3);
arch/powerpc/kvm/book3s_hv_p9_perf.c
160
} else if (vcpu->arch.hfscr & HFSCR_PM) {
arch/powerpc/kvm/book3s_hv_p9_perf.c
182
if (!vcpu->arch.nested)
arch/powerpc/kvm/book3s_hv_p9_perf.c
183
vcpu->arch.hfscr &= ~HFSCR_PM;
arch/powerpc/kvm/book3s_hv_p9_perf.c
41
lp = vcpu->arch.vpa.pinned_addr;
arch/powerpc/kvm/book3s_hv_p9_perf.c
87
if (load_pmu || (vcpu->arch.hfscr & HFSCR_PM)) {
arch/powerpc/kvm/book3s_hv_p9_perf.c
88
mtspr(SPRN_PMC1, vcpu->arch.pmc[0]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
89
mtspr(SPRN_PMC2, vcpu->arch.pmc[1]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
90
mtspr(SPRN_PMC3, vcpu->arch.pmc[2]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
91
mtspr(SPRN_PMC4, vcpu->arch.pmc[3]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
92
mtspr(SPRN_PMC5, vcpu->arch.pmc[4]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
93
mtspr(SPRN_PMC6, vcpu->arch.pmc[5]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
94
mtspr(SPRN_MMCR1, vcpu->arch.mmcr[1]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
95
mtspr(SPRN_MMCR2, vcpu->arch.mmcr[2]);
arch/powerpc/kvm/book3s_hv_p9_perf.c
96
mtspr(SPRN_SDAR, vcpu->arch.sdar);
arch/powerpc/kvm/book3s_hv_p9_perf.c
97
mtspr(SPRN_SIAR, vcpu->arch.siar);
arch/powerpc/kvm/book3s_hv_p9_perf.c
98
mtspr(SPRN_SIER, vcpu->arch.sier[0]);
arch/powerpc/kvm/book3s_hv_ras.c
103
tlbiel_all_lpid(vcpu->kvm->arch.radix);
arch/powerpc/kvm/book3s_hv_ras.c
117
if (vcpu->kvm->arch.fwnmi_enabled) {
arch/powerpc/kvm/book3s_hv_ras.c
137
vcpu->arch.mce_evt = mce_evt;
arch/powerpc/kvm/book3s_hv_ras.c
143
struct kvmppc_vcore *vc = vcpu->arch.vcore;
arch/powerpc/kvm/book3s_hv_ras.c
46
slb = vcpu->arch.slb_shadow.pinned_addr;
arch/powerpc/kvm/book3s_hv_ras.c
52
if ((void *) &slb->save_area[n] > vcpu->arch.slb_shadow.pinned_end)
arch/powerpc/kvm/book3s_hv_ras.c
71
unsigned long srr1 = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_ras.c
76
unsigned long dsisr = vcpu->arch.shregs.dsisr;
arch/powerpc/kvm/book3s_hv_ras.c
86
tlbiel_all_lpid(vcpu->kvm->arch.radix);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1070
entry = &vcpu->arch.mmio_cache.entry[i];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1084
unsigned int index = vcpu->arch.mmio_cache.index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1086
vcpu->arch.mmio_cache.index++;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1087
if (vcpu->arch.mmio_cache.index == MMIO_HPTE_CACHE_SIZE)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1088
vcpu->arch.mmio_cache.index = 0;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1090
return &vcpu->arch.mmio_cache.entry[index];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1126
hash = (vsid ^ ((eaddr & somask) >> pshift)) & kvmppc_hpt_mask(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1137
hpte = (__be64 *)(kvm->arch.hpt.virt + (hash << 7));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1173
hash = hash ^ kvmppc_hpt_mask(&kvm->arch.hpt);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1207
mmio_update = atomic64_read(&kvm->arch.mmio_update);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1222
hpte = (__be64 *)(kvm->arch.hpt.virt + (index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1229
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1241
key = (vcpu->arch.shregs.msr & MSR_PR) ? SLB_VSID_KP : SLB_VSID_KS;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1258
if (data && (vcpu->arch.shregs.msr & MSR_DR)) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1259
unsigned int perm = hpte_get_skey_perm(gr, vcpu->arch.amr);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1267
vcpu->arch.pgfault_addr = addr;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1268
vcpu->arch.pgfault_index = index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1269
vcpu->arch.pgfault_hpte[0] = v;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1270
vcpu->arch.pgfault_hpte[1] = r;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1271
vcpu->arch.pgfault_cache = cache_entry;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1294
if (data && (vcpu->arch.shregs.msr & MSR_IR))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
142
rmap = real_vmalloc_addr(&memslot->arch.rmap[gfn - memslot->base_gfn]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
166
next = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->forw]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
167
prev = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->back]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
243
rmap = &memslot->arch.rmap[slot_fn];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
303
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
307
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
338
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
355
rev = &kvm->arch.hpt.rev[pte_index];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
405
vcpu->arch.pgdir, true,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
406
&vcpu->arch.regs.gpr[4]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
44
if (kvm->arch.online_vcores == 1 && local_paca->kvm_hstate.kvm_vcpu)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
469
"r" (rbvalues[i]), "r" (kvm->arch.lpid));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
472
fixup_tlbie_lpid(rbvalues[i - 1], kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
496
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
498
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
514
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
537
atomic64_inc(&kvm->arch.mmio_update);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
551
&vcpu->arch.regs.gpr[4]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
558
unsigned long *args = &vcpu->arch.regs.gpr[4];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
56
cpumask_setall(&kvm->arch.need_tlb_flush);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
58
cpumask_clear_cpu(cpu, &kvm->arch.need_tlb_flush);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
585
pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
591
hp = (__be64 *) (kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
628
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
637
atomic64_inc(&kvm->arch.mmio_update);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
685
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
688
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
708
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
739
atomic64_inc(&kvm->arch.mmio_update);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
756
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
76
head = &kvm->arch.hpt.rev[i];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
762
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
764
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
79
tail = &kvm->arch.hpt.rev[head->back];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
798
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
801
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
802
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
846
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
849
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
850
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_xics.c
130
struct kvmppc_icp *this_icp = this_vcpu->arch.icp;
arch/powerpc/kvm/book3s_hv_rm_xics.c
136
set_bit(BOOK3S_IRQPRIO_EXTERNAL, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/book3s_hv_rm_xics.c
149
cpu = vcpu->arch.thread_cpu;
arch/powerpc/kvm/book3s_hv_rm_xics.c
170
clear_bit(BOOK3S_IRQPRIO_EXTERNAL, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/book3s_hv_rm_xics.c
208
this_vcpu->arch.icp->rm_dbgstate = new;
arch/powerpc/kvm/book3s_hv_rm_xics.c
209
this_vcpu->arch.icp->rm_dbgtgt = icp->vcpu;
arch/powerpc/kvm/book3s_hv_rm_xics.c
491
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_hv_rm_xics.c
492
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_hv_rm_xics.c
530
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_hv_rm_xics.c
531
struct kvmppc_icp *icp, *this_icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_hv_rm_xics.c
616
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_hv_rm_xics.c
617
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_hv_rm_xics.c
677
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_hv_rm_xics.c
678
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_hv_rm_xics.c
736
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_hv_rm_xics.c
737
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_hv_rm_xics.c
859
xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_hv_rm_xics.c
860
icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_hv_tm.c
103
bescr = vcpu->arch.bescr;
arch/powerpc/kvm/book3s_hv_tm.c
110
vcpu->arch.bescr = bescr;
arch/powerpc/kvm/book3s_hv_tm.c
112
vcpu->arch.shregs.msr = msr;
arch/powerpc/kvm/book3s_hv_tm.c
113
vcpu->arch.cfar = vcpu->arch.regs.nip;
arch/powerpc/kvm/book3s_hv_tm.c
114
vcpu->arch.regs.nip = vcpu->arch.ebbrr;
arch/powerpc/kvm/book3s_hv_tm.c
128
vcpu->arch.shregs.msr = newmsr;
arch/powerpc/kvm/book3s_hv_tm.c
129
vcpu->arch.regs.nip += 4;
arch/powerpc/kvm/book3s_hv_tm.c
135
if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) {
arch/powerpc/kvm/book3s_hv_tm.c
141
if (!(vcpu->arch.hfscr & HFSCR_TM)) {
arch/powerpc/kvm/book3s_hv_tm.c
142
vcpu->arch.hfscr &= ~HFSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
143
vcpu->arch.hfscr |= (u64)FSCR_TM_LG << 56;
arch/powerpc/kvm/book3s_hv_tm.c
144
vcpu->arch.trap = BOOK3S_INTERRUPT_H_FAC_UNAVAIL;
arch/powerpc/kvm/book3s_hv_tm.c
149
vcpu->arch.fscr &= ~FSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
150
vcpu->arch.fscr |= (u64)FSCR_TM_LG << 56;
arch/powerpc/kvm/book3s_hv_tm.c
156
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
arch/powerpc/kvm/book3s_hv_tm.c
166
vcpu->arch.shregs.msr = msr;
arch/powerpc/kvm/book3s_hv_tm.c
167
vcpu->arch.regs.nip += 4;
arch/powerpc/kvm/book3s_hv_tm.c
173
if (!(vcpu->arch.hfscr & HFSCR_TM)) {
arch/powerpc/kvm/book3s_hv_tm.c
174
vcpu->arch.hfscr &= ~HFSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
175
vcpu->arch.hfscr |= (u64)FSCR_TM_LG << 56;
arch/powerpc/kvm/book3s_hv_tm.c
176
vcpu->arch.trap = BOOK3S_INTERRUPT_H_FAC_UNAVAIL;
arch/powerpc/kvm/book3s_hv_tm.c
181
vcpu->arch.fscr &= ~FSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
182
vcpu->arch.fscr |= (u64)FSCR_TM_LG << 56;
arch/powerpc/kvm/book3s_hv_tm.c
19
u64 msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_tm.c
193
if (!(vcpu->arch.orig_texasr & TEXASR_FS)) {
arch/powerpc/kvm/book3s_hv_tm.c
203
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
arch/powerpc/kvm/book3s_hv_tm.c
205
vcpu->arch.shregs.msr &= ~MSR_TS_MASK;
arch/powerpc/kvm/book3s_hv_tm.c
206
vcpu->arch.regs.nip += 4;
arch/powerpc/kvm/book3s_hv_tm.c
21
tfiar = vcpu->arch.regs.nip & ~0x3ull;
arch/powerpc/kvm/book3s_hv_tm.c
213
if (!(vcpu->arch.hfscr & HFSCR_TM)) {
arch/powerpc/kvm/book3s_hv_tm.c
214
vcpu->arch.hfscr &= ~HFSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
215
vcpu->arch.hfscr |= (u64)FSCR_TM_LG << 56;
arch/powerpc/kvm/book3s_hv_tm.c
216
vcpu->arch.trap = BOOK3S_INTERRUPT_H_FAC_UNAVAIL;
arch/powerpc/kvm/book3s_hv_tm.c
221
vcpu->arch.fscr &= ~FSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
222
vcpu->arch.fscr |= (u64)FSCR_TM_LG << 56;
arch/powerpc/kvm/book3s_hv_tm.c
228
if (MSR_TM_ACTIVE(msr) || !(vcpu->arch.texasr & TEXASR_FS)) {
arch/powerpc/kvm/book3s_hv_tm.c
23
if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr))
arch/powerpc/kvm/book3s_hv_tm.c
236
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
arch/powerpc/kvm/book3s_hv_tm.c
238
vcpu->arch.shregs.msr = msr | MSR_TS_S;
arch/powerpc/kvm/book3s_hv_tm.c
239
vcpu->arch.regs.nip += 4;
arch/powerpc/kvm/book3s_hv_tm.c
29
vcpu->arch.tfiar = tfiar;
arch/powerpc/kvm/book3s_hv_tm.c
31
vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr;
arch/powerpc/kvm/book3s_hv_tm.c
44
u32 instr = vcpu->arch.emul_inst;
arch/powerpc/kvm/book3s_hv_tm.c
45
u64 msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_tm.c
56
vcpu->arch.regs.nip -= 4;
arch/powerpc/kvm/book3s_hv_tm.c
72
newmsr = vcpu->arch.shregs.srr1;
arch/powerpc/kvm/book3s_hv_tm.c
78
vcpu->arch.shregs.msr = newmsr;
arch/powerpc/kvm/book3s_hv_tm.c
79
vcpu->arch.cfar = vcpu->arch.regs.nip;
arch/powerpc/kvm/book3s_hv_tm.c
80
vcpu->arch.regs.nip = vcpu->arch.shregs.srr0;
arch/powerpc/kvm/book3s_hv_tm.c
84
if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) {
arch/powerpc/kvm/book3s_hv_tm.c
90
if (!(vcpu->arch.hfscr & HFSCR_EBB)) {
arch/powerpc/kvm/book3s_hv_tm.c
91
vcpu->arch.hfscr &= ~HFSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
92
vcpu->arch.hfscr |= (u64)FSCR_EBB_LG << 56;
arch/powerpc/kvm/book3s_hv_tm.c
93
vcpu->arch.trap = BOOK3S_INTERRUPT_H_FAC_UNAVAIL;
arch/powerpc/kvm/book3s_hv_tm.c
96
if ((msr & MSR_PR) && !(vcpu->arch.fscr & FSCR_EBB)) {
arch/powerpc/kvm/book3s_hv_tm.c
98
vcpu->arch.fscr &= ~FSCR_INTR_CAUSE;
arch/powerpc/kvm/book3s_hv_tm.c
99
vcpu->arch.fscr |= (u64)FSCR_EBB_LG << 56;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
101
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
arch/powerpc/kvm/book3s_hv_tm_builtin.c
115
vcpu->arch.shregs.msr &= ~MSR_TS_MASK; /* go to N state */
arch/powerpc/kvm/book3s_hv_tm_builtin.c
116
vcpu->arch.regs.nip = vcpu->arch.tfhar;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
118
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) | 0xa0000000;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
22
u32 instr = vcpu->arch.emul_inst;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
40
newmsr = vcpu->arch.shregs.srr1;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
45
vcpu->arch.shregs.msr = newmsr;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
46
vcpu->arch.cfar = vcpu->arch.regs.nip - 4;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
47
vcpu->arch.regs.nip = vcpu->arch.shregs.srr0;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
52
msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
53
if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206))
arch/powerpc/kvm/book3s_hv_tm_builtin.c
56
if (!(vcpu->arch.hfscr & HFSCR_EBB) ||
arch/powerpc/kvm/book3s_hv_tm_builtin.c
68
vcpu->arch.shregs.msr = msr;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
69
vcpu->arch.cfar = vcpu->arch.regs.nip - 4;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
70
vcpu->arch.regs.nip = mfspr(SPRN_EBBRR);
arch/powerpc/kvm/book3s_hv_tm_builtin.c
77
msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
84
vcpu->arch.shregs.msr = newmsr;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
90
msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/book3s_hv_tm_builtin.c
92
if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206))
arch/powerpc/kvm/book3s_hv_tm_builtin.c
95
if (!(vcpu->arch.hfscr & HFSCR_TM) || !(msr & MSR_TM))
arch/powerpc/kvm/book3s_hv_tm_builtin.c
99
vcpu->arch.shregs.msr = (msr & ~MSR_TS_MASK) | MSR_TS_T;
arch/powerpc/kvm/book3s_hv_uvmem.c
1055
if (!(kvm->arch.secure_guest & KVMPPC_SECURE_INIT_START))
arch/powerpc/kvm/book3s_hv_uvmem.c
1093
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
1097
ret = uv_page_in(kvm->arch.lpid, page_to_pfn(page) << PAGE_SHIFT,
arch/powerpc/kvm/book3s_hv_uvmem.c
1101
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
263
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
264
list_add(&p->list, &kvm->arch.uvmem_pfns);
arch/powerpc/kvm/book3s_hv_uvmem.c
265
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
277
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
278
list_for_each_entry_safe(p, next, &kvm->arch.uvmem_pfns, list) {
arch/powerpc/kvm/book3s_hv_uvmem.c
286
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
294
list_for_each_entry(p, &kvm->arch.uvmem_pfns, list) {
arch/powerpc/kvm/book3s_hv_uvmem.c
338
list_for_each_entry(p, &kvm->arch.uvmem_pfns, list) {
arch/powerpc/kvm/book3s_hv_uvmem.c
368
list_for_each_entry(iter, &kvm->arch.uvmem_pfns, list)
arch/powerpc/kvm/book3s_hv_uvmem.c
433
uv_unregister_mem_slot(kvm->arch.lpid, memslot->id);
arch/powerpc/kvm/book3s_hv_uvmem.c
449
ret = uv_register_mem_slot(kvm->arch.lpid,
arch/powerpc/kvm/book3s_hv_uvmem.c
472
kvm->arch.secure_guest = KVMPPC_SECURE_INIT_START;
arch/powerpc/kvm/book3s_hv_uvmem.c
482
if (!kvm->arch.svm_enabled)
arch/powerpc/kvm/book3s_hv_uvmem.c
568
ret = uv_page_out(kvm->arch.lpid, pfn << page_shift,
arch/powerpc/kvm/book3s_hv_uvmem.c
594
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
597
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
636
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
653
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
668
if (!(kvm->arch.secure_guest & KVMPPC_SECURE_INIT_START))
arch/powerpc/kvm/book3s_hv_uvmem.c
671
if (kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE)
arch/powerpc/kvm/book3s_hv_uvmem.c
681
kvm->arch.secure_guest = 0;
arch/powerpc/kvm/book3s_hv_uvmem.c
682
uv_svm_terminate(kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv_uvmem.c
780
ret = uv_page_in(kvm->arch.lpid, pfn << page_shift,
arch/powerpc/kvm/book3s_hv_uvmem.c
803
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
825
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
837
if (!(kvm->arch.secure_guest & KVMPPC_SECURE_INIT_START))
arch/powerpc/kvm/book3s_hv_uvmem.c
860
kvm->arch.secure_guest |= KVMPPC_SECURE_INIT_DONE;
arch/powerpc/kvm/book3s_hv_uvmem.c
861
pr_info("LPID %lld went secure\n", kvm->arch.lpid);
arch/powerpc/kvm/book3s_hv_uvmem.c
889
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
902
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
907
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
917
if (!uv_page_in(kvm->arch.lpid, page_to_pfn(page) << page_shift, gpa, 0,
arch/powerpc/kvm/book3s_hv_uvmem.c
923
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
945
if (!(kvm->arch.secure_guest & KVMPPC_SECURE_INIT_START))
arch/powerpc/kvm/book3s_hv_uvmem.c
965
mutex_lock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_hv_uvmem.c
982
mutex_unlock(&kvm->arch.uvmem_lock);
arch/powerpc/kvm/book3s_paired_singles.c
1101
fpd_fadds(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1105
fpd_fsubs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1109
fpd_fdivs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1113
fpd_fres(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1117
fpd_frsqrtes(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1123
fpd_fmuls(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c);
arch/powerpc/kvm/book3s_paired_singles.c
1127
fpd_fmsubs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1131
fpd_fmadds(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1135
fpd_fnmsubs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1139
fpd_fnmadds(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1154
*fpr_d = vcpu->arch.fp.fpscr;
arch/powerpc/kvm/book3s_paired_singles.c
1159
vcpu->arch.fp.fpscr = *fpr_b;
arch/powerpc/kvm/book3s_paired_singles.c
1167
fpd_fcmpu(&vcpu->arch.fp.fpscr, &tmp_cr, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1178
fpd_fcmpo(&vcpu->arch.fp.fpscr, &tmp_cr, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1184
fpd_fneg(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1190
fpd_fabs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1193
fpd_fcpsgn(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1196
fpd_fdiv(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1199
fpd_fadd(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1202
fpd_fsub(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1205
fpd_fctiw(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1208
fpd_fctiwz(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1211
fpd_frsp(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1219
fpd_fsqrt(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1221
fpd_fdiv(&vcpu->arch.fp.fpscr, &cr, fpr_d, (u64*)&one, fpr_d);
arch/powerpc/kvm/book3s_paired_singles.c
1227
fpd_fmul(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c);
arch/powerpc/kvm/book3s_paired_singles.c
1230
fpd_fsel(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1233
fpd_fmsub(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1236
fpd_fmadd(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1239
fpd_fnmsub(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1242
fpd_fnmadd(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
arch/powerpc/kvm/book3s_paired_singles.c
1249
for (i = 0; i < ARRAY_SIZE(vcpu->arch.fp.fpr); i++) {
arch/powerpc/kvm/book3s_paired_singles.c
152
kvm_cvt_df(&VCPU_FPR(vcpu, rt), &vcpu->arch.qpr[rt]);
arch/powerpc/kvm/book3s_paired_singles.c
185
vcpu->arch.paddr_accessed = addr;
arch/powerpc/kvm/book3s_paired_singles.c
202
vcpu->arch.qpr[rs] = *((u32*)tmp);
arch/powerpc/kvm/book3s_paired_singles.c
247
vcpu->arch.paddr_accessed = addr;
arch/powerpc/kvm/book3s_paired_singles.c
277
vcpu->arch.paddr_accessed = addr;
arch/powerpc/kvm/book3s_paired_singles.c
284
vcpu->arch.qpr[rs] = tmp[1];
arch/powerpc/kvm/book3s_paired_singles.c
296
vcpu->arch.qpr[rs] = tmp[1];
arch/powerpc/kvm/book3s_paired_singles.c
314
tmp[1] = vcpu->arch.qpr[rs];
arch/powerpc/kvm/book3s_paired_singles.c
317
vcpu->arch.paddr_accessed = addr;
arch/powerpc/kvm/book3s_paired_singles.c
346
if (!(vcpu->arch.hflags & BOOK3S_HFLAG_PAIRED_SINGLE))
arch/powerpc/kvm/book3s_paired_singles.c
497
u32 *qpr = vcpu->arch.qpr;
arch/powerpc/kvm/book3s_paired_singles.c
513
func(&vcpu->arch.fp.fpscr, &ps0_out, &ps0_in1, &ps0_in2, &ps0_in3);
arch/powerpc/kvm/book3s_paired_singles.c
530
func(&vcpu->arch.fp.fpscr, &qpr[reg_out], &ps1_in1, &ps1_in2, &ps1_in3);
arch/powerpc/kvm/book3s_paired_singles.c
545
u32 *qpr = vcpu->arch.qpr;
arch/powerpc/kvm/book3s_paired_singles.c
562
func(&vcpu->arch.fp.fpscr, &ps0_out, &ps0_in1, &ps0_in2);
arch/powerpc/kvm/book3s_paired_singles.c
578
func(&vcpu->arch.fp.fpscr, &ps1_out, &ps1_in1, &ps1_in2);
arch/powerpc/kvm/book3s_paired_singles.c
595
u32 *qpr = vcpu->arch.qpr;
arch/powerpc/kvm/book3s_paired_singles.c
604
func(&vcpu->arch.fp.fpscr, &ps0_out, &ps0_in);
arch/powerpc/kvm/book3s_paired_singles.c
613
func(&vcpu->arch.fp.fpscr, &qpr[reg_out], &ps1_in);
arch/powerpc/kvm/book3s_paired_singles.c
669
for (i = 0; i < ARRAY_SIZE(vcpu->arch.fp.fpr); i++) {
arch/powerpc/kvm/book3s_paired_singles.c
673
i, f, VCPU_FPR(vcpu, i), i, vcpu->arch.qpr[i]);
arch/powerpc/kvm/book3s_paired_singles.c
761
vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb];
arch/powerpc/kvm/book3s_paired_singles.c
762
vcpu->arch.qpr[ax_rd] ^= 0x80000000;
arch/powerpc/kvm/book3s_paired_singles.c
771
vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb];
arch/powerpc/kvm/book3s_paired_singles.c
781
vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb];
arch/powerpc/kvm/book3s_paired_singles.c
782
vcpu->arch.qpr[ax_rd] |= 0x80000000;
arch/powerpc/kvm/book3s_paired_singles.c
788
vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb];
arch/powerpc/kvm/book3s_paired_singles.c
789
vcpu->arch.qpr[ax_rd] &= ~0x80000000;
arch/powerpc/kvm/book3s_paired_singles.c
796
&vcpu->arch.qpr[ax_rd]);
arch/powerpc/kvm/book3s_paired_singles.c
801
vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb];
arch/powerpc/kvm/book3s_paired_singles.c
806
kvm_cvt_fd(&vcpu->arch.qpr[ax_ra],
arch/powerpc/kvm/book3s_paired_singles.c
810
&vcpu->arch.qpr[ax_rd]);
arch/powerpc/kvm/book3s_paired_singles.c
815
kvm_cvt_fd(&vcpu->arch.qpr[ax_ra],
arch/powerpc/kvm/book3s_paired_singles.c
817
vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rb];
arch/powerpc/kvm/book3s_paired_singles.c
856
vcpu->arch.qpr[ax_rd] = vcpu->arch.qpr[ax_rc];
arch/powerpc/kvm/book3s_pr.c
1013
mtspr(SPRN_TAR, vcpu->arch.tar);
arch/powerpc/kvm/book3s_pr.c
1014
vcpu->arch.shadow_fscr |= FSCR_TAR;
arch/powerpc/kvm/book3s_pr.c
103
vcpu->arch.hflags &= ~BOOK3S_HFLAG_SPLIT_HACK;
arch/powerpc/kvm/book3s_pr.c
1042
if ((vcpu->arch.fscr & FSCR_TAR) && !(fscr & FSCR_TAR)) {
arch/powerpc/kvm/book3s_pr.c
1045
} else if (!(vcpu->arch.fscr & FSCR_TAR) && (fscr & FSCR_TAR)) {
arch/powerpc/kvm/book3s_pr.c
1046
vcpu->arch.fscr = fscr;
arch/powerpc/kvm/book3s_pr.c
1051
vcpu->arch.fscr = fscr;
arch/powerpc/kvm/book3s_pr.c
1087
flags = vcpu->arch.shadow_srr1 & 0x1f0000ull;
arch/powerpc/kvm/book3s_pr.c
115
new_msr = vcpu->arch.intr_msr;
arch/powerpc/kvm/book3s_pr.c
1154
ulong shadow_srr1 = vcpu->arch.shadow_srr1;
arch/powerpc/kvm/book3s_pr.c
1184
} else if (vcpu->arch.mmu.is_dcbz32(vcpu) &&
arch/powerpc/kvm/book3s_pr.c
1185
(!(vcpu->arch.hflags & BOOK3S_HFLAG_DCBZ32))) {
arch/powerpc/kvm/book3s_pr.c
1203
u32 fault_dsisr = vcpu->arch.fault_dsisr;
arch/powerpc/kvm/book3s_pr.c
1283
if (vcpu->arch.papr_enabled) {
arch/powerpc/kvm/book3s_pr.c
1293
if (vcpu->arch.papr_enabled &&
arch/powerpc/kvm/book3s_pr.c
1313
vcpu->arch.hcall_needed = 1;
arch/powerpc/kvm/book3s_pr.c
1315
} else if (vcpu->arch.osi_enabled &&
arch/powerpc/kvm/book3s_pr.c
1325
vcpu->arch.osi_needed = 1;
arch/powerpc/kvm/book3s_pr.c
1348
if (vcpu->arch.hflags & BOOK3S_HFLAG_PAIRED_SINGLE) {
arch/powerpc/kvm/book3s_pr.c
1400
r = kvmppc_handle_fac(vcpu, vcpu->arch.shadow_fscr >> 56);
arch/powerpc/kvm/book3s_pr.c
1418
ulong shadow_srr1 = vcpu->arch.shadow_srr1;
arch/powerpc/kvm/book3s_pr.c
1461
sregs->pvr = vcpu->arch.pvr;
arch/powerpc/kvm/book3s_pr.c
1464
if (vcpu->arch.hflags & BOOK3S_HFLAG_SLB) {
arch/powerpc/kvm/book3s_pr.c
1466
sregs->u.s.ppc64.slb[i].slbe = vcpu->arch.slb[i].orige | i;
arch/powerpc/kvm/book3s_pr.c
1467
sregs->u.s.ppc64.slb[i].slbv = vcpu->arch.slb[i].origv;
arch/powerpc/kvm/book3s_pr.c
1492
if (vcpu->arch.hflags & BOOK3S_HFLAG_SLB) {
arch/powerpc/kvm/book3s_pr.c
1494
vcpu->arch.mmu.slbmte(vcpu, 0, 0);
arch/powerpc/kvm/book3s_pr.c
1495
vcpu->arch.mmu.slbia(vcpu);
arch/powerpc/kvm/book3s_pr.c
1502
vcpu->arch.mmu.slbmte(vcpu, rs, rb);
arch/powerpc/kvm/book3s_pr.c
1508
vcpu->arch.mmu.mtsrin(vcpu, i, sregs->u.s.ppc32.sr[i]);
arch/powerpc/kvm/book3s_pr.c
152
current->thread.kvm_shadow_vcpu = vcpu->arch.shadow_vcpu;
arch/powerpc/kvm/book3s_pr.c
1548
if (vcpu->arch.intr_msr & MSR_LE)
arch/powerpc/kvm/book3s_pr.c
1555
*val = get_reg_val(id, vcpu->arch.tfhar);
arch/powerpc/kvm/book3s_pr.c
1558
*val = get_reg_val(id, vcpu->arch.tfiar);
arch/powerpc/kvm/book3s_pr.c
1561
*val = get_reg_val(id, vcpu->arch.texasr);
arch/powerpc/kvm/book3s_pr.c
1565
vcpu->arch.gpr_tm[id-KVM_REG_PPC_TM_GPR0]);
arch/powerpc/kvm/book3s_pr.c
1574
val->vsxval[j] = vcpu->arch.fp_tm.fpr[i][j];
arch/powerpc/kvm/book3s_pr.c
1577
val->vval = vcpu->arch.vr_tm.vr[i-32];
arch/powerpc/kvm/book3s_pr.c
1584
*val = get_reg_val(id, vcpu->arch.cr_tm);
arch/powerpc/kvm/book3s_pr.c
1587
*val = get_reg_val(id, vcpu->arch.xer_tm);
arch/powerpc/kvm/book3s_pr.c
1590
*val = get_reg_val(id, vcpu->arch.lr_tm);
arch/powerpc/kvm/book3s_pr.c
1593
*val = get_reg_val(id, vcpu->arch.ctr_tm);
arch/powerpc/kvm/book3s_pr.c
1596
*val = get_reg_val(id, vcpu->arch.fp_tm.fpscr);
arch/powerpc/kvm/book3s_pr.c
1599
*val = get_reg_val(id, vcpu->arch.amr_tm);
arch/powerpc/kvm/book3s_pr.c
1602
*val = get_reg_val(id, vcpu->arch.ppr_tm);
arch/powerpc/kvm/book3s_pr.c
1605
*val = get_reg_val(id, vcpu->arch.vrsave_tm);
arch/powerpc/kvm/book3s_pr.c
1609
*val = get_reg_val(id, vcpu->arch.vr_tm.vscr.u[3]);
arch/powerpc/kvm/book3s_pr.c
1614
*val = get_reg_val(id, vcpu->arch.dscr_tm);
arch/powerpc/kvm/book3s_pr.c
1617
*val = get_reg_val(id, vcpu->arch.tar_tm);
arch/powerpc/kvm/book3s_pr.c
1631
vcpu->arch.intr_msr |= MSR_LE;
arch/powerpc/kvm/book3s_pr.c
1633
vcpu->arch.intr_msr &= ~MSR_LE;
arch/powerpc/kvm/book3s_pr.c
1655
vcpu->arch.tfhar = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1658
vcpu->arch.tfiar = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1661
vcpu->arch.texasr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1664
vcpu->arch.gpr_tm[id - KVM_REG_PPC_TM_GPR0] =
arch/powerpc/kvm/book3s_pr.c
1674
vcpu->arch.fp_tm.fpr[i][j] = val->vsxval[j];
arch/powerpc/kvm/book3s_pr.c
1677
vcpu->arch.vr_tm.vr[i-32] = val->vval;
arch/powerpc/kvm/book3s_pr.c
1683
vcpu->arch.cr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1686
vcpu->arch.xer_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1689
vcpu->arch.lr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1692
vcpu->arch.ctr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1695
vcpu->arch.fp_tm.fpscr = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1698
vcpu->arch.amr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1701
vcpu->arch.ppr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1704
vcpu->arch.vrsave_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1708
vcpu->arch.vr.vscr.u[3] = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1713
vcpu->arch.dscr_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1716
vcpu->arch.tar_tm = set_reg_val(id, *val);
arch/powerpc/kvm/book3s_pr.c
1738
vcpu->arch.book3s = vcpu_book3s;
arch/powerpc/kvm/book3s_pr.c
1741
vcpu->arch.shadow_vcpu = kzalloc_obj(*vcpu->arch.shadow_vcpu);
arch/powerpc/kvm/book3s_pr.c
1742
if (!vcpu->arch.shadow_vcpu)
arch/powerpc/kvm/book3s_pr.c
1749
vcpu->arch.shared = (void *)p;
arch/powerpc/kvm/book3s_pr.c
1753
vcpu->arch.shared_big_endian = true;
arch/powerpc/kvm/book3s_pr.c
1755
vcpu->arch.shared_big_endian = false;
arch/powerpc/kvm/book3s_pr.c
1763
vcpu->arch.pvr = 0x3C0301;
arch/powerpc/kvm/book3s_pr.c
1765
vcpu->arch.pvr = mfspr(SPRN_PVR);
arch/powerpc/kvm/book3s_pr.c
1766
vcpu->arch.intr_msr = MSR_SF;
arch/powerpc/kvm/book3s_pr.c
1769
vcpu->arch.pvr = 0x84202;
arch/powerpc/kvm/book3s_pr.c
1770
vcpu->arch.intr_msr = 0;
arch/powerpc/kvm/book3s_pr.c
1772
kvmppc_set_pvr_pr(vcpu, vcpu->arch.pvr);
arch/powerpc/kvm/book3s_pr.c
1773
vcpu->arch.slb_nr = 64;
arch/powerpc/kvm/book3s_pr.c
1775
vcpu->arch.shadow_msr = MSR_USER64 & ~MSR_LE;
arch/powerpc/kvm/book3s_pr.c
1784
free_page((unsigned long)vcpu->arch.shared);
arch/powerpc/kvm/book3s_pr.c
1787
kfree(vcpu->arch.shadow_vcpu);
arch/powerpc/kvm/book3s_pr.c
1800
free_page((unsigned long)vcpu->arch.shared & PAGE_MASK);
arch/powerpc/kvm/book3s_pr.c
1802
kfree(vcpu->arch.shadow_vcpu);
arch/powerpc/kvm/book3s_pr.c
1812
if (!vcpu->arch.sane) {
arch/powerpc/kvm/book3s_pr.c
1952
if (vcpu && (vcpu->arch.hflags & BOOK3S_HFLAG_MULTI_PGSIZE)) {
arch/powerpc/kvm/book3s_pr.c
196
svcpu->gpr[0] = vcpu->arch.regs.gpr[0];
arch/powerpc/kvm/book3s_pr.c
197
svcpu->gpr[1] = vcpu->arch.regs.gpr[1];
arch/powerpc/kvm/book3s_pr.c
198
svcpu->gpr[2] = vcpu->arch.regs.gpr[2];
arch/powerpc/kvm/book3s_pr.c
199
svcpu->gpr[3] = vcpu->arch.regs.gpr[3];
arch/powerpc/kvm/book3s_pr.c
1995
mutex_init(&kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr.c
200
svcpu->gpr[4] = vcpu->arch.regs.gpr[4];
arch/powerpc/kvm/book3s_pr.c
201
svcpu->gpr[5] = vcpu->arch.regs.gpr[5];
arch/powerpc/kvm/book3s_pr.c
2014
WARN_ON(!list_empty(&kvm->arch.spapr_tce_tables));
arch/powerpc/kvm/book3s_pr.c
202
svcpu->gpr[6] = vcpu->arch.regs.gpr[6];
arch/powerpc/kvm/book3s_pr.c
203
svcpu->gpr[7] = vcpu->arch.regs.gpr[7];
arch/powerpc/kvm/book3s_pr.c
204
svcpu->gpr[8] = vcpu->arch.regs.gpr[8];
arch/powerpc/kvm/book3s_pr.c
205
svcpu->gpr[9] = vcpu->arch.regs.gpr[9];
arch/powerpc/kvm/book3s_pr.c
206
svcpu->gpr[10] = vcpu->arch.regs.gpr[10];
arch/powerpc/kvm/book3s_pr.c
207
svcpu->gpr[11] = vcpu->arch.regs.gpr[11];
arch/powerpc/kvm/book3s_pr.c
208
svcpu->gpr[12] = vcpu->arch.regs.gpr[12];
arch/powerpc/kvm/book3s_pr.c
209
svcpu->gpr[13] = vcpu->arch.regs.gpr[13];
arch/powerpc/kvm/book3s_pr.c
210
svcpu->cr = vcpu->arch.regs.ccr;
arch/powerpc/kvm/book3s_pr.c
211
svcpu->xer = vcpu->arch.regs.xer;
arch/powerpc/kvm/book3s_pr.c
212
svcpu->ctr = vcpu->arch.regs.ctr;
arch/powerpc/kvm/book3s_pr.c
213
svcpu->lr = vcpu->arch.regs.link;
arch/powerpc/kvm/book3s_pr.c
214
svcpu->pc = vcpu->arch.regs.nip;
arch/powerpc/kvm/book3s_pr.c
216
svcpu->shadow_fscr = vcpu->arch.shadow_fscr;
arch/powerpc/kvm/book3s_pr.c
222
vcpu->arch.entry_tb = get_tb();
arch/powerpc/kvm/book3s_pr.c
223
vcpu->arch.entry_vtb = get_vtb();
arch/powerpc/kvm/book3s_pr.c
225
vcpu->arch.entry_ic = mfspr(SPRN_IC);
arch/powerpc/kvm/book3s_pr.c
246
smsr |= (guest_msr & vcpu->arch.guest_owned_ext);
arch/powerpc/kvm/book3s_pr.c
260
vcpu->arch.shadow_msr = smsr;
arch/powerpc/kvm/book3s_pr.c
278
vcpu->arch.regs.gpr[0] = svcpu->gpr[0];
arch/powerpc/kvm/book3s_pr.c
279
vcpu->arch.regs.gpr[1] = svcpu->gpr[1];
arch/powerpc/kvm/book3s_pr.c
280
vcpu->arch.regs.gpr[2] = svcpu->gpr[2];
arch/powerpc/kvm/book3s_pr.c
281
vcpu->arch.regs.gpr[3] = svcpu->gpr[3];
arch/powerpc/kvm/book3s_pr.c
282
vcpu->arch.regs.gpr[4] = svcpu->gpr[4];
arch/powerpc/kvm/book3s_pr.c
283
vcpu->arch.regs.gpr[5] = svcpu->gpr[5];
arch/powerpc/kvm/book3s_pr.c
284
vcpu->arch.regs.gpr[6] = svcpu->gpr[6];
arch/powerpc/kvm/book3s_pr.c
285
vcpu->arch.regs.gpr[7] = svcpu->gpr[7];
arch/powerpc/kvm/book3s_pr.c
286
vcpu->arch.regs.gpr[8] = svcpu->gpr[8];
arch/powerpc/kvm/book3s_pr.c
287
vcpu->arch.regs.gpr[9] = svcpu->gpr[9];
arch/powerpc/kvm/book3s_pr.c
288
vcpu->arch.regs.gpr[10] = svcpu->gpr[10];
arch/powerpc/kvm/book3s_pr.c
289
vcpu->arch.regs.gpr[11] = svcpu->gpr[11];
arch/powerpc/kvm/book3s_pr.c
290
vcpu->arch.regs.gpr[12] = svcpu->gpr[12];
arch/powerpc/kvm/book3s_pr.c
291
vcpu->arch.regs.gpr[13] = svcpu->gpr[13];
arch/powerpc/kvm/book3s_pr.c
292
vcpu->arch.regs.ccr = svcpu->cr;
arch/powerpc/kvm/book3s_pr.c
293
vcpu->arch.regs.xer = svcpu->xer;
arch/powerpc/kvm/book3s_pr.c
294
vcpu->arch.regs.ctr = svcpu->ctr;
arch/powerpc/kvm/book3s_pr.c
295
vcpu->arch.regs.link = svcpu->lr;
arch/powerpc/kvm/book3s_pr.c
296
vcpu->arch.regs.nip = svcpu->pc;
arch/powerpc/kvm/book3s_pr.c
297
vcpu->arch.shadow_srr1 = svcpu->shadow_srr1;
arch/powerpc/kvm/book3s_pr.c
298
vcpu->arch.fault_dar = svcpu->fault_dar;
arch/powerpc/kvm/book3s_pr.c
299
vcpu->arch.fault_dsisr = svcpu->fault_dsisr;
arch/powerpc/kvm/book3s_pr.c
300
vcpu->arch.last_inst = svcpu->last_inst;
arch/powerpc/kvm/book3s_pr.c
302
vcpu->arch.shadow_fscr = svcpu->shadow_fscr;
arch/powerpc/kvm/book3s_pr.c
307
vcpu->arch.purr += get_tb() - vcpu->arch.entry_tb;
arch/powerpc/kvm/book3s_pr.c
308
vcpu->arch.spurr += get_tb() - vcpu->arch.entry_tb;
arch/powerpc/kvm/book3s_pr.c
309
to_book3s(vcpu)->vtb += get_vtb() - vcpu->arch.entry_vtb;
arch/powerpc/kvm/book3s_pr.c
311
vcpu->arch.ic += mfspr(SPRN_IC) - vcpu->arch.entry_ic;
arch/powerpc/kvm/book3s_pr.c
327
(vcpu->arch.shadow_srr1 & (MSR_TS_MASK)) !=
arch/powerpc/kvm/book3s_pr.c
330
old_msr |= (vcpu->arch.shadow_srr1 & (MSR_TS_MASK));
arch/powerpc/kvm/book3s_pr.c
346
vcpu->arch.tfhar = mfspr(SPRN_TFHAR);
arch/powerpc/kvm/book3s_pr.c
347
vcpu->arch.texasr = mfspr(SPRN_TEXASR);
arch/powerpc/kvm/book3s_pr.c
348
vcpu->arch.tfiar = mfspr(SPRN_TFIAR);
arch/powerpc/kvm/book3s_pr.c
355
mtspr(SPRN_TFHAR, vcpu->arch.tfhar);
arch/powerpc/kvm/book3s_pr.c
356
mtspr(SPRN_TEXASR, vcpu->arch.texasr);
arch/powerpc/kvm/book3s_pr.c
357
mtspr(SPRN_TFIAR, vcpu->arch.tfiar);
arch/powerpc/kvm/book3s_pr.c
367
ulong ext_diff = (kvmppc_get_msr(vcpu) & ~vcpu->arch.guest_owned_ext) &
arch/powerpc/kvm/book3s_pr.c
404
if (vcpu->arch.fscr & FSCR_TAR)
arch/powerpc/kvm/book3s_pr.c
416
if (vcpu->arch.fscr & FSCR_TAR)
arch/powerpc/kvm/book3s_pr.c
471
if (vcpu->arch.papr_enabled)
arch/powerpc/kvm/book3s_pr.c
494
if (!vcpu->arch.pending_exceptions) {
arch/powerpc/kvm/book3s_pr.c
515
if (!(msr & MSR_PR) && vcpu->arch.magic_page_pa) {
arch/powerpc/kvm/book3s_pr.c
516
struct kvm_vcpu_arch *a = &vcpu->arch;
arch/powerpc/kvm/book3s_pr.c
533
if (vcpu->arch.magic_page_pa &&
arch/powerpc/kvm/book3s_pr.c
536
kvmppc_mmu_pte_flush(vcpu, (uint32_t)vcpu->arch.magic_page_pa,
arch/powerpc/kvm/book3s_pr.c
554
vcpu->arch.hflags &= ~BOOK3S_HFLAG_SLB;
arch/powerpc/kvm/book3s_pr.c
555
vcpu->arch.pvr = pvr;
arch/powerpc/kvm/book3s_pr.c
562
vcpu->arch.cpu_type = KVM_CPU_3S_64;
arch/powerpc/kvm/book3s_pr.c
570
vcpu->arch.cpu_type = KVM_CPU_3S_32;
arch/powerpc/kvm/book3s_pr.c
577
vcpu->arch.hflags &= ~BOOK3S_HFLAG_DCBZ32;
arch/powerpc/kvm/book3s_pr.c
578
if (vcpu->arch.mmu.is_dcbz32(vcpu) && (mfmsr() & MSR_HV) &&
arch/powerpc/kvm/book3s_pr.c
580
vcpu->arch.hflags |= BOOK3S_HFLAG_DCBZ32;
arch/powerpc/kvm/book3s_pr.c
603
vcpu->arch.hflags |= BOOK3S_HFLAG_MULTI_PGSIZE |
arch/powerpc/kvm/book3s_pr.c
610
vcpu->arch.hflags |= BOOK3S_HFLAG_DCBZ32;
arch/powerpc/kvm/book3s_pr.c
625
vcpu->arch.hflags |= BOOK3S_HFLAG_NATIVE_PS;
arch/powerpc/kvm/book3s_pr.c
667
ulong mp_pa = vcpu->arch.magic_page_pa;
arch/powerpc/kvm/book3s_pr.c
694
if (data && (vcpu->arch.fault_dsisr & DSISR_ISSTORE))
arch/powerpc/kvm/book3s_pr.c
699
page_found = vcpu->arch.mmu.xlate(vcpu, eaddr, &pte, data, iswrite);
arch/powerpc/kvm/book3s_pr.c
717
(vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) &&
arch/powerpc/kvm/book3s_pr.c
722
vcpu->arch.mmu.esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid);
arch/powerpc/kvm/book3s_pr.c
735
if (vcpu->arch.mmu.is_dcbz32(vcpu) &&
arch/powerpc/kvm/book3s_pr.c
736
(!(vcpu->arch.hflags & BOOK3S_HFLAG_DCBZ32))) {
arch/powerpc/kvm/book3s_pr.c
754
flags |= vcpu->arch.fault_dsisr & DSISR_ISSTORE;
arch/powerpc/kvm/book3s_pr.c
764
if (data && !(vcpu->arch.fault_dsisr & DSISR_NOHPTE)) {
arch/powerpc/kvm/book3s_pr.c
780
else if (vcpu->arch.mmu.is_dcbz32(vcpu) &&
arch/powerpc/kvm/book3s_pr.c
781
(!(vcpu->arch.hflags & BOOK3S_HFLAG_DCBZ32)))
arch/powerpc/kvm/book3s_pr.c
786
vcpu->arch.paddr_accessed = pte.raddr;
arch/powerpc/kvm/book3s_pr.c
787
vcpu->arch.vaddr_accessed = pte.eaddr;
arch/powerpc/kvm/book3s_pr.c
808
msr &= vcpu->arch.guest_owned_ext;
arch/powerpc/kvm/book3s_pr.c
83
if (vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK)
arch/powerpc/kvm/book3s_pr.c
835
vcpu->arch.guest_owned_ext &= ~(msr | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
843
if (!(vcpu->arch.shadow_fscr & (1ULL << fac))) {
arch/powerpc/kvm/book3s_pr.c
850
vcpu->arch.tar = mfspr(SPRN_TAR);
arch/powerpc/kvm/book3s_pr.c
852
vcpu->arch.shadow_fscr &= ~FSCR_TAR;
arch/powerpc/kvm/book3s_pr.c
865
if (vcpu->arch.hflags & BOOK3S_HFLAG_PAIRED_SINGLE)
arch/powerpc/kvm/book3s_pr.c
891
msr &= ~vcpu->arch.guest_owned_ext;
arch/powerpc/kvm/book3s_pr.c
90
vcpu->arch.hflags |= BOOK3S_HFLAG_SPLIT_HACK;
arch/powerpc/kvm/book3s_pr.c
902
load_fp_state(&vcpu->arch.fp);
arch/powerpc/kvm/book3s_pr.c
904
t->fp_save_area = &vcpu->arch.fp;
arch/powerpc/kvm/book3s_pr.c
912
load_vr_state(&vcpu->arch.vr);
arch/powerpc/kvm/book3s_pr.c
914
t->vr_save_area = &vcpu->arch.vr;
arch/powerpc/kvm/book3s_pr.c
920
vcpu->arch.guest_owned_ext |= msr;
arch/powerpc/kvm/book3s_pr.c
934
lost_ext = vcpu->arch.guest_owned_ext & ~current->thread.regs->msr;
arch/powerpc/kvm/book3s_pr.c
941
load_fp_state(&vcpu->arch.fp);
arch/powerpc/kvm/book3s_pr.c
949
load_vr_state(&vcpu->arch.vr);
arch/powerpc/kvm/book3s_pr.c
96
if (vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) {
arch/powerpc/kvm/book3s_pr.c
962
vcpu->arch.fscr &= ~(0xffULL << 56);
arch/powerpc/kvm/book3s_pr.c
963
vcpu->arch.fscr |= (fac << 56);
arch/powerpc/kvm/book3s_pr.c
993
guest_fac_enabled = (vcpu->arch.fscr & (1ULL << fac));
arch/powerpc/kvm/book3s_pr_papr.c
114
vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false);
arch/powerpc/kvm/book3s_pr_papr.c
121
mutex_unlock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr_papr.c
151
mutex_lock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr_papr.c
202
vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false);
arch/powerpc/kvm/book3s_pr_papr.c
208
mutex_unlock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr_papr.c
224
mutex_lock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr_papr.c
247
vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false);
arch/powerpc/kvm/book3s_pr_papr.c
256
mutex_unlock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr_papr.c
375
!test_bit(cmd/4, vcpu->kvm->arch.enabled_hcalls))
arch/powerpc/kvm/book3s_pr_papr.c
414
if (list_empty(&vcpu->kvm->arch.rtas_tokens))
arch/powerpc/kvm/book3s_pr_papr.c
49
mutex_lock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr_papr.c
494
__set_bit(hcall / 4, kvm->arch.enabled_hcalls);
arch/powerpc/kvm/book3s_pr_papr.c
80
mutex_unlock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_pr_papr.c
96
mutex_lock(&vcpu->kvm->arch.hpt_mutex);
arch/powerpc/kvm/book3s_rtas.c
146
lockdep_assert_held(&kvm->arch.rtas_token_lock);
arch/powerpc/kvm/book3s_rtas.c
148
list_for_each_entry_safe(d, tmp, &kvm->arch.rtas_tokens, list) {
arch/powerpc/kvm/book3s_rtas.c
167
lockdep_assert_held(&kvm->arch.rtas_token_lock);
arch/powerpc/kvm/book3s_rtas.c
169
list_for_each_entry(d, &kvm->arch.rtas_tokens, list) {
arch/powerpc/kvm/book3s_rtas.c
193
list_add_tail(&d->list, &kvm->arch.rtas_tokens);
arch/powerpc/kvm/book3s_rtas.c
206
mutex_lock(&kvm->arch.rtas_token_lock);
arch/powerpc/kvm/book3s_rtas.c
213
mutex_unlock(&kvm->arch.rtas_token_lock);
arch/powerpc/kvm/book3s_rtas.c
258
mutex_lock(&vcpu->kvm->arch.rtas_token_lock);
arch/powerpc/kvm/book3s_rtas.c
261
list_for_each_entry(d, &vcpu->kvm->arch.rtas_tokens, list) {
arch/powerpc/kvm/book3s_rtas.c
269
mutex_unlock(&vcpu->kvm->arch.rtas_token_lock);
arch/powerpc/kvm/book3s_rtas.c
303
list_for_each_entry_safe(d, tmp, &kvm->arch.rtas_tokens, list) {
arch/powerpc/kvm/book3s_xics.c
1066
if (!vcpu->kvm->arch.xics)
arch/powerpc/kvm/book3s_xics.c
1080
vcpu->arch.icp = icp;
arch/powerpc/kvm/book3s_xics.c
1089
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
1103
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
1104
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
1287
struct kvmppc_xics *xics = kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
1370
kvm->arch.xics = NULL;
arch/powerpc/kvm/book3s_xics.c
1387
struct kvmppc_xics **kvm_xics_device = &kvm->arch.xics_device;
arch/powerpc/kvm/book3s_xics.c
1408
if (kvm->arch.xics)
arch/powerpc/kvm/book3s_xics.c
1418
kvm->arch.xics = xics;
arch/powerpc/kvm/book3s_xics.c
1459
if (vcpu->arch.irq_type != KVMPPC_IRQ_DEFAULT)
arch/powerpc/kvm/book3s_xics.c
1464
vcpu->arch.irq_type = KVMPPC_IRQ_XICS;
arch/powerpc/kvm/book3s_xics.c
1471
if (!vcpu->arch.icp)
arch/powerpc/kvm/book3s_xics.c
1473
kfree(vcpu->arch.icp);
arch/powerpc/kvm/book3s_xics.c
1474
vcpu->arch.icp = NULL;
arch/powerpc/kvm/book3s_xics.c
1475
vcpu->arch.irq_type = KVMPPC_IRQ_DEFAULT;
arch/powerpc/kvm/book3s_xics.c
1481
struct kvmppc_xics *xics = kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
1497
struct kvmppc_xics *xics = kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
168
struct kvmppc_xics *xics = kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
198
struct kvmppc_xics *xics = kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
224
struct kvmppc_xics *xics = kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
251
struct kvmppc_xics *xics = kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
589
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
623
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
632
icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
709
icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
724
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
725
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
779
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
780
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
819
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
820
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
850
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
851
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.c
877
struct kvmppc_xics *xics = vcpu->kvm->arch.xics;
arch/powerpc/kvm/book3s_xics.c
882
if (!xics || !vcpu->arch.icp)
arch/powerpc/kvm/book3s_xics.c
960
xics_debugfs_irqmap(m, kvm->arch.pimap);
arch/powerpc/kvm/book3s_xics.c
965
struct kvmppc_icp *icp = vcpu->arch.icp;
arch/powerpc/kvm/book3s_xics.h
122
if (vcpu->arch.icp && nr == vcpu->arch.icp->server_num)
arch/powerpc/kvm/book3s_xics.h
123
return vcpu->arch.icp;
arch/powerpc/kvm/book3s_xive.c
1016
xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
1026
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
1068
if (!vcpu->arch.xive_vcpu)
arch/powerpc/kvm/book3s_xive.c
1072
*server = vcpu->arch.xive_vcpu->server_num;
arch/powerpc/kvm/book3s_xive.c
1186
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1270
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1371
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1393
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1432
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1480
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
1493
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
1494
struct kvmppc_xive *xive = vcpu->kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1515
if (WARN_ON(vcpu->arch.xive_pushed))
arch/powerpc/kvm/book3s_xive.c
1519
vcpu->arch.xive_saved_state.cppr = cppr;
arch/powerpc/kvm/book3s_xive.c
1553
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1651
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1730
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
1732
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1764
if (vcpu->arch.xive_esc_on) {
arch/powerpc/kvm/book3s_xive.c
1765
__raw_readq((void __iomem *)(vcpu->arch.xive_esc_vaddr +
arch/powerpc/kvm/book3s_xive.c
1767
vcpu->arch.xive_esc_on = false;
arch/powerpc/kvm/book3s_xive.c
1775
vcpu->arch.xive_esc_vaddr = 0;
arch/powerpc/kvm/book3s_xive.c
1776
vcpu->arch.xive_esc_raddr = 0;
arch/powerpc/kvm/book3s_xive.c
1798
if (!vcpu->arch.xive_esc_on)
arch/powerpc/kvm/book3s_xive.c
1804
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
1805
struct kvmppc_xive *xive = vcpu->kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
1838
vcpu->arch.xive_cam_word = 0;
arch/powerpc/kvm/book3s_xive.c
1861
vcpu->arch.irq_type = KVMPPC_IRQ_DEFAULT;
arch/powerpc/kvm/book3s_xive.c
1862
vcpu->arch.xive_vcpu = NULL;
arch/powerpc/kvm/book3s_xive.c
1917
if (vcpu->arch.irq_type != KVMPPC_IRQ_DEFAULT)
arch/powerpc/kvm/book3s_xive.c
1933
vcpu->arch.xive_vcpu = xc;
arch/powerpc/kvm/book3s_xive.c
1952
vcpu->arch.xive_saved_state.w01 = cpu_to_be64(0xff000000);
arch/powerpc/kvm/book3s_xive.c
1953
vcpu->arch.xive_cam_word = cpu_to_be32(xc->vp_cam | TM_QW1W2_VO);
arch/powerpc/kvm/book3s_xive.c
2028
vcpu->arch.irq_type = KVMPPC_IRQ_XICS;
arch/powerpc/kvm/book3s_xive.c
2142
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
2310
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
2490
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
2682
kvm->arch.xive = NULL;
arch/powerpc/kvm/book3s_xive.c
2717
&kvm->arch.xive_devices.native :
arch/powerpc/kvm/book3s_xive.c
2718
&kvm->arch.xive_devices.xics_on_xive;
arch/powerpc/kvm/book3s_xive.c
2742
if (kvm->arch.xive)
arch/powerpc/kvm/book3s_xive.c
2774
kvm->arch.xive = xive;
arch/powerpc/kvm/book3s_xive.c
2808
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
289
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
2902
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
338
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
351
xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
458
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
459
struct kvmppc_xive *xive = vcpu->kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
517
struct kvmppc_xive *xive = vcpu->kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
520
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
623
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
633
xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
665
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
673
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
696
if (!tima || !vcpu->arch.xive_cam_word)
arch/powerpc/kvm/book3s_xive.c
701
__raw_writeq(vcpu->arch.xive_saved_state.w01, tima + TM_QW1_OS);
arch/powerpc/kvm/book3s_xive.c
702
__raw_writel(vcpu->arch.xive_cam_word, tima + TM_QW1_OS + TM_WORD2);
arch/powerpc/kvm/book3s_xive.c
703
vcpu->arch.xive_pushed = 1;
arch/powerpc/kvm/book3s_xive.c
713
vcpu->arch.irq_pending = 0;
arch/powerpc/kvm/book3s_xive.c
719
if (vcpu->arch.xive_esc_on) {
arch/powerpc/kvm/book3s_xive.c
720
pq = __raw_readq((void __iomem *)(vcpu->arch.xive_esc_vaddr +
arch/powerpc/kvm/book3s_xive.c
747
vcpu->arch.xive_esc_on = 0;
arch/powerpc/kvm/book3s_xive.c
760
if (!vcpu->arch.xive_pushed)
arch/powerpc/kvm/book3s_xive.c
774
vcpu->arch.xive_saved_state.w01 = __raw_readq(tima + TM_QW1_OS);
arch/powerpc/kvm/book3s_xive.c
777
vcpu->arch.xive_saved_state.lsmfb = 0;
arch/powerpc/kvm/book3s_xive.c
778
vcpu->arch.xive_saved_state.ack = 0xff;
arch/powerpc/kvm/book3s_xive.c
779
vcpu->arch.xive_pushed = 0;
arch/powerpc/kvm/book3s_xive.c
786
void __iomem *esc_vaddr = (void __iomem *)vcpu->arch.xive_esc_vaddr;
arch/powerpc/kvm/book3s_xive.c
794
if (vcpu->arch.xive_esc_on) {
arch/powerpc/kvm/book3s_xive.c
811
vcpu->arch.xive_esc_on = true;
arch/powerpc/kvm/book3s_xive.c
844
vcpu->arch.irq_pending = 1;
arch/powerpc/kvm/book3s_xive.c
846
if (vcpu->arch.ceded || vcpu->arch.nested)
arch/powerpc/kvm/book3s_xive.c
858
vcpu->arch.xive_esc_on = false;
arch/powerpc/kvm/book3s_xive.c
869
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
888
vcpu->kvm->arch.lpid, xc->server_num);
arch/powerpc/kvm/book3s_xive.c
891
vcpu->kvm->arch.lpid, xc->server_num, prio);
arch/powerpc/kvm/book3s_xive.c
922
vcpu->arch.xive_esc_raddr = xd->eoi_page;
arch/powerpc/kvm/book3s_xive.c
923
vcpu->arch.xive_esc_vaddr = (__force u64)xd->eoi_mmio;
arch/powerpc/kvm/book3s_xive.c
937
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive.c
973
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive.c
988
if (!vcpu->arch.xive_vcpu)
arch/powerpc/kvm/book3s_xive.h
205
if (vcpu->arch.xive_vcpu && nr == vcpu->arch.xive_vcpu->server_num)
arch/powerpc/kvm/book3s_xive.h
246
if (vcpu->arch.xive_vcpu && vp_id == vcpu->arch.xive_vcpu->vp_id)
arch/powerpc/kvm/book3s_xive_native.c
1066
kvm->arch.xive = NULL;
arch/powerpc/kvm/book3s_xive_native.c
108
vcpu->arch.xive_cam_word = 0;
arch/powerpc/kvm/book3s_xive_native.c
1098
if (kvm->arch.xive)
arch/powerpc/kvm/book3s_xive_native.c
1126
kvm->arch.xive = xive;
arch/powerpc/kvm/book3s_xive_native.c
1138
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
1149
val->xive_timaval[0] = vcpu->arch.xive_saved_state.w01;
arch/powerpc/kvm/book3s_xive_native.c
1164
vcpu->arch.xive_saved_state.nsr,
arch/powerpc/kvm/book3s_xive_native.c
1165
vcpu->arch.xive_saved_state.cppr,
arch/powerpc/kvm/book3s_xive_native.c
1166
vcpu->arch.xive_saved_state.ipb,
arch/powerpc/kvm/book3s_xive_native.c
1167
vcpu->arch.xive_saved_state.pipr,
arch/powerpc/kvm/book3s_xive_native.c
1168
vcpu->arch.xive_saved_state.w01,
arch/powerpc/kvm/book3s_xive_native.c
1169
(u32) vcpu->arch.xive_cam_word, opal_state);
arch/powerpc/kvm/book3s_xive_native.c
1176
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
1177
struct kvmppc_xive *xive = vcpu->kvm->arch.xive;
arch/powerpc/kvm/book3s_xive_native.c
1189
if (WARN_ON(vcpu->arch.xive_pushed))
arch/powerpc/kvm/book3s_xive_native.c
119
vcpu->arch.irq_type = KVMPPC_IRQ_DEFAULT;
arch/powerpc/kvm/book3s_xive_native.c
1196
vcpu->arch.xive_saved_state.w01 = val->xive_timaval[0];
arch/powerpc/kvm/book3s_xive_native.c
120
vcpu->arch.xive_vcpu = NULL;
arch/powerpc/kvm/book3s_xive_native.c
1224
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
1232
vcpu->arch.xive_saved_state.nsr,
arch/powerpc/kvm/book3s_xive_native.c
1233
vcpu->arch.xive_saved_state.cppr,
arch/powerpc/kvm/book3s_xive_native.c
1234
vcpu->arch.xive_saved_state.ipb,
arch/powerpc/kvm/book3s_xive_native.c
1235
vcpu->arch.xive_saved_state.pipr,
arch/powerpc/kvm/book3s_xive_native.c
1236
be64_to_cpu(vcpu->arch.xive_saved_state.w01),
arch/powerpc/kvm/book3s_xive_native.c
1237
be32_to_cpu(vcpu->arch.xive_cam_word));
arch/powerpc/kvm/book3s_xive_native.c
139
if (vcpu->arch.irq_type != KVMPPC_IRQ_DEFAULT)
arch/powerpc/kvm/book3s_xive_native.c
154
vcpu->arch.xive_vcpu = xc;
arch/powerpc/kvm/book3s_xive_native.c
161
vcpu->arch.irq_type = KVMPPC_IRQ_XIVE;
arch/powerpc/kvm/book3s_xive_native.c
186
vcpu->arch.xive_saved_state.w01 = cpu_to_be64(0xff000000);
arch/powerpc/kvm/book3s_xive_native.c
187
vcpu->arch.xive_cam_word = cpu_to_be32(xc->vp_cam | TM_QW1W2_VO);
arch/powerpc/kvm/book3s_xive_native.c
203
struct kvmppc_xive *xive = kvm->arch.xive;
arch/powerpc/kvm/book3s_xive_native.c
46
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
593
xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
738
xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
76
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
816
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/book3s_xive_native.c
893
struct kvmppc_xive_vcpu *xc = vcpu->arch.xive_vcpu;
arch/powerpc/kvm/booke.c
101
printk("exceptions: %08lx\n", vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
1151
if (vcpu->arch.shared->msr & (MSR_PR | MSR_GS)) {
arch/powerpc/kvm/booke.c
1160
kvmppc_core_queue_program(vcpu, vcpu->arch.fault_esr);
arch/powerpc/kvm/booke.c
1177
if (vcpu->arch.shared->msr & MSR_SPE)
arch/powerpc/kvm/booke.c
119
vcpu->arch.shadow_msr &= ~MSR_SPE;
arch/powerpc/kvm/booke.c
1212
__func__, exit_nr, vcpu->arch.regs.nip);
arch/powerpc/kvm/booke.c
1235
kvmppc_core_queue_data_storage(vcpu, 0, vcpu->arch.fault_dear,
arch/powerpc/kvm/booke.c
1236
vcpu->arch.fault_esr);
arch/powerpc/kvm/booke.c
1242
kvmppc_core_queue_inst_storage(vcpu, vcpu->arch.fault_esr);
arch/powerpc/kvm/booke.c
1248
kvmppc_core_queue_alignment(vcpu, vcpu->arch.fault_dear,
arch/powerpc/kvm/booke.c
1249
vcpu->arch.fault_esr);
arch/powerpc/kvm/booke.c
1255
if (!(vcpu->arch.shared->msr & MSR_PR)) {
arch/powerpc/kvm/booke.c
1269
if (!(vcpu->arch.shared->msr & MSR_PR) &&
arch/powerpc/kvm/booke.c
1284
unsigned long eaddr = vcpu->arch.fault_dear;
arch/powerpc/kvm/booke.c
129
vcpu->arch.shadow_msr |= MSR_SPE;
arch/powerpc/kvm/booke.c
1290
if (!(vcpu->arch.shared->msr & MSR_PR) &&
arch/powerpc/kvm/booke.c
1291
(eaddr & PAGE_MASK) == vcpu->arch.magic_page_ea) {
arch/powerpc/kvm/booke.c
1305
vcpu->arch.fault_dear,
arch/powerpc/kvm/booke.c
1306
vcpu->arch.fault_esr);
arch/powerpc/kvm/booke.c
1331
vcpu->arch.paddr_accessed = gpaddr;
arch/powerpc/kvm/booke.c
1332
vcpu->arch.vaddr_accessed = eaddr;
arch/powerpc/kvm/booke.c
1342
unsigned long eaddr = vcpu->arch.regs.nip;
arch/powerpc/kvm/booke.c
135
if (vcpu->arch.shared->msr & MSR_SPE) {
arch/powerpc/kvm/booke.c
136
if (!(vcpu->arch.shadow_msr & MSR_SPE))
arch/powerpc/kvm/booke.c
138
} else if (vcpu->arch.shadow_msr & MSR_SPE) {
arch/powerpc/kvm/booke.c
1418
u32 old_tsr = vcpu->arch.tsr;
arch/powerpc/kvm/booke.c
1420
vcpu->arch.tsr = new_tsr;
arch/powerpc/kvm/booke.c
1422
if ((old_tsr ^ vcpu->arch.tsr) & (TSR_ENW | TSR_WIS))
arch/powerpc/kvm/booke.c
1431
spin_lock_init(&vcpu->arch.wdt_lock);
arch/powerpc/kvm/booke.c
1432
timer_setup(&vcpu->arch.wdt_timer, kvmppc_watchdog_func, 0);
arch/powerpc/kvm/booke.c
1444
timer_delete_sync(&vcpu->arch.wdt_timer);
arch/powerpc/kvm/booke.c
1453
regs->pc = vcpu->arch.regs.nip;
arch/powerpc/kvm/booke.c
1455
regs->ctr = vcpu->arch.regs.ctr;
arch/powerpc/kvm/booke.c
1456
regs->lr = vcpu->arch.regs.link;
arch/powerpc/kvm/booke.c
1458
regs->msr = vcpu->arch.shared->msr;
arch/powerpc/kvm/booke.c
1461
regs->pid = vcpu->arch.pid;
arch/powerpc/kvm/booke.c
1484
vcpu->arch.regs.nip = regs->pc;
arch/powerpc/kvm/booke.c
1486
vcpu->arch.regs.ctr = regs->ctr;
arch/powerpc/kvm/booke.c
1487
vcpu->arch.regs.link = regs->lr;
arch/powerpc/kvm/booke.c
1516
sregs->u.e.csrr0 = vcpu->arch.csrr0;
arch/powerpc/kvm/booke.c
1517
sregs->u.e.csrr1 = vcpu->arch.csrr1;
arch/powerpc/kvm/booke.c
1518
sregs->u.e.mcsr = vcpu->arch.mcsr;
arch/powerpc/kvm/booke.c
1521
sregs->u.e.tsr = vcpu->arch.tsr;
arch/powerpc/kvm/booke.c
1522
sregs->u.e.tcr = vcpu->arch.tcr;
arch/powerpc/kvm/booke.c
1525
sregs->u.e.vrsave = vcpu->arch.vrsave;
arch/powerpc/kvm/booke.c
1534
vcpu->arch.csrr0 = sregs->u.e.csrr0;
arch/powerpc/kvm/booke.c
1535
vcpu->arch.csrr1 = sregs->u.e.csrr1;
arch/powerpc/kvm/booke.c
1536
vcpu->arch.mcsr = sregs->u.e.mcsr;
arch/powerpc/kvm/booke.c
1539
vcpu->arch.vrsave = sregs->u.e.vrsave;
arch/powerpc/kvm/booke.c
1543
vcpu->arch.dec = sregs->u.e.dec;
arch/powerpc/kvm/booke.c
1559
sregs->u.e.mcsrr0 = vcpu->arch.mcsrr0;
arch/powerpc/kvm/booke.c
1560
sregs->u.e.mcsrr1 = vcpu->arch.mcsrr1;
arch/powerpc/kvm/booke.c
1561
sregs->u.e.decar = vcpu->arch.decar;
arch/powerpc/kvm/booke.c
1562
sregs->u.e.ivpr = vcpu->arch.ivpr;
arch/powerpc/kvm/booke.c
1574
vcpu->arch.mcsrr0 = sregs->u.e.mcsrr0;
arch/powerpc/kvm/booke.c
1575
vcpu->arch.mcsrr1 = sregs->u.e.mcsrr1;
arch/powerpc/kvm/booke.c
1576
vcpu->arch.decar = sregs->u.e.decar;
arch/powerpc/kvm/booke.c
1577
vcpu->arch.ivpr = sregs->u.e.ivpr;
arch/powerpc/kvm/booke.c
1586
sregs->u.e.ivor_low[0] = vcpu->arch.ivor[BOOKE_IRQPRIO_CRITICAL];
arch/powerpc/kvm/booke.c
1587
sregs->u.e.ivor_low[1] = vcpu->arch.ivor[BOOKE_IRQPRIO_MACHINE_CHECK];
arch/powerpc/kvm/booke.c
1588
sregs->u.e.ivor_low[2] = vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE];
arch/powerpc/kvm/booke.c
1589
sregs->u.e.ivor_low[3] = vcpu->arch.ivor[BOOKE_IRQPRIO_INST_STORAGE];
arch/powerpc/kvm/booke.c
1590
sregs->u.e.ivor_low[4] = vcpu->arch.ivor[BOOKE_IRQPRIO_EXTERNAL];
arch/powerpc/kvm/booke.c
1591
sregs->u.e.ivor_low[5] = vcpu->arch.ivor[BOOKE_IRQPRIO_ALIGNMENT];
arch/powerpc/kvm/booke.c
1592
sregs->u.e.ivor_low[6] = vcpu->arch.ivor[BOOKE_IRQPRIO_PROGRAM];
arch/powerpc/kvm/booke.c
1593
sregs->u.e.ivor_low[7] = vcpu->arch.ivor[BOOKE_IRQPRIO_FP_UNAVAIL];
arch/powerpc/kvm/booke.c
1594
sregs->u.e.ivor_low[8] = vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL];
arch/powerpc/kvm/booke.c
1595
sregs->u.e.ivor_low[9] = vcpu->arch.ivor[BOOKE_IRQPRIO_AP_UNAVAIL];
arch/powerpc/kvm/booke.c
1596
sregs->u.e.ivor_low[10] = vcpu->arch.ivor[BOOKE_IRQPRIO_DECREMENTER];
arch/powerpc/kvm/booke.c
1597
sregs->u.e.ivor_low[11] = vcpu->arch.ivor[BOOKE_IRQPRIO_FIT];
arch/powerpc/kvm/booke.c
1598
sregs->u.e.ivor_low[12] = vcpu->arch.ivor[BOOKE_IRQPRIO_WATCHDOG];
arch/powerpc/kvm/booke.c
1599
sregs->u.e.ivor_low[13] = vcpu->arch.ivor[BOOKE_IRQPRIO_DTLB_MISS];
arch/powerpc/kvm/booke.c
1600
sregs->u.e.ivor_low[14] = vcpu->arch.ivor[BOOKE_IRQPRIO_ITLB_MISS];
arch/powerpc/kvm/booke.c
1601
sregs->u.e.ivor_low[15] = vcpu->arch.ivor[BOOKE_IRQPRIO_DEBUG];
arch/powerpc/kvm/booke.c
1610
vcpu->arch.ivor[BOOKE_IRQPRIO_CRITICAL] = sregs->u.e.ivor_low[0];
arch/powerpc/kvm/booke.c
1611
vcpu->arch.ivor[BOOKE_IRQPRIO_MACHINE_CHECK] = sregs->u.e.ivor_low[1];
arch/powerpc/kvm/booke.c
1612
vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE] = sregs->u.e.ivor_low[2];
arch/powerpc/kvm/booke.c
1613
vcpu->arch.ivor[BOOKE_IRQPRIO_INST_STORAGE] = sregs->u.e.ivor_low[3];
arch/powerpc/kvm/booke.c
1614
vcpu->arch.ivor[BOOKE_IRQPRIO_EXTERNAL] = sregs->u.e.ivor_low[4];
arch/powerpc/kvm/booke.c
1615
vcpu->arch.ivor[BOOKE_IRQPRIO_ALIGNMENT] = sregs->u.e.ivor_low[5];
arch/powerpc/kvm/booke.c
1616
vcpu->arch.ivor[BOOKE_IRQPRIO_PROGRAM] = sregs->u.e.ivor_low[6];
arch/powerpc/kvm/booke.c
1617
vcpu->arch.ivor[BOOKE_IRQPRIO_FP_UNAVAIL] = sregs->u.e.ivor_low[7];
arch/powerpc/kvm/booke.c
1618
vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL] = sregs->u.e.ivor_low[8];
arch/powerpc/kvm/booke.c
1619
vcpu->arch.ivor[BOOKE_IRQPRIO_AP_UNAVAIL] = sregs->u.e.ivor_low[9];
arch/powerpc/kvm/booke.c
162
load_fp_state(&vcpu->arch.fp);
arch/powerpc/kvm/booke.c
1620
vcpu->arch.ivor[BOOKE_IRQPRIO_DECREMENTER] = sregs->u.e.ivor_low[10];
arch/powerpc/kvm/booke.c
1621
vcpu->arch.ivor[BOOKE_IRQPRIO_FIT] = sregs->u.e.ivor_low[11];
arch/powerpc/kvm/booke.c
1622
vcpu->arch.ivor[BOOKE_IRQPRIO_WATCHDOG] = sregs->u.e.ivor_low[12];
arch/powerpc/kvm/booke.c
1623
vcpu->arch.ivor[BOOKE_IRQPRIO_DTLB_MISS] = sregs->u.e.ivor_low[13];
arch/powerpc/kvm/booke.c
1624
vcpu->arch.ivor[BOOKE_IRQPRIO_ITLB_MISS] = sregs->u.e.ivor_low[14];
arch/powerpc/kvm/booke.c
1625
vcpu->arch.ivor[BOOKE_IRQPRIO_DEBUG] = sregs->u.e.ivor_low[15];
arch/powerpc/kvm/booke.c
1637
sregs->pvr = vcpu->arch.pvr;
arch/powerpc/kvm/booke.c
164
current->thread.fp_save_area = &vcpu->arch.fp;
arch/powerpc/kvm/booke.c
1641
ret = vcpu->kvm->arch.kvm_ops->get_sregs(vcpu, sregs);
arch/powerpc/kvm/booke.c
1653
if (vcpu->arch.pvr != sregs->pvr)
arch/powerpc/kvm/booke.c
1664
ret = vcpu->kvm->arch.kvm_ops->set_sregs(vcpu, sregs);
arch/powerpc/kvm/booke.c
1678
*val = get_reg_val(id, vcpu->arch.dbg_reg.iac1);
arch/powerpc/kvm/booke.c
1681
*val = get_reg_val(id, vcpu->arch.dbg_reg.iac2);
arch/powerpc/kvm/booke.c
1685
*val = get_reg_val(id, vcpu->arch.dbg_reg.iac3);
arch/powerpc/kvm/booke.c
1688
*val = get_reg_val(id, vcpu->arch.dbg_reg.iac4);
arch/powerpc/kvm/booke.c
1692
*val = get_reg_val(id, vcpu->arch.dbg_reg.dac1);
arch/powerpc/kvm/booke.c
1695
*val = get_reg_val(id, vcpu->arch.dbg_reg.dac2);
arch/powerpc/kvm/booke.c
1704
*val = get_reg_val(id, vcpu->arch.epcr);
arch/powerpc/kvm/booke.c
1708
*val = get_reg_val(id, vcpu->arch.tcr);
arch/powerpc/kvm/booke.c
1711
*val = get_reg_val(id, vcpu->arch.tsr);
arch/powerpc/kvm/booke.c
1717
*val = get_reg_val(id, vcpu->arch.vrsave);
arch/powerpc/kvm/booke.c
1720
r = vcpu->kvm->arch.kvm_ops->get_one_reg(vcpu, id, val);
arch/powerpc/kvm/booke.c
1734
vcpu->arch.dbg_reg.iac1 = set_reg_val(id, *val);
arch/powerpc/kvm/booke.c
1737
vcpu->arch.dbg_reg.iac2 = set_reg_val(id, *val);
arch/powerpc/kvm/booke.c
1741
vcpu->arch.dbg_reg.iac3 = set_reg_val(id, *val);
arch/powerpc/kvm/booke.c
1744
vcpu->arch.dbg_reg.iac4 = set_reg_val(id, *val);
arch/powerpc/kvm/booke.c
1748
vcpu->arch.dbg_reg.dac1 = set_reg_val(id, *val);
arch/powerpc/kvm/booke.c
1751
vcpu->arch.dbg_reg.dac2 = set_reg_val(id, *val);
arch/powerpc/kvm/booke.c
1786
vcpu->arch.vrsave = set_reg_val(id, *val);
arch/powerpc/kvm/booke.c
1789
r = vcpu->kvm->arch.kvm_ops->set_one_reg(vcpu, id, val);
arch/powerpc/kvm/booke.c
1853
vcpu->arch.epcr = new_epcr;
arch/powerpc/kvm/booke.c
1855
vcpu->arch.shadow_epcr &= ~SPRN_EPCR_GICM;
arch/powerpc/kvm/booke.c
1856
if (vcpu->arch.epcr & SPRN_EPCR_ICM)
arch/powerpc/kvm/booke.c
1857
vcpu->arch.shadow_epcr |= SPRN_EPCR_GICM;
arch/powerpc/kvm/booke.c
1864
vcpu->arch.tcr = new_tcr;
arch/powerpc/kvm/booke.c
1871
set_bits(tsr_bits, &vcpu->arch.tsr);
arch/powerpc/kvm/booke.c
1879
clear_bits(tsr_bits, &vcpu->arch.tsr);
arch/powerpc/kvm/booke.c
188
vcpu->arch.shadow_msr &= ~MSR_FP;
arch/powerpc/kvm/booke.c
189
vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_FP;
arch/powerpc/kvm/booke.c
1893
if (vcpu->arch.tcr & TCR_ARE) {
arch/powerpc/kvm/booke.c
1894
vcpu->arch.dec = vcpu->arch.decar;
arch/powerpc/kvm/booke.c
1964
vcpu->arch.shadow_msrp |= MSRP_UCLEP;
arch/powerpc/kvm/booke.c
1966
vcpu->arch.shadow_msrp |= MSRP_DEP;
arch/powerpc/kvm/booke.c
1968
vcpu->arch.shadow_msrp |= MSRP_PMMP;
arch/powerpc/kvm/booke.c
1971
vcpu->arch.shadow_msrp &= ~MSRP_UCLEP;
arch/powerpc/kvm/booke.c
1973
vcpu->arch.shadow_msrp &= ~MSRP_DEP;
arch/powerpc/kvm/booke.c
1975
vcpu->arch.shadow_msrp &= ~MSRP_PMMP;
arch/powerpc/kvm/booke.c
1987
if (!(vcpu->arch.shared->msr & MSR_PR) &&
arch/powerpc/kvm/booke.c
1988
(eaddr & PAGE_MASK) == vcpu->arch.magic_page_ea) {
arch/powerpc/kvm/booke.c
1990
pte->raddr = (vcpu->arch.magic_page_pa & PAGE_MASK) |
arch/powerpc/kvm/booke.c
204
load_vr_state(&vcpu->arch.vr);
arch/powerpc/kvm/booke.c
2041
vcpu->arch.dbg_reg.dbcr0 = 0;
arch/powerpc/kvm/booke.c
2049
vcpu->arch.dbg_reg.dbcr0 = 0;
arch/powerpc/kvm/booke.c
2052
vcpu->arch.dbg_reg.dbcr0 |= DBCR0_IDM | DBCR0_IC;
arch/powerpc/kvm/booke.c
2055
dbg_reg = &(vcpu->arch.dbg_reg);
arch/powerpc/kvm/booke.c
206
current->thread.vr_save_area = &vcpu->arch.vr;
arch/powerpc/kvm/booke.c
2080
uint64_t addr = dbg->arch.bp[n].addr;
arch/powerpc/kvm/booke.c
2081
uint32_t type = dbg->arch.bp[n].type;
arch/powerpc/kvm/booke.c
2126
return kvm->arch.kvm_ops->init_vm(kvm);
arch/powerpc/kvm/booke.c
2134
r = vcpu->kvm->arch.kvm_ops->vcpu_create(vcpu);
arch/powerpc/kvm/booke.c
2139
vcpu->arch.regs.nip = 0;
arch/powerpc/kvm/booke.c
2140
vcpu->arch.shared->pir = vcpu->vcpu_id;
arch/powerpc/kvm/booke.c
2145
vcpu->arch.shadow_msr = MSR_USER | MSR_IS | MSR_DS;
arch/powerpc/kvm/booke.c
2146
vcpu->arch.shadow_pid = 1;
arch/powerpc/kvm/booke.c
2147
vcpu->arch.shared->msr = 0;
arch/powerpc/kvm/booke.c
2152
vcpu->arch.ivpr = 0x55550000;
arch/powerpc/kvm/booke.c
2154
vcpu->arch.ivor[i] = 0x7700 | i * 4;
arch/powerpc/kvm/booke.c
2160
vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu);
arch/powerpc/kvm/booke.c
2167
vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu);
arch/powerpc/kvm/booke.c
2172
kvm->arch.kvm_ops->destroy_vm(kvm);
arch/powerpc/kvm/booke.c
2177
vcpu->kvm->arch.kvm_ops->vcpu_load(vcpu, cpu);
arch/powerpc/kvm/booke.c
2182
vcpu->kvm->arch.kvm_ops->vcpu_put(vcpu);
arch/powerpc/kvm/booke.c
232
vcpu->arch.shadow_msr &= ~MSR_DE;
arch/powerpc/kvm/booke.c
233
vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_DE;
arch/powerpc/kvm/booke.c
243
vcpu->arch.shared->msr |= MSR_DE;
arch/powerpc/kvm/booke.c
245
vcpu->arch.shadow_msr |= MSR_DE;
arch/powerpc/kvm/booke.c
246
vcpu->arch.shared->msr &= ~MSR_DE;
arch/powerpc/kvm/booke.c
257
u32 old_msr = vcpu->arch.shared->msr;
arch/powerpc/kvm/booke.c
263
vcpu->arch.shared->msr = new_msr;
arch/powerpc/kvm/booke.c
275
set_bit(priority, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
281
vcpu->arch.queued_dear = dear_flags;
arch/powerpc/kvm/booke.c
282
vcpu->arch.queued_esr = esr_flags;
arch/powerpc/kvm/booke.c
290
vcpu->arch.queued_dear = dear_flags;
arch/powerpc/kvm/booke.c
291
vcpu->arch.queued_esr = esr_flags;
arch/powerpc/kvm/booke.c
302
vcpu->arch.queued_esr = esr_flags;
arch/powerpc/kvm/booke.c
309
vcpu->arch.queued_dear = dear_flags;
arch/powerpc/kvm/booke.c
310
vcpu->arch.queued_esr = esr_flags;
arch/powerpc/kvm/booke.c
316
vcpu->arch.queued_esr = esr_flags;
arch/powerpc/kvm/booke.c
341
return test_bit(BOOKE_IRQPRIO_DECREMENTER, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
346
clear_bit(BOOKE_IRQPRIO_DECREMENTER, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
362
clear_bit(BOOKE_IRQPRIO_EXTERNAL, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
363
clear_bit(BOOKE_IRQPRIO_EXTERNAL_LEVEL, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
373
clear_bit(BOOKE_IRQPRIO_WATCHDOG, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
383
clear_bit(BOOKE_IRQPRIO_DEBUG, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
394
vcpu->arch.csrr0 = srr0;
arch/powerpc/kvm/booke.c
395
vcpu->arch.csrr1 = srr1;
arch/powerpc/kvm/booke.c
401
vcpu->arch.dsrr0 = srr0;
arch/powerpc/kvm/booke.c
402
vcpu->arch.dsrr1 = srr1;
arch/powerpc/kvm/booke.c
410
vcpu->arch.mcsrr0 = srr0;
arch/powerpc/kvm/booke.c
411
vcpu->arch.mcsrr1 = srr1;
arch/powerpc/kvm/booke.c
421
ulong crit_raw = vcpu->arch.shared->critical;
arch/powerpc/kvm/booke.c
426
ulong new_msr = vcpu->arch.shared->msr;
arch/powerpc/kvm/booke.c
429
if (!(vcpu->arch.shared->msr & MSR_SF)) {
arch/powerpc/kvm/booke.c
437
crit = crit && !(vcpu->arch.shared->msr & MSR_PR);
arch/powerpc/kvm/booke.c
444
if ((priority == BOOKE_IRQPRIO_EXTERNAL) && vcpu->arch.epr_flags)
arch/powerpc/kvm/booke.c
477
allowed = vcpu->arch.shared->msr & MSR_CE;
arch/powerpc/kvm/booke.c
483
allowed = vcpu->arch.shared->msr & MSR_ME;
arch/powerpc/kvm/booke.c
493
allowed = vcpu->arch.shared->msr & MSR_EE;
arch/powerpc/kvm/booke.c
499
allowed = vcpu->arch.shared->msr & MSR_DE;
arch/powerpc/kvm/booke.c
513
set_guest_srr(vcpu, vcpu->arch.regs.nip,
arch/powerpc/kvm/booke.c
514
vcpu->arch.shared->msr);
arch/powerpc/kvm/booke.c
517
set_guest_csrr(vcpu, vcpu->arch.regs.nip,
arch/powerpc/kvm/booke.c
518
vcpu->arch.shared->msr);
arch/powerpc/kvm/booke.c
521
set_guest_dsrr(vcpu, vcpu->arch.regs.nip,
arch/powerpc/kvm/booke.c
522
vcpu->arch.shared->msr);
arch/powerpc/kvm/booke.c
525
set_guest_mcsrr(vcpu, vcpu->arch.regs.nip,
arch/powerpc/kvm/booke.c
526
vcpu->arch.shared->msr);
arch/powerpc/kvm/booke.c
530
vcpu->arch.regs.nip = vcpu->arch.ivpr |
arch/powerpc/kvm/booke.c
531
vcpu->arch.ivor[priority];
arch/powerpc/kvm/booke.c
533
kvmppc_set_esr(vcpu, vcpu->arch.queued_esr);
arch/powerpc/kvm/booke.c
535
kvmppc_set_dar(vcpu, vcpu->arch.queued_dear);
arch/powerpc/kvm/booke.c
537
if (vcpu->arch.epr_flags & KVMPPC_EPR_USER)
arch/powerpc/kvm/booke.c
539
else if (vcpu->arch.epr_flags & KVMPPC_EPR_KERNEL) {
arch/powerpc/kvm/booke.c
540
BUG_ON(vcpu->arch.irq_type != KVMPPC_IRQ_MPIC);
arch/powerpc/kvm/booke.c
547
if (vcpu->arch.epcr & SPRN_EPCR_ICM)
arch/powerpc/kvm/booke.c
553
clear_bit(priority, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/booke.c
562
if (vcpu->arch.pending_exceptions & BOOKE_IRQMASK_EE)
arch/powerpc/kvm/booke.c
564
if (vcpu->arch.pending_exceptions & BOOKE_IRQMASK_CE)
arch/powerpc/kvm/booke.c
566
if (vcpu->arch.pending_exceptions & BOOKE_IRQPRIO_MACHINE_CHECK)
arch/powerpc/kvm/booke.c
582
u32 period = TCR_GET_WP(vcpu->arch.tcr);
arch/powerpc/kvm/booke.c
613
if ((vcpu->arch.tsr & (TSR_ENW | TSR_WIS)) != (TSR_ENW | TSR_WIS))
arch/powerpc/kvm/booke.c
616
spin_lock_irqsave(&vcpu->arch.wdt_lock, flags);
arch/powerpc/kvm/booke.c
623
mod_timer(&vcpu->arch.wdt_timer, jiffies + nr_jiffies);
arch/powerpc/kvm/booke.c
625
timer_delete(&vcpu->arch.wdt_timer);
arch/powerpc/kvm/booke.c
626
spin_unlock_irqrestore(&vcpu->arch.wdt_lock, flags);
arch/powerpc/kvm/booke.c
631
struct kvm_vcpu *vcpu = timer_container_of(vcpu, t, arch.wdt_timer);
arch/powerpc/kvm/booke.c
636
new_tsr = tsr = vcpu->arch.tsr;
arch/powerpc/kvm/booke.c
648
} while (cmpxchg(&vcpu->arch.tsr, tsr, new_tsr) != tsr);
arch/powerpc/kvm/booke.c
660
if (final && (vcpu->arch.tcr & TCR_WRC_MASK) &&
arch/powerpc/kvm/booke.c
661
vcpu->arch.watchdog_enabled) {
arch/powerpc/kvm/booke.c
679
if ((vcpu->arch.tcr & TCR_DIE) && (vcpu->arch.tsr & TSR_DIS))
arch/powerpc/kvm/booke.c
684
if ((vcpu->arch.tcr & TCR_WIE) && (vcpu->arch.tsr & TSR_WIS))
arch/powerpc/kvm/booke.c
692
unsigned long *pending = &vcpu->arch.pending_exceptions;
arch/powerpc/kvm/booke.c
706
vcpu->arch.shared->int_pending = !!*pending;
arch/powerpc/kvm/booke.c
722
if (vcpu->arch.shared->msr & MSR_WE) {
arch/powerpc/kvm/booke.c
752
vcpu->arch.epr_needed = true;
arch/powerpc/kvm/booke.c
765
if (!vcpu->arch.sane) {
arch/powerpc/kvm/booke.c
800
debug = vcpu->arch.dbg_reg;
arch/powerpc/kvm/booke.c
803
current->thread.debug = vcpu->arch.dbg_reg;
arch/powerpc/kvm/booke.c
805
vcpu->arch.pgdir = vcpu->kvm->mm->pgd;
arch/powerpc/kvm/booke.c
848
__func__, vcpu->arch.regs.nip, vcpu->arch.last_inst);
arch/powerpc/kvm/booke.c
852
vcpu->run->hw.hardware_exit_reason |= vcpu->arch.last_inst;
arch/powerpc/kvm/booke.c
867
struct debug_reg *dbg_reg = &(vcpu->arch.dbg_reg);
arch/powerpc/kvm/booke.c
868
u32 dbsr = vcpu->arch.dbsr;
arch/powerpc/kvm/booke.c
881
if (dbsr && (vcpu->arch.shared->msr & MSR_DE) &&
arch/powerpc/kvm/booke.c
882
(vcpu->arch.dbg_reg.dbcr0 & DBCR0_IDM))
arch/powerpc/kvm/booke.c
886
if ((dbsr & DBSR_TIE) && !(vcpu->arch.shared->msr & MSR_DE))
arch/powerpc/kvm/booke.c
896
vcpu->arch.dbsr = 0;
arch/powerpc/kvm/booke.c
897
run->debug.arch.status = 0;
arch/powerpc/kvm/booke.c
898
run->debug.arch.address = vcpu->arch.regs.nip;
arch/powerpc/kvm/booke.c
901
run->debug.arch.status |= KVMPPC_DEBUG_BREAKPOINT;
arch/powerpc/kvm/booke.c
904
run->debug.arch.status |= KVMPPC_DEBUG_WATCH_WRITE;
arch/powerpc/kvm/booke.c
906
run->debug.arch.status |= KVMPPC_DEBUG_WATCH_READ;
arch/powerpc/kvm/booke.c
908
run->debug.arch.address = dbg_reg->dac1;
arch/powerpc/kvm/booke.c
910
run->debug.arch.address = dbg_reg->dac2;
arch/powerpc/kvm/booke.c
94
printk("pc: %08lx msr: %08llx\n", vcpu->arch.regs.nip,
arch/powerpc/kvm/booke.c
95
vcpu->arch.shared->msr);
arch/powerpc/kvm/booke.c
96
printk("lr: %08lx ctr: %08lx\n", vcpu->arch.regs.link,
arch/powerpc/kvm/booke.c
97
vcpu->arch.regs.ctr);
arch/powerpc/kvm/booke.c
978
vcpu->arch.dbsr = mfspr(SPRN_DBSR);
arch/powerpc/kvm/booke.c
98
printk("srr0: %08llx srr1: %08llx\n", vcpu->arch.shared->srr0,
arch/powerpc/kvm/booke.c
99
vcpu->arch.shared->srr1);
arch/powerpc/kvm/booke.c
993
__func__, vcpu->arch.regs.nip);
arch/powerpc/kvm/booke_emulate.c
127
vcpu->arch.shared->dar = spr_val;
arch/powerpc/kvm/booke_emulate.c
130
vcpu->arch.shared->esr = spr_val;
arch/powerpc/kvm/booke_emulate.c
133
vcpu->arch.csrr0 = spr_val;
arch/powerpc/kvm/booke_emulate.c
136
vcpu->arch.csrr1 = spr_val;
arch/powerpc/kvm/booke_emulate.c
139
vcpu->arch.dsrr0 = spr_val;
arch/powerpc/kvm/booke_emulate.c
142
vcpu->arch.dsrr1 = spr_val;
arch/powerpc/kvm/booke_emulate.c
153
vcpu->arch.dbg_reg.iac1 = spr_val;
arch/powerpc/kvm/booke_emulate.c
164
vcpu->arch.dbg_reg.iac2 = spr_val;
arch/powerpc/kvm/booke_emulate.c
176
vcpu->arch.dbg_reg.iac3 = spr_val;
arch/powerpc/kvm/booke_emulate.c
187
vcpu->arch.dbg_reg.iac4 = spr_val;
arch/powerpc/kvm/booke_emulate.c
199
vcpu->arch.dbg_reg.dac1 = spr_val;
arch/powerpc/kvm/booke_emulate.c
210
vcpu->arch.dbg_reg.dac2 = spr_val;
arch/powerpc/kvm/booke_emulate.c
225
vcpu->arch.dbg_reg.dbcr0 = spr_val;
arch/powerpc/kvm/booke_emulate.c
236
vcpu->arch.dbg_reg.dbcr1 = spr_val;
arch/powerpc/kvm/booke_emulate.c
247
vcpu->arch.dbg_reg.dbcr2 = spr_val;
arch/powerpc/kvm/booke_emulate.c
257
vcpu->arch.dbsr &= ~spr_val;
arch/powerpc/kvm/booke_emulate.c
258
if (!(vcpu->arch.dbsr & ~DBSR_IDE))
arch/powerpc/kvm/booke_emulate.c
26
vcpu->arch.regs.nip = vcpu->arch.shared->srr0;
arch/powerpc/kvm/booke_emulate.c
269
if (vcpu->arch.tcr & TCR_WRC_MASK) {
arch/powerpc/kvm/booke_emulate.c
27
kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1);
arch/powerpc/kvm/booke_emulate.c
271
spr_val |= vcpu->arch.tcr & TCR_WRC_MASK;
arch/powerpc/kvm/booke_emulate.c
277
vcpu->arch.decar = spr_val;
arch/powerpc/kvm/booke_emulate.c
298
vcpu->arch.ivpr = spr_val;
arch/powerpc/kvm/booke_emulate.c
304
vcpu->arch.ivor[BOOKE_IRQPRIO_CRITICAL] = spr_val;
arch/powerpc/kvm/booke_emulate.c
307
vcpu->arch.ivor[BOOKE_IRQPRIO_MACHINE_CHECK] = spr_val;
arch/powerpc/kvm/booke_emulate.c
310
vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE] = spr_val;
arch/powerpc/kvm/booke_emulate.c
316
vcpu->arch.ivor[BOOKE_IRQPRIO_INST_STORAGE] = spr_val;
arch/powerpc/kvm/booke_emulate.c
319
vcpu->arch.ivor[BOOKE_IRQPRIO_EXTERNAL] = spr_val;
arch/powerpc/kvm/booke_emulate.c
32
vcpu->arch.regs.nip = vcpu->arch.dsrr0;
arch/powerpc/kvm/booke_emulate.c
322
vcpu->arch.ivor[BOOKE_IRQPRIO_ALIGNMENT] = spr_val;
arch/powerpc/kvm/booke_emulate.c
325
vcpu->arch.ivor[BOOKE_IRQPRIO_PROGRAM] = spr_val;
arch/powerpc/kvm/booke_emulate.c
328
vcpu->arch.ivor[BOOKE_IRQPRIO_FP_UNAVAIL] = spr_val;
arch/powerpc/kvm/booke_emulate.c
33
kvmppc_set_msr(vcpu, vcpu->arch.dsrr1);
arch/powerpc/kvm/booke_emulate.c
331
vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL] = spr_val;
arch/powerpc/kvm/booke_emulate.c
337
vcpu->arch.ivor[BOOKE_IRQPRIO_AP_UNAVAIL] = spr_val;
arch/powerpc/kvm/booke_emulate.c
340
vcpu->arch.ivor[BOOKE_IRQPRIO_DECREMENTER] = spr_val;
arch/powerpc/kvm/booke_emulate.c
343
vcpu->arch.ivor[BOOKE_IRQPRIO_FIT] = spr_val;
arch/powerpc/kvm/booke_emulate.c
346
vcpu->arch.ivor[BOOKE_IRQPRIO_WATCHDOG] = spr_val;
arch/powerpc/kvm/booke_emulate.c
349
vcpu->arch.ivor[BOOKE_IRQPRIO_DTLB_MISS] = spr_val;
arch/powerpc/kvm/booke_emulate.c
352
vcpu->arch.ivor[BOOKE_IRQPRIO_ITLB_MISS] = spr_val;
arch/powerpc/kvm/booke_emulate.c
355
vcpu->arch.ivor[BOOKE_IRQPRIO_DEBUG] = spr_val;
arch/powerpc/kvm/booke_emulate.c
358
vcpu->arch.mcsr &= ~spr_val;
arch/powerpc/kvm/booke_emulate.c
364
mtspr(SPRN_EPCR, vcpu->arch.shadow_epcr);
arch/powerpc/kvm/booke_emulate.c
373
current->thread.debug = vcpu->arch.dbg_reg;
arch/powerpc/kvm/booke_emulate.c
374
switch_booke_debug_regs(&vcpu->arch.dbg_reg);
arch/powerpc/kvm/booke_emulate.c
38
vcpu->arch.regs.nip = vcpu->arch.csrr0;
arch/powerpc/kvm/booke_emulate.c
385
*spr_val = vcpu->arch.ivpr;
arch/powerpc/kvm/booke_emulate.c
388
*spr_val = vcpu->arch.shared->dar;
arch/powerpc/kvm/booke_emulate.c
39
kvmppc_set_msr(vcpu, vcpu->arch.csrr1);
arch/powerpc/kvm/booke_emulate.c
391
*spr_val = vcpu->arch.shared->esr;
arch/powerpc/kvm/booke_emulate.c
394
*spr_val = vcpu->arch.epr;
arch/powerpc/kvm/booke_emulate.c
397
*spr_val = vcpu->arch.csrr0;
arch/powerpc/kvm/booke_emulate.c
400
*spr_val = vcpu->arch.csrr1;
arch/powerpc/kvm/booke_emulate.c
403
*spr_val = vcpu->arch.dsrr0;
arch/powerpc/kvm/booke_emulate.c
406
*spr_val = vcpu->arch.dsrr1;
arch/powerpc/kvm/booke_emulate.c
409
*spr_val = vcpu->arch.dbg_reg.iac1;
arch/powerpc/kvm/booke_emulate.c
412
*spr_val = vcpu->arch.dbg_reg.iac2;
arch/powerpc/kvm/booke_emulate.c
416
*spr_val = vcpu->arch.dbg_reg.iac3;
arch/powerpc/kvm/booke_emulate.c
419
*spr_val = vcpu->arch.dbg_reg.iac4;
arch/powerpc/kvm/booke_emulate.c
423
*spr_val = vcpu->arch.dbg_reg.dac1;
arch/powerpc/kvm/booke_emulate.c
426
*spr_val = vcpu->arch.dbg_reg.dac2;
arch/powerpc/kvm/booke_emulate.c
429
*spr_val = vcpu->arch.dbg_reg.dbcr0;
arch/powerpc/kvm/booke_emulate.c
434
*spr_val = vcpu->arch.dbg_reg.dbcr1;
arch/powerpc/kvm/booke_emulate.c
437
*spr_val = vcpu->arch.dbg_reg.dbcr2;
arch/powerpc/kvm/booke_emulate.c
440
*spr_val = vcpu->arch.dbsr;
arch/powerpc/kvm/booke_emulate.c
443
*spr_val = vcpu->arch.tsr;
arch/powerpc/kvm/booke_emulate.c
446
*spr_val = vcpu->arch.tcr;
arch/powerpc/kvm/booke_emulate.c
450
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_CRITICAL];
arch/powerpc/kvm/booke_emulate.c
453
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_MACHINE_CHECK];
arch/powerpc/kvm/booke_emulate.c
456
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE];
arch/powerpc/kvm/booke_emulate.c
459
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_INST_STORAGE];
arch/powerpc/kvm/booke_emulate.c
462
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_EXTERNAL];
arch/powerpc/kvm/booke_emulate.c
465
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALIGNMENT];
arch/powerpc/kvm/booke_emulate.c
468
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_PROGRAM];
arch/powerpc/kvm/booke_emulate.c
471
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_FP_UNAVAIL];
arch/powerpc/kvm/booke_emulate.c
474
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL];
arch/powerpc/kvm/booke_emulate.c
477
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_AP_UNAVAIL];
arch/powerpc/kvm/booke_emulate.c
480
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DECREMENTER];
arch/powerpc/kvm/booke_emulate.c
483
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_FIT];
arch/powerpc/kvm/booke_emulate.c
486
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_WATCHDOG];
arch/powerpc/kvm/booke_emulate.c
489
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DTLB_MISS];
arch/powerpc/kvm/booke_emulate.c
492
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ITLB_MISS];
arch/powerpc/kvm/booke_emulate.c
495
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DEBUG];
arch/powerpc/kvm/booke_emulate.c
498
*spr_val = vcpu->arch.mcsr;
arch/powerpc/kvm/booke_emulate.c
502
*spr_val = vcpu->arch.epcr;
arch/powerpc/kvm/booke_emulate.c
80
kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr);
arch/powerpc/kvm/booke_emulate.c
90
vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE)
arch/powerpc/kvm/booke_emulate.c
96
vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE)
arch/powerpc/kvm/e500.c
139
vcpu_e500->vcpu.arch.shadow_pid = kvmppc_e500_get_sid(vcpu_e500,
arch/powerpc/kvm/e500.c
143
vcpu_e500->vcpu.arch.shadow_pid1 = kvmppc_e500_get_sid(vcpu_e500,
arch/powerpc/kvm/e500.c
225
if (vcpu->arch.pid != pid) {
arch/powerpc/kvm/e500.c
226
vcpu_e500->pid[0] = vcpu->arch.pid = pid;
arch/powerpc/kvm/e500.c
310
if (vcpu->arch.shadow_msr & MSR_SPE)
arch/powerpc/kvm/e500.c
353
vcpu->arch.pvr = mfspr(SPRN_PVR);
arch/powerpc/kvm/e500.c
356
vcpu->arch.cpu_type = KVM_CPU_E500V2;
arch/powerpc/kvm/e500.c
375
sregs->u.e.ivor_high[0] = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
arch/powerpc/kvm/e500.c
376
sregs->u.e.ivor_high[1] = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
arch/powerpc/kvm/e500.c
377
sregs->u.e.ivor_high[2] = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
arch/powerpc/kvm/e500.c
379
vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
arch/powerpc/kvm/e500.c
406
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL] =
arch/powerpc/kvm/e500.c
408
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA] =
arch/powerpc/kvm/e500.c
410
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND] =
arch/powerpc/kvm/e500.c
415
vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] =
arch/powerpc/kvm/e500.c
451
vcpu->arch.shared = (void*)__get_free_page(GFP_KERNEL|__GFP_ZERO);
arch/powerpc/kvm/e500.c
452
if (!vcpu->arch.shared) {
arch/powerpc/kvm/e500.c
470
free_page((unsigned long)vcpu->arch.shared);
arch/powerpc/kvm/e500.h
208
return vcpu->arch.pid & 0xff;
arch/powerpc/kvm/e500.h
213
return !!(vcpu->arch.shared->msr & (MSR_IS | MSR_DS));
arch/powerpc/kvm/e500.h
218
return !!(vcpu->arch.shared->msr & MSR_PR);
arch/powerpc/kvm/e500.h
223
return (vcpu->arch.shared->mas6 >> 16) & 0xff;
arch/powerpc/kvm/e500.h
228
return vcpu->arch.shared->mas6 & 0x1;
arch/powerpc/kvm/e500.h
237
return (vcpu->arch.shared->mas0 >> 28) & 0x1;
arch/powerpc/kvm/e500.h
242
return vcpu->arch.shared->mas0 & 0xfff;
arch/powerpc/kvm/e500.h
247
return (vcpu->arch.shared->mas0 >> 16) & 0xfff;
arch/powerpc/kvm/e500.h
261
if (get_tlb_ts(tlbe) != !!(vcpu->arch.shared->msr & MSR_IS))
arch/powerpc/kvm/e500.h
305
return get_thread_specific_lpid(vcpu->kvm->arch.lpid);
arch/powerpc/kvm/e500.h
314
unsigned int tidseld = (vcpu->arch.shared->mas4 >> 16) & 0xf;
arch/powerpc/kvm/e500.h
329
has_ftr = ((vcpu->arch.mmucfg & MMUCFG_MAVN) == MMUCFG_MAVN_V2);
arch/powerpc/kvm/e500_emulate.c
225
vcpu->arch.shared->mas0 = spr_val;
arch/powerpc/kvm/e500_emulate.c
228
vcpu->arch.shared->mas1 = spr_val;
arch/powerpc/kvm/e500_emulate.c
231
vcpu->arch.shared->mas2 = spr_val;
arch/powerpc/kvm/e500_emulate.c
234
vcpu->arch.shared->mas7_3 &= ~(u64)0xffffffff;
arch/powerpc/kvm/e500_emulate.c
235
vcpu->arch.shared->mas7_3 |= spr_val;
arch/powerpc/kvm/e500_emulate.c
238
vcpu->arch.shared->mas4 = spr_val;
arch/powerpc/kvm/e500_emulate.c
241
vcpu->arch.shared->mas6 = spr_val;
arch/powerpc/kvm/e500_emulate.c
244
vcpu->arch.shared->mas7_3 &= (u64)0xffffffff;
arch/powerpc/kvm/e500_emulate.c
245
vcpu->arch.shared->mas7_3 |= (u64)spr_val << 32;
arch/powerpc/kvm/e500_emulate.c
273
vcpu->arch.pwrmgtcr0 = spr_val;
arch/powerpc/kvm/e500_emulate.c
286
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL] = spr_val;
arch/powerpc/kvm/e500_emulate.c
289
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA] = spr_val;
arch/powerpc/kvm/e500_emulate.c
292
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND] = spr_val;
arch/powerpc/kvm/e500_emulate.c
297
vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL] = spr_val;
arch/powerpc/kvm/e500_emulate.c
300
vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST] = spr_val;
arch/powerpc/kvm/e500_emulate.c
304
vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] = spr_val;
arch/powerpc/kvm/e500_emulate.c
308
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] = spr_val;
arch/powerpc/kvm/e500_emulate.c
311
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] = spr_val;
arch/powerpc/kvm/e500_emulate.c
338
*spr_val = vcpu->arch.shared->mas0;
arch/powerpc/kvm/e500_emulate.c
341
*spr_val = vcpu->arch.shared->mas1;
arch/powerpc/kvm/e500_emulate.c
344
*spr_val = vcpu->arch.shared->mas2;
arch/powerpc/kvm/e500_emulate.c
347
*spr_val = (u32)vcpu->arch.shared->mas7_3;
arch/powerpc/kvm/e500_emulate.c
350
*spr_val = vcpu->arch.shared->mas4;
arch/powerpc/kvm/e500_emulate.c
353
*spr_val = vcpu->arch.shared->mas6;
arch/powerpc/kvm/e500_emulate.c
356
*spr_val = vcpu->arch.shared->mas7_3 >> 32;
arch/powerpc/kvm/e500_emulate.c
360
*spr_val = vcpu->arch.decar;
arch/powerpc/kvm/e500_emulate.c
363
*spr_val = vcpu->arch.tlbcfg[0];
arch/powerpc/kvm/e500_emulate.c
366
*spr_val = vcpu->arch.tlbcfg[1];
arch/powerpc/kvm/e500_emulate.c
371
*spr_val = vcpu->arch.tlbps[0];
arch/powerpc/kvm/e500_emulate.c
376
*spr_val = vcpu->arch.tlbps[1];
arch/powerpc/kvm/e500_emulate.c
399
*spr_val = vcpu->arch.mmucfg;
arch/powerpc/kvm/e500_emulate.c
408
*spr_val = vcpu->arch.eptcfg;
arch/powerpc/kvm/e500_emulate.c
412
*spr_val = vcpu->arch.pwrmgtcr0;
arch/powerpc/kvm/e500_emulate.c
418
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
arch/powerpc/kvm/e500_emulate.c
421
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
arch/powerpc/kvm/e500_emulate.c
424
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
arch/powerpc/kvm/e500_emulate.c
429
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL];
arch/powerpc/kvm/e500_emulate.c
432
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST];
arch/powerpc/kvm/e500_emulate.c
436
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
arch/powerpc/kvm/e500_emulate.c
440
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
arch/powerpc/kvm/e500_emulate.c
443
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
arch/powerpc/kvm/e500_emulate.c
53
ulong param = vcpu->arch.regs.gpr[rb];
arch/powerpc/kvm/e500_emulate.c
59
clear_bit(prio, &vcpu->arch.pending_exceptions);
arch/powerpc/kvm/e500_emulate.c
65
ulong param = vcpu->arch.regs.gpr[rb];
arch/powerpc/kvm/e500_emulate.c
75
int cpir = cvcpu->arch.shared->pir;
arch/powerpc/kvm/e500_emulate.c
77
set_bit(prio, &cvcpu->arch.pending_exceptions);
arch/powerpc/kvm/e500_emulate.c
94
vcpu->run->debug.arch.address = vcpu->arch.regs.nip;
arch/powerpc/kvm/e500_emulate.c
95
vcpu->run->debug.arch.status = 0;
arch/powerpc/kvm/e500_mmu.c
134
tlbsel = (vcpu->arch.shared->mas4 >> 28) & 0x1;
arch/powerpc/kvm/e500_mmu.c
136
tsized = (vcpu->arch.shared->mas4 >> 7) & 0x1f;
arch/powerpc/kvm/e500_mmu.c
138
vcpu->arch.shared->mas0 = MAS0_TLBSEL(tlbsel) | MAS0_ESEL(victim)
arch/powerpc/kvm/e500_mmu.c
140
vcpu->arch.shared->mas1 = MAS1_VALID | (as ? MAS1_TS : 0)
arch/powerpc/kvm/e500_mmu.c
143
vcpu->arch.shared->mas2 = (eaddr & MAS2_EPN)
arch/powerpc/kvm/e500_mmu.c
144
| (vcpu->arch.shared->mas4 & MAS2_ATTRIB_MASK);
arch/powerpc/kvm/e500_mmu.c
145
vcpu->arch.shared->mas7_3 &= MAS3_U0 | MAS3_U1 | MAS3_U2 | MAS3_U3;
arch/powerpc/kvm/e500_mmu.c
146
vcpu->arch.shared->mas6 = (vcpu->arch.shared->mas6 & MAS6_SPID1)
arch/powerpc/kvm/e500_mmu.c
332
vcpu->arch.shared->mas0 &= ~MAS0_NV(~0);
arch/powerpc/kvm/e500_mmu.c
333
vcpu->arch.shared->mas0 |= MAS0_NV(vcpu_e500->gtlb_nv[tlbsel]);
arch/powerpc/kvm/e500_mmu.c
334
vcpu->arch.shared->mas1 = gtlbe->mas1;
arch/powerpc/kvm/e500_mmu.c
335
vcpu->arch.shared->mas2 = gtlbe->mas2;
arch/powerpc/kvm/e500_mmu.c
336
vcpu->arch.shared->mas7_3 = gtlbe->mas7_3;
arch/powerpc/kvm/e500_mmu.c
360
vcpu->arch.shared->mas0 = MAS0_TLBSEL(tlbsel) | MAS0_ESEL(esel)
arch/powerpc/kvm/e500_mmu.c
362
vcpu->arch.shared->mas1 = gtlbe->mas1;
arch/powerpc/kvm/e500_mmu.c
363
vcpu->arch.shared->mas2 = gtlbe->mas2;
arch/powerpc/kvm/e500_mmu.c
364
vcpu->arch.shared->mas7_3 = gtlbe->mas7_3;
arch/powerpc/kvm/e500_mmu.c
369
tlbsel = vcpu->arch.shared->mas4 >> 28 & 0x1;
arch/powerpc/kvm/e500_mmu.c
372
vcpu->arch.shared->mas0 = MAS0_TLBSEL(tlbsel)
arch/powerpc/kvm/e500_mmu.c
375
vcpu->arch.shared->mas1 =
arch/powerpc/kvm/e500_mmu.c
376
(vcpu->arch.shared->mas6 & MAS6_SPID0)
arch/powerpc/kvm/e500_mmu.c
377
| ((vcpu->arch.shared->mas6 & MAS6_SAS) ? MAS1_TS : 0)
arch/powerpc/kvm/e500_mmu.c
378
| (vcpu->arch.shared->mas4 & MAS4_TSIZED(~0));
arch/powerpc/kvm/e500_mmu.c
379
vcpu->arch.shared->mas2 &= MAS2_EPN;
arch/powerpc/kvm/e500_mmu.c
380
vcpu->arch.shared->mas2 |= vcpu->arch.shared->mas4 &
arch/powerpc/kvm/e500_mmu.c
382
vcpu->arch.shared->mas7_3 &= MAS3_U0 | MAS3_U1 |
arch/powerpc/kvm/e500_mmu.c
410
gtlbe->mas1 = vcpu->arch.shared->mas1;
arch/powerpc/kvm/e500_mmu.c
411
gtlbe->mas2 = vcpu->arch.shared->mas2;
arch/powerpc/kvm/e500_mmu.c
412
if (!(vcpu->arch.shared->msr & MSR_CM))
arch/powerpc/kvm/e500_mmu.c
414
gtlbe->mas7_3 = vcpu->arch.shared->mas7_3;
arch/powerpc/kvm/e500_mmu.c
416
trace_kvm_booke206_gtlb_write(vcpu->arch.shared->mas0, gtlbe->mas1,
arch/powerpc/kvm/e500_mmu.c
497
unsigned int as = !!(vcpu->arch.shared->msr & MSR_IS);
arch/powerpc/kvm/e500_mmu.c
504
unsigned int as = !!(vcpu->arch.shared->msr & MSR_DS);
arch/powerpc/kvm/e500_mmu.c
511
unsigned int as = !!(vcpu->arch.shared->msr & MSR_IS);
arch/powerpc/kvm/e500_mmu.c
513
kvmppc_e500_deliver_tlb_miss(vcpu, vcpu->arch.regs.nip, as);
arch/powerpc/kvm/e500_mmu.c
518
unsigned int as = !!(vcpu->arch.shared->msr & MSR_DS);
arch/powerpc/kvm/e500_mmu.c
520
kvmppc_e500_deliver_tlb_miss(vcpu, vcpu->arch.fault_dear, as);
arch/powerpc/kvm/e500_mmu.c
569
sregs->u.e.mas0 = vcpu->arch.shared->mas0;
arch/powerpc/kvm/e500_mmu.c
570
sregs->u.e.mas1 = vcpu->arch.shared->mas1;
arch/powerpc/kvm/e500_mmu.c
571
sregs->u.e.mas2 = vcpu->arch.shared->mas2;
arch/powerpc/kvm/e500_mmu.c
572
sregs->u.e.mas7_3 = vcpu->arch.shared->mas7_3;
arch/powerpc/kvm/e500_mmu.c
573
sregs->u.e.mas4 = vcpu->arch.shared->mas4;
arch/powerpc/kvm/e500_mmu.c
574
sregs->u.e.mas6 = vcpu->arch.shared->mas6;
arch/powerpc/kvm/e500_mmu.c
576
sregs->u.e.mmucfg = vcpu->arch.mmucfg;
arch/powerpc/kvm/e500_mmu.c
577
sregs->u.e.tlbcfg[0] = vcpu->arch.tlbcfg[0];
arch/powerpc/kvm/e500_mmu.c
578
sregs->u.e.tlbcfg[1] = vcpu->arch.tlbcfg[1];
arch/powerpc/kvm/e500_mmu.c
586
vcpu->arch.shared->mas0 = sregs->u.e.mas0;
arch/powerpc/kvm/e500_mmu.c
587
vcpu->arch.shared->mas1 = sregs->u.e.mas1;
arch/powerpc/kvm/e500_mmu.c
588
vcpu->arch.shared->mas2 = sregs->u.e.mas2;
arch/powerpc/kvm/e500_mmu.c
589
vcpu->arch.shared->mas7_3 = sregs->u.e.mas7_3;
arch/powerpc/kvm/e500_mmu.c
590
vcpu->arch.shared->mas4 = sregs->u.e.mas4;
arch/powerpc/kvm/e500_mmu.c
591
vcpu->arch.shared->mas6 = sregs->u.e.mas6;
arch/powerpc/kvm/e500_mmu.c
605
*val = get_reg_val(id, vcpu->arch.shared->mas0);
arch/powerpc/kvm/e500_mmu.c
608
*val = get_reg_val(id, vcpu->arch.shared->mas1);
arch/powerpc/kvm/e500_mmu.c
611
*val = get_reg_val(id, vcpu->arch.shared->mas2);
arch/powerpc/kvm/e500_mmu.c
614
*val = get_reg_val(id, vcpu->arch.shared->mas7_3);
arch/powerpc/kvm/e500_mmu.c
617
*val = get_reg_val(id, vcpu->arch.shared->mas4);
arch/powerpc/kvm/e500_mmu.c
620
*val = get_reg_val(id, vcpu->arch.shared->mas6);
arch/powerpc/kvm/e500_mmu.c
623
*val = get_reg_val(id, vcpu->arch.mmucfg);
arch/powerpc/kvm/e500_mmu.c
626
*val = get_reg_val(id, vcpu->arch.eptcfg);
arch/powerpc/kvm/e500_mmu.c
633
*val = get_reg_val(id, vcpu->arch.tlbcfg[i]);
arch/powerpc/kvm/e500_mmu.c
640
*val = get_reg_val(id, vcpu->arch.tlbps[i]);
arch/powerpc/kvm/e500_mmu.c
658
vcpu->arch.shared->mas0 = set_reg_val(id, *val);
arch/powerpc/kvm/e500_mmu.c
661
vcpu->arch.shared->mas1 = set_reg_val(id, *val);
arch/powerpc/kvm/e500_mmu.c
664
vcpu->arch.shared->mas2 = set_reg_val(id, *val);
arch/powerpc/kvm/e500_mmu.c
667
vcpu->arch.shared->mas7_3 = set_reg_val(id, *val);
arch/powerpc/kvm/e500_mmu.c
670
vcpu->arch.shared->mas4 = set_reg_val(id, *val);
arch/powerpc/kvm/e500_mmu.c
673
vcpu->arch.shared->mas6 = set_reg_val(id, *val);
arch/powerpc/kvm/e500_mmu.c
678
if (reg != vcpu->arch.mmucfg)
arch/powerpc/kvm/e500_mmu.c
684
if (reg != vcpu->arch.eptcfg)
arch/powerpc/kvm/e500_mmu.c
695
if (reg != vcpu->arch.tlbcfg[i])
arch/powerpc/kvm/e500_mmu.c
705
if (reg != vcpu->arch.tlbps[i])
arch/powerpc/kvm/e500_mmu.c
71
esel += gtlb0_set_base(vcpu_e500, vcpu->arch.shared->mas2);
arch/powerpc/kvm/e500_mmu.c
720
vcpu->arch.tlbcfg[0] &= ~(TLBnCFG_N_ENTRY | TLBnCFG_ASSOC);
arch/powerpc/kvm/e500_mmu.c
722
vcpu->arch.tlbcfg[0] |= params->tlb_sizes[0];
arch/powerpc/kvm/e500_mmu.c
723
vcpu->arch.tlbcfg[0] |= params->tlb_ways[0] << TLBnCFG_ASSOC_SHIFT;
arch/powerpc/kvm/e500_mmu.c
725
vcpu->arch.tlbcfg[1] &= ~(TLBnCFG_N_ENTRY | TLBnCFG_ASSOC);
arch/powerpc/kvm/e500_mmu.c
726
vcpu->arch.tlbcfg[1] |= params->tlb_sizes[1];
arch/powerpc/kvm/e500_mmu.c
727
vcpu->arch.tlbcfg[1] |= params->tlb_ways[1] << TLBnCFG_ASSOC_SHIFT;
arch/powerpc/kvm/e500_mmu.c
870
vcpu->arch.mmucfg = mfspr(SPRN_MMUCFG) & ~MMUCFG_LPIDSIZE;
arch/powerpc/kvm/e500_mmu.c
873
vcpu->arch.tlbcfg[0] = mfspr(SPRN_TLB0CFG) &
arch/powerpc/kvm/e500_mmu.c
875
vcpu->arch.tlbcfg[0] |= params[0].entries;
arch/powerpc/kvm/e500_mmu.c
876
vcpu->arch.tlbcfg[0] |= params[0].ways << TLBnCFG_ASSOC_SHIFT;
arch/powerpc/kvm/e500_mmu.c
878
vcpu->arch.tlbcfg[1] = mfspr(SPRN_TLB1CFG) &
arch/powerpc/kvm/e500_mmu.c
880
vcpu->arch.tlbcfg[1] |= params[1].entries;
arch/powerpc/kvm/e500_mmu.c
881
vcpu->arch.tlbcfg[1] |= params[1].ways << TLBnCFG_ASSOC_SHIFT;
arch/powerpc/kvm/e500_mmu.c
884
vcpu->arch.tlbps[0] = mfspr(SPRN_TLB0PS);
arch/powerpc/kvm/e500_mmu.c
885
vcpu->arch.tlbps[1] = mfspr(SPRN_TLB1PS);
arch/powerpc/kvm/e500_mmu.c
887
vcpu->arch.mmucfg &= ~MMUCFG_LRAT;
arch/powerpc/kvm/e500_mmu.c
890
vcpu->arch.eptcfg = 0;
arch/powerpc/kvm/e500_mmu.c
891
vcpu->arch.tlbcfg[0] &= ~TLBnCFG_PT;
arch/powerpc/kvm/e500_mmu.c
892
vcpu->arch.tlbcfg[1] &= ~TLBnCFG_IND;
arch/powerpc/kvm/e500_mmu_host.c
134
__write_host_tlbe(stlbe, mas0, vcpu_e500->vcpu.kvm->arch.lpid);
arch/powerpc/kvm/e500_mmu_host.c
139
vcpu_e500->vcpu.kvm->arch.lpid);
arch/powerpc/kvm/e500_mmu_host.c
165
ulong shared_page = ((ulong)vcpu->arch.shared) & PAGE_MASK;
arch/powerpc/kvm/e500_mmu_host.c
177
magic.mas2 = vcpu->arch.magic_page_ea | MAS2_M;
arch/powerpc/kvm/e500_mmu_host.c
308
u32 pr = vcpu->arch.shared->msr & MSR_PR;
arch/powerpc/kvm/e500_mmu_host.c
370
pgdir = vcpu_e500->vcpu.arch.pgdir;
arch/powerpc/kvm/e500_mmu_host.c
601
addr_space = (vcpu->arch.shared->msr & MSR_IS) >> MSR_IR_LG;
arch/powerpc/kvm/e500_mmu_host.c
604
mtspr(SPRN_MAS6, (vcpu->arch.pid << MAS6_SPID_SHIFT) | addr_space);
arch/powerpc/kvm/e500_mmu_host.c
631
pr = vcpu->arch.shared->msr & MSR_PR;
arch/powerpc/kvm/e500mc.c
107
vcpu->arch.pid = pid;
arch/powerpc/kvm/e500mc.c
124
mtspr(SPRN_EPCR, vcpu->arch.shadow_epcr);
arch/powerpc/kvm/e500mc.c
126
mtspr(SPRN_MSRP, vcpu->arch.shadow_msrp);
arch/powerpc/kvm/e500mc.c
127
vcpu->arch.eplc = EPC_EGS | (get_lpid(vcpu) << EPC_ELPID_SHIFT);
arch/powerpc/kvm/e500mc.c
128
vcpu->arch.epsc = vcpu->arch.eplc;
arch/powerpc/kvm/e500mc.c
129
mtspr(SPRN_EPLC, vcpu->arch.eplc);
arch/powerpc/kvm/e500mc.c
130
mtspr(SPRN_EPSC, vcpu->arch.epsc);
arch/powerpc/kvm/e500mc.c
132
mtspr(SPRN_GIVPR, vcpu->arch.ivpr);
arch/powerpc/kvm/e500mc.c
133
mtspr(SPRN_GIVOR2, vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE]);
arch/powerpc/kvm/e500mc.c
134
mtspr(SPRN_GIVOR8, vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL]);
arch/powerpc/kvm/e500mc.c
135
mtspr(SPRN_GSPRG0, (unsigned long)vcpu->arch.shared->sprg0);
arch/powerpc/kvm/e500mc.c
136
mtspr(SPRN_GSPRG1, (unsigned long)vcpu->arch.shared->sprg1);
arch/powerpc/kvm/e500mc.c
137
mtspr(SPRN_GSPRG2, (unsigned long)vcpu->arch.shared->sprg2);
arch/powerpc/kvm/e500mc.c
138
mtspr(SPRN_GSPRG3, (unsigned long)vcpu->arch.shared->sprg3);
arch/powerpc/kvm/e500mc.c
140
mtspr(SPRN_GSRR0, vcpu->arch.shared->srr0);
arch/powerpc/kvm/e500mc.c
141
mtspr(SPRN_GSRR1, vcpu->arch.shared->srr1);
arch/powerpc/kvm/e500mc.c
143
mtspr(SPRN_GEPR, vcpu->arch.epr);
arch/powerpc/kvm/e500mc.c
144
mtspr(SPRN_GDEAR, vcpu->arch.shared->dar);
arch/powerpc/kvm/e500mc.c
145
mtspr(SPRN_GESR, vcpu->arch.shared->esr);
arch/powerpc/kvm/e500mc.c
147
if (vcpu->arch.oldpir != mfspr(SPRN_PIR) ||
arch/powerpc/kvm/e500mc.c
156
vcpu->arch.eplc = mfspr(SPRN_EPLC);
arch/powerpc/kvm/e500mc.c
157
vcpu->arch.epsc = mfspr(SPRN_EPSC);
arch/powerpc/kvm/e500mc.c
159
vcpu->arch.shared->sprg0 = mfspr(SPRN_GSPRG0);
arch/powerpc/kvm/e500mc.c
160
vcpu->arch.shared->sprg1 = mfspr(SPRN_GSPRG1);
arch/powerpc/kvm/e500mc.c
161
vcpu->arch.shared->sprg2 = mfspr(SPRN_GSPRG2);
arch/powerpc/kvm/e500mc.c
162
vcpu->arch.shared->sprg3 = mfspr(SPRN_GSPRG3);
arch/powerpc/kvm/e500mc.c
164
vcpu->arch.shared->srr0 = mfspr(SPRN_GSRR0);
arch/powerpc/kvm/e500mc.c
165
vcpu->arch.shared->srr1 = mfspr(SPRN_GSRR1);
arch/powerpc/kvm/e500mc.c
167
vcpu->arch.epr = mfspr(SPRN_GEPR);
arch/powerpc/kvm/e500mc.c
168
vcpu->arch.shared->dar = mfspr(SPRN_GDEAR);
arch/powerpc/kvm/e500mc.c
169
vcpu->arch.shared->esr = mfspr(SPRN_GESR);
arch/powerpc/kvm/e500mc.c
171
vcpu->arch.oldpir = mfspr(SPRN_PIR);
arch/powerpc/kvm/e500mc.c
204
vcpu->arch.shadow_epcr = SPRN_EPCR_DSIGS | SPRN_EPCR_DGTMI | \
arch/powerpc/kvm/e500mc.c
207
vcpu->arch.shadow_epcr |= SPRN_EPCR_ICM;
arch/powerpc/kvm/e500mc.c
209
vcpu->arch.shadow_msrp = MSRP_UCLEP | MSRP_PMMP;
arch/powerpc/kvm/e500mc.c
211
vcpu->arch.pvr = mfspr(SPRN_PVR);
arch/powerpc/kvm/e500mc.c
214
vcpu->arch.cpu_type = KVM_CPU_E500MC;
arch/powerpc/kvm/e500mc.c
236
vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
arch/powerpc/kvm/e500mc.c
237
sregs->u.e.ivor_high[4] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
arch/powerpc/kvm/e500mc.c
238
sregs->u.e.ivor_high[5] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
arch/powerpc/kvm/e500mc.c
263
vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] =
arch/powerpc/kvm/e500mc.c
268
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] =
arch/powerpc/kvm/e500mc.c
270
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] =
arch/powerpc/kvm/e500mc.c
284
*val = get_reg_val(id, vcpu->arch.sprg9);
arch/powerpc/kvm/e500mc.c
300
vcpu->arch.sprg9 = set_reg_val(id, *val);
arch/powerpc/kvm/e500mc.c
318
vcpu->arch.oldpir = 0xffffffff;
arch/powerpc/kvm/e500mc.c
324
vcpu->arch.shared = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
arch/powerpc/kvm/e500mc.c
325
if (!vcpu->arch.shared) {
arch/powerpc/kvm/e500mc.c
341
free_page((unsigned long)vcpu->arch.shared);
arch/powerpc/kvm/e500mc.c
361
kvm->arch.lpid = lpid;
arch/powerpc/kvm/e500mc.c
367
int lpid = kvm->arch.lpid;
arch/powerpc/kvm/emulate.c
115
emulated = vcpu->kvm->arch.kvm_ops->emulate_mtspr(vcpu, sprn,
arch/powerpc/kvm/emulate.c
141
spr_val = vcpu->arch.pvr;
arch/powerpc/kvm/emulate.c
176
emulated = vcpu->kvm->arch.kvm_ops->emulate_mfspr(vcpu, sprn,
arch/powerpc/kvm/emulate.c
223
vcpu->arch.shared->esr | ESR_PTR);
arch/powerpc/kvm/emulate.c
239
vcpu->arch.shared->esr | ESR_PTR);
arch/powerpc/kvm/emulate.c
276
vcpu->run->debug.arch.status = 0;
arch/powerpc/kvm/emulate.c
277
vcpu->run->debug.arch.address = kvmppc_get_pc(vcpu);
arch/powerpc/kvm/emulate.c
290
emulated = vcpu->kvm->arch.kvm_ops->emulate_op(vcpu, inst,
arch/powerpc/kvm/emulate.c
31
pr_debug("mtDEC: %lx\n", vcpu->arch.dec);
arch/powerpc/kvm/emulate.c
32
hrtimer_try_to_cancel(&vcpu->arch.dec_timer);
arch/powerpc/kvm/emulate.c
41
if (vcpu->arch.dec == 0)
arch/powerpc/kvm/emulate.c
51
dec_time = vcpu->arch.dec;
arch/powerpc/kvm/emulate.c
58
hrtimer_start(&vcpu->arch.dec_timer,
arch/powerpc/kvm/emulate.c
60
vcpu->arch.dec_jiffies = get_tb();
arch/powerpc/kvm/emulate.c
65
u64 jd = tb - vcpu->arch.dec_jiffies;
arch/powerpc/kvm/emulate.c
68
if (vcpu->arch.dec < jd)
arch/powerpc/kvm/emulate.c
72
return vcpu->arch.dec - jd;
arch/powerpc/kvm/emulate.c
94
vcpu->arch.dec = (u32) spr_val;
arch/powerpc/kvm/emulate_loadstore.c
114
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
124
vcpu->arch.mmio_sp64_extend = 1;
arch/powerpc/kvm/emulate_loadstore.c
134
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
144
vcpu->arch.vaddr_accessed &= ~((unsigned long)size - 1);
arch/powerpc/kvm/emulate_loadstore.c
145
vcpu->arch.paddr_accessed &= ~((unsigned long)size - 1);
arch/powerpc/kvm/emulate_loadstore.c
148
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
151
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
154
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
157
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
162
vcpu->arch.mmio_vmx_offset =
arch/powerpc/kvm/emulate_loadstore.c
163
(vcpu->arch.vaddr_accessed & 0xf)/size;
arch/powerpc/kvm/emulate_loadstore.c
166
vcpu->arch.mmio_vmx_copy_nums = 2;
arch/powerpc/kvm/emulate_loadstore.c
171
vcpu->arch.mmio_vmx_copy_nums = 1;
arch/powerpc/kvm/emulate_loadstore.c
191
vcpu->arch.mmio_sp64_extend = 1;
arch/powerpc/kvm/emulate_loadstore.c
195
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
198
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
202
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
205
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
212
vcpu->arch.mmio_vsx_copy_nums = 1;
arch/powerpc/kvm/emulate_loadstore.c
215
vcpu->arch.mmio_vsx_copy_nums =
arch/powerpc/kvm/emulate_loadstore.c
233
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
246
if (vcpu->kvm->arch.kvm_ops->giveup_ext)
arch/powerpc/kvm/emulate_loadstore.c
247
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu,
arch/powerpc/kvm/emulate_loadstore.c
251
vcpu->arch.mmio_sp64_extend = 1;
arch/powerpc/kvm/emulate_loadstore.c
257
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
267
vcpu->arch.vaddr_accessed &= ~((unsigned long)size - 1);
arch/powerpc/kvm/emulate_loadstore.c
268
vcpu->arch.paddr_accessed &= ~((unsigned long)size - 1);
arch/powerpc/kvm/emulate_loadstore.c
270
if (vcpu->kvm->arch.kvm_ops->giveup_ext)
arch/powerpc/kvm/emulate_loadstore.c
271
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu,
arch/powerpc/kvm/emulate_loadstore.c
274
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
277
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
280
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
283
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
288
vcpu->arch.mmio_vmx_offset =
arch/powerpc/kvm/emulate_loadstore.c
289
(vcpu->arch.vaddr_accessed & 0xf)/size;
arch/powerpc/kvm/emulate_loadstore.c
292
vcpu->arch.mmio_vmx_copy_nums = 2;
arch/powerpc/kvm/emulate_loadstore.c
296
vcpu->arch.mmio_vmx_copy_nums = 1;
arch/powerpc/kvm/emulate_loadstore.c
315
if (vcpu->kvm->arch.kvm_ops->giveup_ext)
arch/powerpc/kvm/emulate_loadstore.c
316
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu,
arch/powerpc/kvm/emulate_loadstore.c
320
vcpu->arch.mmio_sp64_extend = 1;
arch/powerpc/kvm/emulate_loadstore.c
323
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
326
vcpu->arch.mmio_copy_type =
arch/powerpc/kvm/emulate_loadstore.c
333
vcpu->arch.mmio_vsx_copy_nums = 1;
arch/powerpc/kvm/emulate_loadstore.c
336
vcpu->arch.mmio_vsx_copy_nums =
arch/powerpc/kvm/emulate_loadstore.c
85
vcpu->arch.mmio_vsx_copy_nums = 0;
arch/powerpc/kvm/emulate_loadstore.c
86
vcpu->arch.mmio_vsx_offset = 0;
arch/powerpc/kvm/emulate_loadstore.c
87
vcpu->arch.mmio_copy_type = KVMPPC_VSX_COPY_NONE;
arch/powerpc/kvm/emulate_loadstore.c
88
vcpu->arch.mmio_sp64_extend = 0;
arch/powerpc/kvm/emulate_loadstore.c
89
vcpu->arch.mmio_sign_extend = 0;
arch/powerpc/kvm/emulate_loadstore.c
90
vcpu->arch.mmio_vmx_copy_nums = 0;
arch/powerpc/kvm/emulate_loadstore.c
91
vcpu->arch.mmio_vmx_offset = 0;
arch/powerpc/kvm/emulate_loadstore.c
92
vcpu->arch.mmio_host_swabbed = 0;
arch/powerpc/kvm/emulate_loadstore.c
95
vcpu->arch.regs.msr = kvmppc_get_msr(vcpu);
arch/powerpc/kvm/emulate_loadstore.c
96
if (analyse_instr(&op, &vcpu->arch.regs, inst) == 0) {
arch/powerpc/kvm/mpic.c
117
return vcpu ? vcpu->arch.irq_cpu_id : -1;
arch/powerpc/kvm/mpic.c
1178
struct openpic *opp = vcpu->arch.mpic;
arch/powerpc/kvm/mpic.c
1179
int cpu = vcpu->arch.irq_cpu_id;
arch/powerpc/kvm/mpic.c
1635
dev->kvm->arch.mpic = NULL;
arch/powerpc/kvm/mpic.c
1661
if (dev->kvm->arch.mpic)
arch/powerpc/kvm/mpic.c
1714
dev->kvm->arch.mpic = opp;
arch/powerpc/kvm/mpic.c
1751
if (vcpu->arch.irq_type) {
arch/powerpc/kvm/mpic.c
1759
vcpu->arch.mpic = opp;
arch/powerpc/kvm/mpic.c
1760
vcpu->arch.irq_cpu_id = cpu;
arch/powerpc/kvm/mpic.c
1761
vcpu->arch.irq_type = KVMPPC_IRQ_MPIC;
arch/powerpc/kvm/mpic.c
1765
vcpu->arch.epr_flags |= KVMPPC_EPR_KERNEL;
arch/powerpc/kvm/mpic.c
1779
BUG_ON(!opp->dst[vcpu->arch.irq_cpu_id].vcpu);
arch/powerpc/kvm/mpic.c
1781
opp->dst[vcpu->arch.irq_cpu_id].vcpu = NULL;
arch/powerpc/kvm/mpic.c
1795
struct openpic *opp = kvm->arch.mpic;
arch/powerpc/kvm/mpic.c
1809
struct openpic *opp = kvm->arch.mpic;
arch/powerpc/kvm/mpic.c
1818
openpic_msi_write(kvm->arch.mpic, MSIIR_OFFSET, e->msi.data);
arch/powerpc/kvm/mpic.c
254
pr_debug("%s: cpu %d output %d\n", __func__, dst->vcpu->arch.irq_cpu_id,
arch/powerpc/kvm/mpic.c
272
pr_debug("%s: cpu %d output %d\n", __func__, dst->vcpu->arch.irq_cpu_id,
arch/powerpc/kvm/powerpc.c
1037
vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1038
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/powerpc.c
1053
vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1054
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/powerpc.c
1069
vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1070
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/powerpc.c
1085
vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1086
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/powerpc.c
1135
if (!vcpu->arch.mmio_host_swabbed) {
arch/powerpc/kvm/powerpc.c
1152
if ((vcpu->arch.mmio_sp64_extend) && (run->mmio.len == 4))
arch/powerpc/kvm/powerpc.c
1155
if (vcpu->arch.mmio_sign_extend) {
arch/powerpc/kvm/powerpc.c
1171
switch (vcpu->arch.io_gpr & KVM_MMIO_REG_EXT_MASK) {
arch/powerpc/kvm/powerpc.c
1173
kvmppc_set_gpr(vcpu, vcpu->arch.io_gpr, gpr);
arch/powerpc/kvm/powerpc.c
1176
if (vcpu->kvm->arch.kvm_ops->giveup_ext)
arch/powerpc/kvm/powerpc.c
1177
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_FP);
arch/powerpc/kvm/powerpc.c
1179
kvmppc_set_fpr(vcpu, vcpu->arch.io_gpr & KVM_MMIO_REG_MASK, gpr);
arch/powerpc/kvm/powerpc.c
1183
vcpu->arch.qpr[vcpu->arch.io_gpr & KVM_MMIO_REG_MASK] = gpr;
arch/powerpc/kvm/powerpc.c
1186
kvmppc_set_fpr(vcpu, vcpu->arch.io_gpr & KVM_MMIO_REG_MASK, gpr);
arch/powerpc/kvm/powerpc.c
1187
vcpu->arch.qpr[vcpu->arch.io_gpr & KVM_MMIO_REG_MASK] = gpr;
arch/powerpc/kvm/powerpc.c
1192
if (vcpu->kvm->arch.kvm_ops->giveup_ext)
arch/powerpc/kvm/powerpc.c
1193
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VSX);
arch/powerpc/kvm/powerpc.c
1195
if (vcpu->arch.mmio_copy_type == KVMPPC_VSX_COPY_DWORD)
arch/powerpc/kvm/powerpc.c
1197
else if (vcpu->arch.mmio_copy_type == KVMPPC_VSX_COPY_WORD)
arch/powerpc/kvm/powerpc.c
1199
else if (vcpu->arch.mmio_copy_type ==
arch/powerpc/kvm/powerpc.c
1202
else if (vcpu->arch.mmio_copy_type ==
arch/powerpc/kvm/powerpc.c
1209
if (vcpu->kvm->arch.kvm_ops->giveup_ext)
arch/powerpc/kvm/powerpc.c
1210
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VEC);
arch/powerpc/kvm/powerpc.c
1212
if (vcpu->arch.mmio_copy_type == KVMPPC_VMX_COPY_DWORD)
arch/powerpc/kvm/powerpc.c
1214
else if (vcpu->arch.mmio_copy_type == KVMPPC_VMX_COPY_WORD)
arch/powerpc/kvm/powerpc.c
1216
else if (vcpu->arch.mmio_copy_type ==
arch/powerpc/kvm/powerpc.c
1219
else if (vcpu->arch.mmio_copy_type ==
arch/powerpc/kvm/powerpc.c
1228
kvm_vcpu_write_guest(vcpu, vcpu->arch.nested_io_gpr, &gpr,
arch/powerpc/kvm/powerpc.c
1255
run->mmio.phys_addr = vcpu->arch.paddr_accessed;
arch/powerpc/kvm/powerpc.c
1259
vcpu->arch.io_gpr = rt;
arch/powerpc/kvm/powerpc.c
1260
vcpu->arch.mmio_host_swabbed = host_swabbed;
arch/powerpc/kvm/powerpc.c
1263
vcpu->arch.mmio_sign_extend = sign_extend;
arch/powerpc/kvm/powerpc.c
1305
if (vcpu->arch.mmio_vsx_copy_nums > 4)
arch/powerpc/kvm/powerpc.c
1308
while (vcpu->arch.mmio_vsx_copy_nums) {
arch/powerpc/kvm/powerpc.c
1315
vcpu->arch.paddr_accessed += vcpu->run->mmio.len;
arch/powerpc/kvm/powerpc.c
1317
vcpu->arch.mmio_vsx_copy_nums--;
arch/powerpc/kvm/powerpc.c
1318
vcpu->arch.mmio_vsx_offset++;
arch/powerpc/kvm/powerpc.c
1342
run->mmio.phys_addr = vcpu->arch.paddr_accessed;
arch/powerpc/kvm/powerpc.c
1348
if ((vcpu->arch.mmio_sp64_extend) && (bytes == 4))
arch/powerpc/kvm/powerpc.c
1390
int copy_type = vcpu->arch.mmio_copy_type;
arch/powerpc/kvm/powerpc.c
1396
kvmppc_get_vsr_dword_offset(vcpu->arch.mmio_vsx_offset);
arch/powerpc/kvm/powerpc.c
1413
kvmppc_get_vsr_word_offset(vcpu->arch.mmio_vsx_offset);
arch/powerpc/kvm/powerpc.c
144
struct kvm_vcpu_arch_shared *shared = vcpu->arch.shared;
arch/powerpc/kvm/powerpc.c
1445
vcpu->arch.io_gpr = rs;
arch/powerpc/kvm/powerpc.c
1448
if (vcpu->arch.mmio_vsx_copy_nums > 4)
arch/powerpc/kvm/powerpc.c
1451
while (vcpu->arch.mmio_vsx_copy_nums) {
arch/powerpc/kvm/powerpc.c
1461
vcpu->arch.paddr_accessed += vcpu->run->mmio.len;
arch/powerpc/kvm/powerpc.c
1463
vcpu->arch.mmio_vsx_copy_nums--;
arch/powerpc/kvm/powerpc.c
1464
vcpu->arch.mmio_vsx_offset++;
arch/powerpc/kvm/powerpc.c
1476
vcpu->arch.paddr_accessed += run->mmio.len;
arch/powerpc/kvm/powerpc.c
1479
emulated = kvmppc_handle_vsx_load(vcpu, vcpu->arch.io_gpr,
arch/powerpc/kvm/powerpc.c
1480
run->mmio.len, 1, vcpu->arch.mmio_sign_extend);
arch/powerpc/kvm/powerpc.c
1483
vcpu->arch.io_gpr, run->mmio.len, 1);
arch/powerpc/kvm/powerpc.c
1511
if (vcpu->arch.mmio_vmx_copy_nums > 2)
arch/powerpc/kvm/powerpc.c
1514
while (vcpu->arch.mmio_vmx_copy_nums) {
arch/powerpc/kvm/powerpc.c
1521
vcpu->arch.paddr_accessed += vcpu->run->mmio.len;
arch/powerpc/kvm/powerpc.c
1522
vcpu->arch.mmio_vmx_copy_nums--;
arch/powerpc/kvm/powerpc.c
1523
vcpu->arch.mmio_vmx_offset++;
arch/powerpc/kvm/powerpc.c
1536
kvmppc_get_vmx_dword_offset(vcpu, vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1554
kvmppc_get_vmx_word_offset(vcpu, vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1572
kvmppc_get_vmx_hword_offset(vcpu, vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1590
kvmppc_get_vmx_byte_offset(vcpu, vcpu->arch.mmio_vmx_offset);
arch/powerpc/kvm/powerpc.c
1608
if (vcpu->arch.mmio_vmx_copy_nums > 2)
arch/powerpc/kvm/powerpc.c
1611
vcpu->arch.io_gpr = rs;
arch/powerpc/kvm/powerpc.c
1613
while (vcpu->arch.mmio_vmx_copy_nums) {
arch/powerpc/kvm/powerpc.c
1614
switch (vcpu->arch.mmio_copy_type) {
arch/powerpc/kvm/powerpc.c
1641
vcpu->arch.paddr_accessed += vcpu->run->mmio.len;
arch/powerpc/kvm/powerpc.c
1642
vcpu->arch.mmio_vmx_copy_nums--;
arch/powerpc/kvm/powerpc.c
1643
vcpu->arch.mmio_vmx_offset++;
arch/powerpc/kvm/powerpc.c
1655
vcpu->arch.paddr_accessed += run->mmio.len;
arch/powerpc/kvm/powerpc.c
1659
vcpu->arch.io_gpr, run->mmio.len, 1);
arch/powerpc/kvm/powerpc.c
1662
vcpu->arch.io_gpr, run->mmio.len, 1);
arch/powerpc/kvm/powerpc.c
1793
if (vcpu->arch.mmio_vsx_copy_nums > 0) {
arch/powerpc/kvm/powerpc.c
1794
vcpu->arch.mmio_vsx_copy_nums--;
arch/powerpc/kvm/powerpc.c
1795
vcpu->arch.mmio_vsx_offset++;
arch/powerpc/kvm/powerpc.c
1798
if (vcpu->arch.mmio_vsx_copy_nums > 0) {
arch/powerpc/kvm/powerpc.c
1807
if (vcpu->arch.mmio_vmx_copy_nums > 0) {
arch/powerpc/kvm/powerpc.c
1808
vcpu->arch.mmio_vmx_copy_nums--;
arch/powerpc/kvm/powerpc.c
1809
vcpu->arch.mmio_vmx_offset++;
arch/powerpc/kvm/powerpc.c
1812
if (vcpu->arch.mmio_vmx_copy_nums > 0) {
arch/powerpc/kvm/powerpc.c
1820
} else if (vcpu->arch.osi_needed) {
arch/powerpc/kvm/powerpc.c
1826
vcpu->arch.osi_needed = 0;
arch/powerpc/kvm/powerpc.c
1827
} else if (vcpu->arch.hcall_needed) {
arch/powerpc/kvm/powerpc.c
1833
vcpu->arch.hcall_needed = 0;
arch/powerpc/kvm/powerpc.c
1835
} else if (vcpu->arch.epr_needed) {
arch/powerpc/kvm/powerpc.c
1837
vcpu->arch.epr_needed = 0;
arch/powerpc/kvm/powerpc.c
186
if (vcpu->arch.intr_msr & MSR_LE)
arch/powerpc/kvm/powerpc.c
188
if (shared_big_endian != vcpu->arch.shared_big_endian)
arch/powerpc/kvm/powerpc.c
1890
vcpu->arch.osi_enabled = true;
arch/powerpc/kvm/powerpc.c
1894
vcpu->arch.papr_enabled = true;
arch/powerpc/kvm/powerpc.c
1899
vcpu->arch.epr_flags |= KVMPPC_EPR_USER;
arch/powerpc/kvm/powerpc.c
190
vcpu->arch.shared_big_endian = shared_big_endian;
arch/powerpc/kvm/powerpc.c
1901
vcpu->arch.epr_flags &= ~KVMPPC_EPR_USER;
arch/powerpc/kvm/powerpc.c
1906
vcpu->arch.watchdog_enabled = true;
arch/powerpc/kvm/powerpc.c
1986
vcpu->kvm->arch.fwnmi_enabled = true;
arch/powerpc/kvm/powerpc.c
199
vcpu->arch.disable_kernel_nx = true;
arch/powerpc/kvm/powerpc.c
2003
if (kvm->arch.mpic)
arch/powerpc/kvm/powerpc.c
2007
if (kvm->arch.xics || kvm->arch.xive)
arch/powerpc/kvm/powerpc.c
203
vcpu->arch.magic_page_pa = param1 & ~0xfffULL;
arch/powerpc/kvm/powerpc.c
204
vcpu->arch.magic_page_ea = param2 & ~0xfffULL;
arch/powerpc/kvm/powerpc.c
211
if ((vcpu->arch.magic_page_pa & 0xf000) !=
arch/powerpc/kvm/powerpc.c
212
((ulong)vcpu->arch.shared & 0xf000)) {
arch/powerpc/kvm/powerpc.c
213
void *old_shared = vcpu->arch.shared;
arch/powerpc/kvm/powerpc.c
2139
ret = ret || (kvm->arch.mpic != NULL);
arch/powerpc/kvm/powerpc.c
214
ulong shared = (ulong)vcpu->arch.shared;
arch/powerpc/kvm/powerpc.c
2142
ret = ret || (kvm->arch.xics != NULL);
arch/powerpc/kvm/powerpc.c
2143
ret = ret || (kvm->arch.xive != NULL);
arch/powerpc/kvm/powerpc.c
218
shared |= vcpu->arch.magic_page_pa & 0xf000;
arch/powerpc/kvm/powerpc.c
2182
set_bit(hcall / 4, kvm->arch.enabled_hcalls);
arch/powerpc/kvm/powerpc.c
2184
clear_bit(hcall / 4, kvm->arch.enabled_hcalls);
arch/powerpc/kvm/powerpc.c
2193
if (kvm->arch.kvm_ops->set_smt_mode)
arch/powerpc/kvm/powerpc.c
2194
r = kvm->arch.kvm_ops->set_smt_mode(kvm, mode, flags);
arch/powerpc/kvm/powerpc.c
2201
!kvm->arch.kvm_ops->enable_nested)
arch/powerpc/kvm/powerpc.c
2203
r = kvm->arch.kvm_ops->enable_nested(kvm);
arch/powerpc/kvm/powerpc.c
2209
if (!is_kvmppc_hv_enabled(kvm) || !kvm->arch.kvm_ops->enable_svm)
arch/powerpc/kvm/powerpc.c
221
vcpu->arch.shared = new_shared;
arch/powerpc/kvm/powerpc.c
2211
r = kvm->arch.kvm_ops->enable_svm(kvm);
arch/powerpc/kvm/powerpc.c
2215
if (!is_kvmppc_hv_enabled(kvm) || !kvm->arch.kvm_ops->enable_dawr1)
arch/powerpc/kvm/powerpc.c
2217
r = kvm->arch.kvm_ops->enable_dawr1(kvm);
arch/powerpc/kvm/powerpc.c
2412
r = kvm->arch.kvm_ops->get_smmu_info(kvm, &info);
arch/powerpc/kvm/powerpc.c
2428
if (!kvm->arch.kvm_ops->configure_mmu)
arch/powerpc/kvm/powerpc.c
2433
r = kvm->arch.kvm_ops->configure_mmu(kvm, &cfg);
arch/powerpc/kvm/powerpc.c
2441
if (!kvm->arch.kvm_ops->get_rmmu_info)
arch/powerpc/kvm/powerpc.c
2443
r = kvm->arch.kvm_ops->get_rmmu_info(kvm, &info);
arch/powerpc/kvm/powerpc.c
2460
if (!kvm->arch.kvm_ops->svm_off)
arch/powerpc/kvm/powerpc.c
2463
r = kvm->arch.kvm_ops->svm_off(kvm);
arch/powerpc/kvm/powerpc.c
2468
r = kvm->arch.kvm_ops->arch_vm_ioctl(filp, ioctl, arg);
arch/powerpc/kvm/powerpc.c
2517
if (vcpu->kvm->arch.kvm_ops->create_vcpu_debugfs)
arch/powerpc/kvm/powerpc.c
2518
vcpu->kvm->arch.kvm_ops->create_vcpu_debugfs(vcpu, debugfs_dentry);
arch/powerpc/kvm/powerpc.c
2523
if (kvm->arch.kvm_ops->create_vm_debugfs)
arch/powerpc/kvm/powerpc.c
2524
kvm->arch.kvm_ops->create_vm_debugfs(kvm);
arch/powerpc/kvm/powerpc.c
258
if (!vcpu->arch.pvr)
arch/powerpc/kvm/powerpc.c
262
if ((vcpu->arch.cpu_type != KVM_CPU_3S_64) && vcpu->arch.papr_enabled)
arch/powerpc/kvm/powerpc.c
266
if (!vcpu->arch.papr_enabled && is_kvmppc_hv_enabled(vcpu->kvm))
arch/powerpc/kvm/powerpc.c
277
vcpu->arch.sane = r;
arch/powerpc/kvm/powerpc.c
326
vcpu->arch.vaddr_accessed, dsisr);
arch/powerpc/kvm/powerpc.c
351
ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK;
arch/powerpc/kvm/powerpc.c
357
if (vcpu->kvm->arch.kvm_ops && vcpu->kvm->arch.kvm_ops->store_to_eaddr)
arch/powerpc/kvm/powerpc.c
358
r = vcpu->kvm->arch.kvm_ops->store_to_eaddr(vcpu, eaddr, ptr,
arch/powerpc/kvm/powerpc.c
378
void *magic = vcpu->arch.shared;
arch/powerpc/kvm/powerpc.c
394
ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK;
arch/powerpc/kvm/powerpc.c
400
if (vcpu->kvm->arch.kvm_ops && vcpu->kvm->arch.kvm_ops->load_from_eaddr)
arch/powerpc/kvm/powerpc.c
401
rc = vcpu->kvm->arch.kvm_ops->load_from_eaddr(vcpu, eaddr, ptr,
arch/powerpc/kvm/powerpc.c
424
void *magic = vcpu->arch.shared;
arch/powerpc/kvm/powerpc.c
469
kvm->arch.kvm_ops = kvm_ops;
arch/powerpc/kvm/powerpc.c
499
module_put(kvm->arch.kvm_ops->owner);
arch/powerpc/kvm/powerpc.c
52
return !!(v->arch.pending_exceptions) || kvm_request_pending(v);
arch/powerpc/kvm/powerpc.c
591
if (kvm->arch.emul_smt_mode > 1)
arch/powerpc/kvm/powerpc.c
592
r = kvm->arch.emul_smt_mode;
arch/powerpc/kvm/powerpc.c
594
r = kvm->arch.smt_mode;
arch/powerpc/kvm/powerpc.c
750
vcpu = container_of(timer, struct kvm_vcpu, arch.dec_timer);
arch/powerpc/kvm/powerpc.c
760
hrtimer_setup(&vcpu->arch.dec_timer, kvmppc_decrementer_wakeup, CLOCK_REALTIME,
arch/powerpc/kvm/powerpc.c
764
mutex_init(&vcpu->arch.exit_timing_lock);
arch/powerpc/kvm/powerpc.c
774
rcuwait_init(&vcpu->arch.wait);
arch/powerpc/kvm/powerpc.c
775
vcpu->arch.waitp = &vcpu->arch.wait;
arch/powerpc/kvm/powerpc.c
790
hrtimer_cancel(&vcpu->arch.dec_timer);
arch/powerpc/kvm/powerpc.c
792
switch (vcpu->arch.irq_type) {
arch/powerpc/kvm/powerpc.c
794
kvmppc_mpic_disconnect_vcpu(vcpu->arch.mpic, vcpu);
arch/powerpc/kvm/powerpc.c
827
mtspr(SPRN_VRSAVE, vcpu->arch.vrsave);
arch/powerpc/kvm/powerpc.c
836
vcpu->arch.vrsave = mfspr(SPRN_VRSAVE);
arch/powerpc/kvm/powerpc.c
859
if (kvm->arch.kvm_ops->irq_bypass_add_producer)
arch/powerpc/kvm/powerpc.c
860
return kvm->arch.kvm_ops->irq_bypass_add_producer(cons, prod);
arch/powerpc/kvm/powerpc.c
872
if (kvm->arch.kvm_ops->irq_bypass_del_producer)
arch/powerpc/kvm/powerpc.c
873
kvm->arch.kvm_ops->irq_bypass_del_producer(cons, prod);
arch/powerpc/kvm/powerpc.c
912
int offset = kvmppc_get_vsr_dword_offset(vcpu->arch.mmio_vsx_offset);
arch/powerpc/kvm/powerpc.c
913
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/powerpc.c
931
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/powerpc.c
948
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/powerpc.c
968
int offset = kvmppc_get_vsr_word_offset(vcpu->arch.mmio_vsx_offset);
arch/powerpc/kvm/powerpc.c
969
int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
arch/powerpc/kvm/timing.c
139
min = vcpu->arch.timing_min_duration[i];
arch/powerpc/kvm/timing.c
141
max = vcpu->arch.timing_max_duration[i];
arch/powerpc/kvm/timing.c
143
sum = vcpu->arch.timing_sum_duration[i];
arch/powerpc/kvm/timing.c
145
sum_quad = vcpu->arch.timing_sum_quad_duration[i];
arch/powerpc/kvm/timing.c
150
vcpu->arch.timing_count_type[i],
arch/powerpc/kvm/timing.c
27
mutex_lock(&vcpu->arch.exit_timing_lock);
arch/powerpc/kvm/timing.c
29
vcpu->arch.last_exit_type = 0xDEAD;
arch/powerpc/kvm/timing.c
31
vcpu->arch.timing_count_type[i] = 0;
arch/powerpc/kvm/timing.c
32
vcpu->arch.timing_max_duration[i] = 0;
arch/powerpc/kvm/timing.c
33
vcpu->arch.timing_min_duration[i] = 0xFFFFFFFF;
arch/powerpc/kvm/timing.c
34
vcpu->arch.timing_sum_duration[i] = 0;
arch/powerpc/kvm/timing.c
35
vcpu->arch.timing_sum_quad_duration[i] = 0;
arch/powerpc/kvm/timing.c
37
vcpu->arch.timing_last_exit = 0;
arch/powerpc/kvm/timing.c
38
vcpu->arch.timing_exit.tv64 = 0;
arch/powerpc/kvm/timing.c
39
vcpu->arch.timing_last_enter.tv64 = 0;
arch/powerpc/kvm/timing.c
41
mutex_unlock(&vcpu->arch.exit_timing_lock);
arch/powerpc/kvm/timing.c
48
mutex_lock(&vcpu->arch.exit_timing_lock);
arch/powerpc/kvm/timing.c
50
vcpu->arch.timing_count_type[type]++;
arch/powerpc/kvm/timing.c
53
old = vcpu->arch.timing_sum_duration[type];
arch/powerpc/kvm/timing.c
54
vcpu->arch.timing_sum_duration[type] += duration;
arch/powerpc/kvm/timing.c
55
if (unlikely(old > vcpu->arch.timing_sum_duration[type])) {
arch/powerpc/kvm/timing.c
58
__func__, old, vcpu->arch.timing_sum_duration[type],
arch/powerpc/kvm/timing.c
59
type, vcpu->arch.timing_count_type[type]);
arch/powerpc/kvm/timing.c
63
old = vcpu->arch.timing_sum_quad_duration[type];
arch/powerpc/kvm/timing.c
64
vcpu->arch.timing_sum_quad_duration[type] += (duration*duration);
arch/powerpc/kvm/timing.c
65
if (unlikely(old > vcpu->arch.timing_sum_quad_duration[type])) {
arch/powerpc/kvm/timing.c
69
vcpu->arch.timing_sum_quad_duration[type],
arch/powerpc/kvm/timing.c
70
type, vcpu->arch.timing_count_type[type]);
arch/powerpc/kvm/timing.c
74
if (unlikely(duration < vcpu->arch.timing_min_duration[type]))
arch/powerpc/kvm/timing.c
75
vcpu->arch.timing_min_duration[type] = duration;
arch/powerpc/kvm/timing.c
76
if (unlikely(duration > vcpu->arch.timing_max_duration[type]))
arch/powerpc/kvm/timing.c
77
vcpu->arch.timing_max_duration[type] = duration;
arch/powerpc/kvm/timing.c
79
mutex_unlock(&vcpu->arch.exit_timing_lock);
arch/powerpc/kvm/timing.c
84
u64 exit = vcpu->arch.timing_last_exit;
arch/powerpc/kvm/timing.c
85
u64 enter = vcpu->arch.timing_last_enter.tv64;
arch/powerpc/kvm/timing.c
88
vcpu->arch.timing_last_exit = vcpu->arch.timing_exit.tv64;
arch/powerpc/kvm/timing.c
90
if (unlikely(vcpu->arch.last_exit_type == 0xDEAD || exit == 0))
arch/powerpc/kvm/timing.c
94
add_exit_timing(vcpu, (enter - exit), vcpu->arch.last_exit_type);
arch/powerpc/kvm/timing.c
96
add_exit_timing(vcpu, (vcpu->arch.timing_last_exit - enter),
arch/powerpc/kvm/timing.h
22
vcpu->arch.last_exit_type = type;
arch/powerpc/kvm/trace_booke.h
192
__entry->pending = vcpu->arch.pending_exceptions;
arch/powerpc/kvm/trace_booke.h
54
__entry->msr = vcpu->arch.shared->msr;
arch/powerpc/kvm/trace_booke.h
55
__entry->last_inst = vcpu->arch.last_inst;
arch/powerpc/kvm/trace_hv.h
255
__entry->ceded = vcpu->arch.ceded;
arch/powerpc/kvm/trace_hv.h
256
__entry->pending_exceptions = vcpu->arch.pending_exceptions;
arch/powerpc/kvm/trace_hv.h
279
__entry->trap = vcpu->arch.trap;
arch/powerpc/kvm/trace_hv.h
280
__entry->ceded = vcpu->arch.ceded;
arch/powerpc/kvm/trace_hv.h
282
__entry->msr = vcpu->arch.shregs.msr;
arch/powerpc/kvm/trace_hv.h
444
__entry->n_runnable = vcpu->arch.vcore->n_runnable;
arch/powerpc/kvm/trace_hv.h
508
__entry->ret = vcpu->arch.ret;
arch/powerpc/kvm/trace_pr.h
235
__entry->srr1 = vcpu->arch.shadow_srr1;
arch/powerpc/kvm/trace_pr.h
236
__entry->last_inst = vcpu->arch.last_inst;
arch/powerpc/platforms/powernv/vas-trace.h
111
#define TRACE_INCLUDE_PATH ../../arch/powerpc/platforms/powernv
arch/riscv/include/asm/kvm_aia.h
78
#define kvm_riscv_aia_initialized(k) ((k)->arch.aia.initialized)
arch/riscv/include/asm/kvm_aia.h
80
#define irqchip_in_kernel(k) ((k)->arch.aia.in_kernel)
arch/riscv/include/asm/kvm_gstage.h
104
gstage->vmid = READ_ONCE(kvm->arch.vmid.vmid);
arch/riscv/include/asm/kvm_gstage.h
105
gstage->pgd = kvm->arch.pgd;
arch/riscv/include/asm/kvm_gstage.h
106
gstage->pgd_levels = kvm->arch.pgd_levels;
arch/riscv/include/asm/kvm_vcpu_pmu.h
63
#define vcpu_to_pmu(vcpu) (&(vcpu)->arch.pmu_context)
arch/riscv/include/asm/kvm_vcpu_pmu.h
64
#define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu_context))
arch/riscv/include/asm/kvm_vcpu_sbi_fwft.h
32
#define vcpu_to_fwft(vcpu) (&(vcpu)->arch.fwft_context)
arch/riscv/kernel/machine_kexec.c
151
struct kimage_arch *internal = &image->arch;
arch/riscv/kernel/machine_kexec.c
33
struct kimage_arch *internal = &image->arch;
arch/riscv/kernel/machine_kexec_file.c
28
kvfree(image->arch.fdt);
arch/riscv/kernel/machine_kexec_file.c
29
image->arch.fdt = NULL;
arch/riscv/kernel/machine_kexec_file.c
366
image->arch.fdt = fdt;
arch/riscv/kernel/module-sections.c
131
mod->arch.plt.shdr = sechdrs + i;
arch/riscv/kernel/module-sections.c
133
mod->arch.got.shdr = sechdrs + i;
arch/riscv/kernel/module-sections.c
135
mod->arch.got_plt.shdr = sechdrs + i;
arch/riscv/kernel/module-sections.c
138
if (!mod->arch.plt.shdr) {
arch/riscv/kernel/module-sections.c
142
if (!mod->arch.got.shdr) {
arch/riscv/kernel/module-sections.c
146
if (!mod->arch.got_plt.shdr) {
arch/riscv/kernel/module-sections.c
16
struct mod_section *got_sec = &mod->arch.got;
arch/riscv/kernel/module-sections.c
192
mod->arch.plt.shdr->sh_type = SHT_NOBITS;
arch/riscv/kernel/module-sections.c
193
mod->arch.plt.shdr->sh_flags = SHF_EXECINSTR | SHF_ALLOC;
arch/riscv/kernel/module-sections.c
194
mod->arch.plt.shdr->sh_addralign = L1_CACHE_BYTES;
arch/riscv/kernel/module-sections.c
195
mod->arch.plt.shdr->sh_size = (num_plts + 1) * sizeof(struct plt_entry);
arch/riscv/kernel/module-sections.c
196
mod->arch.plt.num_entries = 0;
arch/riscv/kernel/module-sections.c
197
mod->arch.plt.max_entries = num_plts;
arch/riscv/kernel/module-sections.c
199
mod->arch.got.shdr->sh_type = SHT_NOBITS;
arch/riscv/kernel/module-sections.c
200
mod->arch.got.shdr->sh_flags = SHF_ALLOC;
arch/riscv/kernel/module-sections.c
201
mod->arch.got.shdr->sh_addralign = L1_CACHE_BYTES;
arch/riscv/kernel/module-sections.c
202
mod->arch.got.shdr->sh_size = (num_gots + 1) * sizeof(struct got_entry);
arch/riscv/kernel/module-sections.c
203
mod->arch.got.num_entries = 0;
arch/riscv/kernel/module-sections.c
204
mod->arch.got.max_entries = num_gots;
arch/riscv/kernel/module-sections.c
206
mod->arch.got_plt.shdr->sh_type = SHT_NOBITS;
arch/riscv/kernel/module-sections.c
207
mod->arch.got_plt.shdr->sh_flags = SHF_ALLOC;
arch/riscv/kernel/module-sections.c
208
mod->arch.got_plt.shdr->sh_addralign = L1_CACHE_BYTES;
arch/riscv/kernel/module-sections.c
209
mod->arch.got_plt.shdr->sh_size = (num_plts + 1) * sizeof(struct got_entry);
arch/riscv/kernel/module-sections.c
210
mod->arch.got_plt.num_entries = 0;
arch/riscv/kernel/module-sections.c
211
mod->arch.got_plt.max_entries = num_plts;
arch/riscv/kernel/module-sections.c
35
struct mod_section *got_plt_sec = &mod->arch.got_plt;
arch/riscv/kernel/module-sections.c
37
struct mod_section *plt_sec = &mod->arch.plt;
arch/riscv/kvm/aia.c
100
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/aia.c
106
ncsr_write(CSR_HVIPH, vcpu->arch.aia_context.guest_csr.hviph);
arch/riscv/kvm/aia.c
113
struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
arch/riscv/kvm/aia.c
148
struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
arch/riscv/kvm/aia.c
185
struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
arch/riscv/kvm/aia.c
188
if (!riscv_isa_extension_available(vcpu->arch.isa, SSAIA))
arch/riscv/kvm/aia.c
206
struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
arch/riscv/kvm/aia.c
209
if (!riscv_isa_extension_available(vcpu->arch.isa, SSAIA))
arch/riscv/kvm/aia.c
221
WRITE_ONCE(vcpu->arch.irqs_pending_mask[1], 0);
arch/riscv/kvm/aia.c
51
struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
arch/riscv/kvm/aia.c
57
if (READ_ONCE(vcpu->arch.irqs_pending_mask[1])) {
arch/riscv/kvm/aia.c
58
mask = xchg_acquire(&vcpu->arch.irqs_pending_mask[1], 0);
arch/riscv/kvm/aia.c
59
val = READ_ONCE(vcpu->arch.irqs_pending[1]) & mask;
arch/riscv/kvm/aia.c
68
struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
arch/riscv/kvm/aia.c
83
if (READ_ONCE(vcpu->arch.irqs_pending[1]) &
arch/riscv/kvm/aia.c
84
(vcpu->arch.aia_context.guest_csr.vsieh & upper_32_bits(mask)))
arch/riscv/kvm/aia.c
88
seip = vcpu->arch.guest_csr.vsie;
arch/riscv/kvm/aia_aplic.c
250
struct aplic *aplic = kvm->arch.aia.aplic_state;
arch/riscv/kvm/aia_aplic.c
283
struct aplic *aplic = kvm->arch.aia.aplic_state;
arch/riscv/kvm/aia_aplic.c
392
struct aplic *aplic = kvm->arch.aia.aplic_state;
arch/riscv/kvm/aia_aplic.c
450
addr - vcpu->kvm->arch.aia.aplic_addr,
arch/riscv/kvm/aia_aplic.c
457
struct aplic *aplic = kvm->arch.aia.aplic_state;
arch/riscv/kvm/aia_aplic.c
522
addr - vcpu->kvm->arch.aia.aplic_addr,
arch/riscv/kvm/aia_aplic.c
535
if (!kvm->arch.aia.aplic_state)
arch/riscv/kvm/aia_aplic.c
549
if (!kvm->arch.aia.aplic_state)
arch/riscv/kvm/aia_aplic.c
564
if (!kvm->arch.aia.aplic_state)
arch/riscv/kvm/aia_aplic.c
580
if (!kvm->arch.aia.nr_sources)
arch/riscv/kvm/aia_aplic.c
587
kvm->arch.aia.aplic_state = aplic;
arch/riscv/kvm/aia_aplic.c
590
aplic->nr_irqs = kvm->arch.aia.nr_sources + 1;
arch/riscv/kvm/aia_aplic.c
604
kvm->arch.aia.aplic_addr,
arch/riscv/kvm/aia_aplic.c
625
kvm->arch.aia.aplic_state = NULL;
arch/riscv/kvm/aia_aplic.c
632
struct aplic *aplic = kvm->arch.aia.aplic_state;
arch/riscv/kvm/aia_aplic.c
643
kvm->arch.aia.aplic_state = NULL;
arch/riscv/kvm/aia_device.c
147
struct kvm_aia *aia = &kvm->arch.aia;
arch/riscv/kvm/aia_device.c
173
vcpu_aia = &vcpu->arch.aia_context;
arch/riscv/kvm/aia_device.c
232
struct kvm_aia *aia = &kvm->arch.aia;
arch/riscv/kvm/aia_device.c
258
vaia = &vcpu->arch.aia_context;
arch/riscv/kvm/aia_device.c
285
kvm->arch.aia.initialized = true;
arch/riscv/kvm/aia_device.c
34
if (vcpu->arch.ran_atleast_once)
arch/riscv/kvm/aia_device.c
39
kvm->arch.aia.in_kernel = true;
arch/riscv/kvm/aia_device.c
508
struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
arch/riscv/kvm/aia_device.c
524
struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
arch/riscv/kvm/aia_device.c
54
struct kvm_aia *aia = &kvm->arch.aia;
arch/riscv/kvm/aia_device.c
563
if (vcpu->arch.aia_context.hart_index == hart_index)
arch/riscv/kvm/aia_device.c
578
struct kvm_aia *aia = &kvm->arch.aia;
arch/riscv/kvm/aia_device.c
594
ippn = vcpu->arch.aia_context.imsic_addr >>
arch/riscv/kvm/aia_device.c
618
struct kvm_aia *aia = &kvm->arch.aia;
arch/riscv/kvm/aia_imsic.c
1006
imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
1016
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
1031
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
1098
if (!kvm->arch.aia.nr_ids)
arch/riscv/kvm/aia_imsic.c
1105
vcpu->arch.aia_context.imsic_state = imsic;
arch/riscv/kvm/aia_imsic.c
1108
imsic->nr_msis = kvm->arch.aia.nr_ids + 1;
arch/riscv/kvm/aia_imsic.c
1129
vcpu->arch.aia_context.imsic_addr,
arch/riscv/kvm/aia_imsic.c
1142
vcpu->arch.aia_context.imsic_state = NULL;
arch/riscv/kvm/aia_imsic.c
1150
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
1164
vcpu->arch.aia_context.imsic_state = NULL;
arch/riscv/kvm/aia_imsic.c
616
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
640
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
661
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
682
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
722
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
739
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
761
kvm_riscv_mmu_iounmap(vcpu->kvm, vcpu->arch.aia_context.imsic_addr,
arch/riscv/kvm/aia_imsic.c
792
struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
arch/riscv/kvm/aia_imsic.c
797
if (kvm->arch.aia.mode == KVM_DEV_RISCV_AIA_MODE_EMUL)
arch/riscv/kvm/aia_imsic.c
819
if (kvm->arch.aia.mode == KVM_DEV_RISCV_AIA_MODE_HWACCEL) {
arch/riscv/kvm/aia_imsic.c
849
ret = kvm_riscv_mmu_ioremap(kvm, vcpu->arch.aia_context.imsic_addr,
arch/riscv/kvm/aia_imsic.c
889
vcpu->arch.guest_context.hstatus &= ~HSTATUS_VGEIN;
arch/riscv/kvm/aia_imsic.c
891
vcpu->arch.guest_context.hstatus |=
arch/riscv/kvm/aia_imsic.c
909
struct imsic *imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/aia_imsic.c
963
imsic = vcpu->arch.aia_context.imsic_state;
arch/riscv/kvm/mmu.c
174
kvm_riscv_gstage_gpa_size(kvm->arch.pgd_levels) >> PAGE_SHIFT)
arch/riscv/kvm/mmu.c
235
if (!kvm->arch.pgd)
arch/riscv/kvm/mmu.c
255
if (!kvm->arch.pgd)
arch/riscv/kvm/mmu.c
275
if (!kvm->arch.pgd)
arch/riscv/kvm/mmu.c
435
struct kvm_mmu_memory_cache *pcache = &vcpu->arch.mmu_page_cache;
arch/riscv/kvm/mmu.c
448
ret = kvm_mmu_topup_memory_cache(pcache, kvm->arch.pgd_levels);
arch/riscv/kvm/mmu.c
540
if (kvm->arch.pgd != NULL) {
arch/riscv/kvm/mmu.c
549
kvm->arch.pgd = page_to_virt(pgd_page);
arch/riscv/kvm/mmu.c
550
kvm->arch.pgd_phys = page_to_phys(pgd_page);
arch/riscv/kvm/mmu.c
551
kvm->arch.pgd_levels = kvm_riscv_gstage_max_pgd_levels;
arch/riscv/kvm/mmu.c
562
if (kvm->arch.pgd) {
arch/riscv/kvm/mmu.c
565
kvm_riscv_gstage_gpa_size(kvm->arch.pgd_levels), false);
arch/riscv/kvm/mmu.c
566
pgd = READ_ONCE(kvm->arch.pgd);
arch/riscv/kvm/mmu.c
567
kvm->arch.pgd = NULL;
arch/riscv/kvm/mmu.c
568
kvm->arch.pgd_phys = 0;
arch/riscv/kvm/mmu.c
569
kvm->arch.pgd_levels = 0;
arch/riscv/kvm/mmu.c
579
struct kvm_arch *ka = &vcpu->kvm->arch;
arch/riscv/kvm/mmu.c
64
ret = kvm_mmu_topup_memory_cache(&pcache, kvm->arch.pgd_levels);
arch/riscv/kvm/tlb.c
166
vcpu->arch.last_exit_cpu == vcpu->cpu)
arch/riscv/kvm/tlb.c
180
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/tlb.c
199
struct kvm_vmid *v = &vcpu->kvm->arch.vmid;
arch/riscv/kvm/tlb.c
210
struct kvm_vmid *v = &vcpu->kvm->arch.vmid;
arch/riscv/kvm/tlb.c
223
struct kvm_vcpu_arch *varch = &vcpu->arch;
arch/riscv/kvm/tlb.c
248
struct kvm_vcpu_arch *varch = &vcpu->arch;
arch/riscv/kvm/tlb.c
465
PAGE_SHIFT, READ_ONCE(kvm->arch.vmid.vmid));
arch/riscv/kvm/trace.h
25
__entry->pc = vcpu->arch.guest_context.sepc;
arch/riscv/kvm/vcpu.c
107
bitmap_zero(vcpu->arch.irqs_pending, KVM_RISCV_VCPU_NR_IRQS);
arch/riscv/kvm/vcpu.c
108
bitmap_zero(vcpu->arch.irqs_pending_mask, KVM_RISCV_VCPU_NR_IRQS);
arch/riscv/kvm/vcpu.c
112
vcpu->arch.hfence_head = 0;
arch/riscv/kvm/vcpu.c
113
vcpu->arch.hfence_tail = 0;
arch/riscv/kvm/vcpu.c
114
memset(vcpu->arch.hfence_queue, 0, sizeof(vcpu->arch.hfence_queue));
arch/riscv/kvm/vcpu.c
133
spin_lock_init(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu.c
136
vcpu->arch.ran_atleast_once = false;
arch/riscv/kvm/vcpu.c
138
vcpu->arch.mmu_page_cache.gfp_zero = __GFP_ZERO;
arch/riscv/kvm/vcpu.c
139
bitmap_zero(vcpu->arch.isa, RISCV_ISA_EXT_MAX);
arch/riscv/kvm/vcpu.c
148
vcpu->arch.mvendorid = sbi_get_mvendorid();
arch/riscv/kvm/vcpu.c
149
vcpu->arch.marchid = sbi_get_marchid();
arch/riscv/kvm/vcpu.c
150
vcpu->arch.mimpid = sbi_get_mimpid();
arch/riscv/kvm/vcpu.c
153
spin_lock_init(&vcpu->arch.hfence_lock);
arch/riscv/kvm/vcpu.c
155
spin_lock_init(&vcpu->arch.reset_state.lock);
arch/riscv/kvm/vcpu.c
206
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
arch/riscv/kvm/vcpu.c
220
!kvm_riscv_vcpu_stopped(vcpu) && !vcpu->arch.pause);
arch/riscv/kvm/vcpu.c
230
return (vcpu->arch.guest_context.sstatus & SR_SPP) ? true : false;
arch/riscv/kvm/vcpu.c
236
return vcpu->arch.guest_context.sepc;
arch/riscv/kvm/vcpu.c
353
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
356
if (READ_ONCE(vcpu->arch.irqs_pending_mask[0])) {
arch/riscv/kvm/vcpu.c
357
mask = xchg_acquire(&vcpu->arch.irqs_pending_mask[0], 0);
arch/riscv/kvm/vcpu.c
358
val = READ_ONCE(vcpu->arch.irqs_pending[0]) & mask;
arch/riscv/kvm/vcpu.c
371
struct kvm_vcpu_arch *v = &vcpu->arch;
arch/riscv/kvm/vcpu.c
372
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
419
set_bit(irq, vcpu->arch.irqs_pending);
arch/riscv/kvm/vcpu.c
421
set_bit(irq, vcpu->arch.irqs_pending_mask);
arch/riscv/kvm/vcpu.c
442
clear_bit(irq, vcpu->arch.irqs_pending);
arch/riscv/kvm/vcpu.c
444
set_bit(irq, vcpu->arch.irqs_pending_mask);
arch/riscv/kvm/vcpu.c
453
ie = ((vcpu->arch.guest_csr.vsie & VSIP_VALID_MASK)
arch/riscv/kvm/vcpu.c
455
ie |= vcpu->arch.guest_csr.vsie & ~IRQ_LOCAL_MASK &
arch/riscv/kvm/vcpu.c
457
if (READ_ONCE(vcpu->arch.irqs_pending[0]) & ie)
arch/riscv/kvm/vcpu.c
466
WRITE_ONCE(vcpu->arch.mp_state.mp_state, KVM_MP_STATE_STOPPED);
arch/riscv/kvm/vcpu.c
473
spin_lock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu.c
475
spin_unlock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu.c
480
WRITE_ONCE(vcpu->arch.mp_state.mp_state, KVM_MP_STATE_RUNNABLE);
arch/riscv/kvm/vcpu.c
486
spin_lock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu.c
488
spin_unlock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu.c
493
return READ_ONCE(vcpu->arch.mp_state.mp_state) == KVM_MP_STATE_STOPPED;
arch/riscv/kvm/vcpu.c
499
*mp_state = READ_ONCE(vcpu->arch.mp_state);
arch/riscv/kvm/vcpu.c
509
spin_lock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu.c
513
WRITE_ONCE(vcpu->arch.mp_state, *mp_state);
arch/riscv/kvm/vcpu.c
519
if (vcpu->kvm->arch.mp_state_reset)
arch/riscv/kvm/vcpu.c
528
spin_unlock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu.c
547
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
560
else if (vcpu->arch.last_exit_cpu == cpu && !vcpu->arch.csr_dirty)
arch/riscv/kvm/vcpu.c
563
vcpu->arch.csr_dirty = false;
arch/riscv/kvm/vcpu.c
59
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
60
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu.c
602
kvm_riscv_vcpu_host_fp_save(&vcpu->arch.host_context);
arch/riscv/kvm/vcpu.c
603
kvm_riscv_vcpu_guest_fp_restore(&vcpu->arch.guest_context,
arch/riscv/kvm/vcpu.c
604
vcpu->arch.isa);
arch/riscv/kvm/vcpu.c
605
kvm_riscv_vcpu_host_vector_save(&vcpu->arch.host_context);
arch/riscv/kvm/vcpu.c
606
kvm_riscv_vcpu_guest_vector_restore(&vcpu->arch.guest_context,
arch/riscv/kvm/vcpu.c
607
vcpu->arch.isa);
arch/riscv/kvm/vcpu.c
617
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
623
kvm_riscv_vcpu_guest_fp_save(&vcpu->arch.guest_context,
arch/riscv/kvm/vcpu.c
624
vcpu->arch.isa);
arch/riscv/kvm/vcpu.c
625
kvm_riscv_vcpu_host_fp_restore(&vcpu->arch.host_context);
arch/riscv/kvm/vcpu.c
628
kvm_riscv_vcpu_guest_vector_save(&vcpu->arch.guest_context,
arch/riscv/kvm/vcpu.c
629
vcpu->arch.isa);
arch/riscv/kvm/vcpu.c
630
kvm_riscv_vcpu_host_vector_restore(&vcpu->arch.host_context);
arch/riscv/kvm/vcpu.c
65
memset(&vcpu->arch.smstateen_csr, 0, sizeof(vcpu->arch.smstateen_csr));
arch/riscv/kvm/vcpu.c
671
(!kvm_riscv_vcpu_stopped(vcpu)) && (!vcpu->arch.pause),
arch/riscv/kvm/vcpu.c
675
if (kvm_riscv_vcpu_stopped(vcpu) || vcpu->arch.pause) {
arch/riscv/kvm/vcpu.c
714
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
722
struct kvm_vcpu_smstateen_csr *smcsr = &vcpu->arch.smstateen_csr;
arch/riscv/kvm/vcpu.c
723
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
725
vcpu->arch.host_scounteren = csr_swap(CSR_SCOUNTEREN, csr->scounteren);
arch/riscv/kvm/vcpu.c
726
vcpu->arch.host_senvcfg = csr_swap(CSR_SENVCFG, csr->senvcfg);
arch/riscv/kvm/vcpu.c
728
vcpu->arch.host_sstateen0 = csr_swap(CSR_SSTATEEN0, smcsr->sstateen0);
arch/riscv/kvm/vcpu.c
733
struct kvm_vcpu_smstateen_csr *smcsr = &vcpu->arch.smstateen_csr;
arch/riscv/kvm/vcpu.c
734
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu.c
736
csr->scounteren = csr_swap(CSR_SCOUNTEREN, vcpu->arch.host_scounteren);
arch/riscv/kvm/vcpu.c
737
csr->senvcfg = csr_swap(CSR_SENVCFG, vcpu->arch.host_senvcfg);
arch/riscv/kvm/vcpu.c
739
smcsr->sstateen0 = csr_swap(CSR_SSTATEEN0, vcpu->arch.host_sstateen0);
arch/riscv/kvm/vcpu.c
753
struct kvm_cpu_context *gcntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu.c
754
struct kvm_cpu_context *hcntx = &vcpu->arch.host_context;
arch/riscv/kvm/vcpu.c
793
__kvm_riscv_nacl_switch_to(&vcpu->arch, SBI_EXT_NACL,
arch/riscv/kvm/vcpu.c
812
__kvm_riscv_switch_to(&vcpu->arch);
arch/riscv/kvm/vcpu.c
824
vcpu->arch.last_exit_cpu = vcpu->cpu;
arch/riscv/kvm/vcpu.c
835
if (!vcpu->arch.ran_atleast_once)
arch/riscv/kvm/vcpu.c
839
vcpu->arch.ran_atleast_once = true;
arch/riscv/kvm/vcpu.c
920
if (kvm_riscv_gstage_vmid_ver_changed(&vcpu->kvm->arch.vmid) ||
arch/riscv/kvm/vcpu.c
95
vcpu->arch.last_exit_cpu = -1;
arch/riscv/kvm/vcpu_config.c
23
vcpu->arch.cfg.hedeleg = KVM_HEDELEG_DEFAULT;
arch/riscv/kvm/vcpu_config.c
24
vcpu->arch.cfg.hideleg = KVM_HIDELEG_DEFAULT;
arch/riscv/kvm/vcpu_config.c
29
struct kvm_vcpu_config *cfg = &vcpu->arch.cfg;
arch/riscv/kvm/vcpu_config.c
36
vcpu->arch.csr_dirty = true;
arch/riscv/kvm/vcpu_config.c
41
const unsigned long *isa = vcpu->arch.isa;
arch/riscv/kvm/vcpu_config.c
42
struct kvm_vcpu_config *cfg = &vcpu->arch.cfg;
arch/riscv/kvm/vcpu_config.c
76
struct kvm_vcpu_config *cfg = &vcpu->arch.cfg;
arch/riscv/kvm/vcpu_exit.c
146
if (vcpu->arch.guest_context.sstatus & SR_SPP)
arch/riscv/kvm/vcpu_exit.c
166
vcpu->arch.guest_context.sepc = ncsr_read(CSR_VSTVEC);
arch/riscv/kvm/vcpu_exit.c
169
vcpu->arch.guest_context.sstatus |= SR_SPP;
arch/riscv/kvm/vcpu_exit.c
176
if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) {
arch/riscv/kvm/vcpu_exit.c
229
if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV)
arch/riscv/kvm/vcpu_exit.c
235
if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV)
arch/riscv/kvm/vcpu_exit.c
239
if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV)
arch/riscv/kvm/vcpu_exit.c
254
vcpu->arch.guest_context.sepc,
arch/riscv/kvm/vcpu_exit.c
255
vcpu->arch.guest_context.sstatus,
arch/riscv/kvm/vcpu_exit.c
256
vcpu->arch.guest_context.hstatus);
arch/riscv/kvm/vcpu_exit.c
74
old_hstatus = csr_swap(CSR_HSTATUS, vcpu->arch.guest_context.hstatus);
arch/riscv/kvm/vcpu_fp.c
105
riscv_isa_extension_available(vcpu->arch.isa, d)) {
arch/riscv/kvm/vcpu_fp.c
132
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_fp.c
141
riscv_isa_extension_available(vcpu->arch.isa, f)) {
arch/riscv/kvm/vcpu_fp.c
154
riscv_isa_extension_available(vcpu->arch.isa, d)) {
arch/riscv/kvm/vcpu_fp.c
21
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_fp.c
24
if (riscv_isa_extension_available(vcpu->arch.isa, f) ||
arch/riscv/kvm/vcpu_fp.c
25
riscv_isa_extension_available(vcpu->arch.isa, d))
arch/riscv/kvm/vcpu_fp.c
83
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_fp.c
92
riscv_isa_extension_available(vcpu->arch.isa, f)) {
arch/riscv/kvm/vcpu_insn.c
106
if (!riscv_isa_extension_available(vcpu->arch.isa, ZKR))
arch/riscv/kvm/vcpu_insn.c
131
if (vcpu->arch.csr_decode.return_handled)
arch/riscv/kvm/vcpu_insn.c
133
vcpu->arch.csr_decode.return_handled = 1;
arch/riscv/kvm/vcpu_insn.c
136
insn = vcpu->arch.csr_decode.insn;
arch/riscv/kvm/vcpu_insn.c
138
SET_RD(insn, &vcpu->arch.guest_context,
arch/riscv/kvm/vcpu_insn.c
142
vcpu->arch.guest_context.sepc += INSN_LEN(insn);
arch/riscv/kvm/vcpu_insn.c
152
ulong rs1_val = GET_RS1(insn, &vcpu->arch.guest_context);
arch/riscv/kvm/vcpu_insn.c
187
vcpu->arch.csr_decode.insn = insn;
arch/riscv/kvm/vcpu_insn.c
188
vcpu->arch.csr_decode.return_handled = 0;
arch/riscv/kvm/vcpu_insn.c
292
vcpu->arch.guest_context.sepc += INSN_LEN(insn);
arch/riscv/kvm/vcpu_insn.c
337
ct = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_insn.c
36
utrap.sepc = vcpu->arch.guest_context.sepc;
arch/riscv/kvm/vcpu_insn.c
381
struct kvm_cpu_context *ct = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_insn.c
458
vcpu->arch.mmio_decode.insn = insn;
arch/riscv/kvm/vcpu_insn.c
459
vcpu->arch.mmio_decode.insn_len = insn_len;
arch/riscv/kvm/vcpu_insn.c
460
vcpu->arch.mmio_decode.shift = shift;
arch/riscv/kvm/vcpu_insn.c
461
vcpu->arch.mmio_decode.len = len;
arch/riscv/kvm/vcpu_insn.c
462
vcpu->arch.mmio_decode.return_handled = 0;
arch/riscv/kvm/vcpu_insn.c
509
struct kvm_cpu_context *ct = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_insn.c
52
utrap.sepc = vcpu->arch.guest_context.sepc;
arch/riscv/kvm/vcpu_insn.c
537
data = GET_RS2(insn, &vcpu->arch.guest_context);
arch/riscv/kvm/vcpu_insn.c
553
data64 = GET_RS2S(insn, &vcpu->arch.guest_context);
arch/riscv/kvm/vcpu_insn.c
557
data64 = GET_RS2C(insn, &vcpu->arch.guest_context);
arch/riscv/kvm/vcpu_insn.c
561
data32 = GET_RS2S(insn, &vcpu->arch.guest_context);
arch/riscv/kvm/vcpu_insn.c
565
data32 = GET_RS2C(insn, &vcpu->arch.guest_context);
arch/riscv/kvm/vcpu_insn.c
575
vcpu->arch.mmio_decode.insn = insn;
arch/riscv/kvm/vcpu_insn.c
576
vcpu->arch.mmio_decode.insn_len = insn_len;
arch/riscv/kvm/vcpu_insn.c
577
vcpu->arch.mmio_decode.shift = 0;
arch/riscv/kvm/vcpu_insn.c
578
vcpu->arch.mmio_decode.len = len;
arch/riscv/kvm/vcpu_insn.c
579
vcpu->arch.mmio_decode.return_handled = 0;
arch/riscv/kvm/vcpu_insn.c
636
if (vcpu->arch.mmio_decode.return_handled)
arch/riscv/kvm/vcpu_insn.c
639
vcpu->arch.mmio_decode.return_handled = 1;
arch/riscv/kvm/vcpu_insn.c
640
insn = vcpu->arch.mmio_decode.insn;
arch/riscv/kvm/vcpu_insn.c
645
len = vcpu->arch.mmio_decode.len;
arch/riscv/kvm/vcpu_insn.c
646
shift = vcpu->arch.mmio_decode.shift;
arch/riscv/kvm/vcpu_insn.c
651
SET_RD(insn, &vcpu->arch.guest_context,
arch/riscv/kvm/vcpu_insn.c
656
SET_RD(insn, &vcpu->arch.guest_context,
arch/riscv/kvm/vcpu_insn.c
661
SET_RD(insn, &vcpu->arch.guest_context,
arch/riscv/kvm/vcpu_insn.c
666
SET_RD(insn, &vcpu->arch.guest_context,
arch/riscv/kvm/vcpu_insn.c
675
vcpu->arch.guest_context.sepc += vcpu->arch.mmio_decode.insn_len;
arch/riscv/kvm/vcpu_insn.c
86
kvm_vcpu_on_spin(vcpu, vcpu->arch.guest_context.sstatus & SR_SPP);
arch/riscv/kvm/vcpu_onereg.c
118
if (reg_val == (vcpu->arch.isa[0] & KVM_RISCV_BASE_ISA_MASK))
arch/riscv/kvm/vcpu_onereg.c
121
if (!vcpu->arch.ran_atleast_once) {
arch/riscv/kvm/vcpu_onereg.c
138
reg_val = (vcpu->arch.isa[0] & ~KVM_RISCV_BASE_ISA_MASK) |
arch/riscv/kvm/vcpu_onereg.c
140
vcpu->arch.isa[0] = reg_val;
arch/riscv/kvm/vcpu_onereg.c
165
if (reg_val == vcpu->arch.mvendorid)
arch/riscv/kvm/vcpu_onereg.c
167
if (!vcpu->arch.ran_atleast_once)
arch/riscv/kvm/vcpu_onereg.c
168
vcpu->arch.mvendorid = reg_val;
arch/riscv/kvm/vcpu_onereg.c
173
if (reg_val == vcpu->arch.marchid)
arch/riscv/kvm/vcpu_onereg.c
175
if (!vcpu->arch.ran_atleast_once)
arch/riscv/kvm/vcpu_onereg.c
176
vcpu->arch.marchid = reg_val;
arch/riscv/kvm/vcpu_onereg.c
181
if (reg_val == vcpu->arch.mimpid)
arch/riscv/kvm/vcpu_onereg.c
183
if (!vcpu->arch.ran_atleast_once)
arch/riscv/kvm/vcpu_onereg.c
184
vcpu->arch.mimpid = reg_val;
arch/riscv/kvm/vcpu_onereg.c
202
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
238
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
277
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu_onereg.c
299
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu_onereg.c
31
set_bit(guest_ext, vcpu->arch.isa);
arch/riscv/kvm/vcpu_onereg.c
315
WRITE_ONCE(vcpu->arch.irqs_pending_mask[0], 0);
arch/riscv/kvm/vcpu_onereg.c
324
struct kvm_vcpu_smstateen_csr *csr = &vcpu->arch.smstateen_csr;
arch/riscv/kvm/vcpu_onereg.c
328
if (!riscv_isa_extension_available(vcpu->arch.isa, SMSTATEEN))
arch/riscv/kvm/vcpu_onereg.c
343
struct kvm_vcpu_smstateen_csr *csr = &vcpu->arch.smstateen_csr;
arch/riscv/kvm/vcpu_onereg.c
347
if (!riscv_isa_extension_available(vcpu->arch.isa, SMSTATEEN))
arch/riscv/kvm/vcpu_onereg.c
433
vcpu->arch.csr_dirty = true;
arch/riscv/kvm/vcpu_onereg.c
450
if (__riscv_isa_extension_available(vcpu->arch.isa, guest_ext))
arch/riscv/kvm/vcpu_onereg.c
467
if (reg_val == test_bit(guest_ext, vcpu->arch.isa))
arch/riscv/kvm/vcpu_onereg.c
470
if (!vcpu->arch.ran_atleast_once) {
arch/riscv/kvm/vcpu_onereg.c
477
set_bit(guest_ext, vcpu->arch.isa);
arch/riscv/kvm/vcpu_onereg.c
480
clear_bit(guest_ext, vcpu->arch.isa);
arch/riscv/kvm/vcpu_onereg.c
50
reg_val = vcpu->arch.isa[0] & KVM_RISCV_BASE_ISA_MASK;
arch/riscv/kvm/vcpu_onereg.c
679
if (riscv_isa_extension_available(vcpu->arch.isa, SSAIA))
arch/riscv/kvm/vcpu_onereg.c
68
reg_val = vcpu->arch.mvendorid;
arch/riscv/kvm/vcpu_onereg.c
681
if (riscv_isa_extension_available(vcpu->arch.isa, SMSTATEEN))
arch/riscv/kvm/vcpu_onereg.c
708
if (riscv_isa_extension_available(vcpu->arch.isa, SSAIA)) {
arch/riscv/kvm/vcpu_onereg.c
71
reg_val = vcpu->arch.marchid;
arch/riscv/kvm/vcpu_onereg.c
726
if (riscv_isa_extension_available(vcpu->arch.isa, SMSTATEEN)) {
arch/riscv/kvm/vcpu_onereg.c
74
reg_val = vcpu->arch.mimpid;
arch/riscv/kvm/vcpu_onereg.c
771
const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
773
if (riscv_isa_extension_available(vcpu->arch.isa, f))
arch/riscv/kvm/vcpu_onereg.c
800
const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
802
if (riscv_isa_extension_available(vcpu->arch.isa, d))
arch/riscv/kvm/vcpu_onereg.c
881
if (!riscv_isa_extension_available(vcpu->arch.isa, v))
arch/riscv/kvm/vcpu_onereg.c
891
const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
119
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
131
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
133
vcpu->arch.sbi_context.return_handled = 0;
arch/riscv/kvm/vcpu_sbi.c
158
spin_lock(&tmp->arch.mp_state_lock);
arch/riscv/kvm/vcpu_sbi.c
159
WRITE_ONCE(tmp->arch.mp_state.mp_state, KVM_MP_STATE_STOPPED);
arch/riscv/kvm/vcpu_sbi.c
160
spin_unlock(&tmp->arch.mp_state_lock);
arch/riscv/kvm/vcpu_sbi.c
174
spin_lock(&vcpu->arch.reset_state.lock);
arch/riscv/kvm/vcpu_sbi.c
175
vcpu->arch.reset_state.pc = pc;
arch/riscv/kvm/vcpu_sbi.c
176
vcpu->arch.reset_state.a1 = a1;
arch/riscv/kvm/vcpu_sbi.c
177
spin_unlock(&vcpu->arch.reset_state.lock);
arch/riscv/kvm/vcpu_sbi.c
184
struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
arch/riscv/kvm/vcpu_sbi.c
185
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
186
struct kvm_vcpu_reset_state *reset_state = &vcpu->arch.reset_state;
arch/riscv/kvm/vcpu_sbi.c
190
spin_lock(&vcpu->arch.reset_state.lock);
arch/riscv/kvm/vcpu_sbi.c
193
spin_unlock(&vcpu->arch.reset_state.lock);
arch/riscv/kvm/vcpu_sbi.c
201
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
204
if (vcpu->arch.sbi_context.return_handled)
arch/riscv/kvm/vcpu_sbi.c
206
vcpu->arch.sbi_context.return_handled = 1;
arch/riscv/kvm/vcpu_sbi.c
213
vcpu->arch.guest_context.sepc += 4;
arch/riscv/kvm/vcpu_sbi.c
222
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
243
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
337
if (vcpu->arch.ran_atleast_once)
arch/riscv/kvm/vcpu_sbi.c
402
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
449
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
565
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
591
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
654
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
683
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi.c
706
struct kvm_vcpu_sbi_context *scontext = &vcpu->arch.sbi_context;
arch/riscv/kvm/vcpu_sbi_base.c
19
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_base.c
52
*out_val = vcpu->arch.mvendorid;
arch/riscv/kvm/vcpu_sbi_base.c
55
*out_val = vcpu->arch.marchid;
arch/riscv/kvm/vcpu_sbi_base.c
58
*out_val = vcpu->arch.mimpid;
arch/riscv/kvm/vcpu_sbi_fwft.c
103
struct kvm_vcpu_config *cfg = &vcpu->arch.cfg;
arch/riscv/kvm/vcpu_sbi_fwft.c
124
struct kvm_vcpu_config *cfg = &vcpu->arch.cfg;
arch/riscv/kvm/vcpu_sbi_fwft.c
142
if (!riscv_isa_extension_available(vcpu->arch.isa, SMNPM))
arch/riscv/kvm/vcpu_sbi_fwft.c
153
vcpu->arch.cfg.henvcfg &= ~ENVCFG_PMM;
arch/riscv/kvm/vcpu_sbi_fwft.c
181
vcpu->arch.cfg.henvcfg &= ~ENVCFG_PMM;
arch/riscv/kvm/vcpu_sbi_fwft.c
182
vcpu->arch.cfg.henvcfg |= pmm;
arch/riscv/kvm/vcpu_sbi_fwft.c
190
csr_write(CSR_HENVCFG, vcpu->arch.cfg.henvcfg);
arch/riscv/kvm/vcpu_sbi_fwft.c
199
switch (vcpu->arch.cfg.henvcfg & ENVCFG_PMM) {
arch/riscv/kvm/vcpu_sbi_fwft.c
328
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_fwft.c
94
struct kvm_vcpu_config *cfg = &vcpu->arch.cfg;
arch/riscv/kvm/vcpu_sbi_hsm.c
18
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_hsm.c
27
spin_lock(&target_vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu_sbi_hsm.c
39
spin_unlock(&target_vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu_sbi_hsm.c
48
spin_lock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu_sbi_hsm.c
58
spin_unlock(&vcpu->arch.mp_state_lock);
arch/riscv/kvm/vcpu_sbi_hsm.c
65
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_hsm.c
84
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_pmu.c
20
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
107
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vcpu_sbi_replace.c
116
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vcpu_sbi_replace.c
151
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
20
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
51
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
95
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_sta.c
117
vcpu->arch.sta.shmem = shmem;
arch/riscv/kvm/vcpu_sbi_sta.c
118
vcpu->arch.sta.last_steal = current->sched_info.run_delay;
arch/riscv/kvm/vcpu_sbi_sta.c
126
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_sta.c
165
*value = (unsigned long)vcpu->arch.sta.shmem;
arch/riscv/kvm/vcpu_sbi_sta.c
169
*value = upper_32_bits(vcpu->arch.sta.shmem);
arch/riscv/kvm/vcpu_sbi_sta.c
193
gpa_t hi = upper_32_bits(vcpu->arch.sta.shmem);
arch/riscv/kvm/vcpu_sbi_sta.c
203
gpa_t lo = lower_32_bits(vcpu->arch.sta.shmem);
arch/riscv/kvm/vcpu_sbi_sta.c
21
vcpu->arch.sta.shmem = INVALID_GPA;
arch/riscv/kvm/vcpu_sbi_sta.c
218
vcpu->arch.sta.shmem = new_shmem;
arch/riscv/kvm/vcpu_sbi_sta.c
22
vcpu->arch.sta.last_steal = 0;
arch/riscv/kvm/vcpu_sbi_sta.c
27
gpa_t shmem = vcpu->arch.sta.shmem;
arch/riscv/kvm/vcpu_sbi_sta.c
28
u64 last_steal = vcpu->arch.sta.last_steal;
arch/riscv/kvm/vcpu_sbi_sta.c
50
vcpu->arch.sta.shmem = INVALID_GPA;
arch/riscv/kvm/vcpu_sbi_sta.c
70
vcpu->arch.sta.last_steal = READ_ONCE(current->sched_info.run_delay);
arch/riscv/kvm/vcpu_sbi_sta.c
71
steal += vcpu->arch.sta.last_steal - last_steal;
arch/riscv/kvm/vcpu_sbi_sta.c
83
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_sta.c
96
vcpu->arch.sta.shmem = INVALID_GPA;
arch/riscv/kvm/vcpu_sbi_system.c
15
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_v01.c
24
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_v01.c
81
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vcpu_sbi_v01.c
88
vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
arch/riscv/kvm/vcpu_timer.c
106
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
115
struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
arch/riscv/kvm/vcpu_timer.c
116
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
132
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
133
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
144
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
145
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
158
kvm_riscv_vcpu_timer_cancel(&vcpu->arch.timer);
arch/riscv/kvm/vcpu_timer.c
164
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
165
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
204
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
205
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
248
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
276
ret = kvm_riscv_vcpu_timer_cancel(&vcpu->arch.timer);
arch/riscv/kvm/vcpu_timer.c
277
vcpu->arch.timer.init_done = false;
arch/riscv/kvm/vcpu_timer.c
284
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
287
return kvm_riscv_vcpu_timer_cancel(&vcpu->arch.timer);
arch/riscv/kvm/vcpu_timer.c
292
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
304
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
328
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
343
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
379
struct kvm_guest_timer *gt = &kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
47
struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
arch/riscv/kvm/vcpu_timer.c
48
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_timer.c
87
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
88
struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
arch/riscv/kvm/vcpu_vector.c
105
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_vector.c
147
unsigned long *isa = vcpu->arch.isa;
arch/riscv/kvm/vcpu_vector.c
173
unsigned long *isa = vcpu->arch.isa;
arch/riscv/kvm/vcpu_vector.c
187
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_vector.c
22
unsigned long *isa = vcpu->arch.isa;
arch/riscv/kvm/vcpu_vector.c
23
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_vector.c
79
vcpu->arch.guest_context.vector.datap = kzalloc(riscv_v_vsize, GFP_KERNEL);
arch/riscv/kvm/vcpu_vector.c
80
if (!vcpu->arch.guest_context.vector.datap)
arch/riscv/kvm/vcpu_vector.c
83
vcpu->arch.host_context.vector.datap = kzalloc(riscv_v_vsize, GFP_KERNEL);
arch/riscv/kvm/vcpu_vector.c
84
if (!vcpu->arch.host_context.vector.datap) {
arch/riscv/kvm/vcpu_vector.c
85
kfree(vcpu->arch.guest_context.vector.datap);
arch/riscv/kvm/vcpu_vector.c
86
vcpu->arch.guest_context.vector.datap = NULL;
arch/riscv/kvm/vcpu_vector.c
95
kfree(vcpu->arch.guest_context.vector.datap);
arch/riscv/kvm/vcpu_vector.c
96
kfree(vcpu->arch.host_context.vector.datap);
arch/riscv/kvm/vm.c
205
r = kvm_riscv_gstage_gpa_bits(kvm->arch.pgd_levels);
arch/riscv/kvm/vm.c
222
kvm->arch.mp_state_reset = true;
arch/riscv/kvm/vm.c
256
kvm->arch.pgd_levels = new_levels;
arch/riscv/kvm/vmid.c
52
kvm->arch.vmid.vmid_version = 0;
arch/riscv/kvm/vmid.c
53
kvm->arch.vmid.vmid = 0;
arch/riscv/kvm/vmid.c
76
struct kvm_vmid *vmid = &vcpu->kvm->arch.vmid;
arch/s390/kernel/audit.c
32
int audit_classify_arch(int arch)
arch/s390/kernel/ftrace.c
98
next_trampoline = &mod->arch.next_trampoline;
arch/s390/kernel/ftrace.c
99
trampolines_end = mod->arch.trampolines_end;
arch/s390/kernel/machine_kexec_file.c
224
image->arch.ipl_buf = buf.buffer;
arch/s390/kernel/machine_kexec_file.c
387
vfree(image->arch.ipl_buf);
arch/s390/kernel/machine_kexec_file.c
388
image->arch.ipl_buf = NULL;
arch/s390/kernel/module.c
132
me->arch.nsyms = symtab->sh_size / sizeof(Elf_Sym);
arch/s390/kernel/module.c
133
me->arch.syminfo = vmalloc(array_size(sizeof(struct mod_arch_syminfo),
arch/s390/kernel/module.c
134
me->arch.nsyms));
arch/s390/kernel/module.c
135
if (!me->arch.syminfo)
arch/s390/kernel/module.c
139
for (i = 0; i < me->arch.nsyms; i++) {
arch/s390/kernel/module.c
145
me->arch.syminfo[i].got_offset = -1UL;
arch/s390/kernel/module.c
146
me->arch.syminfo[i].plt_offset = -1UL;
arch/s390/kernel/module.c
147
me->arch.syminfo[i].got_initialized = 0;
arch/s390/kernel/module.c
148
me->arch.syminfo[i].plt_initialized = 0;
arch/s390/kernel/module.c
152
me->arch.got_size = me->arch.plt_size = 0;
arch/s390/kernel/module.c
166
me->arch.got_offset = mod_mem->size;
arch/s390/kernel/module.c
167
mod_mem->size += me->arch.got_size;
arch/s390/kernel/module.c
168
me->arch.plt_offset = mod_mem->size;
arch/s390/kernel/module.c
169
if (me->arch.plt_size) {
arch/s390/kernel/module.c
171
me->arch.plt_size += PLT_ENTRY_SIZE;
arch/s390/kernel/module.c
172
mod_mem->size += me->arch.plt_size;
arch/s390/kernel/module.c
239
info = me->arch.syminfo + r_sym;
arch/s390/kernel/module.c
297
me->arch.got_offset +
arch/s390/kernel/module.c
322
me->arch.got_offset - loc;
arch/s390/kernel/module.c
338
plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
arch/s390/kernel/module.c
345
jump_r1 = plt_base + me->arch.plt_size -
arch/s390/kernel/module.c
361
val = me->arch.plt_offset - me->arch.got_offset +
arch/s390/kernel/module.c
369
me->arch.plt_offset +
arch/s390/kernel/module.c
390
((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
arch/s390/kernel/module.c
400
val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
arch/s390/kernel/module.c
45
execmem_free(mod->arch.trampolines_start);
arch/s390/kernel/module.c
485
me->arch.trampolines_start = (struct ftrace_hotpatch_trampoline *)start;
arch/s390/kernel/module.c
486
me->arch.trampolines_end = (struct ftrace_hotpatch_trampoline *)end;
arch/s390/kernel/module.c
487
me->arch.next_trampoline = me->arch.trampolines_start;
arch/s390/kernel/module.c
503
!nospec_disable && me->arch.plt_size) {
arch/s390/kernel/module.c
506
ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
arch/s390/kernel/module.c
507
me->arch.plt_size - PLT_ENTRY_SIZE;
arch/s390/kernel/module.c
55
vfree(mod->arch.syminfo);
arch/s390/kernel/module.c
56
mod->arch.syminfo = NULL;
arch/s390/kernel/module.c
63
info = me->arch.syminfo + ELF_R_SYM (rela->r_info);
arch/s390/kernel/module.c
78
info->got_offset = me->arch.got_size;
arch/s390/kernel/module.c
79
me->arch.got_size += sizeof(void*);
arch/s390/kernel/module.c
90
info->plt_offset = me->arch.plt_size;
arch/s390/kernel/module.c
91
me->arch.plt_size += PLT_ENTRY_SIZE;
arch/s390/kvm/diag.c
111
if (vcpu->arch.pfault_token != KVM_S390_PFAULT_TOKEN_INVALID) {
arch/s390/kvm/diag.c
128
vcpu->arch.pfault_token = parm.token_addr;
arch/s390/kvm/diag.c
129
vcpu->arch.pfault_select = parm.select_mask;
arch/s390/kvm/diag.c
130
vcpu->arch.pfault_compare = parm.compare_mask;
arch/s390/kvm/diag.c
150
if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
arch/s390/kvm/diag.c
153
vcpu->arch.pfault_token = KVM_S390_PFAULT_TOKEN_INVALID;
arch/s390/kvm/diag.c
192
tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4];
arch/s390/kvm/diag.c
234
unsigned int reg = vcpu->arch.sie_block->ipa & 0xf;
arch/s390/kvm/diag.c
272
if (!vcpu->kvm->arch.css_support ||
arch/s390/kvm/diag.c
306
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/diag.c
43
start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4];
arch/s390/kvm/diag.c
44
end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE;
arch/s390/kvm/diag.c
92
u16 rx = (vcpu->arch.sie_block->ipa & 0xf0) >> 4;
arch/s390/kvm/diag.c
93
u16 ry = (vcpu->arch.sie_block->ipa & 0x0f);
arch/s390/kvm/faultin.c
112
rc = gmap_link(mc, kvm->arch.gmap, f, slot);
arch/s390/kvm/faultin.c
46
mc = vcpu->arch.mc;
arch/s390/kvm/faultin.c
52
if (gmap_try_fixup_minor(kvm->arch.gmap, f) == 0)
arch/s390/kvm/gaccess.c
1025
vcpu->arch.pgm.code = rc;
arch/s390/kvm/gaccess.c
1260
union ctlreg0 ctlreg0 = {.val = vcpu->arch.sie_block->gcr[0]};
arch/s390/kvm/gaccess.c
137
return kvm->arch.sca->ipte_control.kh != 0;
arch/s390/kvm/gaccess.c
139
return kvm->arch.ipte_lock_count != 0;
arch/s390/kvm/gaccess.c
146
mutex_lock(&kvm->arch.ipte_mutex);
arch/s390/kvm/gaccess.c
147
kvm->arch.ipte_lock_count++;
arch/s390/kvm/gaccess.c
148
if (kvm->arch.ipte_lock_count > 1)
arch/s390/kvm/gaccess.c
151
ic = &kvm->arch.sca->ipte_control;
arch/s390/kvm/gaccess.c
1592
rc = kvm_s390_mmu_cache_topup(vcpu->arch.mc);
arch/s390/kvm/gaccess.c
1605
rc = _gaccess_do_shadow(vcpu->arch.mc, sg, saddr, walk);
arch/s390/kvm/gaccess.c
162
mutex_unlock(&kvm->arch.ipte_mutex);
arch/s390/kvm/gaccess.c
1664
rc = kvm_s390_mmu_cache_topup(vcpu->arch.mc);
arch/s390/kvm/gaccess.c
169
mutex_lock(&kvm->arch.ipte_mutex);
arch/s390/kvm/gaccess.c
170
kvm->arch.ipte_lock_count--;
arch/s390/kvm/gaccess.c
171
if (kvm->arch.ipte_lock_count)
arch/s390/kvm/gaccess.c
173
ic = &kvm->arch.sca->ipte_control;
arch/s390/kvm/gaccess.c
179
wake_up(&kvm->arch.ipte_wq);
arch/s390/kvm/gaccess.c
181
mutex_unlock(&kvm->arch.ipte_mutex);
arch/s390/kvm/gaccess.c
189
ic = &kvm->arch.sca->ipte_control;
arch/s390/kvm/gaccess.c
206
ic = &kvm->arch.sca->ipte_control;
arch/s390/kvm/gaccess.c
215
wake_up(&kvm->arch.ipte_wq);
arch/s390/kvm/gaccess.c
248
if (vcpu->arch.acrs_loaded)
arch/s390/kvm/gaccess.c
253
asce->val = vcpu->arch.sie_block->gcr[1];
arch/s390/kvm/gaccess.c
256
asce->val = vcpu->arch.sie_block->gcr[7];
arch/s390/kvm/gaccess.c
264
ald_addr = vcpu->arch.sie_block->gcr[5];
arch/s390/kvm/gaccess.c
266
ald_addr = vcpu->arch.sie_block->gcr[2];
arch/s390/kvm/gaccess.c
299
eax = (vcpu->arch.sie_block->gcr[8] >> 16) & 0xffff;
arch/s390/kvm/gaccess.c
337
struct kvm_s390_pgm_info *pgm = &vcpu->arch.pgm;
arch/s390/kvm/gaccess.c
386
teid->as = psw_bits(vcpu->arch.sie_block->gpsw).as;
arch/s390/kvm/gaccess.c
414
struct psw_bits psw = psw_bits(vcpu->arch.sie_block->gpsw);
arch/s390/kvm/gaccess.c
427
asce->val = vcpu->arch.sie_block->gcr[1];
arch/s390/kvm/gaccess.c
430
asce->val = vcpu->arch.sie_block->gcr[7];
arch/s390/kvm/gaccess.c
433
asce->val = vcpu->arch.sie_block->gcr[13];
arch/s390/kvm/gaccess.c
483
ctlreg0.val = vcpu->arch.sie_block->gcr[0];
arch/s390/kvm/gaccess.c
643
union ctlreg0 ctlreg0 = {.val = vcpu->arch.sie_block->gcr[0]};
arch/s390/kvm/gaccess.c
644
psw_t *psw = &vcpu->arch.sie_block->gpsw;
arch/s390/kvm/gaccess.c
660
r = dat_get_storage_key(kvm->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key);
arch/s390/kvm/gaccess.c
673
psw_t *psw = &vcpu->arch.sie_block->gpsw;
arch/s390/kvm/gaccess.c
678
override = vcpu->arch.sie_block->gcr[0];
arch/s390/kvm/gaccess.c
695
return vcpu->arch.sie_block->gcr[0] & CR0_STORAGE_PROTECTION_OVERRIDE;
arch/s390/kvm/gaccess.c
719
r = dat_get_storage_key(vcpu->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key);
arch/s390/kvm/gaccess.c
779
psw_t *psw = &vcpu->arch.sie_block->gpsw;
arch/s390/kvm/gaccess.c
933
psw_t *psw = &vcpu->arch.sie_block->gpsw;
arch/s390/kvm/gaccess.h
281
u8 access_key = psw_bits(vcpu->arch.sie_block->gpsw).key;
arch/s390/kvm/gaccess.h
325
u8 access_key = psw_bits(vcpu->arch.sie_block->gpsw).key;
arch/s390/kvm/gaccess.h
348
u8 access_key = psw_bits(vcpu->arch.sie_block->gpsw).key;
arch/s390/kvm/gaccess.h
89
return _kvm_s390_logical_to_effective(&vcpu->arch.sie_block->gpsw, ga);
arch/s390/kvm/gmap.c
29
return vcpu->arch.sie_block->prog0c & PROG_IN_SIE;
arch/s390/kvm/guestdbg.c
102
u64 *cr9 = &vcpu->arch.sie_block->gcr[9];
arch/s390/kvm/guestdbg.c
103
u64 *cr10 = &vcpu->arch.sie_block->gcr[10];
arch/s390/kvm/guestdbg.c
104
u64 *cr11 = &vcpu->arch.sie_block->gcr[11];
arch/s390/kvm/guestdbg.c
107
if (vcpu->arch.guestdbg.nr_hw_wp <= 0 ||
arch/s390/kvm/guestdbg.c
108
vcpu->arch.guestdbg.hw_wp_info == NULL)
arch/s390/kvm/guestdbg.c
121
for (i = 0; i < vcpu->arch.guestdbg.nr_hw_wp; i++) {
arch/s390/kvm/guestdbg.c
122
start = vcpu->arch.guestdbg.hw_wp_info[i].addr;
arch/s390/kvm/guestdbg.c
123
len = vcpu->arch.guestdbg.hw_wp_info[i].len;
arch/s390/kvm/guestdbg.c
132
vcpu->arch.guestdbg.cr0 = vcpu->arch.sie_block->gcr[0];
arch/s390/kvm/guestdbg.c
133
vcpu->arch.guestdbg.cr9 = vcpu->arch.sie_block->gcr[9];
arch/s390/kvm/guestdbg.c
134
vcpu->arch.guestdbg.cr10 = vcpu->arch.sie_block->gcr[10];
arch/s390/kvm/guestdbg.c
135
vcpu->arch.guestdbg.cr11 = vcpu->arch.sie_block->gcr[11];
arch/s390/kvm/guestdbg.c
140
vcpu->arch.sie_block->gcr[0] = vcpu->arch.guestdbg.cr0;
arch/s390/kvm/guestdbg.c
141
vcpu->arch.sie_block->gcr[9] = vcpu->arch.guestdbg.cr9;
arch/s390/kvm/guestdbg.c
142
vcpu->arch.sie_block->gcr[10] = vcpu->arch.guestdbg.cr10;
arch/s390/kvm/guestdbg.c
143
vcpu->arch.sie_block->gcr[11] = vcpu->arch.guestdbg.cr11;
arch/s390/kvm/guestdbg.c
156
vcpu->arch.sie_block->gcr[0] &= ~CR0_CLOCK_COMPARATOR_SUBMASK;
arch/s390/kvm/guestdbg.c
157
vcpu->arch.sie_block->gcr[9] |= PER_EVENT_IFETCH;
arch/s390/kvm/guestdbg.c
158
vcpu->arch.sie_block->gcr[10] = 0;
arch/s390/kvm/guestdbg.c
159
vcpu->arch.sie_block->gcr[11] = -1UL;
arch/s390/kvm/guestdbg.c
168
if (vcpu->arch.sie_block->gcr[9] & PER_EVENT_NULLIFICATION)
arch/s390/kvm/guestdbg.c
169
vcpu->arch.sie_block->gcr[9] &= ~PER_EVENT_NULLIFICATION;
arch/s390/kvm/guestdbg.c
211
if (dbg->arch.nr_hw_bp <= 0 || !dbg->arch.hw_bp)
arch/s390/kvm/guestdbg.c
213
else if (dbg->arch.nr_hw_bp > MAX_BP_COUNT)
arch/s390/kvm/guestdbg.c
216
bp_data = memdup_array_user(dbg->arch.hw_bp, dbg->arch.nr_hw_bp,
arch/s390/kvm/guestdbg.c
221
for (i = 0; i < dbg->arch.nr_hw_bp; i++) {
arch/s390/kvm/guestdbg.c
249
for (nr_wp = 0, nr_bp = 0, i = 0; i < dbg->arch.nr_hw_bp; i++) {
arch/s390/kvm/guestdbg.c
266
vcpu->arch.guestdbg.nr_hw_bp = nr_bp;
arch/s390/kvm/guestdbg.c
267
vcpu->arch.guestdbg.hw_bp_info = bp_info;
arch/s390/kvm/guestdbg.c
268
vcpu->arch.guestdbg.nr_hw_wp = nr_wp;
arch/s390/kvm/guestdbg.c
269
vcpu->arch.guestdbg.hw_wp_info = wp_info;
arch/s390/kvm/guestdbg.c
283
for (i = 0; i < vcpu->arch.guestdbg.nr_hw_wp; i++) {
arch/s390/kvm/guestdbg.c
284
hw_wp_info = &vcpu->arch.guestdbg.hw_wp_info[i];
arch/s390/kvm/guestdbg.c
288
kfree(vcpu->arch.guestdbg.hw_wp_info);
arch/s390/kvm/guestdbg.c
289
vcpu->arch.guestdbg.hw_wp_info = NULL;
arch/s390/kvm/guestdbg.c
291
kfree(vcpu->arch.guestdbg.hw_bp_info);
arch/s390/kvm/guestdbg.c
292
vcpu->arch.guestdbg.hw_bp_info = NULL;
arch/s390/kvm/guestdbg.c
294
vcpu->arch.guestdbg.nr_hw_wp = 0;
arch/s390/kvm/guestdbg.c
295
vcpu->arch.guestdbg.nr_hw_bp = 0;
arch/s390/kvm/guestdbg.c
312
struct kvm_hw_bp_info_arch *bp_info = vcpu->arch.guestdbg.hw_bp_info;
arch/s390/kvm/guestdbg.c
315
if (vcpu->arch.guestdbg.nr_hw_bp == 0)
arch/s390/kvm/guestdbg.c
318
for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) {
arch/s390/kvm/guestdbg.c
340
if (vcpu->arch.guestdbg.nr_hw_wp == 0)
arch/s390/kvm/guestdbg.c
343
for (i = 0; i < vcpu->arch.guestdbg.nr_hw_wp; i++) {
arch/s390/kvm/guestdbg.c
344
wp_info = &vcpu->arch.guestdbg.hw_wp_info[i];
arch/s390/kvm/guestdbg.c
387
struct kvm_debug_exit_arch *debug_exit = &vcpu->run->debug.arch;
arch/s390/kvm/guestdbg.c
390
unsigned long addr = vcpu->arch.sie_block->gpsw.addr;
arch/s390/kvm/guestdbg.c
394
vcpu->arch.guestdbg.nr_hw_wp > 0) {
arch/s390/kvm/guestdbg.c
403
vcpu->arch.guestdbg.nr_hw_bp > 0) {
arch/s390/kvm/guestdbg.c
409
vcpu->arch.guestdbg.last_bp = addr;
arch/s390/kvm/guestdbg.c
414
if (bp_info && vcpu->arch.guestdbg.last_bp != peraddr) {
arch/s390/kvm/guestdbg.c
438
if (vcpu->arch.sie_block->icptcode == ICPT_PROGI) {
arch/s390/kvm/guestdbg.c
440
*addr = vcpu->arch.sie_block->peraddr;
arch/s390/kvm/guestdbg.c
455
*addr = __rewind_psw(vcpu->arch.sie_block->gpsw,
arch/s390/kvm/guestdbg.c
457
if (vcpu->arch.sie_block->icptstatus & 0x01) {
arch/s390/kvm/guestdbg.c
458
exec_ilen = (vcpu->arch.sie_block->icptstatus & 0x60) >> 4;
arch/s390/kvm/guestdbg.c
491
(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PER)
arch/s390/kvm/guestdbg.c
495
const u64 cr10 = vcpu->arch.sie_block->gcr[10];
arch/s390/kvm/guestdbg.c
496
const u64 cr11 = vcpu->arch.sie_block->gcr[11];
arch/s390/kvm/guestdbg.c
501
.per_address = __rewind_psw(vcpu->arch.sie_block->gpsw, ilen),
arch/s390/kvm/guestdbg.c
518
!(vcpu->arch.sie_block->gcr[9] & PER_EVENT_IFETCH))
arch/s390/kvm/guestdbg.c
535
const u8 perc = vcpu->arch.sie_block->perc;
arch/s390/kvm/guestdbg.c
536
u64 addr = vcpu->arch.sie_block->gpsw.addr;
arch/s390/kvm/guestdbg.c
537
u64 cr9 = vcpu->arch.sie_block->gcr[9];
arch/s390/kvm/guestdbg.c
538
u64 cr10 = vcpu->arch.sie_block->gcr[10];
arch/s390/kvm/guestdbg.c
539
u64 cr11 = vcpu->arch.sie_block->gcr[11];
arch/s390/kvm/guestdbg.c
570
vcpu->arch.sie_block->perc = guest_perc;
arch/s390/kvm/guestdbg.c
573
vcpu->arch.sie_block->iprcc &= ~PGM_PER;
arch/s390/kvm/guestdbg.c
577
#define pssec(vcpu) (vcpu->arch.sie_block->gcr[1] & _ASCE_SPACE_SWITCH)
arch/s390/kvm/guestdbg.c
578
#define hssec(vcpu) (vcpu->arch.sie_block->gcr[13] & _ASCE_SPACE_SWITCH)
arch/s390/kvm/guestdbg.c
579
#define old_ssec(vcpu) ((vcpu->arch.sie_block->tecmc >> 31) & 0x1)
arch/s390/kvm/guestdbg.c
580
#define old_as_is_home(vcpu) !(vcpu->arch.sie_block->tecmc & 0xffff)
arch/s390/kvm/guestdbg.c
586
if (debug_exit_required(vcpu, vcpu->arch.sie_block->perc,
arch/s390/kvm/guestdbg.c
587
vcpu->arch.sie_block->peraddr))
arch/s390/kvm/guestdbg.c
600
if (vcpu->arch.sie_block->iprcc == PGM_SPACE_SWITCH) {
arch/s390/kvm/guestdbg.c
601
vcpu->arch.sie_block->iprcc = 0;
arch/s390/kvm/guestdbg.c
602
new_as = psw_bits(vcpu->arch.sie_block->gpsw).as;
arch/s390/kvm/guestdbg.c
611
vcpu->arch.sie_block->iprcc = PGM_SPACE_SWITCH;
arch/s390/kvm/guestdbg.c
619
vcpu->arch.sie_block->iprcc = PGM_SPACE_SWITCH;
arch/s390/kvm/guestdbg.c
62
u64 *cr9 = &vcpu->arch.sie_block->gcr[9];
arch/s390/kvm/guestdbg.c
63
u64 *cr10 = &vcpu->arch.sie_block->gcr[10];
arch/s390/kvm/guestdbg.c
64
u64 *cr11 = &vcpu->arch.sie_block->gcr[11];
arch/s390/kvm/guestdbg.c
67
if (vcpu->arch.guestdbg.nr_hw_bp <= 0 ||
arch/s390/kvm/guestdbg.c
68
vcpu->arch.guestdbg.hw_bp_info == NULL)
arch/s390/kvm/guestdbg.c
79
for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) {
arch/s390/kvm/guestdbg.c
80
start = vcpu->arch.guestdbg.hw_bp_info[i].addr;
arch/s390/kvm/guestdbg.c
81
len = vcpu->arch.guestdbg.hw_bp_info[i].len;
arch/s390/kvm/intercept.c
111
vcpu->arch.sie_block->ipa,
arch/s390/kvm/intercept.c
112
vcpu->arch.sie_block->ipb);
arch/s390/kvm/intercept.c
114
switch (vcpu->arch.sie_block->ipa >> 8) {
arch/s390/kvm/intercept.c
147
.code = vcpu->arch.sie_block->iprcc,
arch/s390/kvm/intercept.c
152
switch (vcpu->arch.sie_block->iprcc & ~PGM_PER) {
arch/s390/kvm/intercept.c
163
pgm_info.trans_exc_code = vcpu->arch.sie_block->tecmc;
arch/s390/kvm/intercept.c
171
pgm_info.exc_access_id = vcpu->arch.sie_block->eai;
arch/s390/kvm/intercept.c
179
pgm_info.trans_exc_code = vcpu->arch.sie_block->tecmc;
arch/s390/kvm/intercept.c
180
pgm_info.exc_access_id = vcpu->arch.sie_block->eai;
arch/s390/kvm/intercept.c
181
pgm_info.op_access_id = vcpu->arch.sie_block->oai;
arch/s390/kvm/intercept.c
184
pgm_info.mon_class_nr = vcpu->arch.sie_block->mcn;
arch/s390/kvm/intercept.c
185
pgm_info.mon_code = vcpu->arch.sie_block->tecmc;
arch/s390/kvm/intercept.c
189
pgm_info.data_exc_code = vcpu->arch.sie_block->dxc;
arch/s390/kvm/intercept.c
192
pgm_info.trans_exc_code = vcpu->arch.sie_block->tecmc;
arch/s390/kvm/intercept.c
193
pgm_info.exc_access_id = vcpu->arch.sie_block->eai;
arch/s390/kvm/intercept.c
199
if (vcpu->arch.sie_block->iprcc & PGM_PER) {
arch/s390/kvm/intercept.c
200
pgm_info.per_code = vcpu->arch.sie_block->perc;
arch/s390/kvm/intercept.c
201
pgm_info.per_atmid = vcpu->arch.sie_block->peratmid;
arch/s390/kvm/intercept.c
202
pgm_info.per_address = vcpu->arch.sie_block->peraddr;
arch/s390/kvm/intercept.c
203
pgm_info.per_access_id = vcpu->arch.sie_block->peraid;
arch/s390/kvm/intercept.c
221
itdb = phys_to_virt(vcpu->arch.sie_block->itdba);
arch/s390/kvm/intercept.c
230
#define per_event(vcpu) (vcpu->arch.sie_block->iprcc & PGM_PER)
arch/s390/kvm/intercept.c
237
vcpu->arch.sie_block->iprcc != PGM_PER) {
arch/s390/kvm/intercept.c
266
if (vcpu->arch.sie_block->iprcc == 0)
arch/s390/kvm/intercept.c
270
trace_kvm_s390_intercept_prog(vcpu, vcpu->arch.sie_block->iprcc);
arch/s390/kvm/intercept.c
271
if (vcpu->arch.sie_block->iprcc == PGM_SPECIFICATION) {
arch/s390/kvm/intercept.c
28
struct kvm_s390_sie_block *sie_block = vcpu->arch.sie_block;
arch/s390/kvm/intercept.c
305
u16 eic = vcpu->arch.sie_block->eic;
arch/s390/kvm/intercept.c
31
switch (vcpu->arch.sie_block->icptcode) {
arch/s390/kvm/intercept.c
313
newpsw = vcpu->arch.sie_block->gpsw;
arch/s390/kvm/intercept.c
337
irq.u.extcall.code = vcpu->arch.sie_block->extcpuaddr;
arch/s390/kvm/intercept.c
38
ilen = insn_length(vcpu->arch.sie_block->ipa >> 8);
arch/s390/kvm/intercept.c
399
if (vcpu->arch.sie_block->ipa == 0xb254) /* MVPG */
arch/s390/kvm/intercept.c
401
if (vcpu->arch.sie_block->ipa >> 8 == 0xae) /* SIGP */
arch/s390/kvm/intercept.c
453
memcpy(sida_addr(vcpu->arch.sie_block), sctns, PAGE_SIZE);
arch/s390/kvm/intercept.c
475
trace_kvm_s390_handle_operexc(vcpu, vcpu->arch.sie_block->ipa,
arch/s390/kvm/intercept.c
476
vcpu->arch.sie_block->ipb);
arch/s390/kvm/intercept.c
478
if (vcpu->arch.sie_block->ipa == 0xb256)
arch/s390/kvm/intercept.c
48
ilen = vcpu->arch.sie_block->pgmilc & 0x6;
arch/s390/kvm/intercept.c
481
if (vcpu->kvm->arch.user_operexec)
arch/s390/kvm/intercept.c
484
if (vcpu->arch.sie_block->ipa == 0 && vcpu->kvm->arch.user_instr0)
arch/s390/kvm/intercept.c
497
oldpsw = vcpu->arch.sie_block->gpsw;
arch/s390/kvm/intercept.c
510
u32 pref = *(u32 *)sida_addr(vcpu->arch.sie_block);
arch/s390/kvm/intercept.c
519
struct kvm_s390_float_interrupt *fi = &vcpu->kvm->arch.float_int;
arch/s390/kvm/intercept.c
543
struct uv_cb_share *guest_uvcb = sida_addr(vcpu->arch.sie_block);
arch/s390/kvm/intercept.c
56
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/intercept.c
576
if (vcpu->arch.sie_block->ipa == 0xb210)
arch/s390/kvm/intercept.c
578
if (vcpu->arch.sie_block->ipa == 0xb220)
arch/s390/kvm/intercept.c
580
if (vcpu->arch.sie_block->ipa == 0xb9a4)
arch/s390/kvm/intercept.c
582
if (vcpu->arch.sie_block->ipa >> 8 == 0xae) {
arch/s390/kvm/intercept.c
602
if (!(vcpu->arch.sie_block->icptstatus & 0x02))
arch/s390/kvm/intercept.c
606
if (guestdbg_sstep_enabled(vcpu) && vcpu->arch.local_int.pending_irqs)
arch/s390/kvm/intercept.c
619
switch (vcpu->arch.sie_block->icptcode) {
arch/s390/kvm/intercept.c
94
int viwhy = vcpu->arch.sie_block->ipb >> 16;
arch/s390/kvm/interrupt.c
1025
&vcpu->arch.sie_block->gpsw,
arch/s390/kvm/interrupt.c
1028
&vcpu->arch.sie_block->gpsw,
arch/s390/kvm/interrupt.c
103
return !(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_IO);
arch/s390/kvm/interrupt.c
1039
struct kvm_s390_float_interrupt *fi = &vcpu->kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1069
&vcpu->arch.sie_block->gpsw,
arch/s390/kvm/interrupt.c
1072
&vcpu->arch.sie_block->gpsw,
arch/s390/kvm/interrupt.c
108
return !(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_MCHECK);
arch/s390/kvm/interrupt.c
1088
vcpu->arch.sie_block->iictl = IICTL_CODE_IO;
arch/s390/kvm/interrupt.c
1089
vcpu->arch.sie_block->subchannel_id = io->subchannel_id;
arch/s390/kvm/interrupt.c
1090
vcpu->arch.sie_block->subchannel_nr = io->subchannel_nr;
arch/s390/kvm/interrupt.c
1091
vcpu->arch.sie_block->io_int_parm = io->io_int_parm;
arch/s390/kvm/interrupt.c
1092
vcpu->arch.sie_block->io_int_word = io->io_int_word;
arch/s390/kvm/interrupt.c
1101
&vcpu->arch.sie_block->gpsw,
arch/s390/kvm/interrupt.c
1104
&vcpu->arch.sie_block->gpsw,
arch/s390/kvm/interrupt.c
1114
struct kvm_s390_gisa_interrupt *gi = &vcpu->kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
1120
fi = &vcpu->kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1181
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1200
(vcpu->arch.sie_block->gcr[0] & CR0_EXTERNAL_CALL_SUBMASK))
arch/s390/kvm/interrupt.c
121
!(vcpu->arch.sie_block->gcr[0] & CR0_CLOCK_COMPARATOR_SUBMASK))
arch/s390/kvm/interrupt.c
1216
const u64 ckc = vcpu->arch.sie_block->ckc;
arch/s390/kvm/interrupt.c
1220
if (vcpu->arch.sie_block->gcr[0] & CR0_CLOCK_COMPARATOR_SIGN) {
arch/s390/kvm/interrupt.c
1247
struct kvm_s390_gisa_interrupt *gi = &vcpu->kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
1263
vcpu->arch.sie_block->gcr[6] >> 24))
arch/s390/kvm/interrupt.c
1278
hrtimer_start(&vcpu->arch.ckc_timer, sltime, HRTIMER_MODE_REL);
arch/s390/kvm/interrupt.c
1282
vcpu->kvm->arch.float_int.last_sleep_cpu = vcpu->vcpu_idx;
arch/s390/kvm/interrupt.c
1288
hrtimer_cancel(&vcpu->arch.ckc_timer);
arch/s390/kvm/interrupt.c
1309
vcpu = container_of(timer, struct kvm_vcpu, arch.ckc_timer);
arch/s390/kvm/interrupt.c
132
const u64 ckc = vcpu->arch.sie_block->ckc;
arch/s390/kvm/interrupt.c
1324
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1337
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
134
if (vcpu->arch.sie_block->gcr[0] & CR0_CLOCK_COMPARATOR_SIGN) {
arch/s390/kvm/interrupt.c
1421
struct kvm_debug_exit_arch *debug_exit = &vcpu->run->debug.arch;
arch/s390/kvm/interrupt.c
1423
debug_exit->addr = vcpu->arch.sie_block->gpsw.addr;
arch/s390/kvm/interrupt.c
1435
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
146
(vcpu->arch.sie_block->gcr[0] & CR0_CPU_TIMER_SUBMASK);
arch/s390/kvm/interrupt.c
1477
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1494
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1520
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1540
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1566
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1579
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1599
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1631
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1645
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
1660
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1698
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
1733
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
1771
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1802
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1820
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1840
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1854
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
1875
fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
1909
for (sigcpu = kvm->arch.float_int.last_sleep_cpu; ; sigcpu++) {
arch/s390/kvm/interrupt.c
1926
kvm->arch.gisa_int.origin) ||
arch/s390/kvm/interrupt.c
2069
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
2076
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
2083
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
2138
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
2177
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
2198
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
2237
fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
2288
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
2406
return kvm->arch.adapters[id];
arch/s390/kvm/interrupt.c
2425
if (dev->kvm->arch.adapters[adapter_info.id] != NULL)
arch/s390/kvm/interrupt.c
2439
dev->kvm->arch.adapters[adapter->id] = adapter;
arch/s390/kvm/interrupt.c
2461
kfree(kvm->arch.adapters[i]);
arch/s390/kvm/interrupt.c
2524
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
2564
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
2605
struct kvm_s390_float_interrupt *fi = &kvm->arch.float_int;
arch/s390/kvm/interrupt.c
2638
set_bit(GMAP_FLAG_PFAULT_ENABLED, &dev->kvm->arch.gmap->flags);
arch/s390/kvm/interrupt.c
2643
clear_bit(GMAP_FLAG_PFAULT_ENABLED, &dev->kvm->arch.gmap->flags);
arch/s390/kvm/interrupt.c
269
unsigned long pending = vcpu->kvm->arch.float_int.pending_irqs |
arch/s390/kvm/interrupt.c
270
vcpu->arch.local_int.pending_irqs;
arch/s390/kvm/interrupt.c
2702
if (dev->kvm->arch.flic)
arch/s390/kvm/interrupt.c
2704
dev->kvm->arch.flic = dev;
arch/s390/kvm/interrupt.c
2710
dev->kvm->arch.flic = NULL;
arch/s390/kvm/interrupt.c
272
pending &= ~vcpu->kvm->arch.float_int.masked_irqs;
arch/s390/kvm/interrupt.c
278
struct kvm_s390_gisa_interrupt *gi = &vcpu->kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
2914
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
2998
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
3017
store_local_irq(&vcpu->arch.local_int, &irq, irq_type);
arch/s390/kvm/interrupt.c
303
if (!(vcpu->arch.sie_block->gcr[6] & isc_to_isc_bits(i)))
arch/s390/kvm/interrupt.c
3053
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
3057
for_each_set_bit(vcpu_idx, kvm->arch.idle_mask, online_vcpus) {
arch/s390/kvm/interrupt.c
3061
vcpu_isc_mask = (u8)(vcpu->arch.sie_block->gcr[6] >> 24);
arch/s390/kvm/interrupt.c
3130
gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
3141
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
3151
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
3155
gi->origin = &kvm->arch.sie_page2->gisa;
arch/s390/kvm/interrupt.c
3167
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
3180
vcpu->arch.sie_block->gd = gisa_desc;
arch/s390/kvm/interrupt.c
3181
vcpu->arch.sie_block->eca |= ECA_AIV;
arch/s390/kvm/interrupt.c
3183
vcpu->arch.sie_block->gd & 0x3, vcpu->vcpu_id);
arch/s390/kvm/interrupt.c
3190
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
3208
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
3216
vcpu->arch.sie_block->eca &= ~ECA_AIV;
arch/s390/kvm/interrupt.c
3217
vcpu->arch.sie_block->gd = 0U;
arch/s390/kvm/interrupt.c
323
if (!(vcpu->arch.sie_block->gcr[0] & CR0_EXTERNAL_CALL_SUBMASK))
arch/s390/kvm/interrupt.c
3243
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
325
if (!(vcpu->arch.sie_block->gcr[0] & CR0_EMERGENCY_SIGNAL_SUBMASK))
arch/s390/kvm/interrupt.c
327
if (!(vcpu->arch.sie_block->gcr[0] & CR0_CLOCK_COMPARATOR_SUBMASK))
arch/s390/kvm/interrupt.c
3282
struct kvm_s390_gisa_interrupt *gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
329
if (!(vcpu->arch.sie_block->gcr[0] & CR0_CPU_TIMER_SUBMASK))
arch/s390/kvm/interrupt.c
331
if (!(vcpu->arch.sie_block->gcr[0] & CR0_SERVICE_SIGNAL_SUBMASK)) {
arch/s390/kvm/interrupt.c
3323
gi = &kvm->arch.gisa_int;
arch/s390/kvm/interrupt.c
339
vcpu->arch.sie_block->iictl != IICTL_CODE_NONE)
arch/s390/kvm/interrupt.c
347
if (!(vcpu->arch.sie_block->gcr[14] &
arch/s390/kvm/interrupt.c
348
(vcpu->kvm->arch.float_int.mchk.cr14 |
arch/s390/kvm/interrupt.c
349
vcpu->arch.local_int.irq.mchk.cr14)))
arch/s390/kvm/interrupt.c
364
set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask);
arch/s390/kvm/interrupt.c
370
clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask);
arch/s390/kvm/interrupt.c
377
vcpu->arch.sie_block->lctl = 0x0000;
arch/s390/kvm/interrupt.c
378
vcpu->arch.sie_block->ictl &= ~(ICTL_LPSW | ICTL_STCTL | ICTL_PINT);
arch/s390/kvm/interrupt.c
381
vcpu->arch.sie_block->lctl |= (LCTL_CR0 | LCTL_CR9 |
arch/s390/kvm/interrupt.c
383
vcpu->arch.sie_block->ictl |= (ICTL_STCTL | ICTL_PINT);
arch/s390/kvm/interrupt.c
394
vcpu->arch.sie_block->lctl |= LCTL_CR6;
arch/s390/kvm/interrupt.c
404
vcpu->arch.sie_block->lctl |= LCTL_CR0;
arch/s390/kvm/interrupt.c
412
vcpu->arch.sie_block->ictl |= ICTL_LPSW;
arch/s390/kvm/interrupt.c
414
vcpu->arch.sie_block->lctl |= LCTL_CR14;
arch/s390/kvm/interrupt.c
434
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
441
vcpu->arch.sie_block->iictl = IICTL_CODE_EXT;
arch/s390/kvm/interrupt.c
442
vcpu->arch.sie_block->eic = EXT_IRQ_CPU_TIMER;
arch/s390/kvm/interrupt.c
448
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
450
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
458
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
465
vcpu->arch.sie_block->iictl = IICTL_CODE_EXT;
arch/s390/kvm/interrupt.c
466
vcpu->arch.sie_block->eic = EXT_IRQ_CLK_COMP;
arch/s390/kvm/interrupt.c
47
struct esca_block *sca = vcpu->kvm->arch.sca;
arch/s390/kvm/interrupt.c
472
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
474
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
482
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
501
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
503
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
524
vcpu->arch.sie_block->iictl = IICTL_CODE_MCHK;
arch/s390/kvm/interrupt.c
525
vcpu->arch.sie_block->mcic = mchk->mcic;
arch/s390/kvm/interrupt.c
526
vcpu->arch.sie_block->faddr = mchk->failing_storage_address;
arch/s390/kvm/interrupt.c
527
vcpu->arch.sie_block->edc = mchk->ext_damage_code;
arch/s390/kvm/interrupt.c
535
if (cpu_has_gs() && vcpu->arch.gs_enabled)
arch/s390/kvm/interrupt.c
582
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
584
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
599
rc |= put_guest_lc(vcpu, vcpu->arch.sie_block->todpr,
arch/s390/kvm/interrupt.c
603
rc |= put_guest_lc(vcpu, vcpu->arch.sie_block->ckc >> 8,
arch/s390/kvm/interrupt.c
608
&vcpu->arch.sie_block->gcr, 128);
arch/s390/kvm/interrupt.c
622
struct kvm_s390_float_interrupt *fi = &vcpu->kvm->arch.float_int;
arch/s390/kvm/interrupt.c
623
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
63
struct esca_block *sca = vcpu->kvm->arch.sca;
arch/s390/kvm/interrupt.c
673
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
681
vcpu->arch.sie_block->iictl = IICTL_CODE_RESTART;
arch/s390/kvm/interrupt.c
685
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
687
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
695
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
715
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
731
vcpu->arch.sie_block->iictl = IICTL_CODE_EXT;
arch/s390/kvm/interrupt.c
732
vcpu->arch.sie_block->eic = EXT_IRQ_EMERGENCY_SIG;
arch/s390/kvm/interrupt.c
733
vcpu->arch.sie_block->extcpuaddr = cpu_addr;
arch/s390/kvm/interrupt.c
741
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
743
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
749
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
765
vcpu->arch.sie_block->iictl = IICTL_CODE_EXT;
arch/s390/kvm/interrupt.c
766
vcpu->arch.sie_block->eic = EXT_IRQ_EXTERNAL_CALL;
arch/s390/kvm/interrupt.c
767
vcpu->arch.sie_block->extcpuaddr = extcall.code;
arch/s390/kvm/interrupt.c
775
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
776
rc |= read_guest_lc(vcpu, __LC_EXT_NEW_PSW, &vcpu->arch.sie_block->gpsw,
arch/s390/kvm/interrupt.c
785
vcpu->arch.sie_block->iictl = IICTL_CODE_SPECIFICATION;
arch/s390/kvm/interrupt.c
788
vcpu->arch.sie_block->iictl = IICTL_CODE_OPERAND;
arch/s390/kvm/interrupt.c
798
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
arch/s390/kvm/interrupt.c
86
struct esca_block *sca = vcpu->kvm->arch.sca;
arch/s390/kvm/interrupt.c
904
rc |= put_guest_lc(vcpu, vcpu->arch.sie_block->gbea,
arch/s390/kvm/interrupt.c
908
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
910
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
922
vcpu->arch.sie_block->iictl = IICTL_CODE_EXT;
arch/s390/kvm/interrupt.c
923
vcpu->arch.sie_block->eic = EXT_IRQ_SERVICE_SIG;
arch/s390/kvm/interrupt.c
924
vcpu->arch.sie_block->eiparams = parm;
arch/s390/kvm/interrupt.c
931
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
933
&vcpu->arch.sie_block->gpsw, sizeof(psw_t));
arch/s390/kvm/interrupt.c
942
struct kvm_s390_float_interrupt *fi = &vcpu->kvm->arch.float_int;
arch/s390/kvm/interrupt.c
973
struct kvm_s390_float_interrupt *fi = &vcpu->kvm->arch.float_int;
arch/s390/kvm/interrupt.c
98
return !(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_EXT);
arch/s390/kvm/interrupt.c
997
struct kvm_s390_float_interrupt *fi = &vcpu->kvm->arch.float_int;
arch/s390/kvm/kvm-s390.c
1010
if (kvm->arch.mem_limit != KVM_S390_NO_MEM_LIMIT &&
arch/s390/kvm/kvm-s390.c
1011
new_limit > kvm->arch.mem_limit)
arch/s390/kvm/kvm-s390.c
1019
ret = gmap_set_limit(kvm->arch.gmap, gpa_to_gfn(new_limit));
arch/s390/kvm/kvm-s390.c
1022
(void *)kvm->arch.gmap->asce.val);
arch/s390/kvm/kvm-s390.c
1060
kvm->arch.crypto.crycb->aes_wrapping_key_mask,
arch/s390/kvm/kvm-s390.c
1061
sizeof(kvm->arch.crypto.crycb->aes_wrapping_key_mask));
arch/s390/kvm/kvm-s390.c
1062
kvm->arch.crypto.aes_kw = 1;
arch/s390/kvm/kvm-s390.c
1071
kvm->arch.crypto.crycb->dea_wrapping_key_mask,
arch/s390/kvm/kvm-s390.c
1072
sizeof(kvm->arch.crypto.crycb->dea_wrapping_key_mask));
arch/s390/kvm/kvm-s390.c
1073
kvm->arch.crypto.dea_kw = 1;
arch/s390/kvm/kvm-s390.c
1081
kvm->arch.crypto.aes_kw = 0;
arch/s390/kvm/kvm-s390.c
1082
memset(kvm->arch.crypto.crycb->aes_wrapping_key_mask, 0,
arch/s390/kvm/kvm-s390.c
1083
sizeof(kvm->arch.crypto.crycb->aes_wrapping_key_mask));
arch/s390/kvm/kvm-s390.c
1091
kvm->arch.crypto.dea_kw = 0;
arch/s390/kvm/kvm-s390.c
1092
memset(kvm->arch.crypto.crycb->dea_wrapping_key_mask, 0,
arch/s390/kvm/kvm-s390.c
1093
sizeof(kvm->arch.crypto.crycb->dea_wrapping_key_mask));
arch/s390/kvm/kvm-s390.c
1101
kvm->arch.crypto.apie = 1;
arch/s390/kvm/kvm-s390.c
1108
kvm->arch.crypto.apie = 0;
arch/s390/kvm/kvm-s390.c
1123
if (!vcpu->kvm->arch.use_zpci_interp)
arch/s390/kvm/kvm-s390.c
1126
vcpu->arch.sie_block->ecb2 |= ECB2_ZPCI_LSI;
arch/s390/kvm/kvm-s390.c
1127
vcpu->arch.sie_block->ecb3 |= ECB3_AISII + ECB3_AISI;
arch/s390/kvm/kvm-s390.c
1144
kvm->arch.use_zpci_interp = 1;
arch/s390/kvm/kvm-s390.c
1177
if (kvm->arch.migration_mode)
arch/s390/kvm/kvm-s390.c
1183
if (!kvm->arch.use_cmma) {
arch/s390/kvm/kvm-s390.c
1184
kvm->arch.migration_mode = 1;
arch/s390/kvm/kvm-s390.c
1193
gmap_set_cmma_all_dirty(kvm->arch.gmap);
arch/s390/kvm/kvm-s390.c
1194
atomic64_set(&kvm->arch.cmma_dirty_pages, ram_pages);
arch/s390/kvm/kvm-s390.c
1195
kvm->arch.migration_mode = 1;
arch/s390/kvm/kvm-s390.c
1207
if (!kvm->arch.migration_mode)
arch/s390/kvm/kvm-s390.c
1209
kvm->arch.migration_mode = 0;
arch/s390/kvm/kvm-s390.c
1210
if (kvm->arch.use_cmma)
arch/s390/kvm/kvm-s390.c
1239
u64 mig = kvm->arch.migration_mode;
arch/s390/kvm/kvm-s390.c
1342
gtod->tod = clk.tod + kvm->arch.epoch;
arch/s390/kvm/kvm-s390.c
1345
gtod->epoch_idx = clk.ei + kvm->arch.epdx;
arch/s390/kvm/kvm-s390.c
1433
kvm->arch.model.cpuid = proc->cpuid;
arch/s390/kvm/kvm-s390.c
1438
kvm->arch.model.ibc = unblocked_ibc;
arch/s390/kvm/kvm-s390.c
1440
kvm->arch.model.ibc = lowest_ibc;
arch/s390/kvm/kvm-s390.c
1442
kvm->arch.model.ibc = proc->ibc;
arch/s390/kvm/kvm-s390.c
1444
memcpy(kvm->arch.model.fac_list, proc->fac_list,
arch/s390/kvm/kvm-s390.c
1447
kvm->arch.model.ibc,
arch/s390/kvm/kvm-s390.c
1448
kvm->arch.model.cpuid);
arch/s390/kvm/kvm-s390.c
1450
kvm->arch.model.fac_list[0],
arch/s390/kvm/kvm-s390.c
1451
kvm->arch.model.fac_list[1],
arch/s390/kvm/kvm-s390.c
1452
kvm->arch.model.fac_list[2]);
arch/s390/kvm/kvm-s390.c
1478
bitmap_from_arr64(kvm->arch.cpu_feat, data.feat, KVM_S390_VM_CPU_FEAT_NR_BITS);
arch/s390/kvm/kvm-s390.c
1496
if (copy_from_user(&kvm->arch.model.subfuncs, (void __user *)attr->addr,
arch/s390/kvm/kvm-s390.c
1504
((unsigned long *) &kvm->arch.model.subfuncs.plo)[0],
arch/s390/kvm/kvm-s390.c
1505
((unsigned long *) &kvm->arch.model.subfuncs.plo)[1],
arch/s390/kvm/kvm-s390.c
1506
((unsigned long *) &kvm->arch.model.subfuncs.plo)[2],
arch/s390/kvm/kvm-s390.c
1507
((unsigned long *) &kvm->arch.model.subfuncs.plo)[3]);
arch/s390/kvm/kvm-s390.c
1509
((unsigned long *) &kvm->arch.model.subfuncs.ptff)[0],
arch/s390/kvm/kvm-s390.c
1510
((unsigned long *) &kvm->arch.model.subfuncs.ptff)[1]);
arch/s390/kvm/kvm-s390.c
1512
((unsigned long *) &kvm->arch.model.subfuncs.kmac)[0],
arch/s390/kvm/kvm-s390.c
1513
((unsigned long *) &kvm->arch.model.subfuncs.kmac)[1]);
arch/s390/kvm/kvm-s390.c
1515
((unsigned long *) &kvm->arch.model.subfuncs.kmc)[0],
arch/s390/kvm/kvm-s390.c
1516
((unsigned long *) &kvm->arch.model.subfuncs.kmc)[1]);
arch/s390/kvm/kvm-s390.c
1518
((unsigned long *) &kvm->arch.model.subfuncs.km)[0],
arch/s390/kvm/kvm-s390.c
1519
((unsigned long *) &kvm->arch.model.subfuncs.km)[1]);
arch/s390/kvm/kvm-s390.c
1521
((unsigned long *) &kvm->arch.model.subfuncs.kimd)[0],
arch/s390/kvm/kvm-s390.c
1522
((unsigned long *) &kvm->arch.model.subfuncs.kimd)[1]);
arch/s390/kvm/kvm-s390.c
1524
((unsigned long *) &kvm->arch.model.subfuncs.klmd)[0],
arch/s390/kvm/kvm-s390.c
1525
((unsigned long *) &kvm->arch.model.subfuncs.klmd)[1]);
arch/s390/kvm/kvm-s390.c
1527
((unsigned long *) &kvm->arch.model.subfuncs.pckmo)[0],
arch/s390/kvm/kvm-s390.c
1528
((unsigned long *) &kvm->arch.model.subfuncs.pckmo)[1]);
arch/s390/kvm/kvm-s390.c
1530
((unsigned long *) &kvm->arch.model.subfuncs.kmctr)[0],
arch/s390/kvm/kvm-s390.c
1531
((unsigned long *) &kvm->arch.model.subfuncs.kmctr)[1]);
arch/s390/kvm/kvm-s390.c
1533
((unsigned long *) &kvm->arch.model.subfuncs.kmf)[0],
arch/s390/kvm/kvm-s390.c
1534
((unsigned long *) &kvm->arch.model.subfuncs.kmf)[1]);
arch/s390/kvm/kvm-s390.c
1536
((unsigned long *) &kvm->arch.model.subfuncs.kmo)[0],
arch/s390/kvm/kvm-s390.c
1537
((unsigned long *) &kvm->arch.model.subfuncs.kmo)[1]);
arch/s390/kvm/kvm-s390.c
1539
((unsigned long *) &kvm->arch.model.subfuncs.pcc)[0],
arch/s390/kvm/kvm-s390.c
1540
((unsigned long *) &kvm->arch.model.subfuncs.pcc)[1]);
arch/s390/kvm/kvm-s390.c
1542
((unsigned long *) &kvm->arch.model.subfuncs.ppno)[0],
arch/s390/kvm/kvm-s390.c
1543
((unsigned long *) &kvm->arch.model.subfuncs.ppno)[1]);
arch/s390/kvm/kvm-s390.c
1545
((unsigned long *) &kvm->arch.model.subfuncs.kma)[0],
arch/s390/kvm/kvm-s390.c
1546
((unsigned long *) &kvm->arch.model.subfuncs.kma)[1]);
arch/s390/kvm/kvm-s390.c
1548
((unsigned long *) &kvm->arch.model.subfuncs.kdsa)[0],
arch/s390/kvm/kvm-s390.c
1549
((unsigned long *) &kvm->arch.model.subfuncs.kdsa)[1]);
arch/s390/kvm/kvm-s390.c
1551
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[0],
arch/s390/kvm/kvm-s390.c
1552
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[1],
arch/s390/kvm/kvm-s390.c
1553
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[2],
arch/s390/kvm/kvm-s390.c
1554
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[3]);
arch/s390/kvm/kvm-s390.c
1556
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[0],
arch/s390/kvm/kvm-s390.c
1557
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[1],
arch/s390/kvm/kvm-s390.c
1558
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[2],
arch/s390/kvm/kvm-s390.c
1559
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[3]);
arch/s390/kvm/kvm-s390.c
1592
kvm->arch.model.uv_feat_guest.feat = data;
arch/s390/kvm/kvm-s390.c
1631
proc->cpuid = kvm->arch.model.cpuid;
arch/s390/kvm/kvm-s390.c
1632
proc->ibc = kvm->arch.model.ibc;
arch/s390/kvm/kvm-s390.c
1633
memcpy(&proc->fac_list, kvm->arch.model.fac_list,
arch/s390/kvm/kvm-s390.c
1636
kvm->arch.model.ibc,
arch/s390/kvm/kvm-s390.c
1637
kvm->arch.model.cpuid);
arch/s390/kvm/kvm-s390.c
1639
kvm->arch.model.fac_list[0],
arch/s390/kvm/kvm-s390.c
1640
kvm->arch.model.fac_list[1],
arch/s390/kvm/kvm-s390.c
1641
kvm->arch.model.fac_list[2]);
arch/s390/kvm/kvm-s390.c
1661
memcpy(&mach->fac_mask, kvm->arch.model.fac_mask,
arch/s390/kvm/kvm-s390.c
1666
kvm->arch.model.ibc,
arch/s390/kvm/kvm-s390.c
1667
kvm->arch.model.cpuid);
arch/s390/kvm/kvm-s390.c
1688
bitmap_to_arr64(data.feat, kvm->arch.cpu_feat, KVM_S390_VM_CPU_FEAT_NR_BITS);
arch/s390/kvm/kvm-s390.c
1716
if (copy_to_user((void __user *)attr->addr, &kvm->arch.model.subfuncs,
arch/s390/kvm/kvm-s390.c
1721
((unsigned long *) &kvm->arch.model.subfuncs.plo)[0],
arch/s390/kvm/kvm-s390.c
1722
((unsigned long *) &kvm->arch.model.subfuncs.plo)[1],
arch/s390/kvm/kvm-s390.c
1723
((unsigned long *) &kvm->arch.model.subfuncs.plo)[2],
arch/s390/kvm/kvm-s390.c
1724
((unsigned long *) &kvm->arch.model.subfuncs.plo)[3]);
arch/s390/kvm/kvm-s390.c
1726
((unsigned long *) &kvm->arch.model.subfuncs.ptff)[0],
arch/s390/kvm/kvm-s390.c
1727
((unsigned long *) &kvm->arch.model.subfuncs.ptff)[1]);
arch/s390/kvm/kvm-s390.c
1729
((unsigned long *) &kvm->arch.model.subfuncs.kmac)[0],
arch/s390/kvm/kvm-s390.c
1730
((unsigned long *) &kvm->arch.model.subfuncs.kmac)[1]);
arch/s390/kvm/kvm-s390.c
1732
((unsigned long *) &kvm->arch.model.subfuncs.kmc)[0],
arch/s390/kvm/kvm-s390.c
1733
((unsigned long *) &kvm->arch.model.subfuncs.kmc)[1]);
arch/s390/kvm/kvm-s390.c
1735
((unsigned long *) &kvm->arch.model.subfuncs.km)[0],
arch/s390/kvm/kvm-s390.c
1736
((unsigned long *) &kvm->arch.model.subfuncs.km)[1]);
arch/s390/kvm/kvm-s390.c
1738
((unsigned long *) &kvm->arch.model.subfuncs.kimd)[0],
arch/s390/kvm/kvm-s390.c
1739
((unsigned long *) &kvm->arch.model.subfuncs.kimd)[1]);
arch/s390/kvm/kvm-s390.c
1741
((unsigned long *) &kvm->arch.model.subfuncs.klmd)[0],
arch/s390/kvm/kvm-s390.c
1742
((unsigned long *) &kvm->arch.model.subfuncs.klmd)[1]);
arch/s390/kvm/kvm-s390.c
1744
((unsigned long *) &kvm->arch.model.subfuncs.pckmo)[0],
arch/s390/kvm/kvm-s390.c
1745
((unsigned long *) &kvm->arch.model.subfuncs.pckmo)[1]);
arch/s390/kvm/kvm-s390.c
1747
((unsigned long *) &kvm->arch.model.subfuncs.kmctr)[0],
arch/s390/kvm/kvm-s390.c
1748
((unsigned long *) &kvm->arch.model.subfuncs.kmctr)[1]);
arch/s390/kvm/kvm-s390.c
1750
((unsigned long *) &kvm->arch.model.subfuncs.kmf)[0],
arch/s390/kvm/kvm-s390.c
1751
((unsigned long *) &kvm->arch.model.subfuncs.kmf)[1]);
arch/s390/kvm/kvm-s390.c
1753
((unsigned long *) &kvm->arch.model.subfuncs.kmo)[0],
arch/s390/kvm/kvm-s390.c
1754
((unsigned long *) &kvm->arch.model.subfuncs.kmo)[1]);
arch/s390/kvm/kvm-s390.c
1756
((unsigned long *) &kvm->arch.model.subfuncs.pcc)[0],
arch/s390/kvm/kvm-s390.c
1757
((unsigned long *) &kvm->arch.model.subfuncs.pcc)[1]);
arch/s390/kvm/kvm-s390.c
1759
((unsigned long *) &kvm->arch.model.subfuncs.ppno)[0],
arch/s390/kvm/kvm-s390.c
1760
((unsigned long *) &kvm->arch.model.subfuncs.ppno)[1]);
arch/s390/kvm/kvm-s390.c
1762
((unsigned long *) &kvm->arch.model.subfuncs.kma)[0],
arch/s390/kvm/kvm-s390.c
1763
((unsigned long *) &kvm->arch.model.subfuncs.kma)[1]);
arch/s390/kvm/kvm-s390.c
1765
((unsigned long *) &kvm->arch.model.subfuncs.kdsa)[0],
arch/s390/kvm/kvm-s390.c
1766
((unsigned long *) &kvm->arch.model.subfuncs.kdsa)[1]);
arch/s390/kvm/kvm-s390.c
1768
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[0],
arch/s390/kvm/kvm-s390.c
1769
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[1],
arch/s390/kvm/kvm-s390.c
1770
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[2],
arch/s390/kvm/kvm-s390.c
1771
((unsigned long *) &kvm->arch.model.subfuncs.sortl)[3]);
arch/s390/kvm/kvm-s390.c
1773
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[0],
arch/s390/kvm/kvm-s390.c
1774
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[1],
arch/s390/kvm/kvm-s390.c
1775
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[2],
arch/s390/kvm/kvm-s390.c
1776
((unsigned long *) &kvm->arch.model.subfuncs.dfltcc)[3]);
arch/s390/kvm/kvm-s390.c
1858
unsigned long feat = kvm->arch.model.uv_feat_guest.feat;
arch/s390/kvm/kvm-s390.c
1929
sca = kvm->arch.sca;
arch/s390/kvm/kvm-s390.c
1955
topo = kvm->arch.sca->utility.mtcr;
arch/s390/kvm/kvm-s390.c
2106
if (!uses_skeys(kvm->arch.gmap))
arch/s390/kvm/kvm-s390.c
2119
r = dat_get_storage_key(kvm->arch.gmap->asce,
arch/s390/kvm/kvm-s390.c
2162
r = gmap_enable_skeys(kvm->arch.gmap);
arch/s390/kvm/kvm-s390.c
2186
r = dat_set_storage_key(mc, kvm->arch.gmap->asce,
arch/s390/kvm/kvm-s390.c
2213
if (!kvm->arch.use_cmma)
arch/s390/kvm/kvm-s390.c
2220
if (!peek && !kvm->arch.migration_mode)
arch/s390/kvm/kvm-s390.c
2224
if (!args->count || !uses_cmm(kvm->arch.gmap)) {
arch/s390/kvm/kvm-s390.c
2229
if (!peek && !atomic64_read(&kvm->arch.cmma_dirty_pages)) {
arch/s390/kvm/kvm-s390.c
2240
ret = dat_peek_cmma(args->start_gfn, kvm->arch.gmap->asce, &args->count,
arch/s390/kvm/kvm-s390.c
2243
ret = dat_get_cmma(kvm->arch.gmap->asce, &args->start_gfn, &args->count,
arch/s390/kvm/kvm-s390.c
2244
values, &kvm->arch.cmma_dirty_pages);
arch/s390/kvm/kvm-s390.c
2247
if (kvm->arch.migration_mode)
arch/s390/kvm/kvm-s390.c
2248
args->remaining = atomic64_read(&kvm->arch.cmma_dirty_pages);
arch/s390/kvm/kvm-s390.c
2271
if (!kvm->arch.use_cmma)
arch/s390/kvm/kvm-s390.c
2301
r = dat_set_cmma_bits(mc, kvm->arch.gmap->asce, args->start_gfn,
arch/s390/kvm/kvm-s390.c
2306
set_bit(GMAP_FLAG_USES_CMM, &kvm->arch.gmap->flags);
arch/s390/kvm/kvm-s390.c
2446
if (kvm->arch.pv.dumping)
arch/s390/kvm/kvm-s390.c
2460
kvm->arch.pv.dumping = true;
arch/s390/kvm/kvm-s390.c
2468
if (!kvm->arch.pv.dumping)
arch/s390/kvm/kvm-s390.c
2481
if (!kvm->arch.pv.dumping)
arch/s390/kvm/kvm-s390.c
2523
set_bit(GMAP_FLAG_EXPORT_ON_UNMAP, &kvm->arch.gmap->flags);
arch/s390/kvm/kvm-s390.c
2538
set_bit(IRQ_PEND_EXT_SERVICE, &kvm->arch.float_int.masked_irqs);
arch/s390/kvm/kvm-s390.c
2557
clear_bit(IRQ_PEND_EXT_SERVICE, &kvm->arch.float_int.masked_irqs);
arch/s390/kvm/kvm-s390.c
2582
clear_bit(IRQ_PEND_EXT_SERVICE, &kvm->arch.float_int.masked_irqs);
arch/s390/kvm/kvm-s390.c
2864
if (kvm->arch.use_irqchip)
arch/s390/kvm/kvm-s390.c
3039
kvm->arch.crypto.crycbd = virt_to_phys(kvm->arch.crypto.crycb);
arch/s390/kvm/kvm-s390.c
3042
kvm->arch.crypto.crycbd &= ~(CRYCB_FORMAT_MASK);
arch/s390/kvm/kvm-s390.c
3049
kvm->arch.crypto.crycbd |= CRYCB_FORMAT2;
arch/s390/kvm/kvm-s390.c
3051
kvm->arch.crypto.crycbd |= CRYCB_FORMAT1;
arch/s390/kvm/kvm-s390.c
3072
struct kvm_s390_crypto_cb *crycb = kvm->arch.crypto.crycb;
arch/s390/kvm/kvm-s390.c
3076
switch (kvm->arch.crypto.crycbd & CRYCB_FORMAT_MASK) {
arch/s390/kvm/kvm-s390.c
311
kvm_clock_sync_scb(vcpu->arch.sie_block, *delta);
arch/s390/kvm/kvm-s390.c
3123
memset(&kvm->arch.crypto.crycb->apcb0, 0,
arch/s390/kvm/kvm-s390.c
3124
sizeof(kvm->arch.crypto.crycb->apcb0));
arch/s390/kvm/kvm-s390.c
3125
memset(&kvm->arch.crypto.crycb->apcb1, 0,
arch/s390/kvm/kvm-s390.c
3126
sizeof(kvm->arch.crypto.crycb->apcb1));
arch/s390/kvm/kvm-s390.c
313
kvm->arch.epoch = vcpu->arch.sie_block->epoch;
arch/s390/kvm/kvm-s390.c
314
kvm->arch.epdx = vcpu->arch.sie_block->epdx;
arch/s390/kvm/kvm-s390.c
3146
kvm->arch.crypto.crycb = &kvm->arch.sie_page2->crycb;
arch/s390/kvm/kvm-s390.c
3148
init_rwsem(&kvm->arch.crypto.pqap_hook_rwsem);
arch/s390/kvm/kvm-s390.c
3154
kvm->arch.crypto.aes_kw = 1;
arch/s390/kvm/kvm-s390.c
3155
kvm->arch.crypto.dea_kw = 1;
arch/s390/kvm/kvm-s390.c
3156
get_random_bytes(kvm->arch.crypto.crycb->aes_wrapping_key_mask,
arch/s390/kvm/kvm-s390.c
3157
sizeof(kvm->arch.crypto.crycb->aes_wrapping_key_mask));
arch/s390/kvm/kvm-s390.c
3158
get_random_bytes(kvm->arch.crypto.crycb->dea_wrapping_key_mask,
arch/s390/kvm/kvm-s390.c
3159
sizeof(kvm->arch.crypto.crycb->dea_wrapping_key_mask));
arch/s390/kvm/kvm-s390.c
316
if (vcpu->arch.cputm_enabled)
arch/s390/kvm/kvm-s390.c
3164
free_pages_exact(kvm->arch.sca, sizeof(*kvm->arch.sca));
arch/s390/kvm/kvm-s390.c
3165
kvm->arch.sca = NULL;
arch/s390/kvm/kvm-s390.c
317
vcpu->arch.cputm_start += *delta;
arch/s390/kvm/kvm-s390.c
318
if (vcpu->arch.vsie_block)
arch/s390/kvm/kvm-s390.c
3182
mutex_init(&kvm->arch.pv.import_lock);
arch/s390/kvm/kvm-s390.c
319
kvm_clock_sync_scb(vcpu->arch.vsie_block,
arch/s390/kvm/kvm-s390.c
3200
kvm->arch.sca = alloc_pages_exact(sizeof(*kvm->arch.sca), alloc_flags);
arch/s390/kvm/kvm-s390.c
3202
if (!kvm->arch.sca)
arch/s390/kvm/kvm-s390.c
3207
kvm->arch.dbf = debug_register(debug_name, 32, 1, 7 * sizeof(long));
arch/s390/kvm/kvm-s390.c
3208
if (!kvm->arch.dbf)
arch/s390/kvm/kvm-s390.c
3212
kvm->arch.sie_page2 =
arch/s390/kvm/kvm-s390.c
3214
if (!kvm->arch.sie_page2)
arch/s390/kvm/kvm-s390.c
3217
kvm->arch.sie_page2->kvm = kvm;
arch/s390/kvm/kvm-s390.c
3218
kvm->arch.model.fac_list = kvm->arch.sie_page2->fac_list;
arch/s390/kvm/kvm-s390.c
3221
kvm->arch.model.fac_mask[i] = stfle_fac_list[i] &
arch/s390/kvm/kvm-s390.c
3224
kvm->arch.model.fac_list[i] = stfle_fac_list[i] &
arch/s390/kvm/kvm-s390.c
3227
kvm->arch.model.subfuncs = kvm_s390_available_subfunc;
arch/s390/kvm/kvm-s390.c
3230
set_kvm_facility(kvm->arch.model.fac_mask, 138);
arch/s390/kvm/kvm-s390.c
3231
set_kvm_facility(kvm->arch.model.fac_list, 138);
arch/s390/kvm/kvm-s390.c
3233
set_kvm_facility(kvm->arch.model.fac_mask, 74);
arch/s390/kvm/kvm-s390.c
3234
set_kvm_facility(kvm->arch.model.fac_list, 74);
arch/s390/kvm/kvm-s390.c
3236
set_kvm_facility(kvm->arch.model.fac_mask, 147);
arch/s390/kvm/kvm-s390.c
3237
set_kvm_facility(kvm->arch.model.fac_list, 147);
arch/s390/kvm/kvm-s390.c
3241
set_kvm_facility(kvm->arch.model.fac_mask, 65);
arch/s390/kvm/kvm-s390.c
3243
kvm->arch.model.cpuid = kvm_s390_get_initial_cpuid();
arch/s390/kvm/kvm-s390.c
3244
kvm->arch.model.ibc = sclp.ibc & 0x0fff;
arch/s390/kvm/kvm-s390.c
3246
kvm->arch.model.uv_feat_guest.feat = 0;
arch/s390/kvm/kvm-s390.c
3257
mutex_init(&kvm->arch.float_int.ais_lock);
arch/s390/kvm/kvm-s390.c
3258
spin_lock_init(&kvm->arch.float_int.lock);
arch/s390/kvm/kvm-s390.c
3260
INIT_LIST_HEAD(&kvm->arch.float_int.lists[i]);
arch/s390/kvm/kvm-s390.c
3261
init_waitqueue_head(&kvm->arch.ipte_wq);
arch/s390/kvm/kvm-s390.c
3262
mutex_init(&kvm->arch.ipte_mutex);
arch/s390/kvm/kvm-s390.c
3264
debug_register_view(kvm->arch.dbf, &debug_sprintf_view);
arch/s390/kvm/kvm-s390.c
3267
kvm->arch.mem_limit = type & KVM_VM_S390_UCONTROL ? KVM_S390_NO_MEM_LIMIT : sclp.hamax + 1;
arch/s390/kvm/kvm-s390.c
3268
kvm->arch.gmap = gmap_new(kvm, gpa_to_gfn(kvm->arch.mem_limit));
arch/s390/kvm/kvm-s390.c
3269
if (!kvm->arch.gmap)
arch/s390/kvm/kvm-s390.c
3271
clear_bit(GMAP_FLAG_PFAULT_ENABLED, &kvm->arch.gmap->flags);
arch/s390/kvm/kvm-s390.c
3286
set_bit(GMAP_FLAG_IS_UCONTROL, &kvm->arch.gmap->flags);
arch/s390/kvm/kvm-s390.c
3288
struct crst_table *table = dereference_asce(kvm->arch.gmap->asce);
arch/s390/kvm/kvm-s390.c
3293
kvm->arch.use_pfmfi = sclp.has_pfmfi;
arch/s390/kvm/kvm-s390.c
3294
kvm->arch.use_skf = sclp.has_skey;
arch/s390/kvm/kvm-s390.c
3295
spin_lock_init(&kvm->arch.start_stop_lock);
arch/s390/kvm/kvm-s390.c
3299
INIT_LIST_HEAD(&kvm->arch.pv.need_cleanup);
arch/s390/kvm/kvm-s390.c
3300
kvm->arch.pv.set_aside = NULL;
arch/s390/kvm/kvm-s390.c
3305
free_page((unsigned long)kvm->arch.sie_page2);
arch/s390/kvm/kvm-s390.c
3306
debug_unregister(kvm->arch.dbf);
arch/s390/kvm/kvm-s390.c
3325
scoped_guard(spinlock, &vcpu->kvm->arch.gmap->children_lock)
arch/s390/kvm/kvm-s390.c
3326
gmap_remove_child(vcpu->arch.gmap);
arch/s390/kvm/kvm-s390.c
3327
vcpu->arch.gmap = gmap_put(vcpu->arch.gmap);
arch/s390/kvm/kvm-s390.c
3330
if (vcpu->kvm->arch.use_cmma)
arch/s390/kvm/kvm-s390.c
3335
free_page((unsigned long)(vcpu->arch.sie_block));
arch/s390/kvm/kvm-s390.c
3336
kvm_s390_free_mmu_cache(vcpu->arch.mc);
arch/s390/kvm/kvm-s390.c
3358
if (kvm->arch.pv.mmu_notifier.ops)
arch/s390/kvm/kvm-s390.c
3359
mmu_notifier_unregister(&kvm->arch.pv.mmu_notifier, kvm->mm);
arch/s390/kvm/kvm-s390.c
3361
debug_unregister(kvm->arch.dbf);
arch/s390/kvm/kvm-s390.c
3362
free_page((unsigned long)kvm->arch.sie_page2);
arch/s390/kvm/kvm-s390.c
3366
kvm->arch.gmap = gmap_put(kvm->arch.gmap);
arch/s390/kvm/kvm-s390.c
3373
struct esca_block *sca = vcpu->kvm->arch.sca;
arch/s390/kvm/kvm-s390.c
3384
struct esca_block *sca = vcpu->kvm->arch.sca;
arch/s390/kvm/kvm-s390.c
3388
vcpu->arch.sie_block->scaoh = sca_phys >> 32;
arch/s390/kvm/kvm-s390.c
3389
vcpu->arch.sie_block->scaol = sca_phys & ESCA_SCAOL_MASK;
arch/s390/kvm/kvm-s390.c
3390
vcpu->arch.sie_block->ecb2 |= ECB2_ESCA;
arch/s390/kvm/kvm-s390.c
3396
sca->cpu[vcpu->vcpu_id].sda = virt_to_phys(vcpu->arch.sie_block);
arch/s390/kvm/kvm-s390.c
3410
WARN_ON_ONCE(vcpu->arch.cputm_start != 0);
arch/s390/kvm/kvm-s390.c
3411
raw_write_seqcount_begin(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3412
vcpu->arch.cputm_start = get_tod_clock_fast();
arch/s390/kvm/kvm-s390.c
3413
raw_write_seqcount_end(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3419
WARN_ON_ONCE(vcpu->arch.cputm_start == 0);
arch/s390/kvm/kvm-s390.c
3420
raw_write_seqcount_begin(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3421
vcpu->arch.sie_block->cputm -= get_tod_clock_fast() - vcpu->arch.cputm_start;
arch/s390/kvm/kvm-s390.c
3422
vcpu->arch.cputm_start = 0;
arch/s390/kvm/kvm-s390.c
3423
raw_write_seqcount_end(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3429
WARN_ON_ONCE(vcpu->arch.cputm_enabled);
arch/s390/kvm/kvm-s390.c
3430
vcpu->arch.cputm_enabled = true;
arch/s390/kvm/kvm-s390.c
3437
WARN_ON_ONCE(!vcpu->arch.cputm_enabled);
arch/s390/kvm/kvm-s390.c
3439
vcpu->arch.cputm_enabled = false;
arch/s390/kvm/kvm-s390.c
3460
raw_write_seqcount_begin(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3461
if (vcpu->arch.cputm_enabled)
arch/s390/kvm/kvm-s390.c
3462
vcpu->arch.cputm_start = get_tod_clock_fast();
arch/s390/kvm/kvm-s390.c
3463
vcpu->arch.sie_block->cputm = cputm;
arch/s390/kvm/kvm-s390.c
3464
raw_write_seqcount_end(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3474
if (unlikely(!vcpu->arch.cputm_enabled))
arch/s390/kvm/kvm-s390.c
3475
return vcpu->arch.sie_block->cputm;
arch/s390/kvm/kvm-s390.c
3479
seq = raw_read_seqcount(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3485
value = vcpu->arch.sie_block->cputm;
arch/s390/kvm/kvm-s390.c
3487
if (likely(vcpu->arch.cputm_start))
arch/s390/kvm/kvm-s390.c
3488
value -= get_tod_clock_fast() - vcpu->arch.cputm_start;
arch/s390/kvm/kvm-s390.c
3489
} while (read_seqcount_retry(&vcpu->arch.cputm_seqcount, seq & ~1));
arch/s390/kvm/kvm-s390.c
3498
if (vcpu->arch.cputm_enabled && !is_vcpu_idle(vcpu))
arch/s390/kvm/kvm-s390.c
3506
if (vcpu->arch.cputm_enabled && !is_vcpu_idle(vcpu))
arch/s390/kvm/kvm-s390.c
3516
vcpu->arch.sie_block->epoch = vcpu->kvm->arch.epoch;
arch/s390/kvm/kvm-s390.c
3517
vcpu->arch.sie_block->epdx = vcpu->kvm->arch.epdx;
arch/s390/kvm/kvm-s390.c
3521
vcpu->arch.gmap = vcpu->kvm->arch.gmap;
arch/s390/kvm/kvm-s390.c
3524
if (test_kvm_facility(vcpu->kvm, 74) || vcpu->kvm->arch.user_instr0)
arch/s390/kvm/kvm-s390.c
3525
vcpu->arch.sie_block->ictl |= ICTL_OPEREXC;
arch/s390/kvm/kvm-s390.c
3530
if (test_bit_inv(nr, (unsigned long *)&kvm->arch.model.subfuncs.pckmo) &&
arch/s390/kvm/kvm-s390.c
3560
if (!vcpu->kvm->arch.crypto.apie && !test_kvm_facility(vcpu->kvm, 76))
arch/s390/kvm/kvm-s390.c
3563
vcpu->arch.sie_block->crycbd = vcpu->kvm->arch.crypto.crycbd;
arch/s390/kvm/kvm-s390.c
3564
vcpu->arch.sie_block->ecb3 &= ~(ECB3_AES | ECB3_DEA);
arch/s390/kvm/kvm-s390.c
3565
vcpu->arch.sie_block->eca &= ~ECA_APIE;
arch/s390/kvm/kvm-s390.c
3566
vcpu->arch.sie_block->ecd &= ~(ECD_ECC | ECD_HMAC);
arch/s390/kvm/kvm-s390.c
3568
if (vcpu->kvm->arch.crypto.apie)
arch/s390/kvm/kvm-s390.c
3569
vcpu->arch.sie_block->eca |= ECA_APIE;
arch/s390/kvm/kvm-s390.c
3572
if (vcpu->kvm->arch.crypto.aes_kw) {
arch/s390/kvm/kvm-s390.c
3573
vcpu->arch.sie_block->ecb3 |= ECB3_AES;
arch/s390/kvm/kvm-s390.c
3576
vcpu->arch.sie_block->ecd |= ECD_ECC;
arch/s390/kvm/kvm-s390.c
3578
vcpu->arch.sie_block->ecd |= ECD_HMAC;
arch/s390/kvm/kvm-s390.c
3581
if (vcpu->kvm->arch.crypto.dea_kw)
arch/s390/kvm/kvm-s390.c
3582
vcpu->arch.sie_block->ecb3 |= ECB3_DEA;
arch/s390/kvm/kvm-s390.c
3587
free_page((unsigned long)phys_to_virt(vcpu->arch.sie_block->cbrlo));
arch/s390/kvm/kvm-s390.c
3588
vcpu->arch.sie_block->cbrlo = 0;
arch/s390/kvm/kvm-s390.c
3598
vcpu->arch.sie_block->cbrlo = virt_to_phys(cbrlo_page);
arch/s390/kvm/kvm-s390.c
3604
struct kvm_s390_cpu_model *model = &vcpu->kvm->arch.model;
arch/s390/kvm/kvm-s390.c
3606
vcpu->arch.sie_block->ibc = model->ibc;
arch/s390/kvm/kvm-s390.c
3608
vcpu->arch.sie_block->fac = virt_to_phys(model->fac_list);
arch/s390/kvm/kvm-s390.c
3616
atomic_set(&vcpu->arch.sie_block->cpuflags, CPUSTAT_ZARCH |
arch/s390/kvm/kvm-s390.c
3629
vcpu->arch.sie_block->ecb |= ECB_HOSTPROTINT;
arch/s390/kvm/kvm-s390.c
3631
vcpu->arch.sie_block->ecb |= ECB_SRSI;
arch/s390/kvm/kvm-s390.c
3633
vcpu->arch.sie_block->ecb |= ECB_PTF;
arch/s390/kvm/kvm-s390.c
3635
vcpu->arch.sie_block->ecb |= ECB_TE;
arch/s390/kvm/kvm-s390.c
3637
vcpu->arch.sie_block->ecb |= ECB_SPECI;
arch/s390/kvm/kvm-s390.c
3639
if (test_kvm_facility(vcpu->kvm, 8) && vcpu->kvm->arch.use_pfmfi)
arch/s390/kvm/kvm-s390.c
3640
vcpu->arch.sie_block->ecb2 |= ECB2_PFMFI;
arch/s390/kvm/kvm-s390.c
3642
vcpu->arch.sie_block->ecb2 |= ECB2_IEP;
arch/s390/kvm/kvm-s390.c
3643
vcpu->arch.sie_block->eca = ECA_MVPGI | ECA_PROTEXCI;
arch/s390/kvm/kvm-s390.c
3645
vcpu->arch.sie_block->eca |= ECA_CEI;
arch/s390/kvm/kvm-s390.c
3647
vcpu->arch.sie_block->eca |= ECA_IB;
arch/s390/kvm/kvm-s390.c
3649
vcpu->arch.sie_block->eca |= ECA_SII;
arch/s390/kvm/kvm-s390.c
3651
vcpu->arch.sie_block->eca |= ECA_SIGPI;
arch/s390/kvm/kvm-s390.c
3653
vcpu->arch.sie_block->eca |= ECA_VX;
arch/s390/kvm/kvm-s390.c
3654
vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
arch/s390/kvm/kvm-s390.c
3657
vcpu->arch.sie_block->ecd |= ECD_MEF;
arch/s390/kvm/kvm-s390.c
3659
vcpu->arch.sie_block->ecd |= ECD_ETOKENF;
arch/s390/kvm/kvm-s390.c
3660
if (vcpu->arch.sie_block->gd) {
arch/s390/kvm/kvm-s390.c
3661
vcpu->arch.sie_block->eca |= ECA_AIV;
arch/s390/kvm/kvm-s390.c
3663
vcpu->arch.sie_block->gd & 0x3, vcpu->vcpu_id);
arch/s390/kvm/kvm-s390.c
3665
vcpu->arch.sie_block->sdnxo = virt_to_phys(&vcpu->run->s.regs.sdnx) | SDNXC;
arch/s390/kvm/kvm-s390.c
3666
vcpu->arch.sie_block->riccbd = virt_to_phys(&vcpu->run->s.regs.riccb);
arch/s390/kvm/kvm-s390.c
3671
vcpu->arch.sie_block->ictl |= ICTL_ISKE | ICTL_SSKE | ICTL_RRBE;
arch/s390/kvm/kvm-s390.c
3673
if (vcpu->kvm->arch.use_cmma) {
arch/s390/kvm/kvm-s390.c
3678
hrtimer_setup(&vcpu->arch.ckc_timer, kvm_s390_idle_wakeup, CLOCK_MONOTONIC,
arch/s390/kvm/kvm-s390.c
3681
vcpu->arch.sie_block->hpid = HPID_KVM;
arch/s390/kvm/kvm-s390.c
3711
vcpu->arch.mc = kvm_s390_new_mmu_cache();
arch/s390/kvm/kvm-s390.c
3712
if (!vcpu->arch.mc)
arch/s390/kvm/kvm-s390.c
3716
kvm_s390_free_mmu_cache(vcpu->arch.mc);
arch/s390/kvm/kvm-s390.c
3717
vcpu->arch.mc = NULL;
arch/s390/kvm/kvm-s390.c
3721
vcpu->arch.sie_block = &sie_page->sie_block;
arch/s390/kvm/kvm-s390.c
3722
vcpu->arch.sie_block->itdba = virt_to_phys(&sie_page->itdb);
arch/s390/kvm/kvm-s390.c
3725
vcpu->arch.sie_block->mso = 0;
arch/s390/kvm/kvm-s390.c
3726
vcpu->arch.sie_block->msl = sclp.hamax;
arch/s390/kvm/kvm-s390.c
3728
vcpu->arch.sie_block->icpua = vcpu->vcpu_id;
arch/s390/kvm/kvm-s390.c
3729
spin_lock_init(&vcpu->arch.local_int.lock);
arch/s390/kvm/kvm-s390.c
3730
vcpu->arch.sie_block->gd = kvm_s390_get_gisa_desc(vcpu->kvm);
arch/s390/kvm/kvm-s390.c
3731
seqcount_init(&vcpu->arch.cputm_seqcount);
arch/s390/kvm/kvm-s390.c
3733
vcpu->arch.pfault_token = KVM_S390_PFAULT_TOKEN_INVALID;
arch/s390/kvm/kvm-s390.c
3742
vcpu->arch.acrs_loaded = false;
arch/s390/kvm/kvm-s390.c
3762
vcpu->arch.gmap = gmap_new_child(vcpu->kvm->arch.gmap, -1UL);
arch/s390/kvm/kvm-s390.c
3763
if (!vcpu->arch.gmap)
arch/s390/kvm/kvm-s390.c
3768
vcpu->vcpu_id, vcpu, vcpu->arch.sie_block);
arch/s390/kvm/kvm-s390.c
3769
trace_kvm_s390_create_vcpu(vcpu->vcpu_id, vcpu, vcpu->arch.sie_block);
arch/s390/kvm/kvm-s390.c
3780
gmap_remove_child(vcpu->arch.gmap);
arch/s390/kvm/kvm-s390.c
3781
vcpu->arch.gmap = gmap_put(vcpu->arch.gmap);
arch/s390/kvm/kvm-s390.c
3784
free_page((unsigned long)(vcpu->arch.sie_block));
arch/s390/kvm/kvm-s390.c
3790
clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask);
arch/s390/kvm/kvm-s390.c
3796
return !(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE);
arch/s390/kvm/kvm-s390.c
3801
atomic_or(PROG_BLOCK_SIE, &vcpu->arch.sie_block->prog20);
arch/s390/kvm/kvm-s390.c
3807
atomic_andnot(PROG_BLOCK_SIE, &vcpu->arch.sie_block->prog20);
arch/s390/kvm/kvm-s390.c
3812
atomic_or(PROG_REQUEST, &vcpu->arch.sie_block->prog20);
arch/s390/kvm/kvm-s390.c
3818
return atomic_read(&vcpu->arch.sie_block->prog20) &
arch/s390/kvm/kvm-s390.c
3824
atomic_andnot(PROG_REQUEST, &vcpu->arch.sie_block->prog20);
arch/s390/kvm/kvm-s390.c
3835
while (vcpu->arch.sie_block->prog0c & PROG_IN_SIE)
arch/s390/kvm/kvm-s390.c
3871
r = put_user(vcpu->arch.sie_block->todpr,
arch/s390/kvm/kvm-s390.c
3875
r = put_user(vcpu->arch.sie_block->epoch,
arch/s390/kvm/kvm-s390.c
3883
r = put_user(vcpu->arch.sie_block->ckc,
arch/s390/kvm/kvm-s390.c
3887
r = put_user(vcpu->arch.pfault_token,
arch/s390/kvm/kvm-s390.c
3891
r = put_user(vcpu->arch.pfault_compare,
arch/s390/kvm/kvm-s390.c
3895
r = put_user(vcpu->arch.pfault_select,
arch/s390/kvm/kvm-s390.c
3899
r = put_user(vcpu->arch.sie_block->pp,
arch/s390/kvm/kvm-s390.c
3903
r = put_user(vcpu->arch.sie_block->gbea,
arch/s390/kvm/kvm-s390.c
3921
r = get_user(vcpu->arch.sie_block->todpr,
arch/s390/kvm/kvm-s390.c
3925
r = get_user(vcpu->arch.sie_block->epoch,
arch/s390/kvm/kvm-s390.c
3934
r = get_user(vcpu->arch.sie_block->ckc,
arch/s390/kvm/kvm-s390.c
3938
r = get_user(vcpu->arch.pfault_token,
arch/s390/kvm/kvm-s390.c
3940
if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
arch/s390/kvm/kvm-s390.c
3944
r = get_user(vcpu->arch.pfault_compare,
arch/s390/kvm/kvm-s390.c
3948
r = get_user(vcpu->arch.pfault_select,
arch/s390/kvm/kvm-s390.c
3952
r = get_user(vcpu->arch.sie_block->pp,
arch/s390/kvm/kvm-s390.c
3956
r = get_user(vcpu->arch.sie_block->gbea,
arch/s390/kvm/kvm-s390.c
3968
vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_RI;
arch/s390/kvm/kvm-s390.c
3969
vcpu->arch.pfault_token = KVM_S390_PFAULT_TOKEN_INVALID;
arch/s390/kvm/kvm-s390.c
3987
vcpu->arch.sie_block->gpsw.mask = 0;
arch/s390/kvm/kvm-s390.c
3988
vcpu->arch.sie_block->gpsw.addr = 0;
arch/s390/kvm/kvm-s390.c
3991
vcpu->arch.sie_block->ckc = 0;
arch/s390/kvm/kvm-s390.c
3992
memset(vcpu->arch.sie_block->gcr, 0, sizeof(vcpu->arch.sie_block->gcr));
arch/s390/kvm/kvm-s390.c
3993
vcpu->arch.sie_block->gcr[0] = CR0_INITIAL_MASK;
arch/s390/kvm/kvm-s390.c
3994
vcpu->arch.sie_block->gcr[14] = CR14_INITIAL_MASK;
arch/s390/kvm/kvm-s390.c
4015
vcpu->arch.sie_block->gbea = 1;
arch/s390/kvm/kvm-s390.c
4016
vcpu->arch.sie_block->pp = 0;
arch/s390/kvm/kvm-s390.c
4017
vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
arch/s390/kvm/kvm-s390.c
4018
vcpu->arch.sie_block->todpr = 0;
arch/s390/kvm/kvm-s390.c
4060
memcpy(&vcpu->arch.sie_block->gcr, &sregs->crs, sizeof(sregs->crs));
arch/s390/kvm/kvm-s390.c
4072
memcpy(&sregs->crs, &vcpu->arch.sie_block->gcr, sizeof(sregs->crs));
arch/s390/kvm/kvm-s390.c
4159
vcpu->arch.guestdbg.last_bp = 0;
arch/s390/kvm/kvm-s390.c
4232
rc = gmap_ucas_translate(vcpu->arch.mc, vcpu->arch.gmap, gaddr);
arch/s390/kvm/kvm-s390.c
4261
rc = dat_set_prefix_notif_bit(vcpu->kvm->arch.gmap->asce, gfn);
arch/s390/kvm/kvm-s390.c
4290
vcpu->arch.sie_block->ihcpu = 0xffff;
arch/s390/kvm/kvm-s390.c
4311
vcpu->arch.sie_block->ictl |= ICTL_OPEREXC;
arch/s390/kvm/kvm-s390.c
4321
vcpu->arch.sie_block->ecb2 &= ~ECB2_CMMA;
arch/s390/kvm/kvm-s390.c
4330
if (vcpu->kvm->arch.use_cmma && uses_cmm(vcpu->arch.gmap))
arch/s390/kvm/kvm-s390.c
4331
vcpu->arch.sie_block->ecb2 |= ECB2_CMMA;
arch/s390/kvm/kvm-s390.c
4351
kvm->arch.epoch = gtod->tod - clk.tod;
arch/s390/kvm/kvm-s390.c
4352
kvm->arch.epdx = 0;
arch/s390/kvm/kvm-s390.c
4354
kvm->arch.epdx = gtod->epoch_idx - clk.ei;
arch/s390/kvm/kvm-s390.c
4355
if (kvm->arch.epoch > gtod->tod)
arch/s390/kvm/kvm-s390.c
4356
kvm->arch.epdx -= 1;
arch/s390/kvm/kvm-s390.c
4361
vcpu->arch.sie_block->epoch = kvm->arch.epoch;
arch/s390/kvm/kvm-s390.c
4362
vcpu->arch.sie_block->epdx = kvm->arch.epdx;
arch/s390/kvm/kvm-s390.c
4398
trace_kvm_s390_pfault_init(vcpu, work->arch.pfault_token);
arch/s390/kvm/kvm-s390.c
4399
__kvm_inject_pfault_token(vcpu, true, work->arch.pfault_token);
arch/s390/kvm/kvm-s390.c
4407
trace_kvm_s390_pfault_done(vcpu, work->arch.pfault_token);
arch/s390/kvm/kvm-s390.c
4408
__kvm_inject_pfault_token(vcpu, false, work->arch.pfault_token);
arch/s390/kvm/kvm-s390.c
4429
struct kvm_arch_async_pf arch;
arch/s390/kvm/kvm-s390.c
4431
if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
arch/s390/kvm/kvm-s390.c
4433
if ((vcpu->arch.sie_block->gpsw.mask & vcpu->arch.pfault_select) !=
arch/s390/kvm/kvm-s390.c
4434
vcpu->arch.pfault_compare)
arch/s390/kvm/kvm-s390.c
4440
if (!(vcpu->arch.sie_block->gcr[0] & CR0_SERVICE_SIGNAL_SUBMASK))
arch/s390/kvm/kvm-s390.c
4442
if (!pfault_enabled(vcpu->arch.gmap))
arch/s390/kvm/kvm-s390.c
4446
if (read_guest_real(vcpu, vcpu->arch.pfault_token, &arch.pfault_token, 8))
arch/s390/kvm/kvm-s390.c
4449
return kvm_setup_async_pf(vcpu, current->thread.gmap_teid.addr * PAGE_SIZE, hva, &arch);
arch/s390/kvm/kvm-s390.c
4463
vcpu->arch.sie_block->gg14 = vcpu->run->s.regs.gprs[14];
arch/s390/kvm/kvm-s390.c
4464
vcpu->arch.sie_block->gg15 = vcpu->run->s.regs.gprs[15];
arch/s390/kvm/kvm-s390.c
4481
clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask);
arch/s390/kvm/kvm-s390.c
4483
vcpu->arch.sie_block->icptcode = 0;
arch/s390/kvm/kvm-s390.c
4485
cpuflags = atomic_read(&vcpu->arch.sie_block->cpuflags);
arch/s390/kvm/kvm-s390.c
4511
rc = read_guest_instr(vcpu, vcpu->arch.sie_block->gpsw.addr, &opcode, 1);
arch/s390/kvm/kvm-s390.c
4520
pgm_info = vcpu->arch.pgm;
arch/s390/kvm/kvm-s390.c
4539
.attempt_pfault = pfault_enabled(vcpu->arch.gmap),
arch/s390/kvm/kvm-s390.c
4633
vcpu->arch.sie_block->icptcode);
arch/s390/kvm/kvm-s390.c
4634
trace_kvm_s390_sie_exit(vcpu, vcpu->arch.sie_block->icptcode);
arch/s390/kvm/kvm-s390.c
4639
vcpu->run->s.regs.gprs[14] = vcpu->arch.sie_block->gg14;
arch/s390/kvm/kvm-s390.c
4640
vcpu->run->s.regs.gprs[15] = vcpu->arch.sie_block->gg15;
arch/s390/kvm/kvm-s390.c
4643
sie_page = container_of(vcpu->arch.sie_block,
arch/s390/kvm/kvm-s390.c
4651
if (vcpu->arch.sie_block->icptcode > 0) {
arch/s390/kvm/kvm-s390.c
4657
vcpu->run->s390_sieic.icptcode = vcpu->arch.sie_block->icptcode;
arch/s390/kvm/kvm-s390.c
4658
vcpu->run->s390_sieic.ipa = vcpu->arch.sie_block->ipa;
arch/s390/kvm/kvm-s390.c
4659
vcpu->run->s390_sieic.ipb = vcpu->arch.sie_block->ipb;
arch/s390/kvm/kvm-s390.c
4689
struct sie_page *sie_page = (struct sie_page *)vcpu->arch.sie_block;
arch/s390/kvm/kvm-s390.c
4728
sie_return = kvm_s390_enter_exit_sie(vcpu->arch.sie_block,
arch/s390/kvm/kvm-s390.c
4730
vcpu->arch.gmap->asce.val);
arch/s390/kvm/kvm-s390.c
4746
if (vcpu->arch.sie_block->icptcode == ICPT_PV_INSTR ||
arch/s390/kvm/kvm-s390.c
4747
vcpu->arch.sie_block->icptcode == ICPT_PV_PREF) {
arch/s390/kvm/kvm-s390.c
4748
vcpu->arch.sie_block->gpsw.mask &= ~PSW_INT_MASK;
arch/s390/kvm/kvm-s390.c
4771
vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
arch/s390/kvm/kvm-s390.c
4772
vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
arch/s390/kvm/kvm-s390.c
4774
vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
arch/s390/kvm/kvm-s390.c
4775
vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
arch/s390/kvm/kvm-s390.c
4776
vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
arch/s390/kvm/kvm-s390.c
4779
vcpu->arch.pfault_token = kvm_run->s.regs.pft;
arch/s390/kvm/kvm-s390.c
4780
vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
arch/s390/kvm/kvm-s390.c
4781
vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
arch/s390/kvm/kvm-s390.c
4782
if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
arch/s390/kvm/kvm-s390.c
4786
vcpu->arch.diag318_info.val = kvm_run->s.regs.diag318;
arch/s390/kvm/kvm-s390.c
4787
vcpu->arch.sie_block->cpnc = vcpu->arch.diag318_info.cpnc;
arch/s390/kvm/kvm-s390.c
4788
VCPU_EVENT(vcpu, 3, "setting cpnc to %d", vcpu->arch.diag318_info.cpnc);
arch/s390/kvm/kvm-s390.c
4797
!(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
arch/s390/kvm/kvm-s390.c
4799
vcpu->arch.sie_block->ecb3 |= ECB3_RI;
arch/s390/kvm/kvm-s390.c
4808
!vcpu->arch.gs_enabled) {
arch/s390/kvm/kvm-s390.c
4810
vcpu->arch.sie_block->ecb |= ECB_GS;
arch/s390/kvm/kvm-s390.c
4811
vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
arch/s390/kvm/kvm-s390.c
4812
vcpu->arch.gs_enabled = 1;
arch/s390/kvm/kvm-s390.c
4816
vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
arch/s390/kvm/kvm-s390.c
4817
vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
arch/s390/kvm/kvm-s390.c
4823
vcpu->arch.host_gscb = current->thread.gs_cb;
arch/s390/kvm/kvm-s390.c
4824
save_gs_cb(vcpu->arch.host_gscb);
arch/s390/kvm/kvm-s390.c
4826
if (vcpu->arch.gs_enabled) {
arch/s390/kvm/kvm-s390.c
4843
memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
arch/s390/kvm/kvm-s390.c
4849
vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
arch/s390/kvm/kvm-s390.c
4851
save_access_regs(vcpu->arch.host_acrs);
arch/s390/kvm/kvm-s390.c
4853
vcpu->arch.acrs_loaded = true;
arch/s390/kvm/kvm-s390.c
4868
vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
arch/s390/kvm/kvm-s390.c
4869
vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
arch/s390/kvm/kvm-s390.c
4880
kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
arch/s390/kvm/kvm-s390.c
4881
kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
arch/s390/kvm/kvm-s390.c
4882
kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
arch/s390/kvm/kvm-s390.c
4883
kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
arch/s390/kvm/kvm-s390.c
4884
kvm_run->s.regs.diag318 = vcpu->arch.diag318_info.val;
arch/s390/kvm/kvm-s390.c
4888
if (vcpu->arch.gs_enabled)
arch/s390/kvm/kvm-s390.c
4890
current->thread.gs_cb = vcpu->arch.host_gscb;
arch/s390/kvm/kvm-s390.c
4891
restore_gs_cb(vcpu->arch.host_gscb);
arch/s390/kvm/kvm-s390.c
4892
if (!vcpu->arch.host_gscb)
arch/s390/kvm/kvm-s390.c
4894
vcpu->arch.host_gscb = NULL;
arch/s390/kvm/kvm-s390.c
4904
kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
arch/s390/kvm/kvm-s390.c
4905
kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
arch/s390/kvm/kvm-s390.c
4907
memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
arch/s390/kvm/kvm-s390.c
4909
kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
arch/s390/kvm/kvm-s390.c
4910
kvm_run->s.regs.pft = vcpu->arch.pfault_token;
arch/s390/kvm/kvm-s390.c
4911
kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
arch/s390/kvm/kvm-s390.c
4912
kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
arch/s390/kvm/kvm-s390.c
4914
restore_access_regs(vcpu->arch.host_acrs);
arch/s390/kvm/kvm-s390.c
4915
vcpu->arch.acrs_loaded = false;
arch/s390/kvm/kvm-s390.c
4933
if (vcpu->kvm->arch.pv.dumping)
arch/s390/kvm/kvm-s390.c
5039
&vcpu->arch.sie_block->gpsw, 16);
arch/s390/kvm/kvm-s390.c
5045
&vcpu->arch.sie_block->todpr, 4);
arch/s390/kvm/kvm-s390.c
5049
clkcomp = vcpu->arch.sie_block->ckc >> 8;
arch/s390/kvm/kvm-s390.c
5055
&vcpu->arch.sie_block->gcr, 128);
arch/s390/kvm/kvm-s390.c
5105
spin_lock(&vcpu->kvm->arch.start_stop_lock);
arch/s390/kvm/kvm-s390.c
5112
spin_unlock(&vcpu->kvm->arch.start_stop_lock);
arch/s390/kvm/kvm-s390.c
5141
vcpu->arch.sie_block->gpsw.mask &= ~PSW_INT_MASK;
arch/s390/kvm/kvm-s390.c
5147
spin_unlock(&vcpu->kvm->arch.start_stop_lock);
arch/s390/kvm/kvm-s390.c
5161
spin_lock(&vcpu->kvm->arch.start_stop_lock);
arch/s390/kvm/kvm-s390.c
5168
spin_unlock(&vcpu->kvm->arch.start_stop_lock);
arch/s390/kvm/kvm-s390.c
5201
spin_unlock(&vcpu->kvm->arch.start_stop_lock);
arch/s390/kvm/kvm-s390.c
5215
if (!vcpu->kvm->arch.css_support) {
arch/s390/kvm/kvm-s390.c
5216
vcpu->kvm->arch.css_support = 1;
arch/s390/kvm/kvm-s390.c
5240
if (mop->size + mop->sida_offset > sida_size(vcpu->arch.sie_block))
arch/s390/kvm/kvm-s390.c
5245
sida_addr = (char *)sida_addr(vcpu->arch.sie_block) + mop->sida_offset;
arch/s390/kvm/kvm-s390.c
5301
kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm);
arch/s390/kvm/kvm-s390.c
5384
if (!vcpu->kvm->arch.pv.dumping)
arch/s390/kvm/kvm-s390.c
5504
r = gmap_ucas_map(vcpu->arch.gmap, gpa_to_gfn(ucas.user_addr),
arch/s390/kvm/kvm-s390.c
5522
gmap_ucas_unmap(vcpu->arch.gmap, gpa_to_gfn(ucas.vcpu_addr),
arch/s390/kvm/kvm-s390.c
560
union asce asce = kvm->arch.gmap->asce;
arch/s390/kvm/kvm-s390.c
5638
vmf->page = virt_to_page(vcpu->arch.sie_block);
arch/s390/kvm/kvm-s390.c
5673
if ((new->base_gfn + new->npages) * PAGE_SIZE > kvm->arch.mem_limit)
arch/s390/kvm/kvm-s390.c
5677
if (!kvm->arch.migration_mode)
arch/s390/kvm/kvm-s390.c
5716
rc = dat_delete_slot(mc, kvm->arch.gmap->asce, old->base_gfn, old->npages);
arch/s390/kvm/kvm-s390.c
5719
rc = dat_delete_slot(mc, kvm->arch.gmap->asce, old->base_gfn, old->npages);
arch/s390/kvm/kvm-s390.c
5724
rc = dat_create_slot(mc, kvm->arch.gmap->asce, new->base_gfn, new->npages);
arch/s390/kvm/kvm-s390.c
5750
return dat_test_age_gfn(kvm->arch.gmap->asce, range->start, range->end);
arch/s390/kvm/kvm-s390.c
5764
return gmap_age_gfn(kvm->arch.gmap, range->start, range->end);
arch/s390/kvm/kvm-s390.c
5781
return gmap_unmap_gfn_range(kvm->arch.gmap, range->slot, range->start, range->end);
arch/s390/kvm/kvm-s390.c
726
gmap_sync_dirty_log(kvm->arch.gmap, memslot->base_gfn, last_gfn);
arch/s390/kvm/kvm-s390.c
787
kvm->arch.use_irqchip = 1;
arch/s390/kvm/kvm-s390.c
792
kvm->arch.user_sigp = 1;
arch/s390/kvm/kvm-s390.c
800
set_kvm_facility(kvm->arch.model.fac_mask, 129);
arch/s390/kvm/kvm-s390.c
801
set_kvm_facility(kvm->arch.model.fac_list, 129);
arch/s390/kvm/kvm-s390.c
803
set_kvm_facility(kvm->arch.model.fac_mask, 134);
arch/s390/kvm/kvm-s390.c
804
set_kvm_facility(kvm->arch.model.fac_list, 134);
arch/s390/kvm/kvm-s390.c
807
set_kvm_facility(kvm->arch.model.fac_mask, 135);
arch/s390/kvm/kvm-s390.c
808
set_kvm_facility(kvm->arch.model.fac_list, 135);
arch/s390/kvm/kvm-s390.c
811
set_kvm_facility(kvm->arch.model.fac_mask, 148);
arch/s390/kvm/kvm-s390.c
812
set_kvm_facility(kvm->arch.model.fac_list, 148);
arch/s390/kvm/kvm-s390.c
815
set_kvm_facility(kvm->arch.model.fac_mask, 152);
arch/s390/kvm/kvm-s390.c
816
set_kvm_facility(kvm->arch.model.fac_list, 152);
arch/s390/kvm/kvm-s390.c
819
set_kvm_facility(kvm->arch.model.fac_mask, 192);
arch/s390/kvm/kvm-s390.c
820
set_kvm_facility(kvm->arch.model.fac_list, 192);
arch/s390/kvm/kvm-s390.c
823
set_kvm_facility(kvm->arch.model.fac_mask, 198);
arch/s390/kvm/kvm-s390.c
824
set_kvm_facility(kvm->arch.model.fac_list, 198);
arch/s390/kvm/kvm-s390.c
827
set_kvm_facility(kvm->arch.model.fac_mask, 199);
arch/s390/kvm/kvm-s390.c
828
set_kvm_facility(kvm->arch.model.fac_list, 199);
arch/s390/kvm/kvm-s390.c
843
set_kvm_facility(kvm->arch.model.fac_mask, 64);
arch/s390/kvm/kvm-s390.c
844
set_kvm_facility(kvm->arch.model.fac_list, 64);
arch/s390/kvm/kvm-s390.c
856
set_kvm_facility(kvm->arch.model.fac_mask, 72);
arch/s390/kvm/kvm-s390.c
857
set_kvm_facility(kvm->arch.model.fac_list, 72);
arch/s390/kvm/kvm-s390.c
870
set_kvm_facility(kvm->arch.model.fac_mask, 133);
arch/s390/kvm/kvm-s390.c
871
set_kvm_facility(kvm->arch.model.fac_list, 133);
arch/s390/kvm/kvm-s390.c
882
else if (!hpage || kvm->arch.use_cmma || kvm_is_ucontrol(kvm))
arch/s390/kvm/kvm-s390.c
886
set_bit(GMAP_FLAG_ALLOW_HPAGE_1M, &kvm->arch.gmap->flags);
arch/s390/kvm/kvm-s390.c
892
kvm->arch.use_skf = 0;
arch/s390/kvm/kvm-s390.c
893
kvm->arch.use_pfmfi = 0;
arch/s390/kvm/kvm-s390.c
901
kvm->arch.user_stsi = 1;
arch/s390/kvm/kvm-s390.c
906
kvm->arch.user_instr0 = 1;
arch/s390/kvm/kvm-s390.c
916
set_kvm_facility(kvm->arch.model.fac_mask, 11);
arch/s390/kvm/kvm-s390.c
917
set_kvm_facility(kvm->arch.model.fac_list, 11);
arch/s390/kvm/kvm-s390.c
926
kvm->arch.user_operexec = 1;
arch/s390/kvm/kvm-s390.c
932
kvm->arch.allow_vsie_esamode = 1;
arch/s390/kvm/kvm-s390.c
950
kvm->arch.mem_limit);
arch/s390/kvm/kvm-s390.c
951
if (put_user(kvm->arch.mem_limit, (u64 __user *)attr->addr))
arch/s390/kvm/kvm-s390.c
976
kvm->arch.use_cmma = 1;
arch/s390/kvm/kvm-s390.c
978
kvm->arch.use_pfmfi = 0;
arch/s390/kvm/kvm-s390.c
990
if (!kvm->arch.use_cmma)
arch/s390/kvm/kvm-s390.c
995
start_gfn = dat_reset_cmma(kvm->arch.gmap->asce, start_gfn);
arch/s390/kvm/kvm-s390.h
103
return (atomic_read(&vcpu->arch.sie_block->cpuflags) & flags) == flags;
arch/s390/kvm/kvm-s390.h
113
return test_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask);
arch/s390/kvm/kvm-s390.h
119
return test_bit(GMAP_FLAG_IS_UCONTROL, &kvm->arch.gmap->flags);
arch/s390/kvm/kvm-s390.h
130
return vcpu->arch.sie_block->prefix << GUEST_PREFIX_SHIFT;
arch/s390/kvm/kvm-s390.h
137
vcpu->arch.sie_block->prefix = prefix >> GUEST_PREFIX_SHIFT;
arch/s390/kvm/kvm-s390.h
138
vcpu->arch.sie_block->prefix &= GUEST_PREFIX_MASK_ZARCH;
arch/s390/kvm/kvm-s390.h
145
u32 base2 = vcpu->arch.sie_block->ipb >> 28;
arch/s390/kvm/kvm-s390.h
146
u32 disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16);
arch/s390/kvm/kvm-s390.h
156
u32 base1 = vcpu->arch.sie_block->ipb >> 28;
arch/s390/kvm/kvm-s390.h
160
disp1 = sign_extend64(((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16) +
arch/s390/kvm/kvm-s390.h
161
((vcpu->arch.sie_block->ipb & 0xff00) << 4), 19);
arch/s390/kvm/kvm-s390.h
173
u32 base1 = (vcpu->arch.sie_block->ipb & 0xf0000000) >> 28;
arch/s390/kvm/kvm-s390.h
174
u32 disp1 = (vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16;
arch/s390/kvm/kvm-s390.h
175
u32 base2 = (vcpu->arch.sie_block->ipb & 0xf000) >> 12;
arch/s390/kvm/kvm-s390.h
176
u32 disp2 = vcpu->arch.sie_block->ipb & 0x0fff;
arch/s390/kvm/kvm-s390.h
190
*r1 = (vcpu->arch.sie_block->ipb & 0x00f00000) >> 20;
arch/s390/kvm/kvm-s390.h
192
*r2 = (vcpu->arch.sie_block->ipb & 0x000f0000) >> 16;
arch/s390/kvm/kvm-s390.h
197
u32 base2 = vcpu->arch.sie_block->ipb >> 28;
arch/s390/kvm/kvm-s390.h
198
u32 disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16) +
arch/s390/kvm/kvm-s390.h
199
((vcpu->arch.sie_block->ipb & 0xff00) << 4);
arch/s390/kvm/kvm-s390.h
212
u32 base2 = vcpu->arch.sie_block->ipb >> 28;
arch/s390/kvm/kvm-s390.h
213
u32 disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16);
arch/s390/kvm/kvm-s390.h
224
vcpu->arch.sie_block->gpsw.mask &= ~(3UL << 44);
arch/s390/kvm/kvm-s390.h
225
vcpu->arch.sie_block->gpsw.mask |= cc << 44;
arch/s390/kvm/kvm-s390.h
231
return __test_facility(nr, kvm->arch.model.fac_mask) &&
arch/s390/kvm/kvm-s390.h
232
__test_facility(nr, kvm->arch.model.fac_list);
arch/s390/kvm/kvm-s390.h
249
return test_bit_inv(nr, kvm->arch.cpu_feat);
arch/s390/kvm/kvm-s390.h
255
return kvm->arch.user_cpu_state_ctrl != 0;
arch/s390/kvm/kvm-s390.h
260
if (kvm->arch.user_cpu_state_ctrl)
arch/s390/kvm/kvm-s390.h
264
kvm->arch.user_cpu_state_ctrl = 1;
arch/s390/kvm/kvm-s390.h
285
if (!kvm->arch.gisa_int.origin)
arch/s390/kvm/kvm-s390.h
288
gd = virt_to_phys(kvm->arch.gisa_int.origin);
arch/s390/kvm/kvm-s390.h
328
return kvm->arch.pv.handle;
arch/s390/kvm/kvm-s390.h
333
return vcpu->arch.pv.handle;
arch/s390/kvm/kvm-s390.h
412
struct kvm_s390_sie_block *sie_block = vcpu->arch.sie_block;
arch/s390/kvm/kvm-s390.h
423
vcpu->arch.sie_block->icptstatus &= ~0x02;
arch/s390/kvm/kvm-s390.h
502
rc = get_tod_clock_fast() + kvm->arch.epoch;
arch/s390/kvm/kvm-s390.h
537
return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm);
arch/s390/kvm/kvm-s390.h
54
#define IS_TE_ENABLED(vcpu) ((vcpu->arch.sie_block->ecb & ECB_TE))
arch/s390/kvm/kvm-s390.h
57
((*(char *)phys_to_virt((vcpu)->arch.sie_block->itdba) == TDB_FORMAT1))
arch/s390/kvm/kvm-s390.h
64
debug_sprintf_event((d_kvm)->arch.dbf, d_loglevel, d_string "\n", \
arch/s390/kvm/kvm-s390.h
79
debug_sprintf_event(d_kvm->arch.dbf, d_loglevel, d_string "\n", \
arch/s390/kvm/kvm-s390.h
85
debug_sprintf_event(d_vcpu->kvm->arch.dbf, d_loglevel, \
arch/s390/kvm/kvm-s390.h
87
d_vcpu->arch.sie_block->gpsw.mask, d_vcpu->arch.sie_block->gpsw.addr,\
arch/s390/kvm/kvm-s390.h
93
atomic_or(flags, &vcpu->arch.sie_block->cpuflags);
arch/s390/kvm/kvm-s390.h
98
atomic_andnot(flags, &vcpu->arch.sie_block->cpuflags);
arch/s390/kvm/pci.c
298
gaite->gisa = (u32)virt_to_phys(&kvm->arch.sie_page2->gisa);
arch/s390/kvm/pci.c
466
if (!kvm->arch.use_zpci_interp)
arch/s390/kvm/pci.c
479
zdev->gisa = (u32)virt_to_phys(&kvm->arch.sie_page2->gisa);
arch/s390/kvm/pci.c
488
spin_lock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
489
list_add_tail(&zdev->kzdev->entry, &kvm->arch.kzdev_list);
arch/s390/kvm/pci.c
490
spin_unlock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
547
spin_lock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
549
spin_unlock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
560
spin_lock_init(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
561
INIT_LIST_HEAD(&kvm->arch.kzdev_list);
arch/s390/kvm/pci.c
570
spin_lock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
571
WARN_ON_ONCE(!list_empty(&kvm->arch.kzdev_list));
arch/s390/kvm/pci.c
572
spin_unlock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
580
spin_lock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/pci.c
581
list_for_each_entry(kzdev, &kvm->arch.kzdev_list, entry) {
arch/s390/kvm/pci.c
587
spin_unlock(&kvm->arch.kzdev_list_lock);
arch/s390/kvm/priv.c
1024
vcpu->run->s.regs.gprs[reg1] |= vcpu->arch.sie_block->gpsw.mask >> 32;
arch/s390/kvm/priv.c
1028
vcpu->arch.sie_block->gpsw.mask & 0x00000000ffffffffUL;
arch/s390/kvm/priv.c
1057
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1082
return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm);
arch/s390/kvm/priv.c
1098
psw_bits(vcpu->arch.sie_block->gpsw).eaba == PSW_BITS_AMODE_24BIT)
arch/s390/kvm/priv.c
1118
rc = dat_cond_set_storage_key(vcpu->arch.mc, vcpu->arch.gmap->asce,
arch/s390/kvm/priv.c
1125
kvm_s390_mmu_cache_topup(vcpu->arch.mc);
arch/s390/kvm/priv.c
1134
if (psw_bits(vcpu->arch.sie_block->gpsw).eaba == PSW_BITS_AMODE_64BIT) {
arch/s390/kvm/priv.c
1163
entries = (vcpu->arch.sie_block->cbrlo & ~PAGE_MASK) >> 3;
arch/s390/kvm/priv.c
1165
nappended = dat_perform_essa(vcpu->arch.gmap->asce, gfn, orc, &state, &dirtied);
arch/s390/kvm/priv.c
1176
cbrlo = phys_to_virt(vcpu->arch.sie_block->cbrlo & PAGE_MASK);
arch/s390/kvm/priv.c
1181
atomic64_inc(&vcpu->kvm->arch.cmma_dirty_pages);
arch/s390/kvm/priv.c
1196
if (dat_entry_walk(NULL, gpa_to_gfn(cbrl[i]), vcpu->arch.gmap->asce,
arch/s390/kvm/priv.c
1213
int entries = (vcpu->arch.sie_block->cbrlo & ~PAGE_MASK) >> 3;
arch/s390/kvm/priv.c
1219
if (!vcpu->kvm->arch.use_cmma)
arch/s390/kvm/priv.c
1222
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1225
orc = (vcpu->arch.sie_block->ipb & 0xf0000000) >> 28;
arch/s390/kvm/priv.c
1231
if (!vcpu->kvm->arch.migration_mode) {
arch/s390/kvm/priv.c
1241
set_bit(GMAP_FLAG_USES_CMM, &vcpu->arch.gmap->flags);
arch/s390/kvm/priv.c
1251
vcpu->arch.sie_block->ecb2 |= ECB2_CMMA;
arch/s390/kvm/priv.c
1263
vcpu->arch.sie_block->cbrlo &= PAGE_MASK;
arch/s390/kvm/priv.c
1264
cbrlo = phys_to_virt(vcpu->arch.sie_block->cbrlo);
arch/s390/kvm/priv.c
1276
switch (vcpu->arch.sie_block->ipa & 0x00ff) {
arch/s390/kvm/priv.c
1294
int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4;
arch/s390/kvm/priv.c
1295
int reg3 = vcpu->arch.sie_block->ipa & 0x000f;
arch/s390/kvm/priv.c
1303
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
132
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1321
vcpu->arch.sie_block->gcr[reg] &= 0xffffffff00000000ul;
arch/s390/kvm/priv.c
1322
vcpu->arch.sie_block->gcr[reg] |= ctl_array[nr_regs++];
arch/s390/kvm/priv.c
1333
int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4;
arch/s390/kvm/priv.c
1334
int reg3 = vcpu->arch.sie_block->ipa & 0x000f;
arch/s390/kvm/priv.c
1342
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1356
ctl_array[nr_regs++] = vcpu->arch.sie_block->gcr[reg];
arch/s390/kvm/priv.c
1367
int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4;
arch/s390/kvm/priv.c
1368
int reg3 = vcpu->arch.sie_block->ipa & 0x000f;
arch/s390/kvm/priv.c
1376
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1394
vcpu->arch.sie_block->gcr[reg] = ctl_array[nr_regs++];
arch/s390/kvm/priv.c
1405
int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4;
arch/s390/kvm/priv.c
1406
int reg3 = vcpu->arch.sie_block->ipa & 0x000f;
arch/s390/kvm/priv.c
1414
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1428
ctl_array[nr_regs++] = vcpu->arch.sie_block->gcr[reg];
arch/s390/kvm/priv.c
1439
switch (vcpu->arch.sie_block->ipb & 0x000000ff) {
arch/s390/kvm/priv.c
1466
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1472
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_DAT)
arch/s390/kvm/priv.c
1509
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_DAT)
arch/s390/kvm/priv.c
1516
switch (vcpu->arch.sie_block->ipa & 0x00ff) {
arch/s390/kvm/priv.c
1530
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
1538
vcpu->arch.sie_block->todpr = value;
arch/s390/kvm/priv.c
1554
switch (vcpu->arch.sie_block->ipa & 0x00ff) {
arch/s390/kvm/priv.c
170
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
200
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
223
if (vcpu->arch.skey_enabled)
arch/s390/kvm/priv.c
226
rc = gmap_enable_skeys(vcpu->arch.gmap);
arch/s390/kvm/priv.c
233
if (!vcpu->kvm->arch.use_skf)
arch/s390/kvm/priv.c
234
vcpu->arch.sie_block->ictl |= ICTL_ISKE | ICTL_SSKE | ICTL_RRBE;
arch/s390/kvm/priv.c
236
vcpu->arch.sie_block->ictl &= ~(ICTL_ISKE | ICTL_SSKE | ICTL_RRBE);
arch/s390/kvm/priv.c
237
vcpu->arch.skey_enabled = true;
arch/s390/kvm/priv.c
248
if (vcpu->kvm->arch.use_skf) {
arch/s390/kvm/priv.c
266
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
279
rc = dat_get_storage_key(vcpu->arch.gmap->asce, gpa_to_gfn(gaddr), &key);
arch/s390/kvm/priv.c
297
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
310
rc = dat_reset_reference_bit(vcpu->arch.gmap->asce, gpa_to_gfn(gaddr));
arch/s390/kvm/priv.c
325
unsigned char m3 = vcpu->arch.sie_block->ipb >> 28;
arch/s390/kvm/priv.c
333
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
362
rc = dat_cond_set_storage_key(vcpu->arch.mc, vcpu->arch.gmap->asce,
arch/s390/kvm/priv.c
369
kvm_s390_mmu_cache_topup(vcpu->arch.mc);
arch/s390/kvm/priv.c
388
if (psw_bits(vcpu->arch.sie_block->gpsw).eaba == PSW_BITS_AMODE_64BIT)
arch/s390/kvm/priv.c
39
vcpu->arch.sie_block->ecb3 |= ECB3_RI;
arch/s390/kvm/priv.c
401
if (psw_bits(vcpu->arch.sie_block->gpsw).pstate)
arch/s390/kvm/priv.c
403
wait_event(vcpu->kvm->arch.ipte_wq, !ipte_lock_held(vcpu->kvm));
arch/s390/kvm/priv.c
416
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
423
return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm);
arch/s390/kvm/priv.c
454
inti = kvm_s390_get_io_int(vcpu->kvm, vcpu->arch.sie_block->gcr[6], 0);
arch/s390/kvm/priv.c
48
if ((vcpu->arch.sie_block->ipa & 0xf) <= 4)
arch/s390/kvm/priv.c
533
vcpu->run->s390_tsch.ipb = vcpu->arch.sie_block->ipb;
arch/s390/kvm/priv.c
542
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
545
if (vcpu->kvm->arch.css_support) {
arch/s390/kvm/priv.c
550
if (vcpu->arch.sie_block->ipa == 0xb236)
arch/s390/kvm/priv.c
552
if (vcpu->arch.sie_block->ipa == 0xb235)
arch/s390/kvm/priv.c
602
if (!(vcpu->arch.sie_block->eca & ECA_APIE))
arch/s390/kvm/priv.c
618
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
629
if (!(vcpu->kvm->arch.crypto.crycbd & 0x02) && (reg0 & 0x0000c0f0UL))
arch/s390/kvm/priv.c
642
down_read(&vcpu->kvm->arch.crypto.pqap_hook_rwsem);
arch/s390/kvm/priv.c
643
if (vcpu->kvm->arch.crypto.pqap_hook) {
arch/s390/kvm/priv.c
644
pqap_hook = *vcpu->kvm->arch.crypto.pqap_hook;
arch/s390/kvm/priv.c
65
vcpu->arch.sie_block->ecb |= ECB_GS;
arch/s390/kvm/priv.c
652
up_read(&vcpu->kvm->arch.crypto.pqap_hook_rwsem);
arch/s390/kvm/priv.c
655
up_read(&vcpu->kvm->arch.crypto.pqap_hook_rwsem);
arch/s390/kvm/priv.c
66
vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
arch/s390/kvm/priv.c
67
vcpu->arch.gs_enabled = 1;
arch/s390/kvm/priv.c
674
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
681
fac = *vcpu->kvm->arch.model.fac_list >> 32;
arch/s390/kvm/priv.c
715
psw_t *gpsw = &vcpu->arch.sie_block->gpsw;
arch/s390/kvm/priv.c
741
vcpu->arch.sie_block->gbea = iaddr;
arch/s390/kvm/priv.c
754
iaddr = vcpu->arch.sie_block->gpsw.addr - kvm_s390_get_ilen(vcpu);
arch/s390/kvm/priv.c
755
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
76
int code = vcpu->arch.sie_block->ipb & 0xff;
arch/s390/kvm/priv.c
764
vcpu->arch.sie_block->gpsw = new_psw;
arch/s390/kvm/priv.c
765
if (!is_valid_psw(&vcpu->arch.sie_block->gpsw))
arch/s390/kvm/priv.c
767
vcpu->arch.sie_block->gbea = iaddr;
arch/s390/kvm/priv.c
783
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
794
vcpu->arch.sie_block->gpsw = new_psw;
arch/s390/kvm/priv.c
795
if (!is_valid_psw(&vcpu->arch.sie_block->gpsw))
arch/s390/kvm/priv.c
803
u64 stidp_data = vcpu->kvm->arch.model.cpuid;
arch/s390/kvm/priv.c
810
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
877
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
890
!vcpu->kvm->arch.user_stsi))
arch/s390/kvm/priv.c
93
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/priv.c
931
memcpy(sida_addr(vcpu->arch.sie_block), (void *)mem, PAGE_SIZE);
arch/s390/kvm/priv.c
940
if (vcpu->kvm->arch.user_stsi) {
arch/s390/kvm/priv.c
958
switch (vcpu->arch.sie_block->ipa & 0x00ff) {
arch/s390/kvm/pv.c
1030
kvm->arch.pv.dumping = false;
arch/s390/kvm/pv.c
143
scoped_guard(mutex, &kvm->arch.pv.import_lock) {
arch/s390/kvm/pv.c
222
kvm->arch.pv.handle = 0;
arch/s390/kvm/pv.c
223
kvm->arch.pv.guest_len = 0;
arch/s390/kvm/pv.c
224
kvm->arch.pv.stor_base = 0;
arch/s390/kvm/pv.c
225
kvm->arch.pv.stor_var = NULL;
arch/s390/kvm/pv.c
243
free_pages(vcpu->arch.pv.stor_base,
arch/s390/kvm/pv.c
246
free_page((unsigned long)sida_addr(vcpu->arch.sie_block));
arch/s390/kvm/pv.c
247
vcpu->arch.sie_block->pv_handle_cpu = 0;
arch/s390/kvm/pv.c
248
vcpu->arch.sie_block->pv_handle_config = 0;
arch/s390/kvm/pv.c
249
memset(&vcpu->arch.pv, 0, sizeof(vcpu->arch.pv));
arch/s390/kvm/pv.c
250
vcpu->arch.sie_block->sdf = 0;
arch/s390/kvm/pv.c
256
vcpu->arch.sie_block->gbea = 1;
arch/s390/kvm/pv.c
274
vcpu->arch.pv.stor_base = __get_free_pages(GFP_KERNEL_ACCOUNT,
arch/s390/kvm/pv.c
276
if (!vcpu->arch.pv.stor_base)
arch/s390/kvm/pv.c
281
uvcb.num = vcpu->arch.sie_block->icpua;
arch/s390/kvm/pv.c
282
uvcb.state_origin = virt_to_phys(vcpu->arch.sie_block);
arch/s390/kvm/pv.c
283
uvcb.stor_origin = virt_to_phys((void *)vcpu->arch.pv.stor_base);
arch/s390/kvm/pv.c
288
free_pages(vcpu->arch.pv.stor_base,
arch/s390/kvm/pv.c
292
vcpu->arch.sie_block->sidad = virt_to_phys(sida_addr);
arch/s390/kvm/pv.c
310
vcpu->arch.pv.handle = uvcb.cpu_handle;
arch/s390/kvm/pv.c
311
vcpu->arch.sie_block->pv_handle_cpu = uvcb.cpu_handle;
arch/s390/kvm/pv.c
312
vcpu->arch.sie_block->pv_handle_config = kvm_s390_pv_get_handle(vcpu->kvm);
arch/s390/kvm/pv.c
313
vcpu->arch.sie_block->sdf = 2;
arch/s390/kvm/pv.c
321
vfree(kvm->arch.pv.stor_var);
arch/s390/kvm/pv.c
322
free_pages(kvm->arch.pv.stor_base,
arch/s390/kvm/pv.c
333
kvm->arch.pv.stor_var = NULL;
arch/s390/kvm/pv.c
334
kvm->arch.pv.stor_base = __get_free_pages(GFP_KERNEL_ACCOUNT, get_order(base));
arch/s390/kvm/pv.c
335
if (!kvm->arch.pv.stor_base)
arch/s390/kvm/pv.c
348
kvm->arch.pv.guest_len = npages * PAGE_SIZE;
arch/s390/kvm/pv.c
353
kvm->arch.pv.stor_var = vzalloc(vlen);
arch/s390/kvm/pv.c
354
if (!kvm->arch.pv.stor_var)
arch/s390/kvm/pv.c
462
if (kvm->arch.pv.set_aside)
arch/s390/kvm/pv.c
466
if (kvm->arch.gmap->asce.dt == TABLE_TYPE_SEGMENT)
arch/s390/kvm/pv.c
476
priv->stor_var = kvm->arch.pv.stor_var;
arch/s390/kvm/pv.c
477
priv->stor_base = kvm->arch.pv.stor_base;
arch/s390/kvm/pv.c
479
priv->old_gmap_table = (unsigned long)dereference_asce(kvm->arch.gmap->asce);
arch/s390/kvm/pv.c
480
if (s390_replace_asce(kvm->arch.gmap))
arch/s390/kvm/pv.c
489
gmap_pv_destroy_range(kvm->arch.gmap, 0, gpa_to_gfn(SZ_2G), false);
arch/s390/kvm/pv.c
491
kvm->arch.pv.set_aside = priv;
arch/s390/kvm/pv.c
528
s390_replace_asce(kvm->arch.gmap);
arch/s390/kvm/pv.c
574
if (kvm->arch.pv.set_aside) {
arch/s390/kvm/pv.c
575
list_add(kvm->arch.pv.set_aside, &kvm->arch.pv.need_cleanup);
arch/s390/kvm/pv.c
576
kvm->arch.pv.set_aside = NULL;
arch/s390/kvm/pv.c
580
while (!list_empty(&kvm->arch.pv.need_cleanup)) {
arch/s390/kvm/pv.c
581
cur = list_first_entry(&kvm->arch.pv.need_cleanup, typeof(*cur), list);
arch/s390/kvm/pv.c
605
gmap_pv_destroy_range(kvm->arch.gmap, 0, asce_end(kvm->arch.gmap->asce), false);
arch/s390/kvm/pv.c
636
p = kvm->arch.pv.set_aside;
arch/s390/kvm/pv.c
637
kvm->arch.pv.set_aside = NULL;
arch/s390/kvm/pv.c
643
if (gmap_pv_destroy_range(kvm->arch.gmap, 0, asce_end(kvm->arch.gmap->asce), true))
arch/s390/kvm/pv.c
656
list_add(&p->list, &kvm->arch.pv.need_cleanup);
arch/s390/kvm/pv.c
668
struct kvm *kvm = container_of(subscription, struct kvm, arch.pv.mmu_notifier);
arch/s390/kvm/pv.c
682
set_bit(GMAP_FLAG_EXPORT_ON_UNMAP, &kvm->arch.gmap->flags);
arch/s390/kvm/pv.c
699
if (kvm->arch.pv.mmu_notifier.ops != &kvm_s390_pv_mmu_notifier_ops) {
arch/s390/kvm/pv.c
701
kvm->arch.pv.mmu_notifier.ops = &kvm_s390_pv_mmu_notifier_ops;
arch/s390/kvm/pv.c
702
ret = mmu_notifier_register(&kvm->arch.pv.mmu_notifier, kvm->mm);
arch/s390/kvm/pv.c
704
kvm->arch.pv.mmu_notifier.ops = NULL;
arch/s390/kvm/pv.c
715
uvcb.guest_stor_len = kvm->arch.pv.guest_len;
arch/s390/kvm/pv.c
716
uvcb.guest_asce = kvm->arch.gmap->asce.val;
arch/s390/kvm/pv.c
717
uvcb.guest_sca = virt_to_phys(kvm->arch.sca);
arch/s390/kvm/pv.c
719
virt_to_phys((void *)kvm->arch.pv.stor_base);
arch/s390/kvm/pv.c
720
uvcb.conf_virt_stor_origin = (u64)kvm->arch.pv.stor_var;
arch/s390/kvm/pv.c
721
uvcb.flags.ap_allow_instr = kvm->arch.model.uv_feat_guest.ap;
arch/s390/kvm/pv.c
722
uvcb.flags.ap_instr_intr = kvm->arch.model.uv_feat_guest.ap_intr;
arch/s390/kvm/pv.c
724
clear_bit(GMAP_FLAG_ALLOW_HPAGE_1M, &kvm->arch.gmap->flags);
arch/s390/kvm/pv.c
725
gmap_split_huge_pages(kvm->arch.gmap);
arch/s390/kvm/pv.c
734
kvm->arch.pv.handle = uvcb.guest_handle;
arch/s390/kvm/pv.c
851
.cpu_handle = vcpu->arch.pv.handle,
arch/s390/kvm/pv.c
899
.config_handle = kvm->arch.pv.handle,
arch/s390/kvm/sigp.c
243
struct kvm_s390_local_interrupt *li = &dst_vcpu->arch.local_int;
arch/s390/kvm/sigp.c
378
if (!vcpu->kvm->arch.user_sigp)
arch/s390/kvm/sigp.c
427
int r1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4;
arch/s390/kvm/sigp.c
428
int r3 = vcpu->arch.sie_block->ipa & 0x000f;
arch/s390/kvm/sigp.c
435
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/sigp.c
478
int r3 = vcpu->arch.sie_block->ipa & 0x000f;
arch/s390/kvm/sigp.c
76
psw = &dst_vcpu->arch.sie_block->gpsw;
arch/s390/kvm/sigp.c
77
p_asn = dst_vcpu->arch.sie_block->gcr[4] & 0xffff; /* Primary ASN */
arch/s390/kvm/sigp.c
78
s_asn = dst_vcpu->arch.sie_block->gcr[3] & 0xffff; /* Secondary ASN */
arch/s390/kvm/trace.h
27
__entry->pswmask = vcpu->arch.sie_block->gpsw.mask; \
arch/s390/kvm/trace.h
28
__entry->pswaddr = vcpu->arch.sie_block->gpsw.addr; \
arch/s390/kvm/vsie.c
1069
cr0.val = vcpu->arch.sie_block->gcr[0];
arch/s390/kvm/vsie.c
1164
vcpu->arch.sie_block->fpf & FPF_BPBC)
arch/s390/kvm/vsie.c
1173
vcpu->arch.sie_block->prog0c |= PROG_IN_SIE;
arch/s390/kvm/vsie.c
1195
vcpu->arch.sie_block->prog0c &= ~PROG_IN_SIE;
arch/s390/kvm/vsie.c
1241
lockdep_assert_held(&gmap->kvm->arch.gmap->children_lock);
arch/s390/kvm/vsie.c
1260
asce.val = vcpu->arch.sie_block->gcr[1];
arch/s390/kvm/vsie.c
1261
cr0.val = vcpu->arch.sie_block->gcr[0];
arch/s390/kvm/vsie.c
1265
scoped_guard(spinlock, &vcpu->kvm->arch.gmap->children_lock) {
arch/s390/kvm/vsie.c
1283
gmap = gmap_create_shadow(vcpu->arch.mc, vcpu->kvm->arch.gmap, asce, edat);
arch/s390/kvm/vsie.c
1286
scoped_guard(spinlock, &vcpu->kvm->arch.gmap->children_lock) {
arch/s390/kvm/vsie.c
129
if (!(cpuflags & CPUSTAT_ZARCH) && !vcpu->kvm->arch.allow_vsie_esamode)
arch/s390/kvm/vsie.c
1310
WRITE_ONCE(vcpu->arch.vsie_block, &vsie_page->scb_s);
arch/s390/kvm/vsie.c
1321
scb_s->epoch += vcpu->kvm->arch.epoch;
arch/s390/kvm/vsie.c
1324
scb_s->epdx += vcpu->kvm->arch.epdx;
arch/s390/kvm/vsie.c
1325
if (scb_s->epoch < vcpu->kvm->arch.epoch)
arch/s390/kvm/vsie.c
1338
WRITE_ONCE(vcpu->arch.vsie_block, NULL);
arch/s390/kvm/vsie.c
1442
vsie_page = radix_tree_lookup(&kvm->arch.vsie.addr_to_page, addr >> 9);
arch/s390/kvm/vsie.c
1462
mutex_lock(&kvm->arch.vsie.mutex);
arch/s390/kvm/vsie.c
1463
if (kvm->arch.vsie.page_count < nr_vcpus) {
arch/s390/kvm/vsie.c
1466
mutex_unlock(&kvm->arch.vsie.mutex);
arch/s390/kvm/vsie.c
1470
kvm->arch.vsie.pages[kvm->arch.vsie.page_count] = vsie_page;
arch/s390/kvm/vsie.c
1471
kvm->arch.vsie.page_count++;
arch/s390/kvm/vsie.c
1475
vsie_page = kvm->arch.vsie.pages[kvm->arch.vsie.next];
arch/s390/kvm/vsie.c
1478
kvm->arch.vsie.next++;
arch/s390/kvm/vsie.c
1479
kvm->arch.vsie.next %= nr_vcpus;
arch/s390/kvm/vsie.c
1482
radix_tree_delete(&kvm->arch.vsie.addr_to_page,
arch/s390/kvm/vsie.c
1489
if (radix_tree_insert(&kvm->arch.vsie.addr_to_page, addr >> 9, vsie_page)) {
arch/s390/kvm/vsie.c
1491
mutex_unlock(&kvm->arch.vsie.mutex);
arch/s390/kvm/vsie.c
1495
mutex_unlock(&kvm->arch.vsie.mutex);
arch/s390/kvm/vsie.c
1499
scoped_guard(spinlock, &kvm->arch.gmap->children_lock)
arch/s390/kvm/vsie.c
1518
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
arch/s390/kvm/vsie.c
1568
mutex_init(&kvm->arch.vsie.mutex);
arch/s390/kvm/vsie.c
1569
INIT_RADIX_TREE(&kvm->arch.vsie.addr_to_page, GFP_KERNEL_ACCOUNT);
arch/s390/kvm/vsie.c
1578
mutex_lock(&kvm->arch.vsie.mutex);
arch/s390/kvm/vsie.c
1579
for (i = 0; i < kvm->arch.vsie.page_count; i++) {
arch/s390/kvm/vsie.c
1580
vsie_page = kvm->arch.vsie.pages[i];
arch/s390/kvm/vsie.c
1581
scoped_guard(spinlock, &kvm->arch.gmap->children_lock)
arch/s390/kvm/vsie.c
1584
kvm->arch.vsie.pages[i] = NULL;
arch/s390/kvm/vsie.c
1587
radix_tree_delete(&kvm->arch.vsie.addr_to_page,
arch/s390/kvm/vsie.c
1591
kvm->arch.vsie.page_count = 0;
arch/s390/kvm/vsie.c
1592
mutex_unlock(&kvm->arch.vsie.mutex);
arch/s390/kvm/vsie.c
1597
struct kvm_s390_sie_block *scb = READ_ONCE(vcpu->arch.vsie_block);
arch/s390/kvm/vsie.c
328
int fmt_h = vcpu->arch.sie_block->crycbd & CRYCB_FORMAT_MASK;
arch/s390/kvm/vsie.c
333
apie_h = vcpu->arch.sie_block->eca & ECA_APIE;
arch/s390/kvm/vsie.c
348
vcpu->kvm->arch.crypto.crycb,
arch/s390/kvm/vsie.c
356
ecb3_flags = scb_o->ecb3 & vcpu->arch.sie_block->ecb3 &
arch/s390/kvm/vsie.c
358
ecd_flags = scb_o->ecd & vcpu->arch.sie_block->ecd &
arch/s390/kvm/vsie.c
374
vcpu->kvm->arch.crypto.crycb->dea_wrapping_key_mask;
arch/s390/kvm/vsie.c
413
if (vcpu->kvm->arch.model.ibc && new_ibc) {
arch/s390/kvm/vsie.c
419
if (scb_s->ibc > vcpu->kvm->arch.model.ibc)
arch/s390/kvm/vsie.c
420
scb_s->ibc = vcpu->kvm->arch.model.ibc;
arch/sh/include/asm/syscall_32.h
86
int arch = AUDIT_ARCH_SH;
arch/sh/include/asm/syscall_32.h
89
arch |= __AUDIT_ARCH_LE;
arch/sh/include/asm/syscall_32.h
91
return arch;
arch/sh/kernel/dwarf.c
1110
INIT_LIST_HEAD(&me->arch.cie_list);
arch/sh/kernel/dwarf.c
1111
INIT_LIST_HEAD(&me->arch.fde_list);
arch/sh/kernel/dwarf.c
1138
list_for_each_entry_safe(cie, ctmp, &mod->arch.cie_list, link) {
arch/sh/kernel/dwarf.c
1148
list_for_each_entry_safe(fde, ftmp, &mod->arch.fde_list, link) {
arch/sh/kernel/dwarf.c
857
list_add_tail(&cie->link, &mod->arch.cie_list);
arch/sh/kernel/dwarf.c
949
list_add_tail(&fde->link, &mod->arch.fde_list);
arch/sparc/kernel/audit.c
34
int audit_classify_arch(int arch)
arch/sparc/kernel/audit.c
37
if (arch == AUDIT_ARCH_SPARC)
arch/um/include/asm/processor-generic.h
25
struct arch_thread arch;
arch/um/include/asm/processor-generic.h
44
.arch = INIT_ARCH_THREAD, \
arch/um/kernel/exec.c
25
arch_flush_thread(&current->thread.arch);
arch/um/kernel/process.c
165
arch_copy_thread(&current->thread.arch, &p->thread.arch);
arch/um/kernel/skas/stub_exe.c
157
offsetof(struct seccomp_data, arch)),
arch/um/kernel/trap.c
262
current->thread.arch.faultinfo = fi;
arch/um/kernel/trap.c
382
current->thread.arch.faultinfo = fi;
arch/um/kernel/trap.c
386
current->thread.arch.faultinfo = fi;
arch/um/kernel/trap.c
417
current->thread.arch.faultinfo = *fi;
arch/x86/entry/vsyscall/vsyscall_trace.h
28
#define TRACE_INCLUDE_PATH ../../arch/x86/entry/vsyscall/
arch/x86/include/asm/kvm_host.h
2094
((vcpu) && (vcpu)->arch.handling_intr_from_guest && \
arch/x86/include/asm/kvm_host.h
2095
(!!in_nmi() == ((vcpu)->arch.handling_intr_from_guest == KVM_HANDLING_NMI)))
arch/x86/include/asm/kvm_host.h
2373
#define kvm_arch_has_private_mem(kvm) ((kvm)->arch.has_private_mem)
arch/x86/include/asm/kvm_host.h
2376
#define kvm_arch_has_readonly_mem(kvm) (!(kvm)->arch.has_protected_state)
arch/x86/include/asm/kvm_host.h
2428
# define kvm_arch_vcpu_memslots_id(vcpu) ((vcpu)->arch.hflags & HF_SMM_MASK ? 1 : 0)
arch/x86/kernel/alternative.c
211
its_pages_protect(&mod->arch.its_pages);
arch/x86/kernel/alternative.c
219
for (int i = 0; i < mod->arch.its_pages.num; i++) {
arch/x86/kernel/alternative.c
220
void *page = mod->arch.its_pages.pages[i];
arch/x86/kernel/alternative.c
223
kfree(mod->arch.its_pages.pages);
arch/x86/kernel/alternative.c
234
pages = &its_mod->arch.its_pages;
arch/x86/kernel/asm-offsets.c
81
OFFSET(XEN_vcpu_info_arch_cr2, vcpu_info, arch.cr2);
arch/x86/kernel/audit_64.c
33
int audit_classify_arch(int arch)
arch/x86/kernel/audit_64.c
36
if (arch == AUDIT_ARCH_I386)
arch/x86/kernel/machine_kexec_32.c
107
pmd = image->arch.pmd0;
arch/x86/kernel/machine_kexec_32.c
110
image->arch.pgd, pmd, image->arch.pte0,
arch/x86/kernel/machine_kexec_32.c
113
pmd = image->arch.pmd1;
arch/x86/kernel/machine_kexec_32.c
116
image->arch.pgd, pmd, image->arch.pte1,
arch/x86/kernel/machine_kexec_32.c
198
page_list[PA_PGD] = __pa(image->arch.pgd);
arch/x86/kernel/machine_kexec_32.c
45
free_pages((unsigned long)image->arch.pgd, pgd_allocation_order());
arch/x86/kernel/machine_kexec_32.c
46
image->arch.pgd = NULL;
arch/x86/kernel/machine_kexec_32.c
48
free_page((unsigned long)image->arch.pmd0);
arch/x86/kernel/machine_kexec_32.c
49
image->arch.pmd0 = NULL;
arch/x86/kernel/machine_kexec_32.c
50
free_page((unsigned long)image->arch.pmd1);
arch/x86/kernel/machine_kexec_32.c
51
image->arch.pmd1 = NULL;
arch/x86/kernel/machine_kexec_32.c
53
free_page((unsigned long)image->arch.pte0);
arch/x86/kernel/machine_kexec_32.c
54
image->arch.pte0 = NULL;
arch/x86/kernel/machine_kexec_32.c
55
free_page((unsigned long)image->arch.pte1);
arch/x86/kernel/machine_kexec_32.c
56
image->arch.pte1 = NULL;
arch/x86/kernel/machine_kexec_32.c
61
image->arch.pgd = (pgd_t *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
arch/x86/kernel/machine_kexec_32.c
64
image->arch.pmd0 = (pmd_t *)get_zeroed_page(GFP_KERNEL);
arch/x86/kernel/machine_kexec_32.c
65
image->arch.pmd1 = (pmd_t *)get_zeroed_page(GFP_KERNEL);
arch/x86/kernel/machine_kexec_32.c
67
image->arch.pte0 = (pte_t *)get_zeroed_page(GFP_KERNEL);
arch/x86/kernel/machine_kexec_32.c
68
image->arch.pte1 = (pte_t *)get_zeroed_page(GFP_KERNEL);
arch/x86/kernel/machine_kexec_32.c
69
if (!image->arch.pgd ||
arch/x86/kernel/machine_kexec_32.c
71
!image->arch.pmd0 || !image->arch.pmd1 ||
arch/x86/kernel/machine_kexec_32.c
73
!image->arch.pte0 || !image->arch.pte1) {
arch/x86/kernel/machine_kexec_64.c
153
free_page((unsigned long)image->arch.p4d);
arch/x86/kernel/machine_kexec_64.c
154
image->arch.p4d = NULL;
arch/x86/kernel/machine_kexec_64.c
155
free_page((unsigned long)image->arch.pud);
arch/x86/kernel/machine_kexec_64.c
156
image->arch.pud = NULL;
arch/x86/kernel/machine_kexec_64.c
157
free_page((unsigned long)image->arch.pmd);
arch/x86/kernel/machine_kexec_64.c
158
image->arch.pmd = NULL;
arch/x86/kernel/machine_kexec_64.c
159
free_page((unsigned long)image->arch.pte);
arch/x86/kernel/machine_kexec_64.c
160
image->arch.pte = NULL;
arch/x86/kernel/machine_kexec_64.c
186
image->arch.p4d = p4d;
arch/x86/kernel/machine_kexec_64.c
194
image->arch.pud = pud;
arch/x86/kernel/machine_kexec_64.c
202
image->arch.pmd = pmd;
arch/x86/kernel/machine_kexec_64.c
210
image->arch.pte = pte;
arch/x86/kernel/machine_kexec_64.c
251
image->arch.pgd = alloc_pgt_page(image);
arch/x86/kernel/machine_kexec_64.c
252
if (!image->arch.pgd)
arch/x86/kernel/machine_kexec_64.c
267
result = kernel_ident_mapping_init(&info, image->arch.pgd,
arch/x86/kernel/machine_kexec_64.c
283
result = kernel_ident_mapping_init(&info, image->arch.pgd,
arch/x86/kernel/machine_kexec_64.c
294
result = map_efi_systab(&info, image->arch.pgd);
arch/x86/kernel/machine_kexec_64.c
298
result = map_acpi_tables(&info, image->arch.pgd);
arch/x86/kernel/machine_kexec_64.c
302
result = map_mmio_serial(&info, image->arch.pgd);
arch/x86/kernel/machine_kexec_64.c
310
return init_transition_pgtable(image, image->arch.pgd, control_page);
arch/x86/kernel/machine_kexec_64.c
371
kexec_pa_table_page = (unsigned long)__pa(image->arch.pgd);
arch/x86/kernel/unwind_orc.c
119
if (!mod || !mod->arch.orc_unwind || !mod->arch.orc_unwind_ip)
arch/x86/kernel/unwind_orc.c
121
return __orc_find(mod->arch.orc_unwind_ip, mod->arch.orc_unwind,
arch/x86/kernel/unwind_orc.c
122
mod->arch.num_orcs, ip);
arch/x86/kernel/unwind_orc.c
327
mod->arch.orc_unwind_ip = orc_ip;
arch/x86/kernel/unwind_orc.c
328
mod->arch.orc_unwind = orc;
arch/x86/kernel/unwind_orc.c
329
mod->arch.num_orcs = num_entries;
arch/x86/kvm/cpuid.c
169
return fpu_enable_guest_xfd_features(&vcpu->arch.guest_fpu, xfeatures);
arch/x86/kvm/cpuid.c
191
if (nent != vcpu->arch.cpuid_nent)
arch/x86/kvm/cpuid.c
195
orig = &vcpu->arch.cpuid_entries[i];
arch/x86/kvm/cpuid.c
2095
if (vcpu->arch.cpuid_dynamic_bits_dirty)
arch/x86/kvm/cpuid.c
2131
*ecx = vcpu->arch.pvclock_tsc_mul;
arch/x86/kvm/cpuid.c
2132
*edx = vcpu->arch.pvclock_tsc_shift;
arch/x86/kvm/cpuid.c
2134
*eax = vcpu->arch.hw_tsc_khz;
arch/x86/kvm/cpuid.c
293
vcpu->arch.cpuid_dynamic_bits_dirty = false;
arch/x86/kvm/cpuid.c
301
vcpu->arch.apic_base & MSR_IA32_APICBASE_ENABLE);
arch/x86/kvm/cpuid.c
305
vcpu->arch.ia32_misc_enable_msr &
arch/x86/kvm/cpuid.c
317
best->ebx = xstate_required_size(vcpu->arch.xcr0, false);
arch/x86/kvm/cpuid.c
322
best->ebx = xstate_required_size(vcpu->arch.xcr0 |
arch/x86/kvm/cpuid.c
323
vcpu->arch.ia32_xss, true);
arch/x86/kvm/cpuid.c
377
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/cpuid.c
383
memset(vcpu->arch.cpu_caps, 0, sizeof(vcpu->arch.cpu_caps));
arch/x86/kvm/cpuid.c
409
vcpu->arch.cpu_caps[i] = kvm_cpu_caps[i] |
arch/x86/kvm/cpuid.c
411
vcpu->arch.cpu_caps[i] &= cpuid_get_reg_unsafe(entry, cpuid.reg);
arch/x86/kvm/cpuid.c
441
vcpu->arch.guest_supported_xcr0 = cpuid_get_supported_xcr0(vcpu);
arch/x86/kvm/cpuid.c
442
vcpu->arch.guest_supported_xss = cpuid_get_supported_xss(vcpu);
arch/x86/kvm/cpuid.c
444
vcpu->arch.pv_cpuid.features = kvm_apply_cpuid_pv_features_quirk(vcpu);
arch/x86/kvm/cpuid.c
446
vcpu->arch.is_amd_compatible = guest_cpuid_is_amd_or_hygon(vcpu);
arch/x86/kvm/cpuid.c
447
vcpu->arch.maxphyaddr = cpuid_query_maxphyaddr(vcpu);
arch/x86/kvm/cpuid.c
448
vcpu->arch.reserved_gpa_bits = kvm_vcpu_reserved_gpa_bits_raw(vcpu);
arch/x86/kvm/cpuid.c
453
vcpu->arch.cr4_guest_rsvd_bits = __cr4_reserved_bits(__kvm_cpu_cap_has, UNUSED_) |
arch/x86/kvm/cpuid.c
519
if (vcpu->arch.cpuid_dynamic_bits_dirty)
arch/x86/kvm/cpuid.c
533
swap(vcpu->arch.cpuid_entries, e2);
arch/x86/kvm/cpuid.c
534
swap(vcpu->arch.cpuid_nent, nent);
arch/x86/kvm/cpuid.c
536
memcpy(vcpu_caps, vcpu->arch.cpu_caps, sizeof(vcpu_caps));
arch/x86/kvm/cpuid.c
537
BUILD_BUG_ON(sizeof(vcpu_caps) != sizeof(vcpu->arch.cpu_caps));
arch/x86/kvm/cpuid.c
573
vcpu->arch.xen.cpuid = kvm_get_hypervisor_cpuid(vcpu, XEN_SIGNATURE);
arch/x86/kvm/cpuid.c
582
memcpy(vcpu->arch.cpu_caps, vcpu_caps, sizeof(vcpu_caps));
arch/x86/kvm/cpuid.c
583
swap(vcpu->arch.cpuid_entries, e2);
arch/x86/kvm/cpuid.c
584
swap(vcpu->arch.cpuid_nent, nent);
arch/x86/kvm/cpuid.c
661
if (cpuid->nent < vcpu->arch.cpuid_nent)
arch/x86/kvm/cpuid.c
664
if (vcpu->arch.cpuid_dynamic_bits_dirty)
arch/x86/kvm/cpuid.c
667
if (copy_to_user(entries, vcpu->arch.cpuid_entries,
arch/x86/kvm/cpuid.c
668
vcpu->arch.cpuid_nent * sizeof(struct kvm_cpuid_entry2)))
arch/x86/kvm/cpuid.c
671
cpuid->nent = vcpu->arch.cpuid_nent;
arch/x86/kvm/cpuid.h
138
return vcpu->arch.is_amd_compatible;
arch/x86/kvm/cpuid.h
186
return vcpu->arch.msr_platform_info & MSR_PLATFORM_INFO_CPUID_FAULT;
arch/x86/kvm/cpuid.h
191
return vcpu->arch.msr_misc_features_enables &
arch/x86/kvm/cpuid.h
232
if (!vcpu->arch.pv_cpuid.enforce)
arch/x86/kvm/cpuid.h
235
return vcpu->arch.pv_cpuid.features & (1u << kvm_feature);
arch/x86/kvm/cpuid.h
243
vcpu->arch.cpu_caps[x86_leaf] |= __feature_bit(x86_feature);
arch/x86/kvm/cpuid.h
251
vcpu->arch.cpu_caps[x86_leaf] &= ~__feature_bit(x86_feature);
arch/x86/kvm/cpuid.h
277
return vcpu->arch.cpu_caps[x86_leaf] & __feature_bit(x86_feature);
arch/x86/kvm/cpuid.h
39
return kvm_find_cpuid_entry2(vcpu->arch.cpuid_entries, vcpu->arch.cpuid_nent,
arch/x86/kvm/cpuid.h
46
return kvm_find_cpuid_entry2(vcpu->arch.cpuid_entries, vcpu->arch.cpuid_nent,
arch/x86/kvm/cpuid.h
74
return vcpu->arch.maxphyaddr;
arch/x86/kvm/cpuid.h
79
return !(gpa & vcpu->arch.reserved_gpa_bits);
arch/x86/kvm/debugfs.c
120
rmap = slot->arch.rmap[k];
arch/x86/kvm/debugfs.c
18
*val = vcpu->arch.apic->lapic_timer.timer_advance_ns;
arch/x86/kvm/debugfs.c
36
*val = vcpu->arch.tsc_offset;
arch/x86/kvm/debugfs.c
45
*val = vcpu->arch.tsc_scaling_ratio;
arch/x86/kvm/hyperv.c
1367
if (!vcpu->arch.hyperv_enabled ||
arch/x86/kvm/hyperv.c
137
down_write(&vcpu->kvm->arch.apicv_update_lock);
arch/x86/kvm/hyperv.c
152
up_write(&vcpu->kvm->arch.apicv_update_lock);
arch/x86/kvm/hyperv.c
1738
data = (u64)vcpu->arch.virtual_tsc_khz * 1000;
arch/x86/kvm/hyperv.c
1742
vcpu->kvm->arch.apic_bus_cycle_ns);
arch/x86/kvm/hyperv.c
1756
if (!host && !vcpu->arch.hyperv_enabled)
arch/x86/kvm/hyperv.c
1777
if (!host && !vcpu->arch.hyperv_enabled)
arch/x86/kvm/hyperv.c
2311
vcpu->arch.hyperv_enabled = hyperv_enabled;
arch/x86/kvm/hyperv.c
2318
WARN_ON_ONCE(vcpu->arch.hyperv_enabled);
arch/x86/kvm/hyperv.c
2324
if (!vcpu->arch.hyperv_enabled)
arch/x86/kvm/hyperv.c
2694
vcpu->arch.complete_userspace_io = kvm_hv_hypercall_complete_userspace;
arch/x86/kvm/hyperv.c
357
vcpu->arch.complete_userspace_io =
arch/x86/kvm/hyperv.c
495
ret = kvm_irq_delivery_to_apic(vcpu->kvm, vcpu->arch.apic, &irq);
arch/x86/kvm/hyperv.c
912
vcpu->arch.hyperv = NULL;
arch/x86/kvm/hyperv.c
924
return vcpu->arch.pv_eoi.msr_val & KVM_MSR_ENABLED;
arch/x86/kvm/hyperv.c
935
return kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.pv_eoi.data,
arch/x86/kvm/hyperv.c
975
vcpu->arch.hyperv = hv_vcpu;
arch/x86/kvm/hyperv.h
100
return vcpu->arch.hyperv_enabled && to_kvm_hv(vcpu->kvm)->hv_guest_os_id;
arch/x86/kvm/hyperv.h
61
return &kvm->arch.hyperv;
arch/x86/kvm/hyperv.h
66
return vcpu->arch.hyperv;
arch/x86/kvm/hyperv.h
85
return &vcpu->kvm->arch.hyperv.hv_syndbg;
arch/x86/kvm/i8254.c
220
struct kvm_pit *pit = vcpu->kvm->arch.vpit;
arch/x86/kvm/i8254.c
263
if (atomic_read(&kvm->arch.vapics_in_nmi_mode) > 0)
arch/x86/kvm/i8254.c
654
struct kvm_kpit_state *kps = &kvm->arch.vpit->pit_state;
arch/x86/kvm/i8254.c
667
struct kvm_pit *pit = kvm->arch.vpit;
arch/x86/kvm/i8254.c
679
mutex_lock(&kvm->arch.vpit->pit_state.lock);
arch/x86/kvm/i8254.c
680
memcpy(ps->channels, &kvm->arch.vpit->pit_state.channels,
arch/x86/kvm/i8254.c
682
ps->flags = kvm->arch.vpit->pit_state.flags;
arch/x86/kvm/i8254.c
683
mutex_unlock(&kvm->arch.vpit->pit_state.lock);
arch/x86/kvm/i8254.c
693
struct kvm_pit *pit = kvm->arch.vpit;
arch/x86/kvm/i8254.c
712
struct kvm_pit *pit = kvm->arch.vpit;
arch/x86/kvm/i8254.c
804
struct kvm_pit *pit = kvm->arch.vpit;
arch/x86/kvm/i8259.c
193
struct kvm_pic *s = kvm->arch.vpic;
arch/x86/kvm/i8259.c
234
struct kvm_pic *s = kvm->arch.vpic;
arch/x86/kvm/i8259.c
563
struct kvm_pic *s = kvm->arch.vpic;
arch/x86/kvm/i8259.c
622
kvm->arch.vpic = s;
arch/x86/kvm/i8259.c
642
struct kvm_pic *vpic = kvm->arch.vpic;
arch/x86/kvm/i8259.c
653
kvm->arch.vpic = NULL;
arch/x86/kvm/ioapic.c
123
struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
261
struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
297
struct kvm_ioapic *ioapic = kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
317
struct kvm_ioapic *ioapic = kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
503
struct kvm_ioapic *ioapic = kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
543
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/ioapic.c
588
struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
727
kvm->arch.vioapic = ioapic;
arch/x86/kvm/ioapic.c
736
kvm->arch.vioapic = NULL;
arch/x86/kvm/ioapic.c
745
struct kvm_ioapic *ioapic = kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
754
kvm->arch.vioapic = NULL;
arch/x86/kvm/ioapic.c
760
struct kvm_ioapic *ioapic = kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
770
struct kvm_ioapic *ioapic = kvm->arch.vioapic;
arch/x86/kvm/ioapic.c
94
struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic;
arch/x86/kvm/irq.c
117
if (lapic_in_kernel(v) && v->arch.apic->guest_apic_protected)
arch/x86/kvm/irq.c
136
return v->arch.interrupt.nr;
arch/x86/kvm/irq.c
140
return v->kvm->arch.xen.upcall_vector;
arch/x86/kvm/irq.c
206
trace_kvm_msi_set_irq(msg.address_lo | (kvm->arch.x2apic_format ?
arch/x86/kvm/irq.c
209
irq->dest_id = x86_msi_msg_get_destid(&msg, kvm->arch.x2apic_format);
arch/x86/kvm/irq.c
222
return kvm->arch.x2apic_format && (e->msi.address_hi & 0xff);
arch/x86/kvm/irq.c
379
if ((int)vector > vcpu->arch.highest_stale_pending_ioapic_eoi)
arch/x86/kvm/irq.c
380
vcpu->arch.highest_stale_pending_ioapic_eoi = vector;
arch/x86/kvm/irq.c
396
kvm->arch.nr_reserved_ioapic_pins);
arch/x86/kvm/irq.c
44
return v->arch.pending_external_vector != -1;
arch/x86/kvm/irq.c
486
if (!kvm->arch.nr_possible_bypass_irqs++)
arch/x86/kvm/irq.c
49
int vector = vcpu->arch.pending_external_vector;
arch/x86/kvm/irq.c
492
kvm->arch.nr_possible_bypass_irqs--;
arch/x86/kvm/irq.c
51
vcpu->arch.pending_external_vector = -1;
arch/x86/kvm/irq.c
526
kvm->arch.nr_possible_bypass_irqs--;
arch/x86/kvm/irq.c
582
struct kvm_pic *pic = kvm->arch.vpic;
arch/x86/kvm/irq.c
607
struct kvm_pic *pic = kvm->arch.vpic;
arch/x86/kvm/irq.c
73
return v->arch.interrupt.injected;
arch/x86/kvm/irq.c
83
return v->kvm->arch.vpic->output;
arch/x86/kvm/irq.h
108
int mode = kvm->arch.irqchip_mode;
arch/x86/kvm/irq.h
78
int mode = kvm->arch.irqchip_mode;
arch/x86/kvm/irq.h
99
int mode = kvm->arch.irqchip_mode;
arch/x86/kvm/kvm-asm-offsets.c
17
OFFSET(SVM_vcpu_arch_regs, vcpu_svm, vcpu.arch.regs);
arch/x86/kvm/kvm_cache_regs.h
105
return arch___test_and_set_bit(reg, (unsigned long *)&vcpu->arch.regs_avail);
arch/x86/kvm/kvm_cache_regs.h
121
return vcpu->arch.regs[reg];
arch/x86/kvm/kvm_cache_regs.h
130
vcpu->arch.regs[reg] = val;
arch/x86/kvm/kvm_cache_regs.h
161
return vcpu->arch.walk_mmu->pdptrs[index];
arch/x86/kvm/kvm_cache_regs.h
166
vcpu->arch.walk_mmu->pdptrs[index] = value;
arch/x86/kvm/kvm_cache_regs.h
172
if ((tmask & vcpu->arch.cr0_guest_owned_bits) &&
arch/x86/kvm/kvm_cache_regs.h
175
return vcpu->arch.cr0 & mask;
arch/x86/kvm/kvm_cache_regs.h
194
if ((tmask & vcpu->arch.cr4_guest_owned_bits) &&
arch/x86/kvm/kvm_cache_regs.h
197
return vcpu->arch.cr4 & mask;
arch/x86/kvm/kvm_cache_regs.h
212
return vcpu->arch.cr3;
arch/x86/kvm/kvm_cache_regs.h
22
return vcpu->arch.regs[VCPU_REGS_##uname]; \
arch/x86/kvm/kvm_cache_regs.h
228
vcpu->arch.hflags |= HF_GUEST_MASK;
arch/x86/kvm/kvm_cache_regs.h
234
vcpu->arch.hflags &= ~HF_GUEST_MASK;
arch/x86/kvm/kvm_cache_regs.h
236
if (vcpu->arch.load_eoi_exitmap_pending) {
arch/x86/kvm/kvm_cache_regs.h
237
vcpu->arch.load_eoi_exitmap_pending = false;
arch/x86/kvm/kvm_cache_regs.h
246
return vcpu->arch.hflags & HF_GUEST_MASK;
arch/x86/kvm/kvm_cache_regs.h
27
vcpu->arch.regs[VCPU_REGS_##uname] = val; \
arch/x86/kvm/kvm_cache_regs.h
70
return test_bit(reg, (unsigned long *)&vcpu->arch.regs_avail);
arch/x86/kvm/kvm_cache_regs.h
77
return test_bit(reg, (unsigned long *)&vcpu->arch.regs_dirty);
arch/x86/kvm/kvm_cache_regs.h
84
__set_bit(reg, (unsigned long *)&vcpu->arch.regs_avail);
arch/x86/kvm/kvm_cache_regs.h
91
__set_bit(reg, (unsigned long *)&vcpu->arch.regs_avail);
arch/x86/kvm/kvm_cache_regs.h
92
__set_bit(reg, (unsigned long *)&vcpu->arch.regs_dirty);
arch/x86/kvm/kvm_onhyperv.c
114
struct kvm_arch *kvm_arch = &vcpu->kvm->arch;
arch/x86/kvm/kvm_onhyperv.c
118
vcpu->arch.hv_root_tdp = root_tdp;
arch/x86/kvm/kvm_onhyperv.c
40
struct kvm_arch *kvm_arch = &kvm->arch;
arch/x86/kvm/kvm_onhyperv.c
57
root = vcpu->arch.hv_root_tdp;
arch/x86/kvm/kvm_onhyperv.h
23
&vcpu->kvm->arch.hv_pa_pg;
arch/x86/kvm/lapic.c
102
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
1068
if (!vcpu->kvm->arch.x2apic_broadcast_quirk_disabled &&
arch/x86/kvm/lapic.c
1078
struct kvm_lapic *target = vcpu->arch.apic;
arch/x86/kvm/lapic.c
110
switch (kvm->arch.suppress_eoi_broadcast_mode) {
arch/x86/kvm/lapic.c
1110
if (!kvm->arch.disabled_lapic_found) {
arch/x86/kvm/lapic.c
1111
kvm->arch.disabled_lapic_found = true;
arch/x86/kvm/lapic.c
1119
if (kvm->arch.x2apic_broadcast_quirk_disabled) {
arch/x86/kvm/lapic.c
1142
return vcpu1->arch.apic_arb_prio - vcpu2->arch.apic_arb_prio;
arch/x86/kvm/lapic.c
1240
map = rcu_dereference(kvm->arch.apic_map);
arch/x86/kvm/lapic.c
1290
map = rcu_dereference(kvm->arch.apic_map);
arch/x86/kvm/lapic.c
1366
} else if (kvm_apic_sw_enabled(vcpu->arch.apic)) {
arch/x86/kvm/lapic.c
1407
vcpu->arch.apic_arb_prio++;
arch/x86/kvm/lapic.c
1439
vcpu->arch.pv.pv_unhalted = 1;
arch/x86/kvm/lapic.c
144
switch (kvm->arch.suppress_eoi_broadcast_mode) {
arch/x86/kvm/lapic.c
1511
map = rcu_dereference(kvm->arch.apic_map);
arch/x86/kvm/lapic.c
1539
return test_bit(vector, apic->vcpu->arch.ioapic_handled_vectors);
arch/x86/kvm/lapic.c
1555
if (apic->vcpu->arch.highest_stale_pending_ioapic_eoi == vector)
arch/x86/kvm/lapic.c
1569
apic->vcpu->arch.pending_ioapic_eoi = vector;
arch/x86/kvm/lapic.c
1614
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
1670
return div64_u64(ns, (apic->vcpu->kvm->arch.apic_bus_cycle_ns *
arch/x86/kvm/lapic.c
1686
if (apic->vcpu->arch.tpr_access_reporting)
arch/x86/kvm/lapic.c
1903
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
1930
u64 timer_advance_ns = vcpu->arch.apic->lapic_timer.timer_advance_ns;
arch/x86/kvm/lapic.c
1938
if (vcpu->arch.tsc_scaling_ratio == kvm_caps.default_tsc_scaling_ratio) {
arch/x86/kvm/lapic.c
1943
do_div(delay_ns, vcpu->arch.virtual_tsc_khz);
arch/x86/kvm/lapic.c
1951
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
1963
do_div(ns, vcpu->arch.virtual_tsc_khz);
arch/x86/kvm/lapic.c
1968
do_div(ns, vcpu->arch.virtual_tsc_khz);
arch/x86/kvm/lapic.c
1979
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2003
vcpu->arch.apic->lapic_timer.expired_tscdeadline &&
arch/x86/kvm/lapic.c
2004
vcpu->arch.apic->lapic_timer.timer_advance_ns &&
arch/x86/kvm/lapic.c
2048
if (vcpu->arch.apic->lapic_timer.expired_tscdeadline &&
arch/x86/kvm/lapic.c
2049
vcpu->arch.apic->lapic_timer.timer_advance_ns)
arch/x86/kvm/lapic.c
2068
u32 this_tsc_khz = vcpu->arch.virtual_tsc_khz;
arch/x86/kvm/lapic.c
2096
return (u64)tmict * apic->vcpu->kvm->arch.apic_bus_cycle_ns *
arch/x86/kvm/lapic.c
2238
return vcpu->arch.apic->lapic_timer.hv_timer_in_use;
arch/x86/kvm/lapic.c
2323
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2344
restart_apic_timer(vcpu->arch.apic);
arch/x86/kvm/lapic.c
2349
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2360
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2389
atomic_inc(&apic->vcpu->kvm->arch.vapics_in_nmi_mode);
arch/x86/kvm/lapic.c
2391
atomic_dec(&apic->vcpu->kvm->arch.vapics_in_nmi_mode);
arch/x86/kvm/lapic.c
255
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2591
kvm_lapic_reg_write(vcpu->arch.apic, APIC_EOI, 0);
arch/x86/kvm/lapic.c
2658
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2683
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2685
if (!vcpu->arch.apic) {
arch/x86/kvm/lapic.c
2692
if (!(vcpu->arch.apic_base & MSR_IA32_APICBASE_ENABLE))
arch/x86/kvm/lapic.c
2711
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2721
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2733
apic_set_tpr(vcpu->arch.apic, (cr8 & 0x0f) << 4);
arch/x86/kvm/lapic.c
2740
tpr = (u64) kvm_lapic_get_reg(vcpu->arch.apic, APIC_TASKPRI);
arch/x86/kvm/lapic.c
2747
u64 old_value = vcpu->arch.apic_base;
arch/x86/kvm/lapic.c
2748
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2750
vcpu->arch.apic_base = value;
arch/x86/kvm/lapic.c
2753
vcpu->arch.cpuid_dynamic_bits_dirty = true;
arch/x86/kvm/lapic.c
2767
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
arch/x86/kvm/lapic.c
2783
apic->base_address = apic->vcpu->arch.apic_base &
arch/x86/kvm/lapic.c
2798
if (vcpu->arch.apic_base == value)
arch/x86/kvm/lapic.c
2821
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2860
if (kvm->arch.apic_access_memslot_enabled ||
arch/x86/kvm/lapic.c
2861
kvm->arch.apic_access_memslot_inhibited)
arch/x86/kvm/lapic.c
2869
kvm->arch.apic_access_memslot_enabled = true;
arch/x86/kvm/lapic.c
2879
if (!kvm->arch.apic_access_memslot_enabled)
arch/x86/kvm/lapic.c
2886
if (kvm->arch.apic_access_memslot_enabled) {
arch/x86/kvm/lapic.c
2896
kvm->arch.apic_access_memslot_enabled = false;
arch/x86/kvm/lapic.c
2902
kvm->arch.apic_access_memslot_inhibited = true;
arch/x86/kvm/lapic.c
2912
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
2975
vcpu->arch.pv_eoi.msr_val = 0;
arch/x86/kvm/lapic.c
2980
vcpu->arch.apic_arb_prio = 0;
arch/x86/kvm/lapic.c
2981
vcpu->arch.apic_attention = 0;
arch/x86/kvm/lapic.c
2999
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
300
if (vcpu->kvm->arch.x2apic_format) {
arch/x86/kvm/lapic.c
3029
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3068
vcpu->arch.apic = apic;
arch/x86/kvm/lapic.c
3092
vcpu->arch.apic_base = MSR_IA32_APICBASE_ENABLE;
arch/x86/kvm/lapic.c
3114
vcpu->arch.apic = NULL;
arch/x86/kvm/lapic.c
3121
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3137
u32 lvt0 = kvm_lapic_get_reg(vcpu->arch.apic, APIC_LVT0);
arch/x86/kvm/lapic.c
3139
if (!kvm_apic_hw_enabled(vcpu->arch.apic))
arch/x86/kvm/lapic.c
3149
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3159
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3197
if (apic_x2apic_mode(vcpu->arch.apic)) {
arch/x86/kvm/lapic.c
3198
u32 x2apic_id = kvm_x2apic_id(vcpu->arch.apic);
arch/x86/kvm/lapic.c
3203
if (vcpu->kvm->arch.x2apic_format) {
arch/x86/kvm/lapic.c
3248
memcpy(s->regs, vcpu->arch.apic->regs, sizeof(*s));
arch/x86/kvm/lapic.c
3254
apic_set_reg(s->regs, APIC_TMCCT, __apic_read(vcpu->arch.apic, APIC_TMCCT));
arch/x86/kvm/lapic.c
3261
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3274
memcpy(vcpu->arch.apic->regs, s->regs, sizeof(*s));
arch/x86/kvm/lapic.c
3276
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
arch/x86/kvm/lapic.c
3298
vcpu->arch.apic_arb_prio = 0;
arch/x86/kvm/lapic.c
330
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3311
timer = &vcpu->arch.apic->lapic_timer.timer;
arch/x86/kvm/lapic.c
3350
if (test_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention))
arch/x86/kvm/lapic.c
3351
apic_sync_pv_eoi_from_guest(vcpu, vcpu->arch.apic);
arch/x86/kvm/lapic.c
3353
if (!test_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention))
arch/x86/kvm/lapic.c
3356
if (kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.apic->vapic_cache, &data,
arch/x86/kvm/lapic.c
3360
apic_set_tpr(vcpu->arch.apic, data & 0xff);
arch/x86/kvm/lapic.c
3393
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3397
if (!test_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention))
arch/x86/kvm/lapic.c
3409
kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.apic->vapic_cache, &data,
arch/x86/kvm/lapic.c
3417
&vcpu->arch.apic->vapic_cache,
arch/x86/kvm/lapic.c
3420
__set_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention);
arch/x86/kvm/lapic.c
3422
__clear_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention);
arch/x86/kvm/lapic.c
3425
vcpu->arch.apic->vapic_addr = vapic_addr;
arch/x86/kvm/lapic.c
3465
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3476
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3490
return kvm_lapic_msr_write(vcpu->arch.apic, reg, data);
arch/x86/kvm/lapic.c
3498
return kvm_lapic_msr_read(vcpu->arch.apic, reg, data);
arch/x86/kvm/lapic.c
3504
struct gfn_to_hva_cache *ghc = &vcpu->arch.pv_eoi.data;
arch/x86/kvm/lapic.c
3522
vcpu->arch.pv_eoi.msr_val = data;
arch/x86/kvm/lapic.c
3529
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
3565
if (vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED) {
arch/x86/kvm/lapic.c
415
if (atomic_read_acquire(&kvm->arch.apic_map_dirty) == CLEAN)
arch/x86/kvm/lapic.c
421
mutex_lock(&kvm->arch.apic_map_lock);
arch/x86/kvm/lapic.c
431
if (atomic_cmpxchg_acquire(&kvm->arch.apic_map_dirty,
arch/x86/kvm/lapic.c
434
mutex_unlock(&kvm->arch.apic_map_lock);
arch/x86/kvm/lapic.c
449
max_id = max(max_id, kvm_x2apic_id(vcpu->arch.apic));
arch/x86/kvm/lapic.c
500
old = rcu_dereference_protected(kvm->arch.apic_map,
arch/x86/kvm/lapic.c
501
lockdep_is_held(&kvm->arch.apic_map_lock));
arch/x86/kvm/lapic.c
502
rcu_assign_pointer(kvm->arch.apic_map, new);
arch/x86/kvm/lapic.c
507
atomic_cmpxchg_release(&kvm->arch.apic_map_dirty,
arch/x86/kvm/lapic.c
509
mutex_unlock(&kvm->arch.apic_map_lock);
arch/x86/kvm/lapic.c
530
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
arch/x86/kvm/lapic.c
543
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
arch/x86/kvm/lapic.c
549
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
arch/x86/kvm/lapic.c
555
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
arch/x86/kvm/lapic.c
566
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
arch/x86/kvm/lapic.c
601
return KVM_APIC_MAX_NR_LVT_ENTRIES - !(vcpu->arch.mcg_cap & MCG_CMCI_P);
arch/x86/kvm/lapic.c
606
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
624
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
704
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
744
apic_clear_irr(vec, vcpu->arch.apic);
arch/x86/kvm/lapic.c
821
return apic_find_highest_irr(vcpu->arch.apic);
arch/x86/kvm/lapic.c
832
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/lapic.c
878
map = rcu_dereference(kvm->arch.apic_map);
arch/x86/kvm/lapic.c
894
return kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.pv_eoi.data, &val,
arch/x86/kvm/lapic.c
901
return kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.pv_eoi.data, val,
arch/x86/kvm/lapic.c
907
return vcpu->arch.pv_eoi.msr_val & KVM_MSR_ENABLED;
arch/x86/kvm/lapic.c
915
__set_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention);
arch/x86/kvm/lapic.c
935
__clear_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention);
arch/x86/kvm/lapic.c
985
apic_update_ppr(vcpu->arch.apic);
arch/x86/kvm/lapic.h
181
return vcpu->arch.apic;
arch/x86/kvm/lapic.h
190
return apic->vcpu->arch.apic_base & MSR_IA32_APICBASE_ENABLE;
arch/x86/kvm/lapic.h
205
return lapic_in_kernel(vcpu) && kvm_apic_hw_enabled(vcpu->arch.apic);
arch/x86/kvm/lapic.h
210
return kvm_apic_present(vcpu) && kvm_apic_sw_enabled(vcpu->arch.apic);
arch/x86/kvm/lapic.h
215
return apic->vcpu->arch.apic_base & X2APIC_ENABLE;
arch/x86/kvm/lapic.h
220
return lapic_in_kernel(vcpu) && vcpu->arch.apic->apicv_active;
arch/x86/kvm/lapic.h
225
return lapic_in_kernel(vcpu) && vcpu->arch.apic->pending_events;
arch/x86/kvm/lapic.h
236
return lapic_in_kernel(vcpu) && test_bit(KVM_APIC_INIT, &vcpu->arch.apic->pending_events);
arch/x86/kvm/lapic.h
263
return kvm_apic_mode(vcpu->arch.apic_base);
arch/x86/kvm/mmu.h
120
if (likely(vcpu->arch.mmu->root.hpa != INVALID_PAGE))
arch/x86/kvm/mmu.h
150
u64 root_hpa = vcpu->arch.mmu->root.hpa;
arch/x86/kvm/mmu.h
156
vcpu->arch.mmu->root_role.level);
arch/x86/kvm/mmu.h
171
if (!tdp_enabled || mmu == &vcpu->arch.guest_mmu)
arch/x86/kvm/mmu.h
225
pkru_bits = (vcpu->arch.pkru >> (pte_pkey * 2)) & 3;
arch/x86/kvm/mmu.h
249
return smp_load_acquire(&kvm->arch.shadow_root_allocated);
arch/x86/kvm/mmu.h
299
if (mmu != &vcpu->arch.nested_mmu)
arch/x86/kvm/mmu.h
306
return kvm->arch.vm_type == KVM_X86_TDX_VM;
arch/x86/kvm/mmu.h
311
return kvm->arch.gfn_direct_bits;
arch/x86/kvm/mmu/mmu.c
1181
return &slot->arch.rmap[level - PG_LEVEL_4K][idx];
arch/x86/kvm/mmu/mmu.c
1464
if (kvm->arch.cpu_dirty_log_size)
arch/x86/kvm/mmu/mmu.c
1472
return kvm->arch.cpu_dirty_log_size;
arch/x86/kvm/mmu/mmu.c
1706
struct kvm_mmu_memory_cache *cache = &vcpu->arch.mmu_pte_list_desc_cache;
arch/x86/kvm/mmu/mmu.c
1761
return !tdp_mmu_enabled || READ_ONCE(kvm->arch.indirect_shadow_pages);
arch/x86/kvm/mmu/mmu.c
1809
kvm->arch.n_used_mmu_pages++;
arch/x86/kvm/mmu/mmu.c
1815
kvm->arch.n_used_mmu_pages--;
arch/x86/kvm/mmu/mmu.c
2002
struct hlist_head *page_hash = smp_load_acquire(&kvm->arch.mmu_page_hash);
arch/x86/kvm/mmu/mmu.c
2023
union kvm_mmu_page_role root_role = vcpu->arch.mmu->root_role;
arch/x86/kvm/mmu/mmu.c
2047
if (WARN_ON_ONCE(sp->role.direct || !vcpu->arch.mmu->sync_spte ||
arch/x86/kvm/mmu/mmu.c
2060
return vcpu->arch.mmu->sync_spte(vcpu, sp, i);
arch/x86/kvm/mmu/mmu.c
2122
unlikely(sp->mmu_valid_gen != kvm->arch.mmu_valid_gen);
arch/x86/kvm/mmu/mmu.c
2362
sp->mmu_valid_gen = kvm->arch.mmu_valid_gen;
arch/x86/kvm/mmu/mmu.c
2363
list_add(&sp->link, &kvm->arch.active_mmu_pages);
arch/x86/kvm/mmu/mmu.c
2391
BUG_ON(!kvm->arch.mmu_page_hash);
arch/x86/kvm/mmu/mmu.c
2392
sp_list = &kvm->arch.mmu_page_hash[kvm_page_table_hashfn(gfn)];
arch/x86/kvm/mmu/mmu.c
2409
.page_header_cache = &vcpu->arch.mmu_page_header_cache,
arch/x86/kvm/mmu/mmu.c
2410
.shadow_page_cache = &vcpu->arch.mmu_shadow_page_cache,
arch/x86/kvm/mmu/mmu.c
2411
.shadowed_info_cache = &vcpu->arch.mmu_shadowed_info_cache,
arch/x86/kvm/mmu/mmu.c
2482
iterator->level = vcpu->arch.mmu->root_role.level;
arch/x86/kvm/mmu/mmu.c
2485
vcpu->arch.mmu->cpu_role.base.level < PT64_ROOT_4LEVEL &&
arch/x86/kvm/mmu/mmu.c
2486
!vcpu->arch.mmu->root_role.direct)
arch/x86/kvm/mmu/mmu.c
2494
BUG_ON(root != vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/mmu.c
2497
= vcpu->arch.mmu->pae_root[(addr >> 30) & 3];
arch/x86/kvm/mmu/mmu.c
2508
shadow_walk_init_using_root(iterator, vcpu, vcpu->arch.mmu->root.hpa,
arch/x86/kvm/mmu/mmu.c
252
.efer = vcpu->arch.efer,
arch/x86/kvm/mmu/mmu.c
2577
__link_shadow_page(vcpu->kvm, &vcpu->arch.mmu_pte_list_desc_cache, sptep, sp, true);
arch/x86/kvm/mmu/mmu.c
2788
if (list_empty(&kvm->arch.active_mmu_pages))
arch/x86/kvm/mmu/mmu.c
2792
list_for_each_entry_safe_reverse(sp, tmp, &kvm->arch.active_mmu_pages, link) {
arch/x86/kvm/mmu/mmu.c
2818
if (kvm->arch.n_max_mmu_pages > kvm->arch.n_used_mmu_pages)
arch/x86/kvm/mmu/mmu.c
2819
return kvm->arch.n_max_mmu_pages -
arch/x86/kvm/mmu/mmu.c
2820
kvm->arch.n_used_mmu_pages;
arch/x86/kvm/mmu/mmu.c
2856
if (kvm->arch.n_used_mmu_pages > goal_nr_mmu_pages) {
arch/x86/kvm/mmu/mmu.c
2857
kvm_mmu_zap_oldest_mmu_pages(kvm, kvm->arch.n_used_mmu_pages -
arch/x86/kvm/mmu/mmu.c
2860
goal_nr_mmu_pages = kvm->arch.n_used_mmu_pages;
arch/x86/kvm/mmu/mmu.c
2863
kvm->arch.n_max_mmu_pages = goal_nr_mmu_pages;
arch/x86/kvm/mmu/mmu.c
2885
if (!READ_ONCE(kvm->arch.indirect_shadow_pages))
arch/x86/kvm/mmu/mmu.c
2888
if (!vcpu->arch.mmu->root_role.direct) {
arch/x86/kvm/mmu/mmu.c
2908
vcpu->arch.last_retry_eip = kvm_rip_read(vcpu);
arch/x86/kvm/mmu/mmu.c
2909
vcpu->arch.last_retry_addr = cr2_or_gpa;
arch/x86/kvm/mmu/mmu.c
2977
spin_lock(&kvm->arch.mmu_unsync_pages_lock);
arch/x86/kvm/mmu/mmu.c
2995
spin_unlock(&kvm->arch.mmu_unsync_pages_lock);
arch/x86/kvm/mmu/mmu.c
3073
WARN_ON_ONCE(vcpu->arch.mmu->root_role.direct);
arch/x86/kvm/mmu/mmu.c
3585
if (kvm->arch.has_private_mem &&
arch/x86/kvm/mmu/mmu.c
3912
union kvm_mmu_page_role role = vcpu->arch.mmu->root_role;
arch/x86/kvm/mmu/mmu.c
3929
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/mmu/mmu.c
3983
if (kvm->arch.mmu_page_hash)
arch/x86/kvm/mmu/mmu.c
3997
smp_store_release(&kvm->arch.mmu_page_hash, h);
arch/x86/kvm/mmu/mmu.c
4059
smp_store_release(&kvm->arch.shadow_root_allocated, true);
arch/x86/kvm/mmu/mmu.c
4068
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/mmu/mmu.c
4186
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/mmu/mmu.c
4296
if (vcpu->arch.mmu->root_role.direct)
arch/x86/kvm/mmu/mmu.c
4299
if (!VALID_PAGE(vcpu->arch.mmu->root.hpa))
arch/x86/kvm/mmu/mmu.c
4304
if (vcpu->arch.mmu->cpu_role.base.level >= PT64_ROOT_4LEVEL) {
arch/x86/kvm/mmu/mmu.c
4305
hpa_t root = vcpu->arch.mmu->root.hpa;
arch/x86/kvm/mmu/mmu.c
4321
hpa_t root = vcpu->arch.mmu->pae_root[i];
arch/x86/kvm/mmu/mmu.c
4338
if (is_unsync_root(vcpu->arch.mmu->prev_roots[i].hpa))
arch/x86/kvm/mmu/mmu.c
4342
kvm_mmu_free_roots(vcpu->kvm, vcpu->arch.mmu, roots_to_free);
arch/x86/kvm/mmu/mmu.c
4435
rsvd_check = &vcpu->arch.mmu->shadow_zero_check;
arch/x86/kvm/mmu/mmu.c
4519
u32 id = vcpu->arch.apf.id;
arch/x86/kvm/mmu/mmu.c
4522
vcpu->arch.apf.id = 1;
arch/x86/kvm/mmu/mmu.c
4524
return (vcpu->arch.apf.id++ << 12) | vcpu->vcpu_id;
arch/x86/kvm/mmu/mmu.c
4530
struct kvm_arch_async_pf arch;
arch/x86/kvm/mmu/mmu.c
4532
arch.token = alloc_apf_token(vcpu);
arch/x86/kvm/mmu/mmu.c
4533
arch.gfn = fault->gfn;
arch/x86/kvm/mmu/mmu.c
4534
arch.error_code = fault->error_code;
arch/x86/kvm/mmu/mmu.c
4535
arch.direct_map = vcpu->arch.mmu->root_role.direct;
arch/x86/kvm/mmu/mmu.c
4536
if (arch.direct_map)
arch/x86/kvm/mmu/mmu.c
4537
arch.cr3 = (unsigned long)INVALID_GPA;
arch/x86/kvm/mmu/mmu.c
4539
arch.cr3 = kvm_mmu_get_guest_pgd(vcpu, vcpu->arch.mmu);
arch/x86/kvm/mmu/mmu.c
4542
kvm_vcpu_gfn_to_hva(vcpu, fault->gfn), &arch);
arch/x86/kvm/mmu/mmu.c
4549
if (WARN_ON_ONCE(work->arch.error_code & PFERR_PRIVATE_ACCESS))
arch/x86/kvm/mmu/mmu.c
4552
if ((vcpu->arch.mmu->root_role.direct != work->arch.direct_map) ||
arch/x86/kvm/mmu/mmu.c
4560
if (!vcpu->arch.mmu->root_role.direct &&
arch/x86/kvm/mmu/mmu.c
4561
work->arch.cr3 != kvm_mmu_get_guest_pgd(vcpu, vcpu->arch.mmu))
arch/x86/kvm/mmu/mmu.c
4564
r = kvm_mmu_do_page_fault(vcpu, work->cr2_or_gpa, work->arch.error_code,
arch/x86/kvm/mmu/mmu.c
4777
struct kvm_mmu_page *sp = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/mmu.c
4808
if (WARN_ON_ONCE(kvm_mmu_is_dummy_root(vcpu->arch.mmu->root.hpa)))
arch/x86/kvm/mmu/mmu.c
4856
u32 flags = vcpu->arch.apf.host_apf_flags;
arch/x86/kvm/mmu/mmu.c
4884
vcpu->arch.apf.host_apf_flags = 0;
arch/x86/kvm/mmu/mmu.c
4951
if (vcpu->arch.mmu->page_fault != kvm_tdp_page_fault)
arch/x86/kvm/mmu/mmu.c
4995
if (!vcpu->kvm->arch.pre_fault_allowed)
arch/x86/kvm/mmu/mmu.c
5218
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/mmu/mmu.c
5254
struct kvm_mmu_page *sp = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/mmu.c
5384
vcpu->arch.reserved_gpa_bits,
arch/x86/kvm/mmu/mmu.c
5433
vcpu->arch.reserved_gpa_bits, execonly,
arch/x86/kvm/mmu/mmu.c
557
return tdp_mmu_enabled && vcpu->arch.mmu->root_role.direct;
arch/x86/kvm/mmu/mmu.c
5766
if (vcpu->kvm->arch.vm_type == KVM_X86_TDX_VM)
arch/x86/kvm/mmu/mmu.c
5809
struct kvm_mmu *context = &vcpu->arch.root_mmu;
arch/x86/kvm/mmu/mmu.c
5860
struct kvm_mmu *context = &vcpu->arch.root_mmu;
arch/x86/kvm/mmu/mmu.c
5885
struct kvm_mmu *context = &vcpu->arch.guest_mmu;
arch/x86/kvm/mmu/mmu.c
5937
struct kvm_mmu *context = &vcpu->arch.guest_mmu;
arch/x86/kvm/mmu/mmu.c
5965
struct kvm_mmu *context = &vcpu->arch.root_mmu;
arch/x86/kvm/mmu/mmu.c
5977
struct kvm_mmu *g_context = &vcpu->arch.nested_mmu;
arch/x86/kvm/mmu/mmu.c
599
r = kvm_mmu_topup_memory_cache(&vcpu->arch.mmu_pte_list_desc_cache,
arch/x86/kvm/mmu/mmu.c
604
r = kvm_mmu_topup_memory_cache(&vcpu->arch.mmu_external_spt_cache,
arch/x86/kvm/mmu/mmu.c
6041
vcpu->arch.root_mmu.root_role.invalid = 1;
arch/x86/kvm/mmu/mmu.c
6042
vcpu->arch.guest_mmu.root_role.invalid = 1;
arch/x86/kvm/mmu/mmu.c
6043
vcpu->arch.nested_mmu.root_role.invalid = 1;
arch/x86/kvm/mmu/mmu.c
6044
vcpu->arch.root_mmu.cpu_role.ext.valid = 0;
arch/x86/kvm/mmu/mmu.c
6045
vcpu->arch.guest_mmu.cpu_role.ext.valid = 0;
arch/x86/kvm/mmu/mmu.c
6046
vcpu->arch.nested_mmu.cpu_role.ext.valid = 0;
arch/x86/kvm/mmu/mmu.c
6063
r = mmu_topup_memory_caches(vcpu, !vcpu->arch.mmu->root_role.direct);
arch/x86/kvm/mmu/mmu.c
6069
if (vcpu->arch.mmu->root_role.direct)
arch/x86/kvm/mmu/mmu.c
609
r = kvm_mmu_topup_memory_cache(&vcpu->arch.mmu_shadow_page_cache,
arch/x86/kvm/mmu/mmu.c
6096
kvm_mmu_free_roots(kvm, &vcpu->arch.root_mmu, KVM_MMU_ROOTS_ALL);
arch/x86/kvm/mmu/mmu.c
6097
WARN_ON_ONCE(VALID_PAGE(vcpu->arch.root_mmu.root.hpa));
arch/x86/kvm/mmu/mmu.c
6098
kvm_mmu_free_roots(kvm, &vcpu->arch.guest_mmu, KVM_MMU_ROOTS_ALL);
arch/x86/kvm/mmu/mmu.c
6099
WARN_ON_ONCE(VALID_PAGE(vcpu->arch.guest_mmu.root.hpa));
arch/x86/kvm/mmu/mmu.c
614
r = kvm_mmu_topup_memory_cache(&vcpu->arch.mmu_shadowed_info_cache,
arch/x86/kvm/mmu/mmu.c
6149
__kvm_mmu_free_obsolete_roots(vcpu->kvm, &vcpu->arch.root_mmu);
arch/x86/kvm/mmu/mmu.c
6150
__kvm_mmu_free_obsolete_roots(vcpu->kvm, &vcpu->arch.guest_mmu);
arch/x86/kvm/mmu/mmu.c
619
return kvm_mmu_topup_memory_cache(&vcpu->arch.mmu_page_header_cache,
arch/x86/kvm/mmu/mmu.c
625
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_pte_list_desc_cache);
arch/x86/kvm/mmu/mmu.c
626
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_shadow_page_cache);
arch/x86/kvm/mmu/mmu.c
627
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_shadowed_info_cache);
arch/x86/kvm/mmu/mmu.c
6272
if (!vcpu->kvm->arch.indirect_shadow_pages)
arch/x86/kvm/mmu/mmu.c
628
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_external_spt_cache);
arch/x86/kvm/mmu/mmu.c
629
kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_header_cache);
arch/x86/kvm/mmu/mmu.c
6317
bool direct = vcpu->arch.mmu->root_role.direct;
arch/x86/kvm/mmu/mmu.c
6327
if (vcpu->arch.last_retry_eip == kvm_rip_read(vcpu) &&
arch/x86/kvm/mmu/mmu.c
6328
vcpu->arch.last_retry_addr == cr2_or_gpa)
arch/x86/kvm/mmu/mmu.c
6337
vcpu->arch.last_retry_eip = 0;
arch/x86/kvm/mmu/mmu.c
6338
vcpu->arch.last_retry_addr = 0;
arch/x86/kvm/mmu/mmu.c
6411
bool direct = vcpu->arch.mmu->root_role.direct;
arch/x86/kvm/mmu/mmu.c
6413
if (WARN_ON_ONCE(!VALID_PAGE(vcpu->arch.mmu->root.hpa)))
arch/x86/kvm/mmu/mmu.c
6426
vcpu->kvm->arch.vm_type == KVM_X86_SW_PROTECTED_VM &&
arch/x86/kvm/mmu/mmu.c
6508
if (WARN_ON_ONCE(mmu != vcpu->arch.mmu))
arch/x86/kvm/mmu/mmu.c
6541
if (mmu != &vcpu->arch.guest_mmu) {
arch/x86/kvm/mmu/mmu.c
6574
kvm_mmu_invalidate_addr(vcpu, vcpu->arch.walk_mmu, gva, KVM_MMU_ROOTS_ALL);
arch/x86/kvm/mmu/mmu.c
6582
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/mmu/mmu.c
6653
if (!tdp_enabled && mmu == &vcpu->arch.guest_mmu)
arch/x86/kvm/mmu/mmu.c
6699
vcpu->arch.mmu_pte_list_desc_cache.kmem_cache = pte_list_desc_cache;
arch/x86/kvm/mmu/mmu.c
6700
vcpu->arch.mmu_pte_list_desc_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6702
vcpu->arch.mmu_page_header_cache.kmem_cache = mmu_page_header_cache;
arch/x86/kvm/mmu/mmu.c
6703
vcpu->arch.mmu_page_header_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6705
vcpu->arch.mmu_shadow_page_cache.init_value =
arch/x86/kvm/mmu/mmu.c
6707
if (!vcpu->arch.mmu_shadow_page_cache.init_value)
arch/x86/kvm/mmu/mmu.c
6708
vcpu->arch.mmu_shadow_page_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6710
vcpu->arch.mmu = &vcpu->arch.root_mmu;
arch/x86/kvm/mmu/mmu.c
6711
vcpu->arch.walk_mmu = &vcpu->arch.root_mmu;
arch/x86/kvm/mmu/mmu.c
6713
ret = __kvm_mmu_create(vcpu, &vcpu->arch.guest_mmu);
arch/x86/kvm/mmu/mmu.c
6717
ret = __kvm_mmu_create(vcpu, &vcpu->arch.root_mmu);
arch/x86/kvm/mmu/mmu.c
6723
free_mmu_pages(&vcpu->arch.guest_mmu);
arch/x86/kvm/mmu/mmu.c
6739
&kvm->arch.active_mmu_pages, link) {
arch/x86/kvm/mmu/mmu.c
6810
kvm->arch.mmu_valid_gen = kvm->arch.mmu_valid_gen ? 0 : 1;
arch/x86/kvm/mmu/mmu.c
6856
kvm->arch.shadow_mmio_value = shadow_mmio_value;
arch/x86/kvm/mmu/mmu.c
6857
INIT_LIST_HEAD(&kvm->arch.active_mmu_pages);
arch/x86/kvm/mmu/mmu.c
6859
INIT_LIST_HEAD(&kvm->arch.possible_nx_huge_pages[i].pages);
arch/x86/kvm/mmu/mmu.c
6860
spin_lock_init(&kvm->arch.mmu_unsync_pages_lock);
arch/x86/kvm/mmu/mmu.c
6870
kvm->arch.split_page_header_cache.kmem_cache = mmu_page_header_cache;
arch/x86/kvm/mmu/mmu.c
6871
kvm->arch.split_page_header_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6873
kvm->arch.split_shadow_page_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6875
kvm->arch.split_desc_cache.kmem_cache = pte_list_desc_cache;
arch/x86/kvm/mmu/mmu.c
6876
kvm->arch.split_desc_cache.gfp_zero = __GFP_ZERO;
arch/x86/kvm/mmu/mmu.c
6882
kvm_mmu_free_memory_cache(&kvm->arch.split_desc_cache);
arch/x86/kvm/mmu/mmu.c
6883
kvm_mmu_free_memory_cache(&kvm->arch.split_page_header_cache);
arch/x86/kvm/mmu/mmu.c
6884
kvm_mmu_free_memory_cache(&kvm->arch.split_shadow_page_cache);
arch/x86/kvm/mmu/mmu.c
6889
kvfree(kvm->arch.mmu_page_hash);
arch/x86/kvm/mmu/mmu.c
6998
return need_topup(&kvm->arch.split_desc_cache, SPLIT_DESC_CACHE_MIN_NR_OBJECTS) ||
arch/x86/kvm/mmu/mmu.c
6999
need_topup(&kvm->arch.split_page_header_cache, 1) ||
arch/x86/kvm/mmu/mmu.c
7000
need_topup(&kvm->arch.split_shadow_page_cache, 1);
arch/x86/kvm/mmu/mmu.c
7024
r = __kvm_mmu_topup_memory_cache(&kvm->arch.split_desc_cache, capacity,
arch/x86/kvm/mmu/mmu.c
7029
r = kvm_mmu_topup_memory_cache(&kvm->arch.split_page_header_cache, 1);
arch/x86/kvm/mmu/mmu.c
7033
return kvm_mmu_topup_memory_cache(&kvm->arch.split_shadow_page_cache, 1);
arch/x86/kvm/mmu/mmu.c
7056
caches.page_header_cache = &kvm->arch.split_page_header_cache;
arch/x86/kvm/mmu/mmu.c
7057
caches.shadow_page_cache = &kvm->arch.split_shadow_page_cache;
arch/x86/kvm/mmu/mmu.c
7068
struct kvm_mmu_memory_cache *cache = &kvm->arch.split_desc_cache;
arch/x86/kvm/mmu/mmu.c
713
return &slot->arch.lpage_info[level - 2][idx];
arch/x86/kvm/mmu/mmu.c
7364
list_for_each_entry_safe(sp, node, &kvm->arch.active_mmu_pages, link) {
arch/x86/kvm/mmu/mmu.c
7393
if (list_empty(&kvm->arch.active_mmu_pages))
arch/x86/kvm/mmu/mmu.c
7440
return kvm->arch.vm_type == KVM_X86_DEFAULT_VM &&
arch/x86/kvm/mmu/mmu.c
7494
struct vhost_task *nx_thread = READ_ONCE(kvm->arch.nx_huge_page_recovery_thread);
arch/x86/kvm/mmu/mmu.c
755
kvm->arch.indirect_shadow_pages++;
arch/x86/kvm/mmu/mmu.c
7637
mmu_free_root_page(vcpu->kvm, &vcpu->arch.mmu->mirror_root_hpa,
arch/x86/kvm/mmu/mmu.c
7641
free_mmu_pages(&vcpu->arch.root_mmu);
arch/x86/kvm/mmu/mmu.c
7642
free_mmu_pages(&vcpu->arch.guest_mmu);
arch/x86/kvm/mmu/mmu.c
7711
unsigned long pages = READ_ONCE(kvm->arch.possible_nx_huge_pages[mmu_type].nr_pages);
arch/x86/kvm/mmu/mmu.c
7747
spinlock_t *tdp_mmu_pages_lock = &kvm->arch.tdp_mmu_pages_lock;
arch/x86/kvm/mmu/mmu.c
7759
nx_huge_pages = &kvm->arch.possible_nx_huge_pages[mmu_type].pages;
arch/x86/kvm/mmu/mmu.c
7859
remaining_time = kvm->arch.nx_huge_page_last + msecs_to_jiffies(period)
arch/x86/kvm/mmu/mmu.c
7870
kvm->arch.nx_huge_page_last = get_jiffies_64();
arch/x86/kvm/mmu/mmu.c
7877
struct kvm *kvm = container_of(ka, struct kvm, arch);
arch/x86/kvm/mmu/mmu.c
7880
kvm->arch.nx_huge_page_last = get_jiffies_64();
arch/x86/kvm/mmu/mmu.c
7891
WRITE_ONCE(kvm->arch.nx_huge_page_recovery_thread, nx_thread);
arch/x86/kvm/mmu/mmu.c
7900
return call_once(&kvm->arch.nx_once, kvm_mmu_start_lpage_recovery);
arch/x86/kvm/mmu/mmu.c
7905
if (kvm->arch.nx_huge_page_recovery_thread)
arch/x86/kvm/mmu/mmu.c
7906
vhost_task_stop(kvm->arch.nx_huge_page_recovery_thread);
arch/x86/kvm/mmu/mmu.c
794
++kvm->arch.possible_nx_huge_pages[mmu_type].nr_pages;
arch/x86/kvm/mmu/mmu.c
796
&kvm->arch.possible_nx_huge_pages[mmu_type].pages);
arch/x86/kvm/mmu/mmu.c
814
kvm->arch.indirect_shadow_pages--;
arch/x86/kvm/mmu/mmu.c
831
--kvm->arch.possible_nx_huge_pages[mmu_type].nr_pages;
arch/x86/kvm/mmu/mmu_internal.h
168
sp->external_spt = kvm_mmu_memory_cache_alloc(&vcpu->arch.mmu_external_spt_cache);
arch/x86/kvm/mmu/mmu_internal.h
194
return kvm->arch.cpu_dirty_log_size && sp->role.guest_mode;
arch/x86/kvm/mmu/mmu_internal.h
223
return READ_ONCE(nx_huge_pages) && !kvm->arch.disable_nx_huge_pages;
arch/x86/kvm/mmu/mmu_internal.h
353
.is_tdp = likely(vcpu->arch.mmu->page_fault == kvm_tdp_page_fault),
arch/x86/kvm/mmu/mmu_internal.h
366
if (vcpu->arch.mmu->root_role.direct) {
arch/x86/kvm/mmu/mmu_internal.h
383
r = vcpu->arch.mmu->page_fault(vcpu, &fault);
arch/x86/kvm/mmu/mmutrace.h
304
__entry->mmu_valid_gen = kvm->arch.mmu_valid_gen;
arch/x86/kvm/mmu/mmutrace.h
305
__entry->mmu_used_pages = kvm->arch.n_used_mmu_pages;
arch/x86/kvm/mmu/page_track.c
148
return !!READ_ONCE(slot->arch.gfn_write_track[index]);
arch/x86/kvm/mmu/page_track.c
156
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
164
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
175
if (kvm->arch.vm_type == KVM_X86_TDX_VM)
arch/x86/kvm/mmu/page_track.c
210
smp_store_release(&kvm->arch.external_write_tracking_enabled, true);
arch/x86/kvm/mmu/page_track.c
237
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
255
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
279
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
30
return smp_load_acquire(&kvm->arch.external_write_tracking_enabled);
arch/x86/kvm/mmu/page_track.c
302
head = &kvm->arch.track_notifier_head;
arch/x86/kvm/mmu/page_track.c
44
vfree(slot->arch.gfn_write_track);
arch/x86/kvm/mmu/page_track.c
45
slot->arch.gfn_write_track = NULL;
arch/x86/kvm/mmu/page_track.c
51
const size_t size = sizeof(*slot->arch.gfn_write_track);
arch/x86/kvm/mmu/page_track.c
53
if (!slot->arch.gfn_write_track)
arch/x86/kvm/mmu/page_track.c
54
slot->arch.gfn_write_track = __vcalloc(npages, size,
arch/x86/kvm/mmu/page_track.c
57
return slot->arch.gfn_write_track ? 0 : -ENOMEM;
arch/x86/kvm/mmu/page_track.c
82
val = slot->arch.gfn_write_track[index];
arch/x86/kvm/mmu/page_track.c
87
slot->arch.gfn_write_track[index] += count;
arch/x86/kvm/mmu/page_track.h
35
return !hlist_empty(&kvm->arch.track_notifier_head.track_notifier_list);
arch/x86/kvm/mmu/paging_tmpl.h
159
if (PT_HAVE_ACCESSED_DIRTY(vcpu->arch.mmu) &&
arch/x86/kvm/mmu/paging_tmpl.h
163
if (FNAME(is_rsvd_bits_set)(vcpu->arch.mmu, gpte, PG_LEVEL_4K))
arch/x86/kvm/mmu/paging_tmpl.h
517
walker->fault.nested_page_fault = mmu != vcpu->arch.walk_mmu;
arch/x86/kvm/mmu/paging_tmpl.h
527
return FNAME(walk_addr_generic)(walker, vcpu, vcpu->arch.mmu, addr,
arch/x86/kvm/mmu/paging_tmpl.h
543
FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte);
arch/x86/kvm/mmu/paging_tmpl.h
626
top_level = vcpu->arch.mmu->cpu_role.base.level;
arch/x86/kvm/mmu/paging_tmpl.h
638
if (WARN_ON_ONCE(!VALID_PAGE(vcpu->arch.mmu->root.hpa)))
arch/x86/kvm/mmu/paging_tmpl.h
647
if (unlikely(kvm_mmu_is_dummy_root(vcpu->arch.mmu->root.hpa))) {
arch/x86/kvm/mmu/paging_tmpl.h
815
!is_cr0_wp(vcpu->arch.mmu) && !fault->user && fault->slot) {
arch/x86/kvm/mmu/paging_tmpl.h
825
if (is_cr4_smep(vcpu->arch.mmu))
arch/x86/kvm/mmu/paging_tmpl.h
870
WARN_ON_ONCE((addr >> 32) && mmu == vcpu->arch.walk_mmu);
arch/x86/kvm/mmu/paging_tmpl.h
923
FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte);
arch/x86/kvm/mmu/spte.c
143
struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/spte.c
148
WRITE_ONCE(vcpu->kvm->arch.has_mapped_host_mmio, true);
arch/x86/kvm/mmu/spte.c
280
WARN_ONCE(is_rsvd_spte(&vcpu->arch.mmu->shadow_zero_check, spte, level),
arch/x86/kvm/mmu/spte.c
282
get_rsvd_bits(&vcpu->arch.mmu->shadow_zero_check, spte, level));
arch/x86/kvm/mmu/spte.c
477
kvm->arch.shadow_mmio_value = mmio_value;
arch/x86/kvm/mmu/spte.c
99
spte |= vcpu->kvm->arch.shadow_mmio_value | access;
arch/x86/kvm/mmu/spte.h
295
struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/spte.h
300
return READ_ONCE(vcpu->kvm->arch.has_mapped_host_mmio);
arch/x86/kvm/mmu/spte.h
305
return (spte & shadow_mmio_mask) == kvm->arch.shadow_mmio_value &&
arch/x86/kvm/mmu/tdp_mmu.c
1147
list_for_each_entry(root, &kvm->arch.tdp_mmu_roots, link) {
arch/x86/kvm/mmu/tdp_mmu.c
1330
spin_lock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
1333
spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
136
next_root = list_next_or_null_rcu(&kvm->arch.tdp_mmu_roots,
arch/x86/kvm/mmu/tdp_mmu.c
140
next_root = list_first_or_null_rcu(&kvm->arch.tdp_mmu_roots,
arch/x86/kvm/mmu/tdp_mmu.c
148
next_root = list_next_or_null_rcu(&kvm->arch.tdp_mmu_roots,
arch/x86/kvm/mmu/tdp_mmu.c
17
INIT_LIST_HEAD(&kvm->arch.tdp_mmu_roots);
arch/x86/kvm/mmu/tdp_mmu.c
18
spin_lock_init(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
192
list_for_each_entry(_root, &_kvm->arch.tdp_mmu_roots, link) \
arch/x86/kvm/mmu/tdp_mmu.c
1942
struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/tdp_mmu.c
1947
*root_level = vcpu->arch.mmu->root_role.level;
arch/x86/kvm/mmu/tdp_mmu.c
206
list_for_each_entry_rcu(_root, &_kvm->arch.tdp_mmu_roots, link) \
arch/x86/kvm/mmu/tdp_mmu.c
218
sp = kvm_mmu_memory_cache_alloc(&vcpu->arch.mmu_page_header_cache);
arch/x86/kvm/mmu/tdp_mmu.c
219
sp->spt = kvm_mmu_memory_cache_alloc(&vcpu->arch.mmu_shadow_page_cache);
arch/x86/kvm/mmu/tdp_mmu.c
255
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/mmu/tdp_mmu.c
277
spin_lock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
287
list_for_each_entry(root, &kvm->arch.tdp_mmu_roots, link) {
arch/x86/kvm/mmu/tdp_mmu.c
304
list_add_rcu(&root->link, &kvm->arch.tdp_mmu_roots);
arch/x86/kvm/mmu/tdp_mmu.c
307
spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
331
atomic64_inc(&kvm->arch.tdp_mmu_pages);
arch/x86/kvm/mmu/tdp_mmu.c
339
atomic64_dec(&kvm->arch.tdp_mmu_pages);
arch/x86/kvm/mmu/tdp_mmu.c
356
spin_lock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
359
spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
44
KVM_MMU_WARN_ON(atomic64_read(&kvm->arch.tdp_mmu_pages));
arch/x86/kvm/mmu/tdp_mmu.c
46
WARN_ON(!list_empty(&kvm->arch.tdp_mmu_roots));
arch/x86/kvm/mmu/tdp_mmu.c
91
spin_lock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.c
93
spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
arch/x86/kvm/mmu/tdp_mmu.h
52
return root_to_sp(vcpu->arch.mmu->mirror_root_hpa);
arch/x86/kvm/mmu/tdp_mmu.h
54
return root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/tdp_mmu.h
61
return root_to_sp(vcpu->arch.mmu->mirror_root_hpa);
arch/x86/kvm/mmu/tdp_mmu.h
63
return root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mtrr.c
31
return &vcpu->arch.mtrr_state.var[index];
arch/x86/kvm/mtrr.c
33
return &vcpu->arch.mtrr_state.fixed_64k;
arch/x86/kvm/mtrr.c
37
return &vcpu->arch.mtrr_state.fixed_16k[index];
arch/x86/kvm/mtrr.c
47
return &vcpu->arch.mtrr_state.fixed_4k[index];
arch/x86/kvm/mtrr.c
49
return &vcpu->arch.mtrr_state.deftype;
arch/x86/kvm/pmu.c
1267
filter = rcu_replace_pointer(kvm->arch.pmu_event_filter, filter,
arch/x86/kvm/pmu.c
1352
perf_load_guest_lvtpc(kvm_lapic_get_reg(vcpu->arch.apic, APIC_LVTPC));
arch/x86/kvm/pmu.c
513
filter = srcu_dereference(kvm->arch.pmu_event_filter, &kvm->srcu);
arch/x86/kvm/pmu.c
714
vcpu->kvm->arch.kvmclock_offset;
arch/x86/kvm/pmu.c
793
kvm_apic_local_deliver(vcpu->arch.apic, APIC_LVTPC);
arch/x86/kvm/pmu.c
974
if (!vcpu->kvm->arch.enable_pmu)
arch/x86/kvm/pmu.h
10
#define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu))
arch/x86/kvm/pmu.h
11
#define pmc_to_pmu(pmc) (&(pmc)->vcpu->arch.pmu)
arch/x86/kvm/pmu.h
9
#define vcpu_to_pmu(vcpu) (&(vcpu)->arch.pmu)
arch/x86/kvm/smm.c
114
trace_kvm_smm_transition(vcpu->vcpu_id, vcpu->arch.smbase, entering_smm);
arch/x86/kvm/smm.c
117
vcpu->arch.hflags |= HF_SMM_MASK;
arch/x86/kvm/smm.c
119
vcpu->arch.hflags &= ~(HF_SMM_MASK | HF_SMM_INSIDE_NMI_MASK);
arch/x86/kvm/smm.c
129
vcpu->arch.pdptrs_from_userspace = false;
arch/x86/kvm/smm.c
138
vcpu->arch.smi_pending = true;
arch/x86/kvm/smm.c
198
smram->dr6 = (u32)vcpu->arch.dr6;
arch/x86/kvm/smm.c
199
smram->dr7 = (u32)vcpu->arch.dr7;
arch/x86/kvm/smm.c
222
smram->smbase = vcpu->arch.smbase;
arch/x86/kvm/smm.c
240
smram->dr6 = vcpu->arch.dr6;
arch/x86/kvm/smm.c
241
smram->dr7 = vcpu->arch.dr7;
arch/x86/kvm/smm.c
247
smram->smbase = vcpu->arch.smbase;
arch/x86/kvm/smm.c
250
smram->efer = vcpu->arch.efer;
arch/x86/kvm/smm.c
310
if (kvm_vcpu_write_guest(vcpu, vcpu->arch.smbase + 0xfe00, &smram, sizeof(smram)))
arch/x86/kvm/smm.c
314
vcpu->arch.hflags |= HF_SMM_INSIDE_NMI_MASK;
arch/x86/kvm/smm.c
323
cr0 = vcpu->arch.cr0 & ~(X86_CR0_PE | X86_CR0_EM | X86_CR0_TS | X86_CR0_PG);
arch/x86/kvm/smm.c
335
cs.selector = (vcpu->arch.smbase >> 4) & 0xffff;
arch/x86/kvm/smm.c
336
cs.base = vcpu->arch.smbase;
arch/x86/kvm/smm.c
366
vcpu->arch.cpuid_dynamic_bits_dirty = true;
arch/x86/kvm/smm.c
501
vcpu->arch.smbase = smstate->smbase;
arch/x86/kvm/smm.c
534
vcpu->arch.smbase = smstate->smbase;
arch/x86/kvm/smm.c
581
smbase = vcpu->arch.smbase;
arch/x86/kvm/smm.c
587
if ((vcpu->arch.hflags & HF_SMM_INSIDE_NMI_MASK) == 0)
arch/x86/kvm/smm.h
154
return vcpu->arch.hflags & HF_SMM_MASK;
arch/x86/kvm/svm/avic.c
180
if (x2avic_enabled && (!vcpu || apic_x2apic_mode(vcpu->arch.apic)))
arch/x86/kvm/svm/avic.c
189
return min(kvm->arch.max_vcpu_ids - 1, arch_max);
arch/x86/kvm/svm/avic.c
216
if (x2avic_enabled && apic_x2apic_mode(svm->vcpu.arch.apic)) {
arch/x86/kvm/svm/avic.c
373
return __sme_set(__pa(svm->vcpu.arch.apic->regs));
arch/x86/kvm/svm/avic.c
408
vcpu->arch.apic->apicv_active = false;
arch/x86/kvm/svm/avic.c
415
if (WARN_ON_ONCE(!vcpu->arch.apic->regs))
arch/x86/kvm/svm/avic.c
472
vcpu->arch.apic->irr_pending = true;
arch/x86/kvm/svm/avic.c
627
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/svm/avic.c
710
flat = kvm_lapic_get_reg(vcpu->arch.apic, APIC_DFR) == APIC_DFR_FLAT;
arch/x86/kvm/svm/avic.c
729
if (apic_x2apic_mode(vcpu->arch.apic))
arch/x86/kvm/svm/avic.c
740
u32 ldr = kvm_lapic_get_reg(vcpu->arch.apic, APIC_LDR);
arch/x86/kvm/svm/avic.c
741
u32 id = kvm_xapic_id(vcpu->arch.apic);
arch/x86/kvm/svm/avic.c
744
if (apic_x2apic_mode(vcpu->arch.apic))
arch/x86/kvm/svm/avic.c
759
u32 dfr = kvm_lapic_get_reg(vcpu->arch.apic, APIC_DFR);
arch/x86/kvm/svm/nested.c
100
vcpu->arch.mmu->get_pdptr = nested_svm_get_tdp_pdptr;
arch/x86/kvm/svm/nested.c
101
vcpu->arch.mmu->inject_page_fault = nested_svm_inject_npf_exit;
arch/x86/kvm/svm/nested.c
102
vcpu->arch.walk_mmu = &vcpu->arch.nested_mmu;
arch/x86/kvm/svm/nested.c
107
vcpu->arch.mmu = &vcpu->arch.root_mmu;
arch/x86/kvm/svm/nested.c
108
vcpu->arch.walk_mmu = &vcpu->arch.root_mmu;
arch/x86/kvm/svm/nested.c
1138
vmcb01->save.efer = vcpu->arch.efer;
arch/x86/kvm/svm/nested.c
1140
vmcb01->save.cr4 = vcpu->arch.cr4;
arch/x86/kvm/svm/nested.c
1147
vcpu->arch.nested_run_pending = KVM_NESTED_RUN_PENDING;
arch/x86/kvm/svm/nested.c
1151
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/svm/nested.c
1233
vmcb12->save.efer = svm->vcpu.arch.efer;
arch/x86/kvm/svm/nested.c
1236
vmcb12->save.cr2 = vcpu->arch.cr2;
arch/x86/kvm/svm/nested.c
1237
vmcb12->save.cr4 = svm->vcpu.arch.cr4;
arch/x86/kvm/svm/nested.c
1243
vmcb12->save.dr6 = svm->vcpu.arch.dr6;
arch/x86/kvm/svm/nested.c
1358
if (vcpu->arch.nmi_pending) {
arch/x86/kvm/svm/nested.c
1359
vcpu->arch.nmi_pending--;
arch/x86/kvm/svm/nested.c
1373
svm->vcpu.arch.tsc_offset = svm->vcpu.arch.l1_tsc_offset;
arch/x86/kvm/svm/nested.c
1374
if (vmcb01->control.tsc_offset != svm->vcpu.arch.tsc_offset) {
arch/x86/kvm/svm/nested.c
1375
vmcb01->control.tsc_offset = svm->vcpu.arch.tsc_offset;
arch/x86/kvm/svm/nested.c
1380
vcpu->arch.tsc_scaling_ratio != vcpu->arch.l1_tsc_scaling_ratio) {
arch/x86/kvm/svm/nested.c
1381
vcpu->arch.tsc_scaling_ratio = vcpu->arch.l1_tsc_scaling_ratio;
arch/x86/kvm/svm/nested.c
1398
svm->vcpu.arch.dr7 = DR7_FIXED_1;
arch/x86/kvm/svm/nested.c
1416
svm->vcpu.arch.nmi_injected = false;
arch/x86/kvm/svm/nested.c
1504
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/svm/nested.c
1533
msr = svm->vcpu.arch.regs[VCPU_REGS_RCX];
arch/x86/kvm/svm/nested.c
1619
if (!(vcpu->arch.efer & EFER_SVME) || !is_paging(vcpu)) {
arch/x86/kvm/svm/nested.c
1642
struct kvm_queued_exception *ex = &vcpu->arch.exception_vmexit;
arch/x86/kvm/svm/nested.c
1659
vmcb->control.exit_info_2 = vcpu->arch.cr2;
arch/x86/kvm/svm/nested.c
1664
if (vcpu->arch.dr7 & DR7_GD) {
arch/x86/kvm/svm/nested.c
1665
vcpu->arch.dr7 &= ~DR7_GD;
arch/x86/kvm/svm/nested.c
1682
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/svm/nested.c
1689
bool block_nested_exceptions = vcpu->arch.nested_run_pending;
arch/x86/kvm/svm/nested.c
1709
if (vcpu->arch.exception_vmexit.pending) {
arch/x86/kvm/svm/nested.c
1716
if (vcpu->arch.exception.pending) {
arch/x86/kvm/svm/nested.c
1723
if (vcpu->arch.smi_pending && !svm_smi_blocked(vcpu)) {
arch/x86/kvm/svm/nested.c
1733
if (vcpu->arch.nmi_pending && !svm_nmi_blocked(vcpu)) {
arch/x86/kvm/svm/nested.c
1772
svm->vcpu.arch.apf.host_apf_flags)
arch/x86/kvm/svm/nested.c
1793
vcpu->arch.tsc_scaling_ratio =
arch/x86/kvm/svm/nested.c
1794
kvm_calc_nested_tsc_multiplier(vcpu->arch.l1_tsc_scaling_ratio,
arch/x86/kvm/svm/nested.c
1864
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/nested.c
1933
if (!(vcpu->arch.efer & EFER_SVME) && kvm_state->flags &&
arch/x86/kvm/svm/nested.c
2002
vcpu->arch.nested_run_pending = KVM_NESTED_RUN_PENDING_UNTRUSTED;
arch/x86/kvm/svm/nested.c
2004
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/svm/nested.c
2026
ret = nested_svm_load_cr3(&svm->vcpu, vcpu->arch.cr3,
arch/x86/kvm/svm/nested.c
2050
if (!vcpu->arch.pdptrs_from_userspace &&
arch/x86/kvm/svm/nested.c
2057
if (CC(!load_pdptrs(vcpu, vcpu->arch.cr3)))
arch/x86/kvm/svm/nested.c
621
if (vcpu->arch.exception.injected) {
arch/x86/kvm/svm/nested.c
622
nr = vcpu->arch.exception.vector;
arch/x86/kvm/svm/nested.c
625
if (vcpu->arch.exception.has_error_code) {
arch/x86/kvm/svm/nested.c
628
vcpu->arch.exception.error_code;
arch/x86/kvm/svm/nested.c
631
} else if (vcpu->arch.nmi_injected) {
arch/x86/kvm/svm/nested.c
634
} else if (vcpu->arch.interrupt.injected) {
arch/x86/kvm/svm/nested.c
635
nr = vcpu->arch.interrupt.nr;
arch/x86/kvm/svm/nested.c
638
if (vcpu->arch.interrupt.soft)
arch/x86/kvm/svm/nested.c
683
vcpu->arch.cr3 = cr3;
arch/x86/kvm/svm/nested.c
758
svm->vcpu.arch.cr2 = save->cr2;
arch/x86/kvm/svm/nested.c
771
svm->vcpu.arch.dr6 = svm->nested.save.dr6 | DR6_ACTIVE_LOW;
arch/x86/kvm/svm/nested.c
842
svm->vcpu.arch.nmi_pending++;
arch/x86/kvm/svm/nested.c
89
vcpu->arch.mmu = &vcpu->arch.guest_mmu;
arch/x86/kvm/svm/nested.c
901
vcpu->arch.tsc_offset = kvm_calc_nested_tsc_offset(vcpu->arch.l1_tsc_offset,
arch/x86/kvm/svm/nested.c
905
vmcb02->control.tsc_offset = vcpu->arch.tsc_offset;
arch/x86/kvm/svm/nested.c
928
!vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/nested.c
940
!vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/nested.c
99
vcpu->arch.mmu->get_guest_pgd = nested_svm_get_tdp_cr3;
arch/x86/kvm/svm/sev.c
1062
fpstate_set_confidential(&vcpu->arch.guest_fpu);
arch/x86/kvm/svm/sev.c
1063
vcpu->arch.guest_state_protected = true;
arch/x86/kvm/svm/sev.c
2092
dst_vcpu->arch.guest_state_protected = true;
arch/x86/kvm/svm/sev.c
2097
src_vcpu->arch.guest_state_protected = false;
arch/x86/kvm/svm/sev.c
2117
if (!src_vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/sev.c
2144
if (kvm->arch.vm_type != source_kvm->arch.vm_type ||
arch/x86/kvm/svm/sev.c
2292
if (WARN_ON_ONCE(!kvm->arch.default_tsc_khz))
arch/x86/kvm/svm/sev.c
2295
start.desired_tsc_khz = kvm->arch.default_tsc_khz;
arch/x86/kvm/svm/sev.c
2532
svm->vcpu.arch.guest_state_protected = true;
arch/x86/kvm/svm/sev.c
2611
kvm->arch.pre_fault_allowed = true;
arch/x86/kvm/svm/sev.c
2937
switch (kvm->arch.vm_type) {
arch/x86/kvm/svm/sev.c
2942
kvm->arch.has_private_mem = true;
arch/x86/kvm/svm/sev.c
2945
kvm->arch.has_protected_state = true;
arch/x86/kvm/svm/sev.c
2948
kvm->arch.pre_fault_allowed = !kvm->arch.has_private_mem;
arch/x86/kvm/svm/sev.c
2952
WARN_ONCE(1, "Unsupported VM type %u", kvm->arch.vm_type);
arch/x86/kvm/svm/sev.c
3337
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/sev.c
3391
ghcb_set_rax(ghcb, vcpu->arch.regs[VCPU_REGS_RAX]);
arch/x86/kvm/svm/sev.c
3392
ghcb_set_rbx(ghcb, vcpu->arch.regs[VCPU_REGS_RBX]);
arch/x86/kvm/svm/sev.c
3393
ghcb_set_rcx(ghcb, vcpu->arch.regs[VCPU_REGS_RCX]);
arch/x86/kvm/svm/sev.c
3394
ghcb_set_rdx(ghcb, vcpu->arch.regs[VCPU_REGS_RDX]);
arch/x86/kvm/svm/sev.c
3415
memset(vcpu->arch.regs, 0, sizeof(vcpu->arch.regs));
arch/x86/kvm/svm/sev.c
3420
vcpu->arch.regs[VCPU_REGS_RAX] = kvm_ghcb_get_rax_if_valid(svm);
arch/x86/kvm/svm/sev.c
3421
vcpu->arch.regs[VCPU_REGS_RBX] = kvm_ghcb_get_rbx_if_valid(svm);
arch/x86/kvm/svm/sev.c
3422
vcpu->arch.regs[VCPU_REGS_RCX] = kvm_ghcb_get_rcx_if_valid(svm);
arch/x86/kvm/svm/sev.c
3423
vcpu->arch.regs[VCPU_REGS_RDX] = kvm_ghcb_get_rdx_if_valid(svm);
arch/x86/kvm/svm/sev.c
3424
vcpu->arch.regs[VCPU_REGS_RSI] = kvm_ghcb_get_rsi_if_valid(svm);
arch/x86/kvm/svm/sev.c
3480
if (vcpu->arch.regs[VCPU_REGS_RAX] == 0xd)
arch/x86/kvm/svm/sev.c
3655
svm->vcpu.arch.last_vmentry_cpu == cpu)
arch/x86/kvm/svm/sev.c
3826
vcpu->arch.complete_userspace_io = snp_complete_psc_msr;
arch/x86/kvm/svm/sev.c
4004
vcpu->arch.complete_userspace_io = snp_complete_one_psc;
arch/x86/kvm/svm/sev.c
4041
vcpu->arch.pv.pv_unhalted = false;
arch/x86/kvm/svm/sev.c
4122
if (vcpu->arch.regs[VCPU_REGS_RAX] != sev->vmsa_features) {
arch/x86/kvm/svm/sev.c
4124
vcpu->arch.regs[VCPU_REGS_RAX], sev->vmsa_features);
arch/x86/kvm/svm/sev.c
4227
vcpu->arch.regs[VCPU_REGS_RBX] = vcpu->run->snp_req_certs.npages;
arch/x86/kvm/svm/sev.c
4272
data_gpa = vcpu->arch.regs[VCPU_REGS_RAX];
arch/x86/kvm/svm/sev.c
4273
data_npages = vcpu->arch.regs[VCPU_REGS_RBX];
arch/x86/kvm/svm/sev.c
4283
vcpu->arch.complete_userspace_io = snp_complete_req_certs;
arch/x86/kvm/svm/sev.c
4329
vcpu->arch.regs[VCPU_REGS_RAX] = cpuid_fn;
arch/x86/kvm/svm/sev.c
4330
vcpu->arch.regs[VCPU_REGS_RCX] = 0;
arch/x86/kvm/svm/sev.c
4342
cpuid_value = vcpu->arch.regs[VCPU_REGS_RAX];
arch/x86/kvm/svm/sev.c
4344
cpuid_value = vcpu->arch.regs[VCPU_REGS_RBX];
arch/x86/kvm/svm/sev.c
4346
cpuid_value = vcpu->arch.regs[VCPU_REGS_RCX];
arch/x86/kvm/svm/sev.c
4348
cpuid_value = vcpu->arch.regs[VCPU_REGS_RDX];
arch/x86/kvm/svm/sev.c
4643
vcpu->arch.reserved_gpa_bits &= ~(1UL << (best->ebx & 0x3f));
arch/x86/kvm/svm/sev.c
4758
vcpu->arch.guest_tsc_protected = snp_is_secure_tsc_enabled(vcpu->kvm);
arch/x86/kvm/svm/sev.c
5175
if (!vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/sev.c
5248
if (!vcpu->arch.guest_state_protected || !vmsa)
arch/x86/kvm/svm/sev.c
586
if (kvm->arch.vm_type != KVM_X86_DEFAULT_VM)
arch/x86/kvm/svm/sev.c
608
if (kvm->arch.vm_type != KVM_X86_SEV_VM &&
arch/x86/kvm/svm/sev.c
609
kvm->arch.vm_type != KVM_X86_SEV_ES_VM &&
arch/x86/kvm/svm/sev.c
610
kvm->arch.vm_type != KVM_X86_SNP_VM)
arch/x86/kvm/svm/sev.c
616
return __sev_guest_init(kvm, argp, &data, kvm->arch.vm_type);
arch/x86/kvm/svm/sev.c
936
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/sev.c
952
save->rax = svm->vcpu.arch.regs[VCPU_REGS_RAX];
arch/x86/kvm/svm/sev.c
953
save->rbx = svm->vcpu.arch.regs[VCPU_REGS_RBX];
arch/x86/kvm/svm/sev.c
954
save->rcx = svm->vcpu.arch.regs[VCPU_REGS_RCX];
arch/x86/kvm/svm/sev.c
955
save->rdx = svm->vcpu.arch.regs[VCPU_REGS_RDX];
arch/x86/kvm/svm/sev.c
956
save->rsp = svm->vcpu.arch.regs[VCPU_REGS_RSP];
arch/x86/kvm/svm/sev.c
957
save->rbp = svm->vcpu.arch.regs[VCPU_REGS_RBP];
arch/x86/kvm/svm/sev.c
958
save->rsi = svm->vcpu.arch.regs[VCPU_REGS_RSI];
arch/x86/kvm/svm/sev.c
959
save->rdi = svm->vcpu.arch.regs[VCPU_REGS_RDI];
arch/x86/kvm/svm/sev.c
961
save->r8 = svm->vcpu.arch.regs[VCPU_REGS_R8];
arch/x86/kvm/svm/sev.c
962
save->r9 = svm->vcpu.arch.regs[VCPU_REGS_R9];
arch/x86/kvm/svm/sev.c
963
save->r10 = svm->vcpu.arch.regs[VCPU_REGS_R10];
arch/x86/kvm/svm/sev.c
964
save->r11 = svm->vcpu.arch.regs[VCPU_REGS_R11];
arch/x86/kvm/svm/sev.c
965
save->r12 = svm->vcpu.arch.regs[VCPU_REGS_R12];
arch/x86/kvm/svm/sev.c
966
save->r13 = svm->vcpu.arch.regs[VCPU_REGS_R13];
arch/x86/kvm/svm/sev.c
967
save->r14 = svm->vcpu.arch.regs[VCPU_REGS_R14];
arch/x86/kvm/svm/sev.c
968
save->r15 = svm->vcpu.arch.regs[VCPU_REGS_R15];
arch/x86/kvm/svm/sev.c
970
save->rip = svm->vcpu.arch.regs[VCPU_REGS_RIP];
arch/x86/kvm/svm/sev.c
973
save->xcr0 = svm->vcpu.arch.xcr0;
arch/x86/kvm/svm/sev.c
974
save->pkru = svm->vcpu.arch.pkru;
arch/x86/kvm/svm/sev.c
975
save->xss = svm->vcpu.arch.ia32_xss;
arch/x86/kvm/svm/sev.c
976
save->dr6 = svm->vcpu.arch.dr6;
arch/x86/kvm/svm/sev.c
984
if (vcpu->kvm->arch.vm_type != KVM_X86_DEFAULT_VM) {
arch/x86/kvm/svm/sev.c
985
xsave = &vcpu->arch.guest_fpu.fpstate->regs.xsave;
arch/x86/kvm/svm/svm.c
1002
svm->vmcb01.ptr->control.tsc_offset = vcpu->arch.l1_tsc_offset;
arch/x86/kvm/svm/svm.c
1003
svm->vmcb->control.tsc_offset = vcpu->arch.tsc_offset;
arch/x86/kvm/svm/svm.c
1011
__svm_write_tsc_multiplier(vcpu->arch.tsc_scaling_ratio);
arch/x86/kvm/svm/svm.c
1017
return svm->vcpu.arch.smi_pending ||
arch/x86/kvm/svm/svm.c
1018
svm->vcpu.arch.nmi_pending ||
arch/x86/kvm/svm/svm.c
1061
if (!(vcpu->arch.efer & EFER_SVME)) {
arch/x86/kvm/svm/svm.c
1209
save->g_pat = vcpu->arch.pat;
arch/x86/kvm/svm/svm.c
1242
if (vcpu->kvm->arch.bus_lock_detection_enabled)
arch/x86/kvm/svm/svm.c
1264
vcpu->arch.microcode_version = 0x01000065;
arch/x86/kvm/svm/svm.c
1434
__svm_write_tsc_multiplier(vcpu->arch.tsc_scaling_ratio);
arch/x86/kvm/svm/svm.c
1762
if (vcpu->arch.efer & EFER_LME) {
arch/x86/kvm/svm/svm.c
1764
vcpu->arch.efer |= EFER_LMA;
arch/x86/kvm/svm/svm.c
1765
if (!vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/svm.c
1770
vcpu->arch.efer &= ~EFER_LMA;
arch/x86/kvm/svm/svm.c
1771
if (!vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/svm.c
1776
vcpu->arch.cr0 = cr0;
arch/x86/kvm/svm/svm.c
1820
unsigned long old_cr4 = vcpu->arch.cr4;
arch/x86/kvm/svm/svm.c
1822
vcpu->arch.cr4 = cr4;
arch/x86/kvm/svm/svm.c
1834
vcpu->arch.cpuid_dynamic_bits_dirty = true;
arch/x86/kvm/svm/svm.c
1897
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/svm.c
1913
get_debugreg(vcpu->arch.db[0], 0);
arch/x86/kvm/svm/svm.c
1914
get_debugreg(vcpu->arch.db[1], 1);
arch/x86/kvm/svm/svm.c
1915
get_debugreg(vcpu->arch.db[2], 2);
arch/x86/kvm/svm/svm.c
1916
get_debugreg(vcpu->arch.db[3], 3);
arch/x86/kvm/svm/svm.c
1921
vcpu->arch.dr6 = svm->vmcb->save.dr6;
arch/x86/kvm/svm/svm.c
1922
vcpu->arch.dr7 = svm->vmcb->save.dr7;
arch/x86/kvm/svm/svm.c
1923
vcpu->arch.switch_db_regs &= ~KVM_DEBUGREG_WONT_EXIT;
arch/x86/kvm/svm/svm.c
1931
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/svm/svm.c
2026
kvm_run->debug.arch.dr6 = svm->vmcb->save.dr6;
arch/x86/kvm/svm/svm.c
2027
kvm_run->debug.arch.dr7 = svm->vmcb->save.dr7;
arch/x86/kvm/svm/svm.c
2028
kvm_run->debug.arch.pc =
arch/x86/kvm/svm/svm.c
2030
kvm_run->debug.arch.exception = DB_VECTOR;
arch/x86/kvm/svm/svm.c
2043
kvm_run->debug.arch.pc = svm->vmcb->save.cs.base + svm->vmcb->save.rip;
arch/x86/kvm/svm/svm.c
2044
kvm_run->debug.arch.exception = BP_VECTOR;
arch/x86/kvm/svm/svm.c
209
u64 old_efer = vcpu->arch.efer;
arch/x86/kvm/svm/svm.c
210
vcpu->arch.efer = efer;
arch/x86/kvm/svm/svm.c
2246
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/svm/svm.c
2438
vcpu->arch.nmi_injected = false;
arch/x86/kvm/svm/svm.c
252
vcpu->arch.efer = old_efer;
arch/x86/kvm/svm/svm.c
2524
unsigned long cr0 = vcpu->arch.cr0;
arch/x86/kvm/svm/svm.c
2596
val = vcpu->arch.cr2;
arch/x86/kvm/svm/svm.c
2673
vcpu->arch.switch_db_regs |= KVM_DEBUGREG_WONT_EXIT;
arch/x86/kvm/svm/svm.c
2765
return is_sev_es_guest(vcpu) && vcpu->arch.guest_state_protected &&
arch/x86/kvm/svm/svm.c
2776
return vcpu->kvm->arch.has_protected_state ? -EINVAL : 0;
arch/x86/kvm/svm/svm.c
2921
if (svm_dis && (vcpu->arch.efer & EFER_SVME))
arch/x86/kvm/svm/svm.c
2936
return vcpu->kvm->arch.has_protected_state ? -EINVAL : 0;
arch/x86/kvm/svm/svm.c
3237
if (kvm_is_linear_rip(vcpu, vcpu->arch.cui_linear_rip))
arch/x86/kvm/svm/svm.c
3250
vcpu->arch.cui_linear_rip = kvm_get_linear_rip(vcpu);
arch/x86/kvm/svm/svm.c
3251
vcpu->arch.complete_userspace_io = complete_userspace_buslock;
arch/x86/kvm/svm/svm.c
3254
svm->nested.last_bus_lock_rip = vcpu->arch.cui_linear_rip;
arch/x86/kvm/svm/svm.c
3373
vm_type, vcpu->vcpu_id, svm->current_vmcb->ptr, vcpu->arch.last_vmentry_cpu);
arch/x86/kvm/svm/svm.c
3532
vcpu->arch.regs[VCPU_REGS_RBX]);
arch/x86/kvm/svm/svm.c
3534
"rcx:", vcpu->arch.regs[VCPU_REGS_RCX],
arch/x86/kvm/svm/svm.c
3535
"rdx:", vcpu->arch.regs[VCPU_REGS_RDX]);
arch/x86/kvm/svm/svm.c
3537
"rsi:", vcpu->arch.regs[VCPU_REGS_RSI],
arch/x86/kvm/svm/svm.c
3538
"rdi:", vcpu->arch.regs[VCPU_REGS_RDI]);
arch/x86/kvm/svm/svm.c
3540
"rbp:", vcpu->arch.regs[VCPU_REGS_RBP],
arch/x86/kvm/svm/svm.c
3544
"r8:", vcpu->arch.regs[VCPU_REGS_R8],
arch/x86/kvm/svm/svm.c
3545
"r9:", vcpu->arch.regs[VCPU_REGS_R9]);
arch/x86/kvm/svm/svm.c
3547
"r10:", vcpu->arch.regs[VCPU_REGS_R10],
arch/x86/kvm/svm/svm.c
3548
"r11:", vcpu->arch.regs[VCPU_REGS_R11]);
arch/x86/kvm/svm/svm.c
3550
"r12:", vcpu->arch.regs[VCPU_REGS_R12],
arch/x86/kvm/svm/svm.c
3551
"r13:", vcpu->arch.regs[VCPU_REGS_R13]);
arch/x86/kvm/svm/svm.c
3553
"r14:", vcpu->arch.regs[VCPU_REGS_R14],
arch/x86/kvm/svm/svm.c
3554
"r15:", vcpu->arch.regs[VCPU_REGS_R15]);
arch/x86/kvm/svm/svm.c
3650
vcpu->arch.cr0 = svm->vmcb->save.cr0;
arch/x86/kvm/svm/svm.c
3652
vcpu->arch.cr3 = svm->vmcb->save.cr3;
arch/x86/kvm/svm/svm.c
3673
kvm_run->fail_entry.cpu = vcpu->arch.last_vmentry_cpu;
arch/x86/kvm/svm/svm.c
3776
struct kvm_queued_interrupt *intr = &vcpu->arch.interrupt;
arch/x86/kvm/svm/svm.c
3816
if (!is_guest_mode(vcpu) || !vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/svm.c
3852
if (!READ_ONCE(vcpu->arch.apic->apicv_active)) {
arch/x86/kvm/svm/svm.c
386
struct kvm_queued_exception *ex = &vcpu->arch.exception;
arch/x86/kvm/svm/svm.c
3964
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/svm.c
4006
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/svm.c
4144
hpa_t root_tdp = vcpu->arch.mmu->root.hpa;
arch/x86/kvm/svm/svm.c
4229
if (is_guest_mode(vcpu) && vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/svm.c
426
vcpu->arch.osvw.length = (osvw_len >= 3) ? (osvw_len) : 3;
arch/x86/kvm/svm/svm.c
427
vcpu->arch.osvw.status = osvw_status & ~(6ULL);
arch/x86/kvm/svm/svm.c
4280
vcpu->arch.nmi_injected = false;
arch/x86/kvm/svm/svm.c
4297
vcpu->arch.nmi_injected = true;
arch/x86/kvm/svm/svm.c
438
vcpu->arch.osvw.status |= 1;
arch/x86/kvm/svm/svm.c
4421
svm->vmcb->save.rax = vcpu->arch.regs[VCPU_REGS_RAX];
arch/x86/kvm/svm/svm.c
4422
svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP];
arch/x86/kvm/svm/svm.c
4423
svm->vmcb->save.rip = vcpu->arch.regs[VCPU_REGS_RIP];
arch/x86/kvm/svm/svm.c
4457
svm->vmcb->save.cr2 = vcpu->arch.cr2;
arch/x86/kvm/svm/svm.c
4473
svm_set_dr6(vcpu, vcpu->arch.dr6);
arch/x86/kvm/svm/svm.c
4474
else if (likely(!(vcpu->arch.switch_db_regs & KVM_DEBUGREG_WONT_EXIT)))
arch/x86/kvm/svm/svm.c
4486
vcpu->arch.host_debugctl != svm->vmcb->save.dbgctl)
arch/x86/kvm/svm/svm.c
4506
vcpu->arch.cr2 = svm->vmcb->save.cr2;
arch/x86/kvm/svm/svm.c
4507
vcpu->arch.regs[VCPU_REGS_RAX] = svm->vmcb->save.rax;
arch/x86/kvm/svm/svm.c
4508
vcpu->arch.regs[VCPU_REGS_RSP] = svm->vmcb->save.rsp;
arch/x86/kvm/svm/svm.c
4509
vcpu->arch.regs[VCPU_REGS_RIP] = svm->vmcb->save.rip;
arch/x86/kvm/svm/svm.c
4511
vcpu->arch.regs_dirty = 0;
arch/x86/kvm/svm/svm.c
4517
vcpu->arch.host_debugctl != svm->vmcb->save.dbgctl)
arch/x86/kvm/svm/svm.c
4518
update_debugctlmsr(vcpu->arch.host_debugctl);
arch/x86/kvm/svm/svm.c
4534
if (vcpu->arch.nested_run_pending &&
arch/x86/kvm/svm/svm.c
4538
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/svm/svm.c
4554
vcpu->arch.apf.host_apf_flags =
arch/x86/kvm/svm/svm.c
4557
vcpu->arch.regs_avail &= ~SVM_REGS_LAZY_LOAD_SET;
arch/x86/kvm/svm/svm.c
4591
cr3 = vcpu->arch.cr3;
arch/x86/kvm/svm/svm.c
4807
cr0 = vcpu->arch.cr0 & ~SVM_CR0_SELECTIVE_MASK;
arch/x86/kvm/svm/svm.c
4887
vcpu->arch.at_instruction_boundary = true;
arch/x86/kvm/svm/svm.c
4897
vcpu->arch.mcg_cap &= 0x1ff;
arch/x86/kvm/svm/svm.c
4915
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/svm/svm.c
4947
svm->vmcb->save.rax = vcpu->arch.regs[VCPU_REGS_RAX];
arch/x86/kvm/svm/svm.c
4948
svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP];
arch/x86/kvm/svm/svm.c
4949
svm->vmcb->save.rip = vcpu->arch.regs[VCPU_REGS_RIP];
arch/x86/kvm/svm/svm.c
5033
vcpu->arch.nested_run_pending = KVM_NESTED_RUN_PENDING;
arch/x86/kvm/trace.h
17
__vcpu->arch.guest_state_protected ? 0 : kvm_rip_read(__vcpu); \
arch/x86/kvm/trace.h
1971
#define TRACE_INCLUDE_PATH ../../arch/x86/kvm
arch/x86/kvm/trace.h
932
__entry->len = vcpu->arch.emulate_ctxt->fetch.ptr
arch/x86/kvm/trace.h
933
- vcpu->arch.emulate_ctxt->fetch.data;
arch/x86/kvm/trace.h
934
__entry->rip = vcpu->arch.emulate_ctxt->_eip - __entry->len;
arch/x86/kvm/trace.h
936
vcpu->arch.emulate_ctxt->fetch.data,
arch/x86/kvm/trace.h
938
__entry->flags = kei_decode_mode(vcpu->arch.emulate_ctxt->mode);
arch/x86/kvm/vmx/common.h
62
return kvm->arch.vm_type == KVM_X86_TDX_VM;
arch/x86/kvm/vmx/hyperv.h
51
return vcpu->arch.hyperv_enabled &&
arch/x86/kvm/vmx/nested.c
1191
vcpu->arch.cr3 = cr3;
arch/x86/kvm/vmx/nested.c
2263
if (vcpu->arch.virtual_tsc_khz == 0)
arch/x86/kvm/vmx/nested.c
2268
do_div(preemption_timeout, vcpu->arch.virtual_tsc_khz);
arch/x86/kvm/vmx/nested.c
2276
if (vmx->vcpu.arch.nested_run_pending &&
arch/x86/kvm/vmx/nested.c
2280
return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/nested.c
2282
return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/nested.c
2326
vmcs_write32(NOTIFY_WINDOW, kvm->arch.notify_window);
arch/x86/kvm/vmx/nested.c
2516
if (vmx->vcpu.arch.nested_run_pending) {
arch/x86/kvm/vmx/nested.c
2624
if (kvm_mpx_supported() && vmx->vcpu.arch.nested_run_pending &&
arch/x86/kvm/vmx/nested.c
2721
if (vcpu->arch.nested_run_pending &&
arch/x86/kvm/vmx/nested.c
2727
kvm_set_dr(vcpu, 7, vcpu->arch.dr7);
arch/x86/kvm/vmx/nested.c
2731
if (!vcpu->arch.nested_run_pending ||
arch/x86/kvm/vmx/nested.c
2737
if (kvm_mpx_supported() && (!vcpu->arch.nested_run_pending ||
arch/x86/kvm/vmx/nested.c
2747
vcpu->arch.cr0_guest_owned_bits &= ~vmcs12->cr0_guest_host_mask;
arch/x86/kvm/vmx/nested.c
2748
vmcs_writel(CR0_GUEST_HOST_MASK, ~vcpu->arch.cr0_guest_owned_bits);
arch/x86/kvm/vmx/nested.c
2750
if (vcpu->arch.nested_run_pending &&
arch/x86/kvm/vmx/nested.c
2753
vcpu->arch.pat = vmcs12->guest_ia32_pat;
arch/x86/kvm/vmx/nested.c
2755
vmcs_write64(GUEST_IA32_PAT, vcpu->arch.pat);
arch/x86/kvm/vmx/nested.c
2758
vcpu->arch.tsc_offset = kvm_calc_nested_tsc_offset(
arch/x86/kvm/vmx/nested.c
2759
vcpu->arch.l1_tsc_offset,
arch/x86/kvm/vmx/nested.c
2763
vcpu->arch.tsc_scaling_ratio = kvm_calc_nested_tsc_multiplier(
arch/x86/kvm/vmx/nested.c
2764
vcpu->arch.l1_tsc_scaling_ratio,
arch/x86/kvm/vmx/nested.c
2767
vmcs_write64(TSC_OFFSET, vcpu->arch.tsc_offset);
arch/x86/kvm/vmx/nested.c
2769
vmcs_write64(TSC_MULTIPLIER, vcpu->arch.tsc_scaling_ratio);
arch/x86/kvm/vmx/nested.c
2787
vcpu->arch.efer = nested_vmx_calc_efer(vmx, vmcs12);
arch/x86/kvm/vmx/nested.c
2789
vmx_set_efer(vcpu, vcpu->arch.efer);
arch/x86/kvm/vmx/nested.c
3125
!!(vcpu->arch.efer & EFER_LMA)))
arch/x86/kvm/vmx/nested.c
313
vcpu->arch.regs_avail = ~VMX_REGS_LAZY_LOAD_SET;
arch/x86/kvm/vmx/nested.c
319
vcpu->arch.regs_dirty = 0;
arch/x86/kvm/vmx/nested.c
3352
if (vcpu->arch.nested_run_pending &&
arch/x86/kvm/vmx/nested.c
3425
if (!vcpu->arch.pdptrs_from_userspace &&
arch/x86/kvm/vmx/nested.c
3432
if (CC(!load_pdptrs(vcpu, vcpu->arch.cr3)))
arch/x86/kvm/vmx/nested.c
3630
if (!vcpu->arch.nested_run_pending ||
arch/x86/kvm/vmx/nested.c
3634
(!vcpu->arch.nested_run_pending ||
arch/x86/kvm/vmx/nested.c
3638
if (!vcpu->arch.nested_run_pending ||
arch/x86/kvm/vmx/nested.c
3657
vmcs_writel(GUEST_CR3, vcpu->arch.cr3);
arch/x86/kvm/vmx/nested.c
367
kvm_mmu_free_roots(vcpu->kvm, &vcpu->arch.guest_mmu, KVM_MMU_ROOTS_ALL);
arch/x86/kvm/vmx/nested.c
3749
vcpu->arch.tsc_offset -= vmcs12->tsc_offset;
arch/x86/kvm/vmx/nested.c
3847
vcpu->arch.nested_run_pending = KVM_NESTED_RUN_PENDING;
arch/x86/kvm/vmx/nested.c
3879
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/vmx/nested.c
3884
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/vmx/nested.c
3894
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/vmx/nested.c
3924
/*1*/ (vmcs_readl(GUEST_CR0) & vcpu->arch.cr0_guest_owned_bits) |
arch/x86/kvm/vmx/nested.c
3927
vcpu->arch.cr0_guest_owned_bits));
arch/x86/kvm/vmx/nested.c
3934
/*1*/ (vmcs_readl(GUEST_CR4) & vcpu->arch.cr4_guest_owned_bits) |
arch/x86/kvm/vmx/nested.c
3937
vcpu->arch.cr4_guest_owned_bits));
arch/x86/kvm/vmx/nested.c
3968
} else if (vcpu->arch.exception.injected) {
arch/x86/kvm/vmx/nested.c
3969
nr = vcpu->arch.exception.vector;
arch/x86/kvm/vmx/nested.c
3974
vcpu->arch.event_exit_inst_len;
arch/x86/kvm/vmx/nested.c
3979
if (vcpu->arch.exception.has_error_code) {
arch/x86/kvm/vmx/nested.c
3982
vcpu->arch.exception.error_code;
arch/x86/kvm/vmx/nested.c
3986
} else if (vcpu->arch.nmi_injected) {
arch/x86/kvm/vmx/nested.c
3989
} else if (vcpu->arch.interrupt.injected) {
arch/x86/kvm/vmx/nested.c
3990
nr = vcpu->arch.interrupt.nr;
arch/x86/kvm/vmx/nested.c
3993
if (vcpu->arch.interrupt.soft) {
arch/x86/kvm/vmx/nested.c
3996
vcpu->arch.event_exit_inst_len;
arch/x86/kvm/vmx/nested.c
403
cached_root = &vcpu->arch.mmu->prev_roots[i];
arch/x86/kvm/vmx/nested.c
4051
struct kvm_queued_exception *ex = &vcpu->arch.exception_vmexit;
arch/x86/kvm/vmx/nested.c
4059
exit_qual = vcpu->arch.cr2;
arch/x86/kvm/vmx/nested.c
4061
exit_qual = vcpu->arch.dr6;
arch/x86/kvm/vmx/nested.c
410
kvm_mmu_invalidate_addr(vcpu, vcpu->arch.mmu, addr, roots);
arch/x86/kvm/vmx/nested.c
4146
pending_dbg = vmx_get_pending_dbg_trap(&vcpu->arch.exception);
arch/x86/kvm/vmx/nested.c
4284
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/vmx/nested.c
4291
bool block_nested_exceptions = vcpu->arch.nested_run_pending;
arch/x86/kvm/vmx/nested.c
4319
if (vcpu->arch.mp_state != KVM_MP_STATE_INIT_RECEIVED)
arch/x86/kvm/vmx/nested.c
4333
if (vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED) {
arch/x86/kvm/vmx/nested.c
4351
if (vcpu->arch.exception_vmexit.pending &&
arch/x86/kvm/vmx/nested.c
4352
!vmx_is_low_priority_db_trap(&vcpu->arch.exception_vmexit)) {
arch/x86/kvm/vmx/nested.c
4360
if (vcpu->arch.exception.pending &&
arch/x86/kvm/vmx/nested.c
4361
!vmx_is_low_priority_db_trap(&vcpu->arch.exception)) {
arch/x86/kvm/vmx/nested.c
4375
if (vcpu->arch.exception_vmexit.pending) {
arch/x86/kvm/vmx/nested.c
4383
if (vcpu->arch.exception.pending) {
arch/x86/kvm/vmx/nested.c
4396
if (vcpu->arch.smi_pending && !is_smm(vcpu)) {
arch/x86/kvm/vmx/nested.c
4402
if (vcpu->arch.nmi_pending && !vmx_nmi_blocked(vcpu)) {
arch/x86/kvm/vmx/nested.c
4415
vcpu->arch.nmi_pending = 0;
arch/x86/kvm/vmx/nested.c
4502
value = ktime_to_ns(remaining) * vcpu->arch.virtual_tsc_khz;
arch/x86/kvm/vmx/nested.c
4651
if (vcpu->arch.mp_state == KVM_MP_STATE_HALTED)
arch/x86/kvm/vmx/nested.c
4653
else if (vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED)
arch/x86/kvm/vmx/nested.c
4660
!vcpu->arch.nested_run_pending)
arch/x86/kvm/vmx/nested.c
4698
vmcs12->guest_dr7 = vcpu->arch.dr7;
arch/x86/kvm/vmx/nested.c
4701
vmcs12->guest_ia32_efer = vcpu->arch.efer;
arch/x86/kvm/vmx/nested.c
4782
vcpu->arch.efer = vmcs12->host_ia32_efer;
arch/x86/kvm/vmx/nested.c
4784
vcpu->arch.efer |= (EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/nested.c
4786
vcpu->arch.efer &= ~(EFER_LMA | EFER_LME);
arch/x86/kvm/vmx/nested.c
4787
vmx_set_efer(vcpu, vcpu->arch.efer);
arch/x86/kvm/vmx/nested.c
4801
vcpu->arch.cr0_guest_owned_bits = vmx_l1_guest_owned_cr0_bits();
arch/x86/kvm/vmx/nested.c
4805
vcpu->arch.cr4_guest_owned_bits = ~vmcs_readl(CR4_GUEST_HOST_MASK);
arch/x86/kvm/vmx/nested.c
483
vcpu->arch.mmu = &vcpu->arch.guest_mmu;
arch/x86/kvm/vmx/nested.c
4845
vcpu->arch.pat = vmcs12->host_ia32_pat;
arch/x86/kvm/vmx/nested.c
485
vcpu->arch.mmu->get_guest_pgd = nested_ept_get_eptp;
arch/x86/kvm/vmx/nested.c
486
vcpu->arch.mmu->inject_page_fault = nested_ept_inject_page_fault;
arch/x86/kvm/vmx/nested.c
487
vcpu->arch.mmu->get_pdptr = kvm_pdptr_read;
arch/x86/kvm/vmx/nested.c
489
vcpu->arch.walk_mmu = &vcpu->arch.nested_mmu;
arch/x86/kvm/vmx/nested.c
494
vcpu->arch.mmu = &vcpu->arch.root_mmu;
arch/x86/kvm/vmx/nested.c
4943
vcpu->arch.pat = vmcs_read64(GUEST_IA32_PAT);
arch/x86/kvm/vmx/nested.c
495
vcpu->arch.walk_mmu = &vcpu->arch.root_mmu;
arch/x86/kvm/vmx/nested.c
4967
vcpu->arch.cr0_guest_owned_bits = vmx_l1_guest_owned_cr0_bits();
arch/x86/kvm/vmx/nested.c
4970
vcpu->arch.cr4_guest_owned_bits = ~vmcs_readl(CR4_GUEST_HOST_MASK);
arch/x86/kvm/vmx/nested.c
4974
vcpu->arch.cr3 = vmcs_readl(GUEST_CR3);
arch/x86/kvm/vmx/nested.c
5090
vcpu->arch.tsc_offset = vcpu->arch.l1_tsc_offset;
arch/x86/kvm/vmx/nested.c
5092
vcpu->arch.tsc_scaling_ratio = vcpu->arch.l1_tsc_scaling_ratio;
arch/x86/kvm/vmx/nested.c
5133
vcpu->arch.nmi_injected = false;
arch/x86/kvm/vmx/nested.c
5148
vmcs_write64(TSC_OFFSET, vcpu->arch.tsc_offset);
arch/x86/kvm/vmx/nested.c
5150
vmcs_write64(TSC_MULTIPLIER, vcpu->arch.tsc_scaling_ratio);
arch/x86/kvm/vmx/nested.c
5181
if (kvm_cpu_has_injectable_intr(vcpu) || vcpu->arch.nmi_pending)
arch/x86/kvm/vmx/nested.c
5557
kvm_mmu_free_roots(vcpu->kvm, &vcpu->arch.guest_mmu, KVM_MMU_ROOTS_ALL);
arch/x86/kvm/vmx/nested.c
6002
mmu = &vcpu->arch.guest_mmu;
arch/x86/kvm/vmx/nested.c
6130
kvm_mmu_free_guest_mode_roots(vcpu->kvm, &vcpu->arch.root_mmu);
arch/x86/kvm/vmx/nested.c
6466
return vcpu->arch.apf.host_apf_flags ||
arch/x86/kvm/vmx/nested.c
6778
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/vmx/nested.c
6853
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/vmx/nested.c
7012
vcpu->arch.nested_run_pending = KVM_NESTED_RUN_PENDING_UNTRUSTED;
arch/x86/kvm/vmx/nested.c
7014
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/vmx/nested.c
7059
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/vmx/nested.c
996
if (CC(vcpu->arch.apic_base & X2APIC_ENABLE && e->index >> 8 == 0x8))
arch/x86/kvm/vmx/pmu_intel.h
14
return vcpu->arch.perf_capabilities;
arch/x86/kvm/vmx/posted_intr.c
155
READ_ONCE(kvm->arch.nr_possible_bypass_irqs);
arch/x86/kvm/vmx/sgx.c
166
if (!vcpu->kvm->arch.sgx_provisioning_allowed &&
arch/x86/kvm/vmx/sgx.c
185
xfrm & ~(vcpu->arch.guest_supported_xcr0 | XFEATURE_MASK_FPSSE) ||
arch/x86/kvm/vmx/sgx.c
445
if (!vcpu->kvm->arch.sgx_provisioning_allowed)
arch/x86/kvm/vmx/tdx.c
1043
if (vcpu->arch.host_pkru != 0)
arch/x86/kvm/vmx/tdx.c
1044
wrpkru(vcpu->arch.host_pkru);
arch/x86/kvm/vmx/tdx.c
1089
if (pi_test_pir(kvm_lapic_get_reg(vcpu->arch.apic, APIC_LVTT) &
arch/x86/kvm/vmx/tdx.c
1096
if (vcpu->arch.host_debugctl & ~TDX_DEBUGCTL_PRESERVED)
arch/x86/kvm/vmx/tdx.c
1097
update_debugctlmsr(vcpu->arch.host_debugctl);
arch/x86/kvm/vmx/tdx.c
1101
vcpu->arch.regs_avail &= TDX_REGS_AVAIL_SET;
arch/x86/kvm/vmx/tdx.c
1126
vcpu->arch.nmi_injected = false;
arch/x86/kvm/vmx/tdx.c
1136
vcpu->arch.nmi_pending = 0;
arch/x86/kvm/vmx/tdx.c
1236
tdx->vcpu.arch.complete_userspace_io = tdx_complete_vmcall_map_gpa;
arch/x86/kvm/vmx/tdx.c
1331
vcpu->arch.pio.count = 0;
arch/x86/kvm/vmx/tdx.c
1337
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/vmx/tdx.c
1341
ret = ctxt->ops->pio_in_emulated(ctxt, vcpu->arch.pio.size,
arch/x86/kvm/vmx/tdx.c
1342
vcpu->arch.pio.port, &val, 1);
arch/x86/kvm/vmx/tdx.c
1354
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/vmx/tdx.c
1379
vcpu->arch.complete_userspace_io = write ? tdx_complete_pio_out :
arch/x86/kvm/vmx/tdx.c
1470
vcpu->arch.complete_userspace_io = tdx_complete_mmio_read;
arch/x86/kvm/vmx/tdx.c
1523
vcpu->arch.complete_userspace_io = tdx_complete_get_td_vm_call_info;
arch/x86/kvm/vmx/tdx.c
1556
vcpu->arch.complete_userspace_io = tdx_complete_simple;
arch/x86/kvm/vmx/tdx.c
1577
vcpu->arch.complete_userspace_io = tdx_complete_simple;
arch/x86/kvm/vmx/tdx.c
1623
if (KVM_BUG_ON(kvm->arch.pre_fault_allowed, kvm) ||
arch/x86/kvm/vmx/tdx.c
1996
WARN_ON_ONCE(vcpu->arch.guest_state_protected &&
arch/x86/kvm/vmx/tdx.c
2000
vcpu->run->fail_entry.cpu = vcpu->arch.last_vmentry_cpu;
arch/x86/kvm/vmx/tdx.c
2156
if (vcpu->arch.mcg_cap & MCG_LMCE_P)
arch/x86/kvm/vmx/tdx.c
2160
if (!msr->host_initiated && !(vcpu->arch.mcg_cap & MCG_LMCE_P))
arch/x86/kvm/vmx/tdx.c
2162
msr->data = vcpu->arch.mcg_ext_ctl;
arch/x86/kvm/vmx/tdx.c
2176
if ((!msr->host_initiated && !(vcpu->arch.mcg_cap & MCG_LMCE_P)) ||
arch/x86/kvm/vmx/tdx.c
2179
vcpu->arch.mcg_ext_ctl = msr->data;
arch/x86/kvm/vmx/tdx.c
2345
td_params->tsc_frequency = TDX_TSC_KHZ_TO_25MHZ(kvm->arch.default_tsc_khz);
arch/x86/kvm/vmx/tdx.c
2753
kvm->arch.gfn_direct_bits = TDX_SHARED_BIT_PWL_5;
arch/x86/kvm/vmx/tdx.c
2755
kvm->arch.gfn_direct_bits = TDX_SHARED_BIT_PWL_4;
arch/x86/kvm/vmx/tdx.c
2817
kvm->arch.pre_fault_allowed = true;
arch/x86/kvm/vmx/tdx.c
2942
vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE;
arch/x86/kvm/vmx/tdx.c
632
kvm->arch.has_protected_state = true;
arch/x86/kvm/vmx/tdx.c
637
kvm->arch.has_protected_eoi = true;
arch/x86/kvm/vmx/tdx.c
638
kvm->arch.has_private_mem = true;
arch/x86/kvm/vmx/tdx.c
639
kvm->arch.disabled_quirks |= KVM_X86_QUIRK_IGNORE_GUEST_PAT;
arch/x86/kvm/vmx/tdx.c
688
fpstate_set_confidential(&vcpu->arch.guest_fpu);
arch/x86/kvm/vmx/tdx.c
689
vcpu->arch.apic->guest_apic_protected = true;
arch/x86/kvm/vmx/tdx.c
692
vcpu->arch.efer = EFER_SCE | EFER_LME | EFER_LMA | EFER_NX;
arch/x86/kvm/vmx/tdx.c
694
vcpu->arch.switch_db_regs = KVM_DEBUGREG_AUTO_SWITCH;
arch/x86/kvm/vmx/tdx.c
695
vcpu->arch.cr0_guest_owned_bits = -1ul;
arch/x86/kvm/vmx/tdx.c
696
vcpu->arch.cr4_guest_owned_bits = -1ul;
arch/x86/kvm/vmx/tdx.c
699
vcpu->arch.guest_tsc_protected = true;
arch/x86/kvm/vmx/tdx.c
700
vcpu->arch.tsc_offset = kvm_tdx->tsc_offset;
arch/x86/kvm/vmx/tdx.c
701
vcpu->arch.l1_tsc_offset = vcpu->arch.tsc_offset;
arch/x86/kvm/vmx/tdx.c
702
vcpu->arch.tsc_scaling_ratio = kvm_tdx->tsc_multiplier;
arch/x86/kvm/vmx/tdx.c
703
vcpu->arch.l1_tsc_scaling_ratio = kvm_tdx->tsc_multiplier;
arch/x86/kvm/vmx/tdx.c
705
vcpu->arch.guest_state_protected =
arch/x86/kvm/vmx/tdx.c
709
vcpu->arch.xfd_no_write_intercept = true;
arch/x86/kvm/vmx/vmx.c
1121
u64 guest_efer = vmx->vcpu.arch.efer;
arch/x86/kvm/vmx/vmx.c
1146
(enable_ept && ((vmx->vcpu.arch.efer ^ kvm_host.efer) & EFER_NX))) {
arch/x86/kvm/vmx/vmx.c
1872
(!vcpu->arch.exception.pending ||
arch/x86/kvm/vmx/vmx.c
1873
vcpu->arch.exception.vector == DB_VECTOR) &&
arch/x86/kvm/vmx/vmx.c
1874
(!vcpu->arch.exception_vmexit.pending ||
arch/x86/kvm/vmx/vmx.c
1875
vcpu->arch.exception_vmexit.vector == DB_VECTOR)) {
arch/x86/kvm/vmx/vmx.c
1904
struct kvm_queued_exception *ex = &vcpu->arch.exception;
arch/x86/kvm/vmx/vmx.c
1928
inc_eip = vcpu->arch.event_exit_inst_len;
arch/x86/kvm/vmx/vmx.c
1937
vmx->vcpu.arch.event_exit_inst_len);
arch/x86/kvm/vmx/vmx.c
1975
(vmx->vcpu.arch.efer & EFER_SCE);
arch/x86/kvm/vmx/vmx.c
2025
vmcs_write64(TSC_OFFSET, vcpu->arch.tsc_offset);
arch/x86/kvm/vmx/vmx.c
2030
vmcs_write64(TSC_MULTIPLIER, vcpu->arch.tsc_scaling_ratio);
arch/x86/kvm/vmx/vmx.c
2110
!(vcpu->arch.arch_capabilities & ARCH_CAP_TSX_CTRL_MSR))
arch/x86/kvm/vmx/vmx.c
2147
msr_info->data = vcpu->arch.mcg_ext_ctl;
arch/x86/kvm/vmx/vmx.c
2333
vcpu->arch.xfd_no_write_intercept = true;
arch/x86/kvm/vmx/vmx.c
2429
!(vcpu->arch.arch_capabilities & ARCH_CAP_TSX_CTRL_MSR))
arch/x86/kvm/vmx/vmx.c
2452
vcpu->arch.mcg_ext_ctl = data;
arch/x86/kvm/vmx/vmx.c
2605
vcpu->arch.regs[VCPU_REGS_RSP] = vmcs_readl(GUEST_RSP);
arch/x86/kvm/vmx/vmx.c
2608
vcpu->arch.regs[VCPU_REGS_RIP] = vmcs_readl(GUEST_RIP);
arch/x86/kvm/vmx/vmx.c
2615
guest_owned_bits = vcpu->arch.cr0_guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
2617
vcpu->arch.cr0 &= ~guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
2618
vcpu->arch.cr0 |= vmcs_readl(GUEST_CR0) & guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
2626
vcpu->arch.cr3 = vmcs_readl(GUEST_CR3);
arch/x86/kvm/vmx/vmx.c
2629
guest_owned_bits = vcpu->arch.cr4_guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
2631
vcpu->arch.cr4 &= ~guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
2632
vcpu->arch.cr4 |= vmcs_readl(GUEST_CR4) & guest_owned_bits;
arch/x86/kvm/vmx/vmx.c
3208
vcpu->arch.efer = efer;
arch/x86/kvm/vmx/vmx.c
3239
vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA);
arch/x86/kvm/vmx/vmx.c
3244
vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA);
arch/x86/kvm/vmx/vmx.c
3315
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/vmx/vmx.c
3351
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/vmx/vmx.c
3366
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/vmx/vmx.c
3418
vcpu->arch.cr0 = cr0;
arch/x86/kvm/vmx/vmx.c
3422
if (vcpu->arch.efer & EFER_LME) {
arch/x86/kvm/vmx/vmx.c
3505
guest_cr3 = vcpu->arch.cr3;
arch/x86/kvm/vmx/vmx.c
3563
vcpu->arch.cr4 = cr4;
arch/x86/kvm/vmx/vmx.c
3595
vcpu->arch.cpuid_dynamic_bits_dirty = true;
arch/x86/kvm/vmx/vmx.c
4141
msr_bitmap[read_idx] = ~kvm_lapic_readable_reg_mask(vcpu->arch.apic);
arch/x86/kvm/vmx/vmx.c
4265
if (vcpu->arch.xfd_no_write_intercept)
arch/x86/kvm/vmx/vmx.c
4376
if (!vcpu->arch.apic->apicv_active)
arch/x86/kvm/vmx/vmx.c
4500
vcpu->arch.cr4_guest_owned_bits = KVM_POSSIBLE_CR4_GUEST_BITS &
arch/x86/kvm/vmx/vmx.c
4501
~vcpu->arch.cr4_guest_rsvd_bits;
arch/x86/kvm/vmx/vmx.c
4503
vcpu->arch.cr4_guest_owned_bits &= ~X86_CR4_TLBFLUSH_BITS;
arch/x86/kvm/vmx/vmx.c
4504
vcpu->arch.cr4_guest_owned_bits &= ~X86_CR4_PDPTR_BITS;
arch/x86/kvm/vmx/vmx.c
4507
vcpu->arch.cr4_guest_owned_bits &=
arch/x86/kvm/vmx/vmx.c
4509
vmcs_writel(CR4_GUEST_HOST_MASK, ~vcpu->arch.cr4_guest_owned_bits);
arch/x86/kvm/vmx/vmx.c
4601
if (vmx->vcpu.arch.switch_db_regs & KVM_DEBUGREG_WONT_EXIT)
arch/x86/kvm/vmx/vmx.c
4785
if (!vcpu->kvm->arch.bus_lock_detection_enabled)
arch/x86/kvm/vmx/vmx.c
4796
return get_order(kvm->arch.max_vcpu_ids * sizeof(*to_kvm_vmx(kvm)->pid_table));
arch/x86/kvm/vmx/vmx.c
4868
vmcs_write16(LAST_PID_POINTER_INDEX, kvm->arch.max_vcpu_ids - 1);
arch/x86/kvm/vmx/vmx.c
4878
vmcs_write32(NOTIFY_WINDOW, kvm->arch.notify_window);
arch/x86/kvm/vmx/vmx.c
4901
vmcs_write64(GUEST_IA32_PAT, vmx->vcpu.arch.pat);
arch/x86/kvm/vmx/vmx.c
4908
vmx->vcpu.arch.cr0_guest_owned_bits = vmx_l1_guest_owned_cr0_bits();
arch/x86/kvm/vmx/vmx.c
4909
vmcs_writel(CR0_GUEST_HOST_MASK, ~vmx->vcpu.arch.cr0_guest_owned_bits);
arch/x86/kvm/vmx/vmx.c
4943
__pa(vmx->vcpu.arch.apic->regs));
arch/x86/kvm/vmx/vmx.c
4971
vcpu->arch.microcode_version = 0x100000000ULL;
arch/x86/kvm/vmx/vmx.c
5069
int irq = vcpu->arch.interrupt.nr;
arch/x86/kvm/vmx/vmx.c
5071
trace_kvm_inj_virq(irq, vcpu->arch.interrupt.soft, reinjected);
arch/x86/kvm/vmx/vmx.c
5076
if (vcpu->arch.interrupt.soft)
arch/x86/kvm/vmx/vmx.c
5077
inc_eip = vcpu->arch.event_exit_inst_len;
arch/x86/kvm/vmx/vmx.c
5082
if (vcpu->arch.interrupt.soft) {
arch/x86/kvm/vmx/vmx.c
5085
vmx->vcpu.arch.event_exit_inst_len);
arch/x86/kvm/vmx/vmx.c
517
if ((vcpu->arch.arch_capabilities & ARCH_CAP_FB_CLEAR) ||
arch/x86/kvm/vmx/vmx.c
5173
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/vmx/vmx.c
518
((vcpu->arch.arch_capabilities & ARCH_CAP_MDS_NO) &&
arch/x86/kvm/vmx/vmx.c
519
(vcpu->arch.arch_capabilities & ARCH_CAP_TAA_NO) &&
arch/x86/kvm/vmx/vmx.c
520
(vcpu->arch.arch_capabilities & ARCH_CAP_PSDP_NO) &&
arch/x86/kvm/vmx/vmx.c
5200
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/vmx/vmx.c
521
(vcpu->arch.arch_capabilities & ARCH_CAP_FBSDP_NO) &&
arch/x86/kvm/vmx/vmx.c
522
(vcpu->arch.arch_capabilities & ARCH_CAP_SBDR_SSDP_NO)))
arch/x86/kvm/vmx/vmx.c
5247
to_vmx(vcpu)->vcpu.arch.event_exit_inst_len =
arch/x86/kvm/vmx/vmx.c
5277
if (vcpu->arch.halt_request) {
arch/x86/kvm/vmx/vmx.c
5278
vcpu->arch.halt_request = 0;
arch/x86/kvm/vmx/vmx.c
5323
return vcpu->arch.guest_fpu.fpstate->xfd &&
arch/x86/kvm/vmx/vmx.c
5331
if (vcpu->arch.apf.host_apf_flags)
arch/x86/kvm/vmx/vmx.c
5395
is_xfd_nm_fault(vcpu) ? vcpu->arch.guest_fpu.xfd_err : 0);
arch/x86/kvm/vmx/vmx.c
5444
vcpu->run->internal.data[3] = vcpu->arch.last_vmentry_cpu;
arch/x86/kvm/vmx/vmx.c
5494
kvm_run->debug.arch.dr6 = dr6 | DR6_ACTIVE_LOW;
arch/x86/kvm/vmx/vmx.c
5495
kvm_run->debug.arch.dr7 = vmcs_readl(GUEST_DR7);
arch/x86/kvm/vmx/vmx.c
5503
vmx->vcpu.arch.event_exit_inst_len =
arch/x86/kvm/vmx/vmx.c
5506
kvm_run->debug.arch.pc = kvm_get_linear_rip(vcpu);
arch/x86/kvm/vmx/vmx.c
5507
kvm_run->debug.arch.exception = ex_no;
arch/x86/kvm/vmx/vmx.c
5736
vcpu->run->debug.arch.dr6 = DR6_BD | DR6_ACTIVE_LOW;
arch/x86/kvm/vmx/vmx.c
5737
vcpu->run->debug.arch.dr7 = dr7;
arch/x86/kvm/vmx/vmx.c
5738
vcpu->run->debug.arch.pc = kvm_get_linear_rip(vcpu);
arch/x86/kvm/vmx/vmx.c
5739
vcpu->run->debug.arch.exception = DB_VECTOR;
arch/x86/kvm/vmx/vmx.c
5756
vcpu->arch.switch_db_regs |= KVM_DEBUGREG_WONT_EXIT;
arch/x86/kvm/vmx/vmx.c
5774
get_debugreg(vcpu->arch.db[0], 0);
arch/x86/kvm/vmx/vmx.c
5775
get_debugreg(vcpu->arch.db[1], 1);
arch/x86/kvm/vmx/vmx.c
5776
get_debugreg(vcpu->arch.db[2], 2);
arch/x86/kvm/vmx/vmx.c
5777
get_debugreg(vcpu->arch.db[3], 3);
arch/x86/kvm/vmx/vmx.c
5778
get_debugreg(vcpu->arch.dr6, 6);
arch/x86/kvm/vmx/vmx.c
5779
vcpu->arch.dr7 = vmcs_readl(GUEST_DR7);
arch/x86/kvm/vmx/vmx.c
5781
vcpu->arch.switch_db_regs &= ~KVM_DEBUGREG_WONT_EXIT;
arch/x86/kvm/vmx/vmx.c
5888
vcpu->arch.nmi_injected = false;
arch/x86/kvm/vmx/vmx.c
6012
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/vmx/vmx.c
6021
(kvm_is_exception_pending(vcpu) || vcpu->arch.exception.injected);
arch/x86/kvm/vmx/vmx.c
6055
if (vcpu->arch.halt_request) {
arch/x86/kvm/vmx/vmx.c
6056
vcpu->arch.halt_request = 0;
arch/x86/kvm/vmx/vmx.c
6260
if (vcpu->kvm->arch.notify_vmexit_flags & KVM_X86_NOTIFY_VMEXIT_USER ||
arch/x86/kvm/vmx/vmx.c
6498
vmx->loaded_vmcs->vmcs, vcpu->arch.last_vmentry_cpu);
arch/x86/kvm/vmx/vmx.c
6537
vcpu->arch.efer | (EFER_LMA | EFER_LME));
arch/x86/kvm/vmx/vmx.c
6540
vcpu->arch.efer & ~(EFER_LMA | EFER_LME));
arch/x86/kvm/vmx/vmx.c
6696
if (KVM_BUG_ON(vcpu->arch.nested_run_pending, vcpu->kvm))
arch/x86/kvm/vmx/vmx.c
6749
vcpu->run->fail_entry.cpu = vcpu->arch.last_vmentry_cpu;
arch/x86/kvm/vmx/vmx.c
6758
vcpu->run->fail_entry.cpu = vcpu->arch.last_vmentry_cpu;
arch/x86/kvm/vmx/vmx.c
6779
vcpu->arch.nmi_pending) {
arch/x86/kvm/vmx/vmx.c
6905
else if (VALID_PAGE(vcpu->arch.root_mmu.root.hpa))
arch/x86/kvm/vmx/vmx.c
6906
vmx_flush_tlb_ept_root(vcpu->arch.root_mmu.root.hpa);
arch/x86/kvm/vmx/vmx.c
7105
rdmsrq(MSR_IA32_XFD_ERR, vcpu->arch.guest_fpu.xfd_err);
arch/x86/kvm/vmx/vmx.c
7112
vcpu->arch.apf.host_apf_flags = kvm_read_and_reset_apf_flags();
arch/x86/kvm/vmx/vmx.c
7143
vcpu->arch.at_instruction_boundary = true;
arch/x86/kvm/vmx/vmx.c
7243
vcpu->arch.nmi_injected = false;
arch/x86/kvm/vmx/vmx.c
7257
vcpu->arch.nmi_injected = true;
arch/x86/kvm/vmx/vmx.c
7266
vcpu->arch.event_exit_inst_len = vmcs_read32(instr_len_field);
arch/x86/kvm/vmx/vmx.c
7280
vcpu->arch.event_exit_inst_len = vmcs_read32(instr_len_field);
arch/x86/kvm/vmx/vmx.c
7468
if (vcpu->arch.cr2 != native_read_cr2())
arch/x86/kvm/vmx/vmx.c
7469
native_write_cr2(vcpu->arch.cr2);
arch/x86/kvm/vmx/vmx.c
7471
vmx->fail = __vmx_vcpu_run(vmx, (unsigned long *)&vcpu->arch.regs,
arch/x86/kvm/vmx/vmx.c
7474
vcpu->arch.cr2 = native_read_cr2();
arch/x86/kvm/vmx/vmx.c
7475
vcpu->arch.regs_avail &= ~VMX_REGS_LAZY_LOAD_SET;
arch/x86/kvm/vmx/vmx.c
7538
vmcs_writel(GUEST_RSP, vcpu->arch.regs[VCPU_REGS_RSP]);
arch/x86/kvm/vmx/vmx.c
7540
vmcs_writel(GUEST_RIP, vcpu->arch.regs[VCPU_REGS_RIP]);
arch/x86/kvm/vmx/vmx.c
7541
vcpu->arch.regs_dirty = 0;
arch/x86/kvm/vmx/vmx.c
7544
set_debugreg(vcpu->arch.dr6, 6);
arch/x86/kvm/vmx/vmx.c
7601
if (vcpu->arch.host_debugctl)
arch/x86/kvm/vmx/vmx.c
7602
update_debugctlmsr(vcpu->arch.host_debugctl);
arch/x86/kvm/vmx/vmx.c
7624
if (vcpu->arch.nested_run_pending &&
arch/x86/kvm/vmx/vmx.c
7628
vcpu->arch.nested_run_pending = 0;
arch/x86/kvm/vmx/vmx.c
7795
kvm->arch.cpu_dirty_log_size = PML_LOG_NR_ENTRIES;
arch/x86/kvm/vmx/vmx.c
8310
struct kvm_timer *ktimer = &vcpu->arch.apic->lapic_timer;
arch/x86/kvm/vmx/vmx.c
8325
if (vcpu->arch.l1_tsc_scaling_ratio != kvm_caps.default_tsc_scaling_ratio &&
arch/x86/kvm/vmx/vmx.c
8328
vcpu->arch.l1_tsc_scaling_ratio, &delta_tsc))
arch/x86/kvm/vmx/vmx.c
8373
if (vcpu->arch.mcg_cap & MCG_LMCE_P)
arch/x86/kvm/vmx/vmx.c
8385
if (vcpu->arch.nested_run_pending)
arch/x86/kvm/vmx/vmx.c
8430
vcpu->arch.nested_run_pending = KVM_NESTED_RUN_PENDING;
arch/x86/kvm/vmx/vmx.c
8526
(unsigned long *)&vcpu->arch.pmu.global_status);
arch/x86/kvm/vmx/vmx.c
950
if (vcpu->arch.xfd_no_write_intercept)
arch/x86/kvm/vmx/vmx.h
407
val |= vcpu->arch.host_debugctl & VMX_HOST_OWNED_DEBUGCTL_BITS;
arch/x86/kvm/vmx/vmx.h
420
if (!((val ^ vcpu->arch.host_debugctl) & VMX_HOST_OWNED_DEBUGCTL_BITS))
arch/x86/kvm/x86.c
1014
atomic_inc(&vcpu->arch.nmi_queued);
arch/x86/kvm/x86.c
10298
if (vcpu->arch.tsc_always_catchup)
arch/x86/kvm/x86.c
10342
return (READ_ONCE(kvm->arch.apicv_inhibit_reasons) == 0);
arch/x86/kvm/x86.c
10348
ulong vm_reasons = READ_ONCE(vcpu->kvm->arch.apicv_inhibit_reasons);
arch/x86/kvm/x86.c
10376
set_or_clear_apicv_inhibit(&kvm->arch.apicv_inhibit_reasons, reason, true);
arch/x86/kvm/x86.c
10378
init_rwsem(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
10392
map = rcu_dereference(vcpu->kvm->arch.apic_map);
arch/x86/kvm/x86.c
1050
return (vcpu->arch.apf.msr_en_val & mask) == mask;
arch/x86/kvm/x86.c
10518
vcpu->arch.complete_userspace_io = complete_hypercall;
arch/x86/kvm/x86.c
1055
return vcpu->arch.reserved_gpa_bits | rsvd_bits(5, 8) | rsvd_bits(1, 2);
arch/x86/kvm/x86.c
10581
kvm_run->apic_base = vcpu->arch.apic_base;
arch/x86/kvm/x86.c
10603
if (vcpu->arch.apic->apicv_active)
arch/x86/kvm/x86.c
10606
if (!vcpu->arch.apic->vapic_addr)
arch/x86/kvm/x86.c
1063
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
10639
vcpu->arch.exception.has_error_code &= is_protmode(vcpu);
arch/x86/kvm/x86.c
10641
trace_kvm_inj_exception(vcpu->arch.exception.vector,
arch/x86/kvm/x86.c
10642
vcpu->arch.exception.has_error_code,
arch/x86/kvm/x86.c
10643
vcpu->arch.exception.error_code,
arch/x86/kvm/x86.c
10644
vcpu->arch.exception.injected);
arch/x86/kvm/x86.c
10727
if (vcpu->arch.exception.injected)
arch/x86/kvm/x86.c
10731
else if (vcpu->arch.nmi_injected)
arch/x86/kvm/x86.c
10733
else if (vcpu->arch.interrupt.injected)
arch/x86/kvm/x86.c
10741
WARN_ON_ONCE(vcpu->arch.exception.injected &&
arch/x86/kvm/x86.c
10742
vcpu->arch.exception.pending);
arch/x86/kvm/x86.c
10761
WARN_ON_ONCE(vcpu->arch.exception_vmexit.injected ||
arch/x86/kvm/x86.c
10762
vcpu->arch.exception_vmexit.pending);
arch/x86/kvm/x86.c
10771
if (vcpu->arch.exception.pending) {
arch/x86/kvm/x86.c
10782
if (exception_type(vcpu->arch.exception.vector) == EXCPT_FAULT)
arch/x86/kvm/x86.c
10786
if (vcpu->arch.exception.vector == DB_VECTOR &&
arch/x86/kvm/x86.c
10787
vcpu->arch.dr7 & DR7_GD) {
arch/x86/kvm/x86.c
10788
vcpu->arch.dr7 &= ~DR7_GD;
arch/x86/kvm/x86.c
10794
vcpu->arch.exception.pending = false;
arch/x86/kvm/x86.c
10795
vcpu->arch.exception.injected = true;
arch/x86/kvm/x86.c
10816
if (vcpu->arch.smi_pending) {
arch/x86/kvm/x86.c
10822
vcpu->arch.smi_pending = false;
arch/x86/kvm/x86.c
10823
++vcpu->arch.smi_count;
arch/x86/kvm/x86.c
10831
if (vcpu->arch.nmi_pending) {
arch/x86/kvm/x86.c
10837
--vcpu->arch.nmi_pending;
arch/x86/kvm/x86.c
10838
vcpu->arch.nmi_injected = true;
arch/x86/kvm/x86.c
10843
if (vcpu->arch.nmi_pending)
arch/x86/kvm/x86.c
10882
WARN_ON_ONCE(vcpu->arch.exception.pending ||
arch/x86/kvm/x86.c
10883
vcpu->arch.exception_vmexit.pending);
arch/x86/kvm/x86.c
10908
if (kvm_x86_call(get_nmi_mask)(vcpu) || vcpu->arch.nmi_injected)
arch/x86/kvm/x86.c
10920
vcpu->arch.nmi_pending += atomic_xchg(&vcpu->arch.nmi_queued, 0);
arch/x86/kvm/x86.c
10921
vcpu->arch.nmi_pending = min(vcpu->arch.nmi_pending, limit);
arch/x86/kvm/x86.c
10923
if (vcpu->arch.nmi_pending &&
arch/x86/kvm/x86.c
10925
vcpu->arch.nmi_pending--;
arch/x86/kvm/x86.c
10927
if (vcpu->arch.nmi_pending)
arch/x86/kvm/x86.c
10934
return vcpu->arch.nmi_pending +
arch/x86/kvm/x86.c
10951
struct kvm_lapic *apic = vcpu->arch.apic;
arch/x86/kvm/x86.c
10957
down_read(&vcpu->kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
10982
up_read(&vcpu->kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
11002
if (apic_x2apic_mode(vcpu->arch.apic) &&
arch/x86/kvm/x86.c
11014
lockdep_assert_held_write(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
11019
old = new = kvm->arch.apicv_inhibit_reasons;
arch/x86/kvm/x86.c
1102
vcpu->arch.pdptrs_from_userspace = false;
arch/x86/kvm/x86.c
11025
atomic_read(&kvm->arch.apicv_nr_irq_window_req));
arch/x86/kvm/x86.c
11041
kvm->arch.apicv_inhibit_reasons = new;
arch/x86/kvm/x86.c
11050
kvm->arch.apicv_inhibit_reasons = new;
arch/x86/kvm/x86.c
11060
down_write(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
11062
up_write(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
11085
if (READ_ONCE(kvm->arch.apicv_inhibit_reasons) & ~BIT(APICV_INHIBIT_REASON_IRQWIN)) {
arch/x86/kvm/x86.c
11086
guard(rwsem_read)(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
11087
if (READ_ONCE(kvm->arch.apicv_inhibit_reasons) & ~BIT(APICV_INHIBIT_REASON_IRQWIN)) {
arch/x86/kvm/x86.c
11088
atomic_add(add, &kvm->arch.apicv_nr_irq_window_req);
arch/x86/kvm/x86.c
11099
guard(rwsem_write)(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
11100
if (atomic_add_return(add, &kvm->arch.apicv_nr_irq_window_req) == inc)
arch/x86/kvm/x86.c
11110
bitmap_zero(vcpu->arch.ioapic_handled_vectors, 256);
arch/x86/kvm/x86.c
11111
vcpu->arch.highest_stale_pending_ioapic_eoi = -1;
arch/x86/kvm/x86.c
11116
kvm_scan_ioapic_routes(vcpu, vcpu->arch.ioapic_handled_vectors);
arch/x86/kvm/x86.c
11119
kvm_ioapic_scan_entry(vcpu, vcpu->arch.ioapic_handled_vectors);
arch/x86/kvm/x86.c
11123
vcpu->arch.load_eoi_exitmap_pending = true;
arch/x86/kvm/x86.c
11130
if (!kvm_apic_hw_enabled(vcpu->arch.apic))
arch/x86/kvm/x86.c
11138
vcpu->arch.ioapic_handled_vectors,
arch/x86/kvm/x86.c
11145
vcpu, (u64 *)vcpu->arch.ioapic_handled_vectors);
arch/x86/kvm/x86.c
11253
vcpu->arch.apf.halted = true;
arch/x86/kvm/x86.c
11270
BUG_ON(vcpu->arch.pending_ioapic_eoi > 255);
arch/x86/kvm/x86.c
11271
if (test_bit(vcpu->arch.pending_ioapic_eoi,
arch/x86/kvm/x86.c
11272
vcpu->arch.ioapic_handled_vectors)) {
arch/x86/kvm/x86.c
11275
vcpu->arch.pending_ioapic_eoi;
arch/x86/kvm/x86.c
11331
if (vcpu->arch.mp_state != KVM_MP_STATE_RUNNABLE) {
arch/x86/kvm/x86.c
11346
if (vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED) {
arch/x86/kvm/x86.c
11430
if (vcpu->arch.guest_fpu.xfd_err)
arch/x86/kvm/x86.c
11431
wrmsrq(MSR_IA32_XFD_ERR, vcpu->arch.guest_fpu.xfd_err);
arch/x86/kvm/x86.c
11435
if (unlikely(vcpu->arch.switch_db_regs &&
arch/x86/kvm/x86.c
11436
!(vcpu->arch.switch_db_regs & KVM_DEBUGREG_AUTO_SWITCH))) {
arch/x86/kvm/x86.c
11438
set_debugreg(vcpu->arch.eff_db[0], 0);
arch/x86/kvm/x86.c
11439
set_debugreg(vcpu->arch.eff_db[1], 1);
arch/x86/kvm/x86.c
11440
set_debugreg(vcpu->arch.eff_db[2], 2);
arch/x86/kvm/x86.c
11441
set_debugreg(vcpu->arch.eff_db[3], 3);
arch/x86/kvm/x86.c
11443
if (unlikely(vcpu->arch.switch_db_regs & KVM_DEBUGREG_WONT_EXIT))
arch/x86/kvm/x86.c
11456
if ((debug_ctl ^ vcpu->arch.host_debugctl) & kvm_x86_ops.HOST_OWNED_DEBUGCTL &&
arch/x86/kvm/x86.c
11457
!vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
11459
vcpu->arch.host_debugctl = debug_ctl;
arch/x86/kvm/x86.c
11509
if (unlikely(vcpu->arch.switch_db_regs & KVM_DEBUGREG_WONT_EXIT)) {
arch/x86/kvm/x86.c
11511
WARN_ON(vcpu->arch.switch_db_regs & KVM_DEBUGREG_AUTO_SWITCH);
arch/x86/kvm/x86.c
11527
vcpu->arch.last_vmentry_cpu = vcpu->cpu;
arch/x86/kvm/x86.c
11528
vcpu->arch.last_guest_tsc = kvm_read_l1_tsc(vcpu, rdtsc());
arch/x86/kvm/x86.c
11540
if (vcpu->arch.xfd_no_write_intercept)
arch/x86/kvm/x86.c
11545
if (vcpu->arch.guest_fpu.xfd_err)
arch/x86/kvm/x86.c
11594
!vcpu->arch.guest_state_protected)) {
arch/x86/kvm/x86.c
11599
if (unlikely(vcpu->arch.tsc_always_catchup))
arch/x86/kvm/x86.c
11602
if (vcpu->arch.apic_attention)
arch/x86/kvm/x86.c
11615
if (unlikely(vcpu->arch.apic_attention))
arch/x86/kvm/x86.c
11623
return (vcpu->arch.mp_state == KVM_MP_STATE_RUNNABLE &&
arch/x86/kvm/x86.c
11624
!vcpu->arch.apf.halted);
arch/x86/kvm/x86.c
11640
(vcpu->arch.nmi_pending &&
arch/x86/kvm/x86.c
11646
(vcpu->arch.smi_pending &&
arch/x86/kvm/x86.c
11677
return kvm_vcpu_running(vcpu) || vcpu->arch.pv.pv_unhalted ||
arch/x86/kvm/x86.c
11699
if (vcpu->arch.mp_state == KVM_MP_STATE_HALTED)
arch/x86/kvm/x86.c
11732
switch(vcpu->arch.mp_state) {
arch/x86/kvm/x86.c
11738
vcpu->arch.apf.halted = false;
arch/x86/kvm/x86.c
11763
vcpu->arch.at_instruction_boundary = false;
arch/x86/kvm/x86.c
1179
if ((vcpu->arch.efer & EFER_LME) && !is_paging(vcpu) &&
arch/x86/kvm/x86.c
11811
if (kvm_vcpu_has_events(vcpu) || vcpu->arch.pv.pv_unhalted)
arch/x86/kvm/x86.c
11870
return vcpu->arch.preempted_in_kernel;
arch/x86/kvm/x86.c
11875
if (READ_ONCE(vcpu->arch.pv.pv_unhalted))
arch/x86/kvm/x86.c
11895
if (KVM_BUG_ON(!vcpu->arch.pio.count, vcpu->kvm))
arch/x86/kvm/x86.c
1190
if (!(vcpu->arch.efer & EFER_LME) && (cr0 & X86_CR0_PG) &&
arch/x86/kvm/x86.c
11959
vcpu->arch.complete_userspace_io = complete_emulated_mmio;
arch/x86/kvm/x86.c
11966
if (KVM_BUG_ON(vcpu->arch.guest_fpu.fpstate->in_use, vcpu->kvm))
arch/x86/kvm/x86.c
11970
fpu_swap_kvm_fpstate(&vcpu->arch.guest_fpu, true);
arch/x86/kvm/x86.c
11977
if (KVM_BUG_ON(!vcpu->arch.guest_fpu.fpstate->in_use, vcpu->kvm))
arch/x86/kvm/x86.c
11980
fpu_swap_kvm_fpstate(&vcpu->arch.guest_fpu, false);
arch/x86/kvm/x86.c
11991
if (vcpu->arch.nested_run_pending == KVM_NESTED_RUN_PENDING)
arch/x86/kvm/x86.c
11992
vcpu->arch.nested_run_pending = KVM_NESTED_RUN_PENDING_UNTRUSTED;
arch/x86/kvm/x86.c
12000
if (WARN_ON_ONCE(vcpu->arch.mp_state == KVM_MP_STATE_SIPI_RECEIVED))
arch/x86/kvm/x86.c
12007
if (vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED &&
arch/x86/kvm/x86.c
12016
struct kvm_queued_exception *ex = &vcpu->arch.exception;
arch/x86/kvm/x86.c
12031
if (unlikely(vcpu->arch.mp_state == KVM_MP_STATE_UNINITIALIZED)) {
arch/x86/kvm/x86.c
12089
if (vcpu->arch.exception_from_userspace && is_guest_mode(vcpu) &&
arch/x86/kvm/x86.c
12098
vcpu->arch.exception_from_userspace = false;
arch/x86/kvm/x86.c
12100
if (unlikely(vcpu->arch.complete_userspace_io)) {
arch/x86/kvm/x86.c
12101
int (*cui)(struct kvm_vcpu *) = vcpu->arch.complete_userspace_io;
arch/x86/kvm/x86.c
12102
vcpu->arch.complete_userspace_io = NULL;
arch/x86/kvm/x86.c
12107
WARN_ON_ONCE(vcpu->arch.pio.count);
arch/x86/kvm/x86.c
12124
if (kvm_run->kvm_valid_regs && likely(!vcpu->arch.guest_state_protected))
arch/x86/kvm/x86.c
12136
if (vcpu->arch.emulate_regs_need_sync_to_vcpu) {
arch/x86/kvm/x86.c
12144
emulator_writeback_register_cache(vcpu->arch.emulate_ctxt);
arch/x86/kvm/x86.c
12145
vcpu->arch.emulate_regs_need_sync_to_vcpu = false;
arch/x86/kvm/x86.c
12172
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
12173
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
1218
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
12184
vcpu->arch.emulate_regs_need_sync_from_vcpu = true;
arch/x86/kvm/x86.c
12185
vcpu->arch.emulate_regs_need_sync_to_vcpu = false;
arch/x86/kvm/x86.c
12209
vcpu->arch.exception.pending = false;
arch/x86/kvm/x86.c
12210
vcpu->arch.exception_vmexit.pending = false;
arch/x86/kvm/x86.c
12217
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
12218
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
12231
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
1224
if (vcpu->arch.xcr0 != kvm_host.xcr0)
arch/x86/kvm/x86.c
12253
sregs->cr2 = vcpu->arch.cr2;
arch/x86/kvm/x86.c
1226
load_guest ? vcpu->arch.xcr0 : kvm_host.xcr0);
arch/x86/kvm/x86.c
12260
sregs->efer = vcpu->arch.efer;
arch/x86/kvm/x86.c
12261
sregs->apic_base = vcpu->arch.apic_base;
arch/x86/kvm/x86.c
12268
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
12271
if (vcpu->arch.interrupt.injected && !vcpu->arch.interrupt.soft)
arch/x86/kvm/x86.c
12272
set_bit(vcpu->arch.interrupt.nr,
arch/x86/kvm/x86.c
12282
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
1229
vcpu->arch.ia32_xss != kvm_host.xss)
arch/x86/kvm/x86.c
12297
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
12298
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
1230
wrmsrq(MSR_IA32_XSS, load_guest ? vcpu->arch.ia32_xss : kvm_host.xss);
arch/x86/kvm/x86.c
12320
if ((vcpu->arch.mp_state == KVM_MP_STATE_HALTED ||
arch/x86/kvm/x86.c
12321
vcpu->arch.mp_state == KVM_MP_STATE_AP_RESET_HOLD) &&
arch/x86/kvm/x86.c
12322
vcpu->arch.pv.pv_unhalted)
arch/x86/kvm/x86.c
12325
mp_state->mp_state = vcpu->arch.mp_state;
arch/x86/kvm/x86.c
1235
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
12364
set_bit(KVM_APIC_SIPI, &vcpu->arch.apic->pending_events);
arch/x86/kvm/x86.c
12379
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
1239
vcpu->arch.pkru != vcpu->arch.host_pkru &&
arch/x86/kvm/x86.c
1240
((vcpu->arch.xcr0 & XFEATURE_MASK_PKRU) ||
arch/x86/kvm/x86.c
1242
wrpkru(vcpu->arch.pkru);
arch/x86/kvm/x86.c
12463
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
1247
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
12473
vcpu->arch.cr2 = sregs->cr2;
arch/x86/kvm/x86.c
12475
vcpu->arch.cr3 = sregs->cr3;
arch/x86/kvm/x86.c
12481
*mmu_reset_needed |= vcpu->arch.efer != sregs->efer;
arch/x86/kvm/x86.c
1251
((vcpu->arch.xcr0 & XFEATURE_MASK_PKRU) ||
arch/x86/kvm/x86.c
1253
vcpu->arch.pkru = rdpkru();
arch/x86/kvm/x86.c
1254
if (vcpu->arch.pkru != vcpu->arch.host_pkru)
arch/x86/kvm/x86.c
1255
wrpkru(vcpu->arch.host_pkru);
arch/x86/kvm/x86.c
12557
if (valid_pdptrs && (!pae || vcpu->arch.guest_state_protected))
arch/x86/kvm/x86.c
12571
vcpu->arch.pdptrs_from_userspace = true;
arch/x86/kvm/x86.c
12585
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
12586
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
12604
down_write(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
12613
up_write(&kvm->arch.apicv_update_lock);
arch/x86/kvm/x86.c
1262
return vcpu->arch.guest_supported_xcr0 & XFEATURE_MASK_USER_DYNAMIC;
arch/x86/kvm/x86.c
12622
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
12629
if (kvm_is_exception_pending(vcpu) || vcpu->arch.exception.injected)
arch/x86/kvm/x86.c
12649
vcpu->arch.eff_db[i] = dbg->arch.debugreg[i];
arch/x86/kvm/x86.c
12650
vcpu->arch.guest_debug_dr7 = dbg->arch.debugreg[7];
arch/x86/kvm/x86.c
12653
vcpu->arch.eff_db[i] = vcpu->arch.db[i];
arch/x86/kvm/x86.c
12658
vcpu->arch.singlestep_rip = kvm_get_linear_rip(vcpu);
arch/x86/kvm/x86.c
1269
u64 old_xcr0 = vcpu->arch.xcr0;
arch/x86/kvm/x86.c
12705
if (fpstate_is_confidential(&vcpu->arch.guest_fpu))
arch/x86/kvm/x86.c
12706
return vcpu->kvm->arch.has_protected_state ? -EINVAL : 0;
arch/x86/kvm/x86.c
12710
fxsave = &vcpu->arch.guest_fpu.fpstate->regs.fxsave;
arch/x86/kvm/x86.c
12728
if (fpstate_is_confidential(&vcpu->arch.guest_fpu))
arch/x86/kvm/x86.c
12729
return vcpu->kvm->arch.has_protected_state ? -EINVAL : 0;
arch/x86/kvm/x86.c
12733
fxsave = &vcpu->arch.guest_fpu.fpstate->regs.fxsave;
arch/x86/kvm/x86.c
12802
if (!kvm->arch.max_vcpu_ids)
arch/x86/kvm/x86.c
12803
kvm->arch.max_vcpu_ids = KVM_MAX_VCPU_IDS;
arch/x86/kvm/x86.c
12805
if (id >= kvm->arch.max_vcpu_ids)
arch/x86/kvm/x86.c
12817
if (enable_mediated_pmu && kvm->arch.enable_pmu &&
arch/x86/kvm/x86.c
12818
!kvm->arch.created_mediated_pmu) {
arch/x86/kvm/x86.c
12825
kvm->arch.created_mediated_pmu = true;
arch/x86/kvm/x86.c
12827
kvm->arch.enable_pmu = false;
arch/x86/kvm/x86.c
12838
vcpu->arch.last_vmentry_cpu = -1;
arch/x86/kvm/x86.c
12839
vcpu->arch.regs_avail = ~0;
arch/x86/kvm/x86.c
12840
vcpu->arch.regs_dirty = ~0;
arch/x86/kvm/x86.c
12842
kvm_gpc_init(&vcpu->arch.pv_time, vcpu->kvm);
arch/x86/kvm/x86.c
1285
valid_bits = vcpu->arch.guest_supported_xcr0 | XFEATURE_MASK_FP;
arch/x86/kvm/x86.c
12862
vcpu->arch.pio_data = page_address(page);
arch/x86/kvm/x86.c
12864
vcpu->arch.mce_banks = kcalloc(KVM_MAX_MCE_BANKS * 4, sizeof(u64),
arch/x86/kvm/x86.c
12866
vcpu->arch.mci_ctl2_banks = kcalloc(KVM_MAX_MCE_BANKS, sizeof(u64),
arch/x86/kvm/x86.c
12868
if (!vcpu->arch.mce_banks || !vcpu->arch.mci_ctl2_banks)
arch/x86/kvm/x86.c
12870
vcpu->arch.mcg_cap = KVM_MAX_MCE_BANKS;
arch/x86/kvm/x86.c
12872
if (!zalloc_cpumask_var(&vcpu->arch.wbinvd_dirty_mask,
arch/x86/kvm/x86.c
12879
if (!fpu_alloc_guest_fpstate(&vcpu->arch.guest_fpu)) {
arch/x86/kvm/x86.c
12887
vcpu->arch.arch_capabilities = kvm_get_arch_capabilities();
arch/x86/kvm/x86.c
12888
vcpu->arch.msr_platform_info = MSR_PLATFORM_INFO_CPUID_FAULT;
arch/x86/kvm/x86.c
12889
vcpu->arch.perf_capabilities = kvm_caps.supported_perf_cap;
arch/x86/kvm/x86.c
12893
vcpu->arch.pending_external_vector = -1;
arch/x86/kvm/x86.c
12894
vcpu->arch.preempted_in_kernel = false;
arch/x86/kvm/x86.c
12897
vcpu->arch.hv_root_tdp = INVALID_PAGE;
arch/x86/kvm/x86.c
12907
kvm_set_tsc_khz(vcpu, vcpu->kvm->arch.default_tsc_khz);
arch/x86/kvm/x86.c
12914
fpu_free_guest_fpstate(&vcpu->arch.guest_fpu);
arch/x86/kvm/x86.c
12916
kmem_cache_free(x86_emulator_cache, vcpu->arch.emulate_ctxt);
arch/x86/kvm/x86.c
12918
free_cpumask_var(vcpu->arch.wbinvd_dirty_mask);
arch/x86/kvm/x86.c
12920
kfree(vcpu->arch.mce_banks);
arch/x86/kvm/x86.c
12921
kfree(vcpu->arch.mci_ctl2_banks);
arch/x86/kvm/x86.c
12922
free_page((unsigned long)vcpu->arch.pio_data);
arch/x86/kvm/x86.c
12939
vcpu->arch.msr_kvm_poll_control = 1;
arch/x86/kvm/x86.c
12958
kmem_cache_free(x86_emulator_cache, vcpu->arch.emulate_ctxt);
arch/x86/kvm/x86.c
12959
free_cpumask_var(vcpu->arch.wbinvd_dirty_mask);
arch/x86/kvm/x86.c
12960
fpu_free_guest_fpstate(&vcpu->arch.guest_fpu);
arch/x86/kvm/x86.c
12965
kfree(vcpu->arch.mce_banks);
arch/x86/kvm/x86.c
12966
kfree(vcpu->arch.mci_ctl2_banks);
arch/x86/kvm/x86.c
12971
free_page((unsigned long)vcpu->arch.pio_data);
arch/x86/kvm/x86.c
12972
kvfree(vcpu->arch.cpuid_entries);
arch/x86/kvm/x86.c
12977
struct fpstate *fpstate = vcpu->arch.guest_fpu.fpstate;
arch/x86/kvm/x86.c
1304
vcpu->arch.xcr0 = xcr0;
arch/x86/kvm/x86.c
13050
vcpu->arch.hflags = 0;
arch/x86/kvm/x86.c
13052
vcpu->arch.smi_pending = 0;
arch/x86/kvm/x86.c
13053
vcpu->arch.smi_count = 0;
arch/x86/kvm/x86.c
13054
atomic_set(&vcpu->arch.nmi_queued, 0);
arch/x86/kvm/x86.c
13055
vcpu->arch.nmi_pending = 0;
arch/x86/kvm/x86.c
13056
vcpu->arch.nmi_injected = false;
arch/x86/kvm/x86.c
13060
memset(vcpu->arch.db, 0, sizeof(vcpu->arch.db));
arch/x86/kvm/x86.c
13062
vcpu->arch.dr6 = DR6_ACTIVE_LOW;
arch/x86/kvm/x86.c
13063
vcpu->arch.dr7 = DR7_FIXED_1;
arch/x86/kvm/x86.c
13066
vcpu->arch.cr2 = 0;
arch/x86/kvm/x86.c
13069
vcpu->arch.apf.msr_en_val = 0;
arch/x86/kvm/x86.c
1307
vcpu->arch.cpuid_dynamic_bits_dirty = true;
arch/x86/kvm/x86.c
13070
vcpu->arch.apf.msr_int_val = 0;
arch/x86/kvm/x86.c
13071
vcpu->arch.st.msr_val = 0;
arch/x86/kvm/x86.c
13077
vcpu->arch.apf.halted = false;
arch/x86/kvm/x86.c
13082
vcpu->arch.smbase = 0x30000;
arch/x86/kvm/x86.c
13084
vcpu->arch.pat = MSR_IA32_CR_PAT_DEFAULT;
arch/x86/kvm/x86.c
13086
vcpu->arch.msr_misc_features_enables = 0;
arch/x86/kvm/x86.c
13087
vcpu->arch.ia32_misc_enable_msr = MSR_IA32_MISC_ENABLE_PEBS_UNAVAIL |
arch/x86/kvm/x86.c
13095
memset(vcpu->arch.regs, 0, sizeof(vcpu->arch.regs));
arch/x86/kvm/x86.c
13113
vcpu->arch.cr3 = 0;
arch/x86/kvm/x86.c
13207
if (stable && vcpu->arch.last_host_tsc > local_tsc) {
arch/x86/kvm/x86.c
13209
if (vcpu->arch.last_host_tsc > max_tsc)
arch/x86/kvm/x86.c
13210
max_tsc = vcpu->arch.last_host_tsc;
arch/x86/kvm/x86.c
13256
kvm->arch.backwards_tsc_observed = true;
arch/x86/kvm/x86.c
13258
vcpu->arch.tsc_offset_adjustment += delta_cyc;
arch/x86/kvm/x86.c
13259
vcpu->arch.last_host_tsc = local_tsc;
arch/x86/kvm/x86.c
13269
kvm->arch.last_tsc_nsec = 0;
arch/x86/kvm/x86.c
13270
kvm->arch.last_tsc_write = 0;
arch/x86/kvm/x86.c
13318
return vcpu->kvm->arch.bsp_vcpu_id == vcpu->vcpu_id;
arch/x86/kvm/x86.c
13324
return (vcpu->arch.apic_base & MSR_IA32_APICBASE_BSP) != 0;
arch/x86/kvm/x86.c
13330
kfree(kvm->arch.hv_pa_pg);
arch/x86/kvm/x86.c
13344
kvm->arch.vm_type = type;
arch/x86/kvm/x86.c
13345
kvm->arch.has_private_mem =
arch/x86/kvm/x86.c
13348
kvm->arch.pre_fault_allowed =
arch/x86/kvm/x86.c
13350
kvm->arch.disabled_quirks = kvm_caps.inapplicable_quirks & kvm_caps.supported_quirks;
arch/x86/kvm/x86.c
13364
atomic_set(&kvm->arch.noncoherent_dma_count, 0);
arch/x86/kvm/x86.c
13366
raw_spin_lock_init(&kvm->arch.tsc_write_lock);
arch/x86/kvm/x86.c
13367
mutex_init(&kvm->arch.apic_map_lock);
arch/x86/kvm/x86.c
13368
seqcount_raw_spinlock_init(&kvm->arch.pvclock_sc, &kvm->arch.tsc_write_lock);
arch/x86/kvm/x86.c
13369
kvm->arch.kvmclock_offset = -get_kvmclock_base_ns();
arch/x86/kvm/x86.c
13371
raw_spin_lock_irqsave(&kvm->arch.tsc_write_lock, flags);
arch/x86/kvm/x86.c
13373
raw_spin_unlock_irqrestore(&kvm->arch.tsc_write_lock, flags);
arch/x86/kvm/x86.c
13375
kvm->arch.default_tsc_khz = max_tsc_khz ? : tsc_khz;
arch/x86/kvm/x86.c
13376
kvm->arch.apic_bus_cycle_ns = APIC_BUS_CYCLE_NS_DEFAULT;
arch/x86/kvm/x86.c
13377
kvm->arch.guest_can_read_msr_platform_info = true;
arch/x86/kvm/x86.c
13378
kvm->arch.enable_pmu = enable_pmu;
arch/x86/kvm/x86.c
13381
spin_lock_init(&kvm->arch.hv_root_tdp_lock);
arch/x86/kvm/x86.c
13382
kvm->arch.hv_root_tdp = INVALID_PAGE;
arch/x86/kvm/x86.c
13396
once_init(&kvm->arch.nx_once);
arch/x86/kvm/x86.c
13515
if (kvm->arch.created_mediated_pmu)
arch/x86/kvm/x86.c
13518
kvm_free_msr_filter(srcu_dereference_check(kvm->arch.msr_filter, &kvm->srcu, 1));
arch/x86/kvm/x86.c
13523
kvfree(rcu_dereference_check(kvm->arch.apic_map, 1));
arch/x86/kvm/x86.c
13524
kfree(srcu_dereference_check(kvm->arch.pmu_event_filter, &kvm->srcu, 1));
arch/x86/kvm/x86.c
13537
vfree(slot->arch.rmap[i]);
arch/x86/kvm/x86.c
13538
slot->arch.rmap[i] = NULL;
arch/x86/kvm/x86.c
13549
vfree(slot->arch.lpage_info[i - 1]);
arch/x86/kvm/x86.c
13550
slot->arch.lpage_info[i - 1] = NULL;
arch/x86/kvm/x86.c
13558
const int sz = sizeof(*slot->arch.rmap[0]);
arch/x86/kvm/x86.c
13565
if (slot->arch.rmap[i])
arch/x86/kvm/x86.c
13568
slot->arch.rmap[i] = __vcalloc(lpages, sz, GFP_KERNEL_ACCOUNT);
arch/x86/kvm/x86.c
13569
if (!slot->arch.rmap[i]) {
arch/x86/kvm/x86.c
13589
memset(&slot->arch, 0, sizeof(slot->arch));
arch/x86/kvm/x86.c
13609
slot->arch.lpage_info[i - 1] = linfo;
arch/x86/kvm/x86.c
13641
vfree(slot->arch.lpage_info[i - 1]);
arch/x86/kvm/x86.c
13642
slot->arch.lpage_info[i - 1] = NULL;
arch/x86/kvm/x86.c
13686
memcpy(&new->arch, &old->arch, sizeof(old->arch));
arch/x86/kvm/x86.c
13698
if (!kvm->arch.cpu_dirty_log_size)
arch/x86/kvm/x86.c
13770
if (kvm->arch.cpu_dirty_log_size) {
arch/x86/kvm/x86.c
13831
if (!kvm->arch.n_requested_mmu_pages &&
arch/x86/kvm/x86.c
13851
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
13861
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
13880
if (vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
13910
kvm_is_linear_rip(vcpu, vcpu->arch.singlestep_rip))
arch/x86/kvm/x86.c
13938
while (vcpu->arch.apf.gfns[key] != ~0)
arch/x86/kvm/x86.c
13941
vcpu->arch.apf.gfns[key] = gfn;
arch/x86/kvm/x86.c
13950
(vcpu->arch.apf.gfns[key] != gfn &&
arch/x86/kvm/x86.c
13951
vcpu->arch.apf.gfns[key] != ~0); i++)
arch/x86/kvm/x86.c
13959
return vcpu->arch.apf.gfns[kvm_async_pf_gfn_slot(vcpu, gfn)] == gfn;
arch/x86/kvm/x86.c
13968
if (WARN_ON_ONCE(vcpu->arch.apf.gfns[i] != gfn))
arch/x86/kvm/x86.c
13972
vcpu->arch.apf.gfns[i] = ~0;
arch/x86/kvm/x86.c
13975
if (vcpu->arch.apf.gfns[j] == ~0)
arch/x86/kvm/x86.c
13977
k = kvm_async_pf_hash_fn(vcpu->arch.apf.gfns[j]);
arch/x86/kvm/x86.c
13984
vcpu->arch.apf.gfns[i] = vcpu->arch.apf.gfns[j];
arch/x86/kvm/x86.c
13993
return kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.apf.data, &reason,
arch/x86/kvm/x86.c
14001
return kvm_write_guest_offset_cached(vcpu->kvm, &vcpu->arch.apf.data,
arch/x86/kvm/x86.c
14010
if (kvm_read_guest_offset_cached(vcpu->kvm, &vcpu->arch.apf.data,
arch/x86/kvm/x86.c
14023
if (!vcpu->arch.apf.send_always &&
arch/x86/kvm/x86.c
14024
(vcpu->arch.guest_state_protected || !kvm_x86_call(get_cpl)(vcpu)))
arch/x86/kvm/x86.c
14032
return vcpu->arch.apf.delivery_as_pf_vmexit;
arch/x86/kvm/x86.c
14065
trace_kvm_async_pf_not_present(work->arch.token, work->cr2_or_gpa);
arch/x86/kvm/x86.c
14066
kvm_add_async_pf_gfn(vcpu, work->arch.gfn);
arch/x86/kvm/x86.c
14074
fault.address = work->arch.token;
arch/x86/kvm/x86.c
1409
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/x86.c
14097
.vector = vcpu->arch.apf.vec
arch/x86/kvm/x86.c
14101
work->arch.token = ~0; /* broadcast wakeup */
arch/x86/kvm/x86.c
14103
kvm_del_async_pf_gfn(vcpu, work->arch.gfn);
arch/x86/kvm/x86.c
14104
trace_kvm_async_pf_ready(work->arch.token, work->cr2_or_gpa);
arch/x86/kvm/x86.c
14108
!apf_put_user_ready(vcpu, work->arch.token)) {
arch/x86/kvm/x86.c
14109
WRITE_ONCE(vcpu->arch.apf.pageready_pending, true);
arch/x86/kvm/x86.c
14113
vcpu->arch.apf.halted = false;
arch/x86/kvm/x86.c
14124
if (!READ_ONCE(vcpu->arch.apf.pageready_pending))
arch/x86/kvm/x86.c
14153
if (atomic_inc_return(&kvm->arch.noncoherent_dma_count) == 1)
arch/x86/kvm/x86.c
14159
if (!atomic_dec_return(&kvm->arch.noncoherent_dma_count))
arch/x86/kvm/x86.c
14165
return atomic_read(&kvm->arch.noncoherent_dma_count);
arch/x86/kvm/x86.c
14171
return (vcpu->arch.msr_kvm_poll_control & 1) == 0;
arch/x86/kvm/x86.c
14228
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
14247
vcpu->arch.walk_mmu->inject_page_fault(vcpu, &fault);
arch/x86/kvm/x86.c
14394
vcpu->arch.complete_userspace_io = complete_sev_es_emulated_mmio;
arch/x86/kvm/x86.c
14433
vcpu->arch.complete_userspace_io = complete_sev_es_emulated_mmio;
arch/x86/kvm/x86.c
14440
vcpu->arch.sev_pio_count -= count;
arch/x86/kvm/x86.c
14441
vcpu->arch.sev_pio_data += count * size;
arch/x86/kvm/x86.c
14449
int size = vcpu->arch.pio.size;
arch/x86/kvm/x86.c
14450
int port = vcpu->arch.pio.port;
arch/x86/kvm/x86.c
14452
vcpu->arch.pio.count = 0;
arch/x86/kvm/x86.c
14453
if (vcpu->arch.sev_pio_count)
arch/x86/kvm/x86.c
14463
min_t(unsigned int, PAGE_SIZE / size, vcpu->arch.sev_pio_count);
arch/x86/kvm/x86.c
14464
int ret = emulator_pio_out(vcpu, size, port, vcpu->arch.sev_pio_data, count);
arch/x86/kvm/x86.c
14472
if (!vcpu->arch.sev_pio_count)
arch/x86/kvm/x86.c
14476
vcpu->arch.complete_userspace_io = complete_sev_es_emulated_outs;
arch/x86/kvm/x86.c
14485
unsigned count = vcpu->arch.pio.count;
arch/x86/kvm/x86.c
14486
int size = vcpu->arch.pio.size;
arch/x86/kvm/x86.c
14487
int port = vcpu->arch.pio.port;
arch/x86/kvm/x86.c
14489
complete_emulator_pio_in(vcpu, vcpu->arch.sev_pio_data);
arch/x86/kvm/x86.c
14491
if (vcpu->arch.sev_pio_count)
arch/x86/kvm/x86.c
14501
min_t(unsigned int, PAGE_SIZE / size, vcpu->arch.sev_pio_count);
arch/x86/kvm/x86.c
14502
if (!emulator_pio_in(vcpu, size, port, vcpu->arch.sev_pio_data, count))
arch/x86/kvm/x86.c
14507
if (!vcpu->arch.sev_pio_count)
arch/x86/kvm/x86.c
14511
vcpu->arch.complete_userspace_io = complete_sev_es_emulated_ins;
arch/x86/kvm/x86.c
14519
vcpu->arch.sev_pio_data = data;
arch/x86/kvm/x86.c
14520
vcpu->arch.sev_pio_count = count;
arch/x86/kvm/x86.c
1480
vcpu->arch.cr3 = cr3;
arch/x86/kvm/x86.c
1506
vcpu->arch.cr8 = cr8;
arch/x86/kvm/x86.c
1516
return vcpu->arch.cr8;
arch/x86/kvm/x86.c
1526
vcpu->arch.eff_db[i] = vcpu->arch.db[i];
arch/x86/kvm/x86.c
1535
dr7 = vcpu->arch.guest_debug_dr7;
arch/x86/kvm/x86.c
1537
dr7 = vcpu->arch.dr7;
arch/x86/kvm/x86.c
1539
vcpu->arch.switch_db_regs &= ~KVM_DEBUGREG_BP_ENABLED;
arch/x86/kvm/x86.c
1541
vcpu->arch.switch_db_regs |= KVM_DEBUGREG_BP_ENABLED;
arch/x86/kvm/x86.c
1559
size_t size = ARRAY_SIZE(vcpu->arch.db);
arch/x86/kvm/x86.c
1563
vcpu->arch.db[array_index_nospec(dr, size)] = val;
arch/x86/kvm/x86.c
1565
vcpu->arch.eff_db[dr] = val;
arch/x86/kvm/x86.c
1571
vcpu->arch.dr6 = (val & DR6_VOLATILE) | kvm_dr6_fixed(vcpu);
arch/x86/kvm/x86.c
1577
vcpu->arch.dr7 = (val & DR7_VOLATILE) | DR7_FIXED_1;
arch/x86/kvm/x86.c
1588
size_t size = ARRAY_SIZE(vcpu->arch.db);
arch/x86/kvm/x86.c
1592
return vcpu->arch.db[array_index_nospec(dr, size)];
arch/x86/kvm/x86.c
1595
return vcpu->arch.dr6;
arch/x86/kvm/x86.c
1598
return vcpu->arch.dr7;
arch/x86/kvm/x86.c
1760
u64 old_efer = vcpu->arch.efer;
arch/x86/kvm/x86.c
1772
(vcpu->arch.efer & EFER_LME) != (efer & EFER_LME))
arch/x86/kvm/x86.c
1777
efer |= vcpu->arch.efer & EFER_LMA;
arch/x86/kvm/x86.c
1816
msr_filter = srcu_dereference(kvm->arch.msr_filter, &kvm->srcu);
arch/x86/kvm/x86.c
2091
kvm_register_write(vcpu, vcpu->arch.cui_rdmsr_imm_reg,
arch/x86/kvm/x86.c
2117
if (!(vcpu->kvm->arch.user_space_msr_mask & msr_reason))
arch/x86/kvm/x86.c
2126
vcpu->arch.complete_userspace_io = completion;
arch/x86/kvm/x86.c
2168
vcpu->arch.cui_rdmsr_imm_reg = reg;
arch/x86/kvm/x86.c
2250
enabled = vcpu->arch.ia32_misc_enable_msr & MSR_IA32_MISC_ENABLE_MWAIT;
arch/x86/kvm/x86.c
2286
if (!lapic_in_kernel(vcpu) || !apic_x2apic_mode(vcpu->arch.apic) ||
arch/x86/kvm/x86.c
2287
kvm_x2apic_icr_write_fast(vcpu->arch.apic, data))
arch/x86/kvm/x86.c
2455
struct kvm_arch *ka = &vcpu->kvm->arch;
arch/x86/kvm/x86.c
2464
vcpu->arch.time = system_time;
arch/x86/kvm/x86.c
2469
kvm_gpc_activate(&vcpu->arch.pv_time, system_time & ~1ULL,
arch/x86/kvm/x86.c
2472
kvm_gpc_deactivate(&vcpu->arch.pv_time);
arch/x86/kvm/x86.c
2540
vcpu->arch.tsc_catchup = 1;
arch/x86/kvm/x86.c
2541
vcpu->arch.tsc_always_catchup = 1;
arch/x86/kvm/x86.c
2577
&vcpu->arch.virtual_tsc_shift,
arch/x86/kvm/x86.c
2578
&vcpu->arch.virtual_tsc_mult);
arch/x86/kvm/x86.c
2579
vcpu->arch.virtual_tsc_khz = user_tsc_khz;
arch/x86/kvm/x86.c
2599
u64 tsc = pvclock_scale_delta(kernel_ns-vcpu->arch.this_tsc_nsec,
arch/x86/kvm/x86.c
2600
vcpu->arch.virtual_tsc_mult,
arch/x86/kvm/x86.c
2601
vcpu->arch.virtual_tsc_shift);
arch/x86/kvm/x86.c
2602
tsc += vcpu->arch.this_tsc_write;
arch/x86/kvm/x86.c
2616
struct kvm_arch *ka = &vcpu->kvm->arch;
arch/x86/kvm/x86.c
2673
tsc = kvm_scale_tsc(rdtsc(), vcpu->arch.l1_tsc_scaling_ratio);
arch/x86/kvm/x86.c
2680
return vcpu->arch.l1_tsc_offset +
arch/x86/kvm/x86.c
2681
kvm_scale_tsc(host_tsc, vcpu->arch.l1_tsc_scaling_ratio);
arch/x86/kvm/x86.c
2712
if (vcpu->arch.guest_tsc_protected)
arch/x86/kvm/x86.c
2716
vcpu->arch.l1_tsc_offset,
arch/x86/kvm/x86.c
2719
vcpu->arch.l1_tsc_offset = l1_offset;
arch/x86/kvm/x86.c
2727
vcpu->arch.tsc_offset = kvm_calc_nested_tsc_offset(
arch/x86/kvm/x86.c
2732
vcpu->arch.tsc_offset = l1_offset;
arch/x86/kvm/x86.c
2739
vcpu->arch.l1_tsc_scaling_ratio = l1_multiplier;
arch/x86/kvm/x86.c
2743
vcpu->arch.tsc_scaling_ratio = kvm_calc_nested_tsc_multiplier(
arch/x86/kvm/x86.c
2747
vcpu->arch.tsc_scaling_ratio = l1_multiplier;
arch/x86/kvm/x86.c
2776
lockdep_assert_held(&kvm->arch.tsc_write_lock);
arch/x86/kvm/x86.c
2778
if (vcpu->arch.guest_tsc_protected)
arch/x86/kvm/x86.c
2782
vcpu->kvm->arch.user_set_tsc = true;
arch/x86/kvm/x86.c
2788
kvm->arch.last_tsc_nsec = ns;
arch/x86/kvm/x86.c
2789
kvm->arch.last_tsc_write = tsc;
arch/x86/kvm/x86.c
2790
kvm->arch.last_tsc_khz = vcpu->arch.virtual_tsc_khz;
arch/x86/kvm/x86.c
2791
kvm->arch.last_tsc_offset = offset;
arch/x86/kvm/x86.c
2793
vcpu->arch.last_guest_tsc = tsc;
arch/x86/kvm/x86.c
2807
kvm->arch.cur_tsc_generation++;
arch/x86/kvm/x86.c
2808
kvm->arch.cur_tsc_nsec = ns;
arch/x86/kvm/x86.c
2809
kvm->arch.cur_tsc_write = tsc;
arch/x86/kvm/x86.c
2810
kvm->arch.cur_tsc_offset = offset;
arch/x86/kvm/x86.c
2811
kvm->arch.nr_vcpus_matched_tsc = 0;
arch/x86/kvm/x86.c
2812
} else if (vcpu->arch.this_tsc_generation != kvm->arch.cur_tsc_generation) {
arch/x86/kvm/x86.c
2813
kvm->arch.nr_vcpus_matched_tsc++;
arch/x86/kvm/x86.c
2817
vcpu->arch.this_tsc_generation = kvm->arch.cur_tsc_generation;
arch/x86/kvm/x86.c
2818
vcpu->arch.this_tsc_nsec = kvm->arch.cur_tsc_nsec;
arch/x86/kvm/x86.c
2819
vcpu->arch.this_tsc_write = kvm->arch.cur_tsc_write;
arch/x86/kvm/x86.c
2833
raw_spin_lock_irqsave(&kvm->arch.tsc_write_lock, flags);
arch/x86/kvm/x86.c
2836
elapsed = ns - kvm->arch.last_tsc_nsec;
arch/x86/kvm/x86.c
2838
if (vcpu->arch.virtual_tsc_khz) {
arch/x86/kvm/x86.c
2845
} else if (kvm->arch.user_set_tsc) {
arch/x86/kvm/x86.c
2846
u64 tsc_exp = kvm->arch.last_tsc_write +
arch/x86/kvm/x86.c
2848
u64 tsc_hz = vcpu->arch.virtual_tsc_khz * 1000LL;
arch/x86/kvm/x86.c
2877
vcpu->arch.virtual_tsc_khz == kvm->arch.last_tsc_khz) {
arch/x86/kvm/x86.c
2879
offset = kvm->arch.cur_tsc_offset;
arch/x86/kvm/x86.c
2889
raw_spin_unlock_irqrestore(&kvm->arch.tsc_write_lock, flags);
arch/x86/kvm/x86.c
2895
u64 tsc_offset = vcpu->arch.l1_tsc_offset;
arch/x86/kvm/x86.c
2901
if (vcpu->arch.l1_tsc_scaling_ratio != kvm_caps.default_tsc_scaling_ratio)
arch/x86/kvm/x86.c
2904
vcpu->arch.l1_tsc_scaling_ratio);
arch/x86/kvm/x86.c
3123
struct kvm_arch *ka = &kvm->arch;
arch/x86/kvm/x86.c
3127
lockdep_assert_held(&kvm->arch.tsc_write_lock);
arch/x86/kvm/x86.c
3159
raw_spin_lock_irq(&kvm->arch.tsc_write_lock);
arch/x86/kvm/x86.c
3160
write_seqcount_begin(&kvm->arch.pvclock_sc);
arch/x86/kvm/x86.c
3173
struct kvm_arch *ka = &kvm->arch;
arch/x86/kvm/x86.c
3214
struct kvm_arch *ka = &kvm->arch;
arch/x86/kvm/x86.c
3249
struct kvm_arch *ka = &kvm->arch;
arch/x86/kvm/x86.c
3319
struct kvm_vcpu_arch *vcpu = &v->arch;
arch/x86/kvm/x86.c
3320
struct kvm_arch *ka = &v->kvm->arch;
arch/x86/kvm/x86.c
3380
v->arch.l1_tsc_scaling_ratio);
arch/x86/kvm/x86.c
3394
hv_clock.system_time = kernel_ns + v->kvm->arch.kvmclock_offset;
arch/x86/kvm/x86.c
3463
struct kvm_arch *ka = &kvm->arch;
arch/x86/kvm/x86.c
3554
return !!(vcpu->arch.msr_hwcr & BIT_ULL(18));
arch/x86/kvm/x86.c
3561
u64 mcg_cap = vcpu->arch.mcg_cap;
arch/x86/kvm/x86.c
3569
vcpu->arch.mcg_status = data;
arch/x86/kvm/x86.c
3577
vcpu->arch.mcg_ctl = data;
arch/x86/kvm/x86.c
3591
vcpu->arch.mci_ctl2_banks[offset] = data;
arch/x86/kvm/x86.c
3623
vcpu->arch.mce_banks[offset] = data;
arch/x86/kvm/x86.c
3650
vcpu->arch.apf.msr_en_val = data;
arch/x86/kvm/x86.c
3658
if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa,
arch/x86/kvm/x86.c
3662
vcpu->arch.apf.send_always = (data & KVM_ASYNC_PF_SEND_ALWAYS);
arch/x86/kvm/x86.c
3663
vcpu->arch.apf.delivery_as_pf_vmexit = data & KVM_ASYNC_PF_DELIVERY_AS_PF_VMEXIT;
arch/x86/kvm/x86.c
3679
vcpu->arch.apf.msr_int_val = data;
arch/x86/kvm/x86.c
3681
vcpu->arch.apf.vec = data & KVM_ASYNC_PF_VEC_MASK;
arch/x86/kvm/x86.c
3688
kvm_gpc_deactivate(&vcpu->arch.pv_time);
arch/x86/kvm/x86.c
3689
vcpu->arch.time = 0;
arch/x86/kvm/x86.c
3750
struct gfn_to_hva_cache *ghc = &vcpu->arch.st.cache;
arch/x86/kvm/x86.c
3753
gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS;
arch/x86/kvm/x86.c
3762
if (!(vcpu->arch.st.msr_val & KVM_MSR_ENABLED))
arch/x86/kvm/x86.c
3805
vcpu->arch.st.preempted = 0;
arch/x86/kvm/x86.c
3819
vcpu->arch.st.preempted = 0;
arch/x86/kvm/x86.c
3833
vcpu->arch.st.last_steal;
arch/x86/kvm/x86.c
3834
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/x86/kvm/x86.c
3886
KVM_BUG_ON(!vcpu->arch.guest_fpu.fpstate->in_use, vcpu->kvm);
arch/x86/kvm/x86.c
3933
vcpu->arch.microcode_version = data;
arch/x86/kvm/x86.c
3939
vcpu->arch.arch_capabilities = data;
arch/x86/kvm/x86.c
3954
if (vcpu->arch.perf_capabilities == data)
arch/x86/kvm/x86.c
3957
vcpu->arch.perf_capabilities = data;
arch/x86/kvm/x86.c
4019
vcpu->arch.msr_hwcr = data;
arch/x86/kvm/x86.c
4031
vcpu->arch.pat = data;
arch/x86/kvm/x86.c
4046
s64 adj = data - vcpu->arch.ia32_tsc_adjust_msr;
arch/x86/kvm/x86.c
4053
vcpu->arch.ia32_tsc_adjust_msr = data;
arch/x86/kvm/x86.c
4057
u64 old_val = vcpu->arch.ia32_misc_enable_msr;
arch/x86/kvm/x86.c
4073
vcpu->arch.ia32_misc_enable_msr = data;
arch/x86/kvm/x86.c
4074
vcpu->arch.cpuid_dynamic_bits_dirty = true;
arch/x86/kvm/x86.c
4076
vcpu->arch.ia32_misc_enable_msr = data;
arch/x86/kvm/x86.c
4083
vcpu->arch.smbase = data;
arch/x86/kvm/x86.c
4086
vcpu->arch.msr_ia32_power_ctl = data;
arch/x86/kvm/x86.c
4091
} else if (!vcpu->arch.guest_tsc_protected) {
arch/x86/kvm/x86.c
4092
u64 adj = kvm_compute_l1_tsc_offset(vcpu, data) - vcpu->arch.l1_tsc_offset;
arch/x86/kvm/x86.c
4094
vcpu->arch.ia32_tsc_adjust_msr += adj;
arch/x86/kvm/x86.c
4101
if (data & ~vcpu->arch.guest_supported_xss)
arch/x86/kvm/x86.c
4103
if (vcpu->arch.ia32_xss == data)
arch/x86/kvm/x86.c
4105
vcpu->arch.ia32_xss = data;
arch/x86/kvm/x86.c
4106
vcpu->arch.cpuid_dynamic_bits_dirty = true;
arch/x86/kvm/x86.c
4111
vcpu->arch.smi_count = data;
arch/x86/kvm/x86.c
4117
vcpu->kvm->arch.wall_clock = data;
arch/x86/kvm/x86.c
4124
vcpu->kvm->arch.wall_clock = data;
arch/x86/kvm/x86.c
4161
smp_store_mb(vcpu->arch.apf.pageready_pending, false);
arch/x86/kvm/x86.c
4176
vcpu->arch.st.msr_val = data;
arch/x86/kvm/x86.c
4200
vcpu->arch.msr_kvm_poll_control = data;
arch/x86/kvm/x86.c
4252
vcpu->arch.osvw.length = data;
arch/x86/kvm/x86.c
4257
vcpu->arch.osvw.status = data;
arch/x86/kvm/x86.c
4262
vcpu->arch.msr_platform_info = data;
arch/x86/kvm/x86.c
4269
vcpu->arch.msr_misc_features_enables = data;
arch/x86/kvm/x86.c
4280
fpu_update_guest_xfd(&vcpu->arch.guest_fpu, data);
arch/x86/kvm/x86.c
4290
vcpu->arch.guest_fpu.xfd_err = data;
arch/x86/kvm/x86.c
4310
u64 mcg_cap = vcpu->arch.mcg_cap;
arch/x86/kvm/x86.c
4320
data = vcpu->arch.mcg_cap;
arch/x86/kvm/x86.c
4325
data = vcpu->arch.mcg_ctl;
arch/x86/kvm/x86.c
4328
data = vcpu->arch.mcg_status;
arch/x86/kvm/x86.c
4339
data = vcpu->arch.mci_ctl2_banks[offset];
arch/x86/kvm/x86.c
4348
data = vcpu->arch.mce_banks[offset];
arch/x86/kvm/x86.c
4400
msr_info->data = vcpu->arch.microcode_version;
arch/x86/kvm/x86.c
4405
msr_info->data = vcpu->arch.arch_capabilities;
arch/x86/kvm/x86.c
4410
msr_info->data = vcpu->arch.perf_capabilities;
arch/x86/kvm/x86.c
4413
msr_info->data = vcpu->arch.msr_ia32_power_ctl;
arch/x86/kvm/x86.c
4428
offset = vcpu->arch.l1_tsc_offset;
arch/x86/kvm/x86.c
4429
ratio = vcpu->arch.l1_tsc_scaling_ratio;
arch/x86/kvm/x86.c
4431
offset = vcpu->arch.tsc_offset;
arch/x86/kvm/x86.c
4432
ratio = vcpu->arch.tsc_scaling_ratio;
arch/x86/kvm/x86.c
4439
msr_info->data = vcpu->arch.pat;
arch/x86/kvm/x86.c
4463
msr_info->data = vcpu->arch.apic_base;
arch/x86/kvm/x86.c
4471
msr_info->data = (u64)vcpu->arch.ia32_tsc_adjust_msr;
arch/x86/kvm/x86.c
4474
msr_info->data = vcpu->arch.ia32_misc_enable_msr;
arch/x86/kvm/x86.c
4479
msr_info->data = vcpu->arch.smbase;
arch/x86/kvm/x86.c
4482
msr_info->data = vcpu->arch.smi_count;
arch/x86/kvm/x86.c
4491
msr_info->data = vcpu->arch.efer;
arch/x86/kvm/x86.c
4497
msr_info->data = vcpu->kvm->arch.wall_clock;
arch/x86/kvm/x86.c
4503
msr_info->data = vcpu->kvm->arch.wall_clock;
arch/x86/kvm/x86.c
4509
msr_info->data = vcpu->arch.time;
arch/x86/kvm/x86.c
4515
msr_info->data = vcpu->arch.time;
arch/x86/kvm/x86.c
4521
msr_info->data = vcpu->arch.apf.msr_en_val;
arch/x86/kvm/x86.c
4527
msr_info->data = vcpu->arch.apf.msr_int_val;
arch/x86/kvm/x86.c
4539
msr_info->data = vcpu->arch.st.msr_val;
arch/x86/kvm/x86.c
4545
msr_info->data = vcpu->arch.pv_eoi.msr_val;
arch/x86/kvm/x86.c
4551
msr_info->data = vcpu->arch.msr_kvm_poll_control;
arch/x86/kvm/x86.c
4566
msr_info->data = vcpu->arch.ia32_xss;
arch/x86/kvm/x86.c
4611
msr_info->data = vcpu->arch.osvw.length;
arch/x86/kvm/x86.c
4616
msr_info->data = vcpu->arch.osvw.status;
arch/x86/kvm/x86.c
4620
!vcpu->kvm->arch.guest_can_read_msr_platform_info)
arch/x86/kvm/x86.c
4622
msr_info->data = vcpu->arch.msr_platform_info;
arch/x86/kvm/x86.c
4625
msr_info->data = vcpu->arch.msr_misc_features_enables;
arch/x86/kvm/x86.c
4628
msr_info->data = vcpu->arch.msr_hwcr;
arch/x86/kvm/x86.c
4636
msr_info->data = vcpu->arch.guest_fpu.fpstate->xfd;
arch/x86/kvm/x86.c
4643
msr_info->data = vcpu->arch.guest_fpu.xfd_err;
arch/x86/kvm/x86.c
4787
return kvm && kvm->arch.has_protected_state ? 0 : KVM_SYNC_X86_VALID_FIELDS;
arch/x86/kvm/x86.c
5179
cpumask_set_cpu(cpu, vcpu->arch.wbinvd_dirty_mask);
arch/x86/kvm/x86.c
5200
vcpu->arch.host_pkru = read_pkru();
arch/x86/kvm/x86.c
5203
if (unlikely(vcpu->arch.tsc_offset_adjustment)) {
arch/x86/kvm/x86.c
5204
adjust_tsc_offset_host(vcpu, vcpu->arch.tsc_offset_adjustment);
arch/x86/kvm/x86.c
5205
vcpu->arch.tsc_offset_adjustment = 0;
arch/x86/kvm/x86.c
5210
s64 tsc_delta = !vcpu->arch.last_host_tsc ? 0 :
arch/x86/kvm/x86.c
5211
rdtsc() - vcpu->arch.last_host_tsc;
arch/x86/kvm/x86.c
5217
vcpu->arch.last_guest_tsc);
arch/x86/kvm/x86.c
5219
if (!vcpu->arch.guest_tsc_protected)
arch/x86/kvm/x86.c
5220
vcpu->arch.tsc_catchup = 1;
arch/x86/kvm/x86.c
5230
if (!vcpu->kvm->arch.use_master_clock || vcpu->cpu == -1)
arch/x86/kvm/x86.c
5242
struct gfn_to_hva_cache *ghc = &vcpu->arch.st.cache;
arch/x86/kvm/x86.c
5246
gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS;
arch/x86/kvm/x86.c
5255
if (!vcpu->arch.at_instruction_boundary) {
arch/x86/kvm/x86.c
5261
if (!(vcpu->arch.st.msr_val & KVM_MSR_ENABLED))
arch/x86/kvm/x86.c
5264
if (vcpu->arch.st.preempted)
arch/x86/kvm/x86.c
5282
vcpu->arch.st.preempted = KVM_VCPU_PREEMPTED;
arch/x86/kvm/x86.c
5297
vcpu->arch.preempted_in_kernel = vcpu->arch.guest_state_protected ||
arch/x86/kvm/x86.c
5313
vcpu->arch.last_host_tsc = rdtsc();
arch/x86/kvm/x86.c
5319
if (vcpu->arch.apic->guest_apic_protected)
arch/x86/kvm/x86.c
5332
if (vcpu->arch.apic->guest_apic_protected)
arch/x86/kvm/x86.c
5393
if (vcpu->arch.pending_external_vector != -1)
arch/x86/kvm/x86.c
5396
vcpu->arch.pending_external_vector = irq->irq;
arch/x86/kvm/x86.c
5413
vcpu->arch.tpr_access_reporting = !!tac->enabled;
arch/x86/kvm/x86.c
5429
vcpu->arch.mcg_cap = mcg_cap;
arch/x86/kvm/x86.c
5432
vcpu->arch.mcg_ctl = ~(u64)0;
arch/x86/kvm/x86.c
5435
vcpu->arch.mce_banks[bank*4] = ~(u64)0;
arch/x86/kvm/x86.c
5437
vcpu->arch.mci_ctl2_banks[bank] = 0;
arch/x86/kvm/x86.c
5466
u64 mcg_cap = vcpu->arch.mcg_cap;
arch/x86/kvm/x86.c
5471
vcpu->arch.mcg_status = mce->mcg_status;
arch/x86/kvm/x86.c
5474
!(vcpu->arch.mci_ctl2_banks[mce->bank] & MCI_CTL2_CMCI_EN))
arch/x86/kvm/x86.c
5478
kvm_apic_local_deliver(vcpu->arch.apic, APIC_LVTCMCI);
arch/x86/kvm/x86.c
5486
u64 mcg_cap = vcpu->arch.mcg_cap;
arch/x86/kvm/x86.c
5488
u64 *banks = vcpu->arch.mce_banks;
arch/x86/kvm/x86.c
5503
vcpu->arch.mcg_ctl != ~(u64)0)
arch/x86/kvm/x86.c
5512
if ((vcpu->arch.mcg_status & MCG_STATUS_MCIP) ||
arch/x86/kvm/x86.c
5521
vcpu->arch.mcg_status = mce->mcg_status;
arch/x86/kvm/x86.c
5545
if (vcpu->arch.exception_vmexit.pending &&
arch/x86/kvm/x86.c
5546
!vcpu->arch.exception.pending &&
arch/x86/kvm/x86.c
5547
!vcpu->arch.exception.injected)
arch/x86/kvm/x86.c
5548
return &vcpu->arch.exception_vmexit;
arch/x86/kvm/x86.c
5550
return &vcpu->arch.exception;
arch/x86/kvm/x86.c
5570
if (!vcpu->kvm->arch.exception_payload_enabled &&
arch/x86/kvm/x86.c
5605
if (!vcpu->kvm->arch.exception_payload_enabled)
arch/x86/kvm/x86.c
5615
vcpu->arch.interrupt.injected && !vcpu->arch.interrupt.soft;
arch/x86/kvm/x86.c
5616
events->interrupt.nr = vcpu->arch.interrupt.nr;
arch/x86/kvm/x86.c
5619
events->nmi.injected = vcpu->arch.nmi_injected;
arch/x86/kvm/x86.c
5627
events->smi.pending = vcpu->arch.smi_pending;
arch/x86/kvm/x86.c
5629
!!(vcpu->arch.hflags & HF_SMM_INSIDE_NMI_MASK);
arch/x86/kvm/x86.c
5636
if (vcpu->kvm->arch.exception_payload_enabled)
arch/x86/kvm/x86.c
5638
if (vcpu->kvm->arch.triple_fault_event) {
arch/x86/kvm/x86.c
5656
if (!vcpu->kvm->arch.exception_payload_enabled)
arch/x86/kvm/x86.c
5681
vcpu->arch.exception_from_userspace = events->exception.pending;
arch/x86/kvm/x86.c
5683
vcpu->arch.exception_vmexit.pending = false;
arch/x86/kvm/x86.c
5685
vcpu->arch.exception.injected = events->exception.injected;
arch/x86/kvm/x86.c
5686
vcpu->arch.exception.pending = events->exception.pending;
arch/x86/kvm/x86.c
5687
vcpu->arch.exception.vector = events->exception.nr;
arch/x86/kvm/x86.c
5688
vcpu->arch.exception.has_error_code = events->exception.has_error_code;
arch/x86/kvm/x86.c
5689
vcpu->arch.exception.error_code = events->exception.error_code;
arch/x86/kvm/x86.c
5690
vcpu->arch.exception.has_payload = events->exception_has_payload;
arch/x86/kvm/x86.c
5691
vcpu->arch.exception.payload = events->exception_payload;
arch/x86/kvm/x86.c
5693
vcpu->arch.interrupt.injected = events->interrupt.injected;
arch/x86/kvm/x86.c
5694
vcpu->arch.interrupt.nr = events->interrupt.nr;
arch/x86/kvm/x86.c
5695
vcpu->arch.interrupt.soft = events->interrupt.soft;
arch/x86/kvm/x86.c
5700
vcpu->arch.nmi_injected = events->nmi.injected;
arch/x86/kvm/x86.c
5702
vcpu->arch.nmi_pending = 0;
arch/x86/kvm/x86.c
5703
atomic_set(&vcpu->arch.nmi_queued, events->nmi.pending);
arch/x86/kvm/x86.c
5711
vcpu->arch.apic->sipi_vector = events->sipi_vector;
arch/x86/kvm/x86.c
5715
if (!!(vcpu->arch.hflags & HF_SMM_MASK) != events->smi.smm) {
arch/x86/kvm/x86.c
5720
vcpu->arch.smi_pending = events->smi.pending;
arch/x86/kvm/x86.c
5724
vcpu->arch.hflags |= HF_SMM_INSIDE_NMI_MASK;
arch/x86/kvm/x86.c
5726
vcpu->arch.hflags &= ~HF_SMM_INSIDE_NMI_MASK;
arch/x86/kvm/x86.c
5737
set_bit(KVM_APIC_INIT, &vcpu->arch.apic->pending_events);
arch/x86/kvm/x86.c
5739
clear_bit(KVM_APIC_INIT, &vcpu->arch.apic->pending_events);
arch/x86/kvm/x86.c
5744
if (!vcpu->kvm->arch.triple_fault_event)
arch/x86/kvm/x86.c
5762
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
5763
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
5770
BUILD_BUG_ON(ARRAY_SIZE(vcpu->arch.db) != ARRAY_SIZE(dbgregs->db));
arch/x86/kvm/x86.c
5771
for (i = 0; i < ARRAY_SIZE(vcpu->arch.db); i++)
arch/x86/kvm/x86.c
5772
dbgregs->db[i] = vcpu->arch.db[i];
arch/x86/kvm/x86.c
5774
dbgregs->dr6 = vcpu->arch.dr6;
arch/x86/kvm/x86.c
5775
dbgregs->dr7 = vcpu->arch.dr7;
arch/x86/kvm/x86.c
5784
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
5785
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
5796
for (i = 0; i < ARRAY_SIZE(vcpu->arch.db); i++)
arch/x86/kvm/x86.c
5797
vcpu->arch.db[i] = dbgregs->db[i];
arch/x86/kvm/x86.c
5800
vcpu->arch.dr6 = dbgregs->dr6;
arch/x86/kvm/x86.c
5801
vcpu->arch.dr7 = dbgregs->dr7;
arch/x86/kvm/x86.c
5823
u64 supported_xcr0 = vcpu->arch.guest_supported_xcr0 |
arch/x86/kvm/x86.c
5826
if (fpstate_is_confidential(&vcpu->arch.guest_fpu))
arch/x86/kvm/x86.c
5827
return vcpu->kvm->arch.has_protected_state ? -EINVAL : 0;
arch/x86/kvm/x86.c
5829
fpu_copy_guest_fpstate_to_uabi(&vcpu->arch.guest_fpu, state, size,
arch/x86/kvm/x86.c
583
vcpu->arch.apf.gfns[i] = ~0;
arch/x86/kvm/x86.c
5830
supported_xcr0, vcpu->arch.pkru);
arch/x86/kvm/x86.c
5846
if (fpstate_is_confidential(&vcpu->arch.guest_fpu))
arch/x86/kvm/x86.c
5847
return vcpu->kvm->arch.has_protected_state ? -EINVAL : 0;
arch/x86/kvm/x86.c
5854
xstate->xsave.header.xfeatures &= ~vcpu->arch.guest_fpu.fpstate->xfd;
arch/x86/kvm/x86.c
5856
return fpu_copy_uabi_to_guest_fpstate(&vcpu->arch.guest_fpu,
arch/x86/kvm/x86.c
5859
&vcpu->arch.pkru);
arch/x86/kvm/x86.c
5865
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
5866
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
5877
guest_xcrs->xcrs[0].value = vcpu->arch.xcr0;
arch/x86/kvm/x86.c
5886
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
5887
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
5916
if (!vcpu->arch.pv_time.active)
arch/x86/kvm/x86.c
5918
vcpu->arch.pvclock_set_guest_stopped_request = true;
arch/x86/kvm/x86.c
5948
if (put_user(vcpu->arch.l1_tsc_offset, uaddr))
arch/x86/kvm/x86.c
5976
raw_spin_lock_irqsave(&kvm->arch.tsc_write_lock, flags);
arch/x86/kvm/x86.c
5978
matched = (vcpu->arch.virtual_tsc_khz &&
arch/x86/kvm/x86.c
5979
kvm->arch.last_tsc_khz == vcpu->arch.virtual_tsc_khz &&
arch/x86/kvm/x86.c
5980
kvm->arch.last_tsc_offset == offset);
arch/x86/kvm/x86.c
5982
tsc = kvm_scale_tsc(rdtsc(), vcpu->arch.l1_tsc_scaling_ratio) + offset;
arch/x86/kvm/x86.c
5986
raw_spin_unlock_irqrestore(&kvm->arch.tsc_write_lock, flags);
arch/x86/kvm/x86.c
6072
vcpu->arch.pv_cpuid.enforce = cap->args[0];
arch/x86/kvm/x86.c
6428
if (vcpu->arch.guest_fpu.uabi_size > sizeof(struct kvm_xsave))
arch/x86/kvm/x86.c
6447
int size = vcpu->arch.guest_fpu.uabi_size;
arch/x86/kvm/x86.c
6460
int size = vcpu->arch.guest_fpu.uabi_size;
arch/x86/kvm/x86.c
6511
if (vcpu->arch.guest_tsc_protected)
arch/x86/kvm/x86.c
6529
r = vcpu->arch.virtual_tsc_khz;
arch/x86/kvm/x86.c
6636
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
6637
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
6653
if (vcpu->kvm->arch.has_protected_state &&
arch/x86/kvm/x86.c
6654
vcpu->arch.guest_state_protected)
arch/x86/kvm/x86.c
6717
kvm->arch.n_requested_mmu_pages = kvm_nr_mmu_pages;
arch/x86/kvm/x86.c
6735
if (!kvm->arch.cpu_dirty_log_size)
arch/x86/kvm/x86.c
6757
kvm->arch.disabled_quirks |= cap->args[0] & kvm_caps.supported_quirks;
arch/x86/kvm/x86.c
6772
kvm->arch.irqchip_mode = KVM_IRQCHIP_SPLIT;
arch/x86/kvm/x86.c
6773
kvm->arch.nr_reserved_ioapic_pins = cap->args[0];
arch/x86/kvm/x86.c
6794
kvm->arch.x2apic_format = true;
arch/x86/kvm/x86.c
6796
kvm->arch.x2apic_broadcast_quirk_disabled = true;
arch/x86/kvm/x86.c
6799
kvm->arch.suppress_eoi_broadcast_mode = KVM_SUPPRESS_EOI_BROADCAST_ENABLED;
arch/x86/kvm/x86.c
6801
kvm->arch.suppress_eoi_broadcast_mode = KVM_SUPPRESS_EOI_BROADCAST_DISABLED;
arch/x86/kvm/x86.c
6829
kvm->arch.guest_can_read_msr_platform_info = cap->args[0];
arch/x86/kvm/x86.c
6833
kvm->arch.exception_payload_enabled = cap->args[0];
arch/x86/kvm/x86.c
6837
kvm->arch.triple_fault_event = cap->args[0];
arch/x86/kvm/x86.c
6844
kvm->arch.user_space_msr_mask = cap->args[0];
arch/x86/kvm/x86.c
6858
kvm->arch.bus_lock_detection_enabled = true;
arch/x86/kvm/x86.c
6872
kvm->arch.sgx_provisioning_allowed = true;
arch/x86/kvm/x86.c
6897
kvm->arch.hypercall_exit_enabled = cap->args[0];
arch/x86/kvm/x86.c
6904
kvm->arch.exit_on_emulation_error = cap->args[0];
arch/x86/kvm/x86.c
6913
if (!kvm->created_vcpus && !kvm->arch.created_mediated_pmu) {
arch/x86/kvm/x86.c
6914
kvm->arch.enable_pmu = !(cap->args[0] & KVM_PMU_CAP_DISABLE);
arch/x86/kvm/x86.c
6925
if (kvm->arch.bsp_vcpu_id > cap->args[0]) {
arch/x86/kvm/x86.c
6927
} else if (kvm->arch.max_vcpu_ids == cap->args[0]) {
arch/x86/kvm/x86.c
6929
} else if (!kvm->arch.max_vcpu_ids) {
arch/x86/kvm/x86.c
6930
kvm->arch.max_vcpu_ids = cap->args[0];
arch/x86/kvm/x86.c
6945
kvm->arch.notify_window = cap->args[0] >> 32;
arch/x86/kvm/x86.c
6946
kvm->arch.notify_vmexit_flags = (u32)cap->args[0];
arch/x86/kvm/x86.c
6975
kvm->arch.disable_nx_huge_pages = true;
arch/x86/kvm/x86.c
7000
kvm->arch.apic_bus_cycle_ns = bus_cycle_ns;
arch/x86/kvm/x86.c
7102
old_filter = rcu_replace_pointer(kvm->arch.msr_filter, new_filter,
arch/x86/kvm/x86.c
7215
struct kvm_arch *ka = &kvm->arch;
arch/x86/kvm/x86.c
7327
if (kvm->arch.has_protected_eoi)
arch/x86/kvm/x86.c
7352
kvm->arch.irqchip_mode = KVM_IRQCHIP_KERNEL;
arch/x86/kvm/x86.c
7369
if (kvm->arch.vpit)
arch/x86/kvm/x86.c
7375
kvm->arch.vpit = kvm_create_pit(kvm, u.pit_config.flags);
arch/x86/kvm/x86.c
7376
if (kvm->arch.vpit)
arch/x86/kvm/x86.c
7428
if (!kvm->arch.vpit)
arch/x86/kvm/x86.c
7445
if (!kvm->arch.vpit)
arch/x86/kvm/x86.c
7454
if (!kvm->arch.vpit)
arch/x86/kvm/x86.c
7471
if (!kvm->arch.vpit)
arch/x86/kvm/x86.c
7484
if (!kvm->arch.vpit)
arch/x86/kvm/x86.c
7496
(kvm->arch.max_vcpu_ids && arg > kvm->arch.max_vcpu_ids))
arch/x86/kvm/x86.c
7499
kvm->arch.bsp_vcpu_id = arg;
arch/x86/kvm/x86.c
7562
WRITE_ONCE(kvm->arch.default_tsc_khz, user_tsc_khz);
arch/x86/kvm/x86.c
7569
r = READ_ONCE(kvm->arch.default_tsc_khz);
arch/x86/kvm/x86.c
7801
!kvm_iodevice_write(vcpu, &vcpu->arch.apic->dev, addr, n, v))
arch/x86/kvm/x86.c
7821
!kvm_iodevice_read(vcpu, &vcpu->arch.apic->dev,
arch/x86/kvm/x86.c
7853
struct kvm_mmu *mmu = vcpu->arch.mmu;
arch/x86/kvm/x86.c
7868
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
7878
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
788
vcpu->arch.dr6 &= ~DR_TRAP_BITS;
arch/x86/kvm/x86.c
7890
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
7899
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
7932
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
7991
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
805
vcpu->arch.dr6 |= DR6_ACTIVE_LOW;
arch/x86/kvm/x86.c
806
vcpu->arch.dr6 |= ex->payload;
arch/x86/kvm/x86.c
807
vcpu->arch.dr6 ^= ex->payload & DR6_ACTIVE_LOW;
arch/x86/kvm/x86.c
8097
struct kvm_mmu *mmu = vcpu->arch.walk_mmu;
arch/x86/kvm/x86.c
8107
!permission_fault(vcpu, vcpu->arch.walk_mmu,
arch/x86/kvm/x86.c
8108
vcpu->arch.mmio_access, 0, access))) {
arch/x86/kvm/x86.c
8109
*gpa = vcpu->arch.mmio_gfn << PAGE_SHIFT |
arch/x86/kvm/x86.c
815
vcpu->arch.dr6 &= ~BIT(12);
arch/x86/kvm/x86.c
8159
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
818
vcpu->arch.cr2 = ex->payload;
arch/x86/kvm/x86.c
831
struct kvm_queued_exception *ex = &vcpu->arch.exception_vmexit;
arch/x86/kvm/x86.c
8426
WARN_ON_ONCE(vcpu->arch.pio.count);
arch/x86/kvm/x86.c
8451
vcpu->arch.pio.port = port;
arch/x86/kvm/x86.c
8452
vcpu->arch.pio.in = in;
arch/x86/kvm/x86.c
8453
vcpu->arch.pio.count = count;
arch/x86/kvm/x86.c
8454
vcpu->arch.pio.size = size;
arch/x86/kvm/x86.c
8457
memset(vcpu->arch.pio_data, 0, size * count);
arch/x86/kvm/x86.c
8459
memcpy(vcpu->arch.pio_data, data, size * count);
arch/x86/kvm/x86.c
8482
int size = vcpu->arch.pio.size;
arch/x86/kvm/x86.c
8483
unsigned int count = vcpu->arch.pio.count;
arch/x86/kvm/x86.c
8484
memcpy(val, vcpu->arch.pio_data, size * count);
arch/x86/kvm/x86.c
8485
trace_kvm_pio(KVM_PIO_IN, vcpu->arch.pio.port, size, count, vcpu->arch.pio_data);
arch/x86/kvm/x86.c
8486
vcpu->arch.pio.count = 0;
arch/x86/kvm/x86.c
8494
if (vcpu->arch.pio.count) {
arch/x86/kvm/x86.c
8542
cpumask_set_cpu(cpu, vcpu->arch.wbinvd_dirty_mask);
arch/x86/kvm/x86.c
8543
wbinvd_on_cpus_mask(vcpu->arch.wbinvd_dirty_mask);
arch/x86/kvm/x86.c
8545
cpumask_clear(vcpu->arch.wbinvd_dirty_mask);
arch/x86/kvm/x86.c
8592
value = vcpu->arch.cr2;
arch/x86/kvm/x86.c
862
if (!vcpu->arch.exception.pending && !vcpu->arch.exception.injected) {
arch/x86/kvm/x86.c
8621
vcpu->arch.cr2 = val;
arch/x86/kvm/x86.c
864
vcpu->arch.exception.pending = true;
arch/x86/kvm/x86.c
865
vcpu->arch.exception.injected = false;
arch/x86/kvm/x86.c
867
vcpu->arch.exception.has_error_code = has_error;
arch/x86/kvm/x86.c
868
vcpu->arch.exception.vector = nr;
arch/x86/kvm/x86.c
869
vcpu->arch.exception.error_code = error_code;
arch/x86/kvm/x86.c
870
vcpu->arch.exception.has_payload = has_payload;
arch/x86/kvm/x86.c
871
vcpu->arch.exception.payload = payload;
arch/x86/kvm/x86.c
876
prev_nr = vcpu->arch.exception.vector;
arch/x86/kvm/x86.c
8811
emul_to_vcpu(ctxt)->arch.halt_request = 1;
arch/x86/kvm/x86.c
8886
*xcr = emul_to_vcpu(ctxt)->arch.xcr0;
arch/x86/kvm/x86.c
890
vcpu->arch.exception.injected = false;
arch/x86/kvm/x86.c
891
vcpu->arch.exception.pending = false;
arch/x86/kvm/x86.c
8995
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
9018
vcpu->arch.emulate_ctxt = ctxt;
arch/x86/kvm/x86.c
9025
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
9046
vcpu->arch.emulate_regs_need_sync_from_vcpu = false;
arch/x86/kvm/x86.c
9051
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
9123
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
9156
run->internal.data[ndata++] = vcpu->arch.last_vmentry_cpu;
arch/x86/kvm/x86.c
9172
vcpu->run->internal.data[1] = vcpu->arch.last_vmentry_cpu;
arch/x86/kvm/x86.c
9188
if (kvm->arch.exit_on_emulation_error ||
arch/x86/kvm/x86.c
9265
kvm_run->debug.arch.dr6 = DR6_BS | DR6_ACTIVE_LOW;
arch/x86/kvm/x86.c
9266
kvm_run->debug.arch.pc = kvm_get_linear_rip(vcpu);
arch/x86/kvm/x86.c
9267
kvm_run->debug.arch.exception = DB_VECTOR;
arch/x86/kvm/x86.c
9339
(vcpu->arch.guest_debug_dr7 & DR7_BP_EN_MASK)) {
arch/x86/kvm/x86.c
9343
vcpu->arch.guest_debug_dr7,
arch/x86/kvm/x86.c
9344
vcpu->arch.eff_db);
arch/x86/kvm/x86.c
9347
kvm_run->debug.arch.dr6 = dr6 | DR6_ACTIVE_LOW;
arch/x86/kvm/x86.c
9348
kvm_run->debug.arch.pc = eip;
arch/x86/kvm/x86.c
9349
kvm_run->debug.arch.exception = DB_VECTOR;
arch/x86/kvm/x86.c
9356
if (unlikely(vcpu->arch.dr7 & DR7_BP_EN_MASK) &&
arch/x86/kvm/x86.c
9360
vcpu->arch.dr7,
arch/x86/kvm/x86.c
9361
vcpu->arch.db);
arch/x86/kvm/x86.c
942
vcpu->arch.exception.injected = true;
arch/x86/kvm/x86.c
943
vcpu->arch.exception.has_error_code = has_error_code;
arch/x86/kvm/x86.c
9434
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
944
vcpu->arch.exception.vector = nr;
arch/x86/kvm/x86.c
945
vcpu->arch.exception.error_code = error_code;
arch/x86/kvm/x86.c
9452
struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt;
arch/x86/kvm/x86.c
946
vcpu->arch.exception.has_payload = false;
arch/x86/kvm/x86.c
947
vcpu->arch.exception.payload = 0;
arch/x86/kvm/x86.c
9564
if (vcpu->arch.emulate_regs_need_sync_from_vcpu) {
arch/x86/kvm/x86.c
9565
vcpu->arch.emulate_regs_need_sync_from_vcpu = false;
arch/x86/kvm/x86.c
9575
if (vcpu->arch.mmu->root_role.direct) {
arch/x86/kvm/x86.c
9609
} else if (vcpu->arch.pio.count) {
arch/x86/kvm/x86.c
9610
if (!vcpu->arch.pio.in) {
arch/x86/kvm/x86.c
9612
vcpu->arch.pio.count = 0;
arch/x86/kvm/x86.c
9615
vcpu->arch.complete_userspace_io = complete_emulated_pio;
arch/x86/kvm/x86.c
9624
vcpu->arch.complete_userspace_io = complete_emulated_mmio;
arch/x86/kvm/x86.c
9625
} else if (vcpu->arch.complete_userspace_io) {
arch/x86/kvm/x86.c
9637
vcpu->arch.emulate_regs_need_sync_to_vcpu = false;
arch/x86/kvm/x86.c
9665
vcpu->arch.emulate_regs_need_sync_to_vcpu = true;
arch/x86/kvm/x86.c
9685
vcpu->arch.pio.count = 0;
arch/x86/kvm/x86.c
9691
vcpu->arch.pio.count = 0;
arch/x86/kvm/x86.c
9693
if (unlikely(!kvm_is_linear_rip(vcpu, vcpu->arch.cui_linear_rip)))
arch/x86/kvm/x86.c
9714
vcpu->arch.complete_userspace_io =
arch/x86/kvm/x86.c
9718
vcpu->arch.cui_linear_rip = kvm_get_linear_rip(vcpu);
arch/x86/kvm/x86.c
9719
vcpu->arch.complete_userspace_io = complete_fast_pio_out;
arch/x86/kvm/x86.c
9729
if (KVM_BUG_ON(vcpu->arch.pio.count != 1, vcpu->kvm))
arch/x86/kvm/x86.c
9732
if (unlikely(!kvm_is_linear_rip(vcpu, vcpu->arch.cui_linear_rip))) {
arch/x86/kvm/x86.c
9733
vcpu->arch.pio.count = 0;
arch/x86/kvm/x86.c
9738
val = (vcpu->arch.pio.size < 4) ? kvm_rax_read(vcpu) : 0;
arch/x86/kvm/x86.c
9761
vcpu->arch.cui_linear_rip = kvm_get_linear_rip(vcpu);
arch/x86/kvm/x86.c
9762
vcpu->arch.complete_userspace_io = complete_fast_pio_in;
arch/x86/kvm/x86.c
996
fault_mmu = fault->nested_page_fault ? vcpu->arch.mmu :
arch/x86/kvm/x86.c
997
vcpu->arch.walk_mmu;
arch/x86/kvm/x86.h
188
return vcpu->arch.last_vmentry_cpu == -1 && !is_guest_mode(vcpu);
arch/x86/kvm/x86.h
198
WARN_ON_ONCE(vcpu->arch.nested_run_pending == KVM_NESTED_RUN_PENDING);
arch/x86/kvm/x86.h
203
vcpu->arch.mp_state = mp_state;
arch/x86/kvm/x86.h
205
vcpu->arch.pv.pv_unhalted = false;
arch/x86/kvm/x86.h
210
return vcpu->arch.exception.pending ||
arch/x86/kvm/x86.h
211
vcpu->arch.exception_vmexit.pending ||
arch/x86/kvm/x86.h
217
vcpu->arch.exception.pending = false;
arch/x86/kvm/x86.h
218
vcpu->arch.exception.injected = false;
arch/x86/kvm/x86.h
219
vcpu->arch.exception_vmexit.pending = false;
arch/x86/kvm/x86.h
225
vcpu->arch.interrupt.injected = true;
arch/x86/kvm/x86.h
226
vcpu->arch.interrupt.soft = soft;
arch/x86/kvm/x86.h
227
vcpu->arch.interrupt.nr = vector;
arch/x86/kvm/x86.h
232
vcpu->arch.interrupt.injected = false;
arch/x86/kvm/x86.h
237
return vcpu->arch.exception.injected || vcpu->arch.interrupt.injected ||
arch/x86/kvm/x86.h
238
vcpu->arch.nmi_injected;
arch/x86/kvm/x86.h
254
return !!(vcpu->arch.efer & EFER_LMA);
arch/x86/kvm/x86.h
264
WARN_ON_ONCE(vcpu->arch.guest_state_protected);
arch/x86/kvm/x86.h
279
return vcpu->arch.guest_state_protected || is_64_bit_mode(vcpu);
arch/x86/kvm/x86.h
293
return vcpu->arch.walk_mmu == &vcpu->arch.nested_mmu;
arch/x86/kvm/x86.h
381
vcpu->arch.mmio_gva = mmu_is_nested(vcpu) ? 0 : gva & PAGE_MASK;
arch/x86/kvm/x86.h
382
vcpu->arch.mmio_access = access;
arch/x86/kvm/x86.h
383
vcpu->arch.mmio_gfn = gfn;
arch/x86/kvm/x86.h
384
vcpu->arch.mmio_gen = gen;
arch/x86/kvm/x86.h
389
return vcpu->arch.mmio_gen == kvm_memslots(vcpu->kvm)->generation;
arch/x86/kvm/x86.h
400
if (gva != MMIO_GVA_ANY && vcpu->arch.mmio_gva != (gva & PAGE_MASK))
arch/x86/kvm/x86.h
403
vcpu->arch.mmio_gva = 0;
arch/x86/kvm/x86.h
408
if (vcpu_match_mmio_gen(vcpu) && vcpu->arch.mmio_gva &&
arch/x86/kvm/x86.h
409
vcpu->arch.mmio_gva == (gva & PAGE_MASK))
arch/x86/kvm/x86.h
417
if (vcpu_match_mmio_gen(vcpu) && vcpu->arch.mmio_gfn &&
arch/x86/kvm/x86.h
418
vcpu->arch.mmio_gfn == gpa >> PAGE_SHIFT)
arch/x86/kvm/x86.h
441
return !(kvm->arch.disabled_quirks & quirk);
arch/x86/kvm/x86.h
556
return pvclock_scale_delta(nsec, vcpu->arch.virtual_tsc_mult,
arch/x86/kvm/x86.h
557
vcpu->arch.virtual_tsc_shift);
arch/x86/kvm/x86.h
576
kvm->arch.disabled_exits |= mask;
arch/x86/kvm/x86.h
581
return kvm->arch.disabled_exits & KVM_X86_DISABLE_EXITS_MWAIT;
arch/x86/kvm/x86.h
586
return kvm->arch.disabled_exits & KVM_X86_DISABLE_EXITS_HLT;
arch/x86/kvm/x86.h
591
return kvm->arch.disabled_exits & KVM_X86_DISABLE_EXITS_PAUSE;
arch/x86/kvm/x86.h
596
return kvm->arch.disabled_exits & KVM_X86_DISABLE_EXITS_CSTATE;
arch/x86/kvm/x86.h
601
return kvm->arch.disabled_exits & KVM_X86_DISABLE_EXITS_APERFMPERF;
arch/x86/kvm/x86.h
606
return kvm->arch.notify_vmexit_flags & KVM_X86_NOTIFY_VMEXIT_ENABLED;
arch/x86/kvm/x86.h
612
WRITE_ONCE(vcpu->arch.handling_intr_from_guest, (u8)intr);
arch/x86/kvm/x86.h
617
WRITE_ONCE(vcpu->arch.handling_intr_from_guest, 0);
arch/x86/kvm/x86.h
622
return vcpu->arch.handling_intr_from_guest == KVM_HANDLING_NMI;
arch/x86/kvm/x86.h
692
return !(cr4 & vcpu->arch.cr4_guest_rsvd_bits);
arch/x86/kvm/x86.h
759
return kvm->arch.hypercall_exit_enabled & BIT(hc_nr);
arch/x86/kvm/xen.c
1045
vcpu->arch.xen.current_runstate = data->u.runstate.state;
arch/x86/kvm/xen.c
1046
vcpu->arch.xen.runstate_entry_time =
arch/x86/kvm/xen.c
1048
vcpu->arch.xen.runstate_times[RUNSTATE_running] =
arch/x86/kvm/xen.c
1050
vcpu->arch.xen.runstate_times[RUNSTATE_runnable] =
arch/x86/kvm/xen.c
1052
vcpu->arch.xen.runstate_times[RUNSTATE_blocked] =
arch/x86/kvm/xen.c
1054
vcpu->arch.xen.runstate_times[RUNSTATE_offline] =
arch/x86/kvm/xen.c
1056
vcpu->arch.xen.last_steal = current->sched_info.run_delay;
arch/x86/kvm/xen.c
1081
(vcpu->arch.xen.runstate_entry_time +
arch/x86/kvm/xen.c
1087
vcpu->arch.xen.runstate_entry_time +=
arch/x86/kvm/xen.c
1089
vcpu->arch.xen.runstate_times[RUNSTATE_running] +=
arch/x86/kvm/xen.c
1091
vcpu->arch.xen.runstate_times[RUNSTATE_runnable] +=
arch/x86/kvm/xen.c
1093
vcpu->arch.xen.runstate_times[RUNSTATE_blocked] +=
arch/x86/kvm/xen.c
1095
vcpu->arch.xen.runstate_times[RUNSTATE_offline] +=
arch/x86/kvm/xen.c
110
if (atomic_read(&vcpu->arch.xen.timer_pending) > 0) {
arch/x86/kvm/xen.c
1100
else if (vcpu->arch.xen.runstate_cache.active)
arch/x86/kvm/xen.c
1109
vcpu->arch.xen.vcpu_id = data->u.vcpu_id;
arch/x86/kvm/xen.c
1123
vcpu->arch.xen.timer_virq = data->u.timer.port;
arch/x86/kvm/xen.c
1136
vcpu->arch.xen.upcall_vector = data->u.vector;
arch/x86/kvm/xen.c
1146
mutex_unlock(&vcpu->kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
115
e.port = vcpu->arch.xen.timer_virq;
arch/x86/kvm/xen.c
1154
mutex_lock(&vcpu->kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
1158
if (kvm_gpc_is_gpa_active(&vcpu->arch.xen.vcpu_info_cache))
arch/x86/kvm/xen.c
1159
data->u.gpa = vcpu->arch.xen.vcpu_info_cache.gpa;
arch/x86/kvm/xen.c
1166
if (kvm_gpc_is_hva_active(&vcpu->arch.xen.vcpu_info_cache))
arch/x86/kvm/xen.c
1167
data->u.hva = vcpu->arch.xen.vcpu_info_cache.uhva;
arch/x86/kvm/xen.c
1174
if (vcpu->arch.xen.vcpu_time_info_cache.active)
arch/x86/kvm/xen.c
1175
data->u.gpa = vcpu->arch.xen.vcpu_time_info_cache.gpa;
arch/x86/kvm/xen.c
1186
if (vcpu->arch.xen.runstate_cache.active) {
arch/x86/kvm/xen.c
1187
data->u.gpa = vcpu->arch.xen.runstate_cache.gpa;
arch/x86/kvm/xen.c
1197
data->u.runstate.state = vcpu->arch.xen.current_runstate;
arch/x86/kvm/xen.c
120
vcpu->arch.xen.timer_expires = 0;
arch/x86/kvm/xen.c
1206
data->u.runstate.state = vcpu->arch.xen.current_runstate;
arch/x86/kvm/xen.c
1208
vcpu->arch.xen.runstate_entry_time;
arch/x86/kvm/xen.c
121
atomic_set(&vcpu->arch.xen.timer_pending, 0);
arch/x86/kvm/xen.c
1210
vcpu->arch.xen.runstate_times[RUNSTATE_running];
arch/x86/kvm/xen.c
1212
vcpu->arch.xen.runstate_times[RUNSTATE_runnable];
arch/x86/kvm/xen.c
1214
vcpu->arch.xen.runstate_times[RUNSTATE_blocked];
arch/x86/kvm/xen.c
1216
vcpu->arch.xen.runstate_times[RUNSTATE_offline];
arch/x86/kvm/xen.c
1225
data->u.vcpu_id = vcpu->arch.xen.vcpu_id;
arch/x86/kvm/xen.c
1239
if (vcpu->arch.xen.timer_expires) {
arch/x86/kvm/xen.c
1240
hrtimer_cancel(&vcpu->arch.xen.timer);
arch/x86/kvm/xen.c
1244
data->u.timer.port = vcpu->arch.xen.timer_virq;
arch/x86/kvm/xen.c
1246
data->u.timer.expires_ns = vcpu->arch.xen.timer_expires;
arch/x86/kvm/xen.c
1256
if (vcpu->arch.xen.timer_expires)
arch/x86/kvm/xen.c
1257
hrtimer_start_expires(&vcpu->arch.xen.timer,
arch/x86/kvm/xen.c
1264
data->u.vector = vcpu->arch.xen.upcall_vector;
arch/x86/kvm/xen.c
1272
mutex_unlock(&vcpu->kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
128
arch.xen.timer);
arch/x86/kvm/xen.c
1284
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
1285
if (kvm->arch.xen.long_mode != lm) {
arch/x86/kvm/xen.c
1286
kvm->arch.xen.long_mode = lm;
arch/x86/kvm/xen.c
1292
if (kvm->arch.xen.shinfo_cache.active &&
arch/x86/kvm/xen.c
1296
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
132
if (atomic_read(&vcpu->arch.xen.timer_pending))
arch/x86/kvm/xen.c
1338
hva_t blob_addr = lm ? kvm->arch.xen.hvm_config.blob_addr_64
arch/x86/kvm/xen.c
1339
: kvm->arch.xen.hvm_config.blob_addr_32;
arch/x86/kvm/xen.c
1340
u8 blob_size = lm ? kvm->arch.xen.hvm_config.blob_size_64
arch/x86/kvm/xen.c
1341
: kvm->arch.xen.hvm_config.blob_size_32;
arch/x86/kvm/xen.c
137
e.port = vcpu->arch.xen.timer_virq;
arch/x86/kvm/xen.c
1391
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
1393
if (xhc->msr && !kvm->arch.xen.hvm_config.msr)
arch/x86/kvm/xen.c
1395
else if (!xhc->msr && kvm->arch.xen.hvm_config.msr)
arch/x86/kvm/xen.c
1398
old_flags = kvm->arch.xen.hvm_config.flags;
arch/x86/kvm/xen.c
1399
memcpy(&kvm->arch.xen.hvm_config, xhc, sizeof(*xhc));
arch/x86/kvm/xen.c
1401
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
1419
if (unlikely(!kvm_is_linear_rip(vcpu, vcpu->arch.xen.hypercall_rip)))
arch/x86/kvm/xen.c
142
vcpu->arch.xen.timer_expires = 0;
arch/x86/kvm/xen.c
1427
if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode)
arch/x86/kvm/xen.c
1437
struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache;
arch/x86/kvm/xen.c
1449
if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) {
arch/x86/kvm/xen.c
146
atomic_inc(&vcpu->arch.xen.timer_pending);
arch/x86/kvm/xen.c
1480
!(vcpu->kvm->arch.xen.hvm_config.flags & KVM_XEN_HVM_CONFIG_EVTCHN_SEND))
arch/x86/kvm/xen.c
1539
vcpu->arch.xen.poll_evtchn = port;
arch/x86/kvm/xen.c
1541
vcpu->arch.xen.poll_evtchn = -1;
arch/x86/kvm/xen.c
1543
set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask);
arch/x86/kvm/xen.c
1549
mod_timer(&vcpu->arch.xen.poll_timer,
arch/x86/kvm/xen.c
1555
timer_delete(&vcpu->arch.xen.poll_timer);
arch/x86/kvm/xen.c
1560
vcpu->arch.xen.poll_evtchn = 0;
arch/x86/kvm/xen.c
1564
clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask);
arch/x86/kvm/xen.c
1574
arch.xen.poll_timer);
arch/x86/kvm/xen.c
1615
if (vcpu->arch.xen.vcpu_id != vcpu_id) {
arch/x86/kvm/xen.c
1647
if (vcpu->arch.xen.vcpu_id != vcpu_id) {
arch/x86/kvm/xen.c
1720
if (params[0] == XENVER_version && vcpu->kvm->arch.xen.xen_version) {
arch/x86/kvm/xen.c
1721
r = vcpu->kvm->arch.xen.xen_version;
arch/x86/kvm/xen.c
1764
vcpu->arch.xen.hypercall_rip = kvm_get_linear_rip(vcpu);
arch/x86/kvm/xen.c
1765
vcpu->arch.complete_userspace_io =
arch/x86/kvm/xen.c
1773
int poll_evtchn = vcpu->arch.xen.poll_evtchn;
arch/x86/kvm/xen.c
1776
test_and_clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask)) {
arch/x86/kvm/xen.c
179
if (hv_clock->tsc_shift != vcpu->arch.pvclock_tsc_shift ||
arch/x86/kvm/xen.c
1794
struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache;
arch/x86/kvm/xen.c
180
hv_clock->tsc_to_system_mul != vcpu->arch.pvclock_tsc_mul)
arch/x86/kvm/xen.c
1823
if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) {
arch/x86/kvm/xen.c
1851
gpc = &vcpu->arch.xen.vcpu_info_cache;
arch/x86/kvm/xen.c
1859
if (!test_and_set_bit(port_word_bit, &vcpu->arch.xen.evtchn_pending_sel))
arch/x86/kvm/xen.c
1864
if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) {
arch/x86/kvm/xen.c
1880
if (kick_vcpu && vcpu->arch.xen.upcall_vector) {
arch/x86/kvm/xen.c
189
struct kvm_vcpu_xen *xen = &vcpu->arch.xen;
arch/x86/kvm/xen.c
1934
struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache;
arch/x86/kvm/xen.c
2071
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2072
evtchnfd = idr_find(&kvm->arch.xen.evtchn_ports, port);
arch/x86/kvm/xen.c
2102
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
216
!vcpu->kvm->arch.use_master_clock)
arch/x86/kvm/xen.c
2165
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2166
ret = idr_alloc(&kvm->arch.xen.evtchn_ports, evtchnfd, port, port + 1,
arch/x86/kvm/xen.c
2168
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2186
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2187
evtchnfd = idr_remove(&kvm->arch.xen.evtchn_ports, port);
arch/x86/kvm/xen.c
2188
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2206
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2213
idr_for_each_entry(&kvm->arch.xen.evtchn_ports, evtchnfd, i)
arch/x86/kvm/xen.c
2218
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2223
idr_for_each_entry(&kvm->arch.xen.evtchn_ports, evtchnfd, i) {
arch/x86/kvm/xen.c
2225
idr_remove(&kvm->arch.xen.evtchn_ports, evtchnfd->send_port);
arch/x86/kvm/xen.c
2227
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2280
evtchnfd = idr_find(&vcpu->kvm->arch.xen.evtchn_ports, send.port);
arch/x86/kvm/xen.c
2299
vcpu->arch.xen.vcpu_id = vcpu->vcpu_idx;
arch/x86/kvm/xen.c
2300
vcpu->arch.xen.poll_evtchn = 0;
arch/x86/kvm/xen.c
2302
timer_setup(&vcpu->arch.xen.poll_timer, cancel_evtchn_poll, 0);
arch/x86/kvm/xen.c
2303
hrtimer_setup(&vcpu->arch.xen.timer, xen_timer_callback, CLOCK_MONOTONIC,
arch/x86/kvm/xen.c
2306
kvm_gpc_init(&vcpu->arch.xen.runstate_cache, vcpu->kvm);
arch/x86/kvm/xen.c
2307
kvm_gpc_init(&vcpu->arch.xen.runstate2_cache, vcpu->kvm);
arch/x86/kvm/xen.c
2308
kvm_gpc_init(&vcpu->arch.xen.vcpu_info_cache, vcpu->kvm);
arch/x86/kvm/xen.c
2309
kvm_gpc_init(&vcpu->arch.xen.vcpu_time_info_cache, vcpu->kvm);
arch/x86/kvm/xen.c
2317
kvm_gpc_deactivate(&vcpu->arch.xen.runstate_cache);
arch/x86/kvm/xen.c
2318
kvm_gpc_deactivate(&vcpu->arch.xen.runstate2_cache);
arch/x86/kvm/xen.c
2319
kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_info_cache);
arch/x86/kvm/xen.c
2320
kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_time_info_cache);
arch/x86/kvm/xen.c
2322
timer_delete_sync(&vcpu->arch.xen.poll_timer);
arch/x86/kvm/xen.c
2327
mutex_init(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
2328
idr_init(&kvm->arch.xen.evtchn_ports);
arch/x86/kvm/xen.c
2329
kvm_gpc_init(&kvm->arch.xen.shinfo_cache, kvm);
arch/x86/kvm/xen.c
2337
kvm_gpc_deactivate(&kvm->arch.xen.shinfo_cache);
arch/x86/kvm/xen.c
2339
idr_for_each_entry(&kvm->arch.xen.evtchn_ports, evtchnfd, i) {
arch/x86/kvm/xen.c
2344
idr_destroy(&kvm->arch.xen.evtchn_ports);
arch/x86/kvm/xen.c
2346
if (kvm->arch.xen.hvm_config.msr)
arch/x86/kvm/xen.c
300
if (vcpu->arch.xen.timer_expires)
arch/x86/kvm/xen.c
301
hrtimer_cancel(&vcpu->arch.xen.timer);
arch/x86/kvm/xen.c
303
atomic_set(&vcpu->arch.xen.timer_pending, 0);
arch/x86/kvm/xen.c
304
vcpu->arch.xen.timer_expires = guest_abs;
arch/x86/kvm/xen.c
307
xen_timer_callback(&vcpu->arch.xen.timer);
arch/x86/kvm/xen.c
309
hrtimer_start(&vcpu->arch.xen.timer,
arch/x86/kvm/xen.c
316
hrtimer_cancel(&vcpu->arch.xen.timer);
arch/x86/kvm/xen.c
317
vcpu->arch.xen.timer_expires = 0;
arch/x86/kvm/xen.c
318
atomic_set(&vcpu->arch.xen.timer_pending, 0);
arch/x86/kvm/xen.c
323
struct kvm_vcpu_xen *vx = &v->arch.xen;
arch/x86/kvm/xen.c
393
if (IS_ENABLED(CONFIG_64BIT) && v->kvm->arch.xen.long_mode) {
arch/x86/kvm/xen.c
40
struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache;
arch/x86/kvm/xen.c
457
if (v->kvm->arch.xen.runstate_update_flag)
arch/x86/kvm/xen.c
521
if (v->kvm->arch.xen.runstate_update_flag) {
arch/x86/kvm/xen.c
589
struct kvm_vcpu_xen *vx = &v->arch.xen;
arch/x86/kvm/xen.c
623
irq.vector = v->arch.xen.upcall_vector;
arch/x86/kvm/xen.c
641
unsigned long evtchn_pending_sel = READ_ONCE(v->arch.xen.evtchn_pending_sel);
arch/x86/kvm/xen.c
642
struct gfn_to_pfn_cache *gpc = &v->arch.xen.vcpu_info_cache;
arch/x86/kvm/xen.c
664
if (IS_ENABLED(CONFIG_64BIT) && v->kvm->arch.xen.long_mode) {
arch/x86/kvm/xen.c
67
BUILD_BUG_ON(offsetof(struct compat_shared_info, arch.wc_sec_hi) != 0x924);
arch/x86/kvm/xen.c
672
"+m" (v->arch.xen.evtchn_pending_sel)
arch/x86/kvm/xen.c
684
"+m" (v->arch.xen.evtchn_pending_sel)
arch/x86/kvm/xen.c
693
if (v->arch.xen.upcall_vector)
arch/x86/kvm/xen.c
699
struct gfn_to_pfn_cache *gpc = &v->arch.xen.vcpu_info_cache;
arch/x86/kvm/xen.c
75
if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) {
arch/x86/kvm/xen.c
756
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
757
kvm->arch.xen.long_mode = !!data->u.long_mode;
arch/x86/kvm/xen.c
765
r = kvm->arch.xen.shinfo_cache.active ?
arch/x86/kvm/xen.c
767
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
775
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
783
kvm_gpc_deactivate(&kvm->arch.xen.shinfo_cache);
arch/x86/kvm/xen.c
786
r = kvm_gpc_activate(&kvm->arch.xen.shinfo_cache,
arch/x86/kvm/xen.c
795
kvm_gpc_deactivate(&kvm->arch.xen.shinfo_cache);
arch/x86/kvm/xen.c
798
r = kvm_gpc_activate_hva(&kvm->arch.xen.shinfo_cache,
arch/x86/kvm/xen.c
805
if (!r && kvm->arch.xen.shinfo_cache.active)
arch/x86/kvm/xen.c
808
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
815
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
816
kvm->arch.xen.upcall_vector = data->u.vector;
arch/x86/kvm/xen.c
817
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
827
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
828
kvm->arch.xen.xen_version = data->u.xen_version;
arch/x86/kvm/xen.c
829
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
838
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
839
kvm->arch.xen.runstate_update_flag = !!data->u.runstate_update_flag;
arch/x86/kvm/xen.c
840
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
85
wc_sec_hi = &shinfo->arch.wc_sec_hi;
arch/x86/kvm/xen.c
855
mutex_lock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
859
data->u.long_mode = kvm->arch.xen.long_mode;
arch/x86/kvm/xen.c
864
if (kvm_gpc_is_gpa_active(&kvm->arch.xen.shinfo_cache))
arch/x86/kvm/xen.c
865
data->u.shared_info.gfn = gpa_to_gfn(kvm->arch.xen.shinfo_cache.gpa);
arch/x86/kvm/xen.c
872
if (kvm_gpc_is_hva_active(&kvm->arch.xen.shinfo_cache))
arch/x86/kvm/xen.c
873
data->u.shared_info.hva = kvm->arch.xen.shinfo_cache.uhva;
arch/x86/kvm/xen.c
880
data->u.vector = kvm->arch.xen.upcall_vector;
arch/x86/kvm/xen.c
885
data->u.xen_version = kvm->arch.xen.xen_version;
arch/x86/kvm/xen.c
894
data->u.runstate_update_flag = kvm->arch.xen.runstate_update_flag;
arch/x86/kvm/xen.c
902
mutex_unlock(&kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
910
mutex_lock(&vcpu->kvm->arch.xen.xen_lock);
arch/x86/kvm/xen.c
924
kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_info_cache);
arch/x86/kvm/xen.c
929
r = kvm_gpc_activate(&vcpu->arch.xen.vcpu_info_cache,
arch/x86/kvm/xen.c
933
kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_info_cache);
arch/x86/kvm/xen.c
938
r = kvm_gpc_activate_hva(&vcpu->arch.xen.vcpu_info_cache,
arch/x86/kvm/xen.c
949
kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_time_info_cache);
arch/x86/kvm/xen.c
954
r = kvm_gpc_activate(&vcpu->arch.xen.vcpu_time_info_cache,
arch/x86/kvm/xen.c
971
kvm_gpc_deactivate(&vcpu->arch.xen.runstate_cache);
arch/x86/kvm/xen.c
972
kvm_gpc_deactivate(&vcpu->arch.xen.runstate2_cache);
arch/x86/kvm/xen.c
981
if (IS_ENABLED(CONFIG_64BIT) && vcpu->kvm->arch.xen.long_mode)
arch/x86/kvm/xen.c
988
r = kvm_gpc_activate(&vcpu->arch.xen.runstate_cache,
arch/x86/kvm/xen.c
995
kvm_gpc_deactivate(&vcpu->arch.xen.runstate2_cache);
arch/x86/kvm/xen.c
999
r = kvm_gpc_activate(&vcpu->arch.xen.runstate2_cache,
arch/x86/kvm/xen.h
100
return !!vcpu->arch.xen.timer_virq;
arch/x86/kvm/xen.h
106
return atomic_read(&vcpu->arch.xen.timer_pending);
arch/x86/kvm/xen.h
226
struct compat_arch_vcpu_info arch;
arch/x86/kvm/xen.h
245
struct compat_arch_shared_info arch;
arch/x86/kvm/xen.h
48
vcpu->arch.xen.vcpu_info_cache.active &&
arch/x86/kvm/xen.h
49
vcpu->arch.xen.upcall_vector && __kvm_xen_has_interrupt(vcpu))
arch/x86/kvm/xen.h
56
vcpu->arch.xen.cpuid.base &&
arch/x86/kvm/xen.h
57
function <= vcpu->arch.xen.cpuid.limit &&
arch/x86/kvm/xen.h
58
function == (vcpu->arch.xen.cpuid.base | XEN_CPUID_LEAF(3));
arch/x86/kvm/xen.h
64
kvm->arch.xen.hvm_config.msr;
arch/x86/kvm/xen.h
72
return msr && msr == kvm->arch.xen.hvm_config.msr;
arch/x86/kvm/xen.h
78
(kvm->arch.xen.hvm_config.flags &
arch/x86/kvm/xen.h
85
vcpu->arch.xen.vcpu_info_cache.active &&
arch/x86/kvm/xen.h
86
vcpu->kvm->arch.xen.upcall_vector)
arch/x86/kvm/xen.h
95
vcpu->arch.xen.evtchn_pending_sel;
arch/x86/um/ptrace_32.c
113
child->thread.arch.debugregs[addr] = data;
arch/x86/um/ptrace_32.c
167
tmp = child->thread.arch.debugregs[addr];
arch/x86/um/ptrace_64.c
125
child->thread.arch.debugregs[addr] = data;
arch/x86/um/ptrace_64.c
187
tmp = child->thread.arch.debugregs[addr];
arch/x86/um/shared/sysdep/stub_32.h
136
stub_seccomp_restore_state(struct stub_data_arch *arch)
arch/x86/um/shared/sysdep/stub_32.h
138
for (int i = 0; i < sizeof(arch->tls) / sizeof(arch->tls[0]); i++) {
arch/x86/um/shared/sysdep/stub_32.h
139
if (arch->sync & (1 << i))
arch/x86/um/shared/sysdep/stub_32.h
141
(unsigned long) &arch->tls[i]);
arch/x86/um/shared/sysdep/stub_32.h
144
arch->sync = 0;
arch/x86/um/shared/sysdep/stub_64.h
140
stub_seccomp_restore_state(struct stub_data_arch *arch)
arch/x86/um/shared/sysdep/stub_64.h
146
if (arch->sync & STUB_SYNC_FS_BASE)
arch/x86/um/shared/sysdep/stub_64.h
147
stub_syscall2(__NR_arch_prctl, ARCH_SET_FS, arch->fs_base);
arch/x86/um/shared/sysdep/stub_64.h
148
if (arch->sync & STUB_SYNC_GS_BASE)
arch/x86/um/shared/sysdep/stub_64.h
149
stub_syscall2(__NR_arch_prctl, ARCH_SET_GS, arch->gs_base);
arch/x86/um/shared/sysdep/stub_64.h
151
arch->sync = 0;
arch/x86/um/signal.c
115
struct faultinfo * fi = &current->thread.arch.faultinfo;
arch/x86/um/tls_32.c
133
&task->thread.arch.tls_array[i - GDT_ENTRY_TLS_MIN];
arch/x86/um/tls_32.c
157
&task->thread.arch.tls_array[i - GDT_ENTRY_TLS_MIN];
arch/x86/um/tls_32.c
205
t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].tls = *info;
arch/x86/um/tls_32.c
206
t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].present = 1;
arch/x86/um/tls_32.c
207
t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].flushed = flushed;
arch/x86/um/tls_32.c
239
if (!t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].present)
arch/x86/um/tls_32.c
242
*info = t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].tls;
arch/x86/um/tls_32.c
250
!t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].flushed)) {
arch/x86/um/tls_32.c
66
if (!t->arch.tls_array[idx].present)
arch/x86/um/tls_32.c
93
&to->thread.arch.tls_array[idx - GDT_ENTRY_TLS_MIN];
arch/x86/xen/enlighten_pv.c
1247
&HYPERVISOR_shared_info->arch.nmi_reason))
arch/x86/xen/enlighten_pv.c
1250
&HYPERVISOR_shared_info->arch.nmi_reason))
arch/x86/xen/mmu_pv.c
1288
this_cpu_read(xen_vcpu)->arch.cr2 = cr2;
arch/x86/xen/p2m.c
273
HYPERVISOR_shared_info->arch.pfn_to_mfn_frame_list_list = ~0UL;
arch/x86/xen/p2m.c
275
HYPERVISOR_shared_info->arch.pfn_to_mfn_frame_list_list =
arch/x86/xen/p2m.c
277
HYPERVISOR_shared_info->arch.max_pfn = xen_p2m_last_pfn;
arch/x86/xen/p2m.c
278
HYPERVISOR_shared_info->arch.p2m_generation = 0;
arch/x86/xen/p2m.c
279
HYPERVISOR_shared_info->arch.p2m_vaddr = (unsigned long)xen_p2m_addr;
arch/x86/xen/p2m.c
280
HYPERVISOR_shared_info->arch.p2m_cr3 =
arch/x86/xen/p2m.c
492
HYPERVISOR_shared_info->arch.p2m_generation++;
arch/x86/xen/p2m.c
497
HYPERVISOR_shared_info->arch.p2m_generation++;
arch/x86/xen/p2m.c
592
HYPERVISOR_shared_info->arch.p2m_generation++;
arch/x86/xen/p2m.c
597
HYPERVISOR_shared_info->arch.p2m_generation++;
arch/x86/xen/p2m.c
612
HYPERVISOR_shared_info->arch.max_pfn = xen_p2m_last_pfn;
drivers/hwtracing/coresight/coresight-etm.h
238
u8 arch;
drivers/hwtracing/coresight/coresight-etm3x-core.c
736
static bool etm_arch_supported(u8 arch)
drivers/hwtracing/coresight/coresight-etm3x-core.c
738
switch (arch) {
drivers/hwtracing/coresight/coresight-etm3x-core.c
781
drvdata->arch = BMVAL(etmidr, 4, 11);
drivers/hwtracing/coresight/coresight-etm3x-core.c
877
if (etm_arch_supported(drvdata->arch) == false)
drivers/hwtracing/coresight/coresight-etm4x-core.c
1254
drvdata->arch = etm_devarch_to_arch(devarch);
drivers/hwtracing/coresight/coresight-etm4x-core.c
1286
drvdata->arch = etm_devarch_to_arch(devarch);
drivers/hwtracing/coresight/coresight-etm4x-core.c
1485
if ((drvdata->arch < ETM_ARCH_V4_3) || (drvdata->nr_resource > 0))
drivers/hwtracing/coresight/coresight-etm4x-core.c
2230
if (!drvdata->arch)
drivers/hwtracing/coresight/coresight-etm4x-core.c
2233
major = ETM_ARCH_MAJOR_VERSION(drvdata->arch);
drivers/hwtracing/coresight/coresight-etm4x-core.c
2234
minor = ETM_ARCH_MINOR_VERSION(drvdata->arch);
drivers/hwtracing/coresight/coresight-etm4x.h
1039
u8 arch;
drivers/hwtracing/coresight/coresight-etm4x.h
1111
return drvdata->arch >= ETM_ARCH_ETE;
drivers/hwtracing/coresight/coresight-etm4x.h
795
#define ETM_ARCH_MAJOR_VERSION(arch) (((arch) >> 4) & 0xfU)
drivers/hwtracing/coresight/coresight-etm4x.h
796
#define ETM_ARCH_MINOR_VERSION(arch) ((arch) & 0xfU)
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1577
vf_count = padap->params.arch.vfcount;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1853
for (i = 0; i < padap->params.arch.vfcount; i++)
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1857
sge_qbase->vfcount = padap->params.arch.vfcount;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2648
if (padap->params.arch.mps_rplc_size > CUDBG_MAX_RPLC_SIZE) {
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2657
tcam->rplc_size = padap->params.arch.mps_rplc_size;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2671
n = padap->params.arch.mps_tcam_size;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
287
len = adap->params.arch.vfcount *
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
357
adap->params.arch.mps_tcam_size;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
794
for (i = 0; i < padap->params.arch.nchan; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
465
struct arch_specific_params arch; /* chip specific params */
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1685
if (adap->params.arch.mps_rplc_size > 128)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1802
if (adap->params.arch.mps_rplc_size > 128) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1856
if (adap->params.arch.mps_rplc_size > 128)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1865
if (adap->params.arch.mps_rplc_size > 128)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2359
int vf, vfcount = adapter->params.arch.vfcount;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3517
for (i = 0; i < adap->params.arch.nchan; i++)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3601
for (i = 0; i < adap->params.arch.nchan; i++) \
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3610
for (i = 0; i < adap->params.arch.nchan; i++) \
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3716
for (i = 0; i < adap->params.arch.nchan; i++)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3740
if (adap->params.arch.nchan == NCHAN)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
827
if (adap->params.arch.nchan == NCHAN) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
4509
u16 cng_ch_bits_log = adap->params.arch.cng_ch_bits_log;
drivers/net/ethernet/chelsio/cxgb4/sge.c
482
u32 val = adap->params.arch.sge_fl_db;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5630
int nchan = adap->params.arch.nchan;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5663
int nchan = adap->params.arch.nchan;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5914
if (adap->params.arch.nchan == NCHAN) {
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5922
if (adap->params.arch.nchan == NCHAN) {
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
6052
for (i = 0; i < adap->params.arch.pm_stats_cnt; i++) {
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
6079
for (i = 0; i < adap->params.arch.pm_stats_cnt; i++) {
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7979
unsigned int max_naddr = adap->params.arch.mps_tcam_size;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8144
unsigned int max_mac_addr = adap->params.arch.mps_tcam_size;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9136
adapter->params.arch.sge_fl_db = DBPRIO_F;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9137
adapter->params.arch.mps_tcam_size =
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9139
adapter->params.arch.mps_rplc_size = 128;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9140
adapter->params.arch.nchan = NCHAN;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9141
adapter->params.arch.pm_stats_cnt = PM_NSTATS;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9142
adapter->params.arch.vfcount = 128;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9146
adapter->params.arch.cng_ch_bits_log = 2;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9150
adapter->params.arch.sge_fl_db = DBPRIO_F | DBTYPE_F;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9151
adapter->params.arch.mps_tcam_size =
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9153
adapter->params.arch.mps_rplc_size = 128;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9154
adapter->params.arch.nchan = NCHAN;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9155
adapter->params.arch.pm_stats_cnt = PM_NSTATS;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9156
adapter->params.arch.vfcount = 128;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9157
adapter->params.arch.cng_ch_bits_log = 2;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9161
adapter->params.arch.sge_fl_db = 0;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9162
adapter->params.arch.mps_tcam_size =
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9164
adapter->params.arch.mps_rplc_size = 256;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9165
adapter->params.arch.nchan = 2;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9166
adapter->params.arch.pm_stats_cnt = T6_PM_NSTATS;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9167
adapter->params.arch.vfcount = 256;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9171
adapter->params.arch.cng_ch_bits_log = 3;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
527
u32 val = adapter->params.arch.sge_fl_db;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h
266
struct arch_specific_params arch; /* chip specific params */
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
1535
unsigned int max_naddr = adapter->params.arch.mps_tcam_size;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
1619
unsigned int max_naddr = adapter->params.arch.mps_tcam_size;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
1698
unsigned int max_mac_addr = adapter->params.arch.mps_tcam_size;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
2162
adapter->params.arch.sge_fl_db = DBPRIO_F;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
2163
adapter->params.arch.mps_tcam_size =
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
2170
adapter->params.arch.sge_fl_db = DBPRIO_F | DBTYPE_F;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
2171
adapter->params.arch.mps_tcam_size =
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
2178
adapter->params.arch.sge_fl_db = 0;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
2179
adapter->params.arch.mps_tcam_size =
drivers/net/ethernet/sfc/falcon/nic.c
180
#define REGISTER(name, arch, min_rev, max_rev) { \
drivers/net/ethernet/sfc/falcon/nic.c
181
arch ## R_ ## min_rev ## max_rev ## _ ## name, \
drivers/net/ethernet/sfc/falcon/nic.c
182
REGISTER_REVISION_ ## arch ## min_rev, \
drivers/net/ethernet/sfc/falcon/nic.c
183
REGISTER_REVISION_ ## arch ## max_rev \
drivers/net/ethernet/sfc/falcon/nic.c
305
#define REGISTER_TABLE_DIMENSIONS(_, offset, arch, min_rev, max_rev, step, rows) { \
drivers/net/ethernet/sfc/falcon/nic.c
307
REGISTER_REVISION_ ## arch ## min_rev, \
drivers/net/ethernet/sfc/falcon/nic.c
308
REGISTER_REVISION_ ## arch ## max_rev, \
drivers/net/ethernet/sfc/falcon/nic.c
311
#define REGISTER_TABLE(name, arch, min_rev, max_rev) \
drivers/net/ethernet/sfc/falcon/nic.c
313
name, arch ## R_ ## min_rev ## max_rev ## _ ## name, \
drivers/net/ethernet/sfc/falcon/nic.c
314
arch, min_rev, max_rev, \
drivers/net/ethernet/sfc/falcon/nic.c
315
arch ## R_ ## min_rev ## max_rev ## _ ## name ## _STEP, \
drivers/net/ethernet/sfc/falcon/nic.c
316
arch ## R_ ## min_rev ## max_rev ## _ ## name ## _ROWS)
drivers/net/ethernet/sfc/nic.c
182
#define REGISTER(name, arch, min_rev, max_rev) { \
drivers/net/ethernet/sfc/nic.c
183
arch ## R_ ## min_rev ## max_rev ## _ ## name, \
drivers/net/ethernet/sfc/nic.c
184
REGISTER_REVISION_ ## arch ## min_rev, \
drivers/net/ethernet/sfc/nic.c
185
REGISTER_REVISION_ ## arch ## max_rev \
drivers/net/ethernet/sfc/nic.c
203
#define REGISTER_TABLE_DIMENSIONS(_, offset, arch, min_rev, max_rev, step, rows) { \
drivers/net/ethernet/sfc/nic.c
205
REGISTER_REVISION_ ## arch ## min_rev, \
drivers/net/ethernet/sfc/nic.c
206
REGISTER_REVISION_ ## arch ## max_rev, \
drivers/net/ethernet/sfc/nic.c
209
#define REGISTER_TABLE(name, arch, min_rev, max_rev) \
drivers/net/ethernet/sfc/nic.c
211
name, arch ## R_ ## min_rev ## max_rev ## _ ## name, \
drivers/net/ethernet/sfc/nic.c
212
arch, min_rev, max_rev, \
drivers/net/ethernet/sfc/nic.c
213
arch ## R_ ## min_rev ## max_rev ## _ ## name ## _STEP, \
drivers/net/ethernet/sfc/nic.c
214
arch ## R_ ## min_rev ## max_rev ## _ ## name ## _ROWS)
drivers/net/ethernet/sfc/siena/nic.c
186
#define REGISTER(name, arch, min_rev, max_rev) { \
drivers/net/ethernet/sfc/siena/nic.c
187
arch ## R_ ## min_rev ## max_rev ## _ ## name, \
drivers/net/ethernet/sfc/siena/nic.c
188
REGISTER_REVISION_ ## arch ## min_rev, \
drivers/net/ethernet/sfc/siena/nic.c
189
REGISTER_REVISION_ ## arch ## max_rev \
drivers/net/ethernet/sfc/siena/nic.c
311
#define REGISTER_TABLE_DIMENSIONS(_, offset, arch, min_rev, max_rev, step, rows) { \
drivers/net/ethernet/sfc/siena/nic.c
313
REGISTER_REVISION_ ## arch ## min_rev, \
drivers/net/ethernet/sfc/siena/nic.c
314
REGISTER_REVISION_ ## arch ## max_rev, \
drivers/net/ethernet/sfc/siena/nic.c
317
#define REGISTER_TABLE(name, arch, min_rev, max_rev) \
drivers/net/ethernet/sfc/siena/nic.c
319
name, arch ## R_ ## min_rev ## max_rev ## _ ## name, \
drivers/net/ethernet/sfc/siena/nic.c
320
arch, min_rev, max_rev, \
drivers/net/ethernet/sfc/siena/nic.c
321
arch ## R_ ## min_rev ## max_rev ## _ ## name ## _STEP, \
drivers/net/ethernet/sfc/siena/nic.c
322
arch ## R_ ## min_rev ## max_rev ## _ ## name ## _ROWS)
drivers/pinctrl/pinctrl-microchip-sgpio.c
102
.arch = SGPIO_ARCH_SPARX5,
drivers/pinctrl/pinctrl-microchip-sgpio.c
190
switch (priv->properties->arch) {
drivers/pinctrl/pinctrl-microchip-sgpio.c
216
switch (priv->properties->arch) {
drivers/pinctrl/pinctrl-microchip-sgpio.c
243
switch (priv->properties->arch) {
drivers/pinctrl/pinctrl-microchip-sgpio.c
292
switch (priv->properties->arch) {
drivers/pinctrl/pinctrl-microchip-sgpio.c
329
switch (priv->properties->arch) {
drivers/pinctrl/pinctrl-microchip-sgpio.c
59
int arch;
drivers/pinctrl/pinctrl-microchip-sgpio.c
92
.arch = SGPIO_ARCH_LUTON,
drivers/pinctrl/pinctrl-microchip-sgpio.c
97
.arch = SGPIO_ARCH_OCELOT,
drivers/s390/char/zcore.c
299
unsigned char arch;
drivers/s390/char/zcore.c
335
rc = memcpy_hsa_kernel(&arch, __LC_AR_MODE_ID, 1);
drivers/s390/char/zcore.c
339
if (arch == ARCH_S390) {
drivers/s390/crypto/vfio_ap_ops.c
1826
if (kvm->arch.crypto.crycbd) {
drivers/s390/crypto/vfio_ap_ops.c
1827
down_write(&kvm->arch.crypto.pqap_hook_rwsem);
drivers/s390/crypto/vfio_ap_ops.c
1828
kvm->arch.crypto.pqap_hook = &matrix_mdev->pqap_hook;
drivers/s390/crypto/vfio_ap_ops.c
1829
up_write(&kvm->arch.crypto.pqap_hook_rwsem);
drivers/s390/crypto/vfio_ap_ops.c
1885
if (kvm && kvm->arch.crypto.crycbd) {
drivers/s390/crypto/vfio_ap_ops.c
1886
down_write(&kvm->arch.crypto.pqap_hook_rwsem);
drivers/s390/crypto/vfio_ap_ops.c
1887
kvm->arch.crypto.pqap_hook = NULL;
drivers/s390/crypto/vfio_ap_ops.c
1888
up_write(&kvm->arch.crypto.pqap_hook_rwsem);
drivers/s390/crypto/vfio_ap_ops.c
455
gisa = kvm->arch.gisa_int.origin;
drivers/s390/crypto/vfio_ap_ops.c
594
if (!(vcpu->arch.sie_block->eca & ECA_AIV)) {
drivers/s390/crypto/vfio_ap_ops.c
596
__func__, apqn, vcpu->arch.sie_block->eca);
drivers/s390/crypto/vfio_ap_ops.c
603
if (!vcpu->kvm->arch.crypto.pqap_hook) {
drivers/s390/crypto/vfio_ap_ops.c
610
matrix_mdev = container_of(vcpu->kvm->arch.crypto.pqap_hook,
drivers/soc/versatile/soc-integrator.c
73
static DEVICE_ATTR_RO(arch);
drivers/usb/dwc2/core.h
631
unsigned arch:2;
drivers/usb/dwc2/debugfs.c
735
print_param(seq, hw, arch);
drivers/usb/dwc2/hcd.c
73
switch (hsotg->hw_params.arch) {
drivers/usb/dwc2/hcd.c
781
hsotg->hw_params.arch == GHWCFG2_EXT_DMA_ARCH) {
drivers/usb/dwc2/params.c
510
bool dma_capable = !(hw->arch == GHWCFG2_SLAVE_ONLY_ARCH);
drivers/usb/dwc2/params.c
823
bool dma_capable = !(hw->arch == GHWCFG2_SLAVE_ONLY_ARCH);
drivers/usb/dwc2/params.c
969
hw->arch = (hwcfg2 & GHWCFG2_ARCHITECTURE_MASK) >>
drivers/video/fbdev/nvidia/nvidia.c
1230
u32 arch = 0;
drivers/video/fbdev/nvidia/nvidia.c
1240
arch = NV_ARCH_10;
drivers/video/fbdev/nvidia/nvidia.c
1245
arch = NV_ARCH_20;
drivers/video/fbdev/nvidia/nvidia.c
1252
arch = NV_ARCH_30;
drivers/video/fbdev/nvidia/nvidia.c
1267
arch = NV_ARCH_40;
drivers/video/fbdev/nvidia/nvidia.c
1270
arch = NV_ARCH_04;
drivers/video/fbdev/nvidia/nvidia.c
1276
return arch;
drivers/video/fbdev/riva/fbdev.c
1852
u32 arch = 0;
drivers/video/fbdev/riva/fbdev.c
1862
arch = NV_ARCH_10;
drivers/video/fbdev/riva/fbdev.c
1867
arch = NV_ARCH_20;
drivers/video/fbdev/riva/fbdev.c
1874
arch = NV_ARCH_30;
drivers/video/fbdev/riva/fbdev.c
1877
arch = NV_ARCH_04;
drivers/video/fbdev/riva/fbdev.c
1880
arch = NV_ARCH_03;
drivers/video/fbdev/riva/fbdev.c
1885
return arch;
fs/binfmt_elf.c
729
struct arch_elf_state *arch,
fs/binfmt_elf.c
764
pr->pr_datasz, ELF_COMPAT, arch);
fs/binfmt_elf.c
776
struct arch_elf_state *arch)
fs/binfmt_elf.c
824
ret = parse_elf_property(note.data, &off, datasz, arch,
include/kvm/arm_arch_timer.h
123
#define vcpu_timer(v) (&(v)->arch.timer_cpu)
include/kvm/arm_arch_timer.h
125
#define vcpu_vtimer(v) (&(v)->arch.timer_cpu.timers[TIMER_VTIMER])
include/kvm/arm_arch_timer.h
126
#define vcpu_ptimer(v) (&(v)->arch.timer_cpu.timers[TIMER_PTIMER])
include/kvm/arm_arch_timer.h
127
#define vcpu_hvtimer(v) (&(v)->arch.timer_cpu.timers[TIMER_HVTIMER])
include/kvm/arm_arch_timer.h
128
#define vcpu_hptimer(v) (&(v)->arch.timer_cpu.timers[TIMER_HPTIMER])
include/kvm/arm_arch_timer.h
131
#define timer_context_to_vcpu(ctx) container_of((ctx), struct kvm_vcpu, arch.timer_cpu.timers[(ctx)->timer_id])
include/kvm/arm_arch_timer.h
132
#define timer_vm_data(ctx) (&(timer_context_to_vcpu(ctx)->kvm->arch.timer_data))
include/kvm/arm_arch_timer.h
135
#define get_vgic_ppi(k, i) (((k)->arch.vgic.vgic_model != KVM_DEV_TYPE_ARM_VGIC_V5) ? \
include/kvm/arm_pmu.h
44
#define kvm_arm_pmu_irq_initialized(v) ((v)->arch.pmu.irq_num != 0)
include/kvm/arm_pmu.h
87
vcpu->arch.pmu.events = *kvm_get_pmu_events(); \
include/kvm/arm_psci.h
32
if (vcpu->kvm->arch.psci_version)
include/kvm/arm_psci.h
33
return vcpu->kvm->arch.psci_version;
include/kvm/arm_vgic.h
108
#define irq_is_sgi(k, i) __irq_is_sgi((k)->arch.vgic.vgic_model, i)
include/kvm/arm_vgic.h
109
#define irq_is_ppi(k, i) __irq_is_ppi((k)->arch.vgic.vgic_model, i)
include/kvm/arm_vgic.h
110
#define irq_is_spi(k, i) __irq_is_spi((k)->arch.vgic.vgic_model, i)
include/kvm/arm_vgic.h
111
#define irq_is_lpi(k, i) __irq_is_lpi((k)->arch.vgic.vgic_model, i)
include/kvm/arm_vgic.h
123
#define __vgic_is_v(k, v) ((k)->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V##v)
include/kvm/arm_vgic.h
577
#define irqchip_in_kernel(k) (!!((k)->arch.vgic.in_kernel))
include/kvm/arm_vgic.h
578
#define vgic_initialized(k) ((k)->arch.vgic.initialized)
include/kvm/arm_vgic.h
583
switch ((k)->arch.vgic.vgic_model) { \
include/kvm/arm_vgic.h
585
__ret &= FIELD_GET(GICV5_HWIRQ_ID, i) < (k)->arch.vgic.nr_spis; \
include/kvm/arm_vgic.h
588
__ret &= (i) < ((k)->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS); \
include/linux/audit.h
138
extern int audit_classify_arch(int arch);
include/linux/audit.h
308
#define audit_is_compat(arch) (!((arch) & __AUDIT_ARCH_64BIT))
include/linux/audit.h
310
#define audit_is_compat(arch) false
include/linux/elf.h
91
struct arch_elf_state *arch)
include/linux/elf.h
97
bool compat, struct arch_elf_state *arch);
include/linux/ftrace.h
788
struct dyn_arch_ftrace arch;
include/linux/hp_sdc.h
293
#error No support for device registration on this arch yet.
include/linux/kexec.h
376
struct kimage_arch arch;
include/linux/kvm_host.h
1736
return vcpu->arch.waitp;
include/linux/kvm_host.h
244
struct kvm_arch_async_pf arch;
include/linux/kvm_host.h
252
unsigned long hva, struct kvm_arch_async_pf *arch);
include/linux/kvm_host.h
388
struct kvm_vcpu_arch arch;
include/linux/kvm_host.h
600
struct kvm_arch_memory_slot arch;
include/linux/kvm_host.h
833
struct kvm_arch arch;
include/linux/mm_types_task.h
77
struct arch_tlbflush_unmap_batch arch;
include/linux/module.h
457
struct mod_arch_specific arch;
include/linux/pe.h
369
struct data_dirent arch; /* reservered */
include/sound/sof/header.h
186
uint32_t arch; /* Identifier of architecture */
include/uapi/linux/kvm.h
272
struct kvm_debug_exit_arch arch;
include/uapi/linux/kvm.h
638
struct kvm_guest_debug_arch arch;
include/uapi/linux/ptrace.h
87
__u32 arch;
include/uapi/linux/seccomp.h
64
__u32 arch;
include/xen/interface/xen.h
534
struct arch_vcpu_info arch;
include/xen/interface/xen.h
587
struct arch_shared_info arch;
kernel/audit.h
146
int arch;
kernel/auditfilter.c
213
struct audit_field *arch = entry->rule.arch_f;
kernel/auditfilter.c
215
if (!arch) {
kernel/auditfilter.c
224
switch (audit_classify_arch(arch->val)) {
kernel/auditsc.c
1018
ctx->arch = 0;
kernel/auditsc.c
159
switch (audit_classify_syscall(ctx->arch, n)) {
kernel/auditsc.c
1668
context->arch, context->major);
kernel/auditsc.c
2013
context->arch = syscall_get_arch(current);
kernel/auditsc.c
551
result = audit_comparator(ctx->arch, f->op, f->val);
kernel/events/uprobes.c
1074
void *insn = &uprobe->arch.insn;
kernel/events/uprobes.c
1075
int size = sizeof(uprobe->arch.insn);
kernel/events/uprobes.c
1114
if (is_trap_insn((uprobe_opcode_t *)&uprobe->arch.insn))
kernel/events/uprobes.c
1117
ret = arch_uprobe_analyze_insn(&uprobe->arch, mm, vaddr);
kernel/events/uprobes.c
1170
ret = set_swbp(&uprobe->arch, vma, vaddr);
kernel/events/uprobes.c
1185
return set_orig_insn(&uprobe->arch, vma, vaddr);
kernel/events/uprobes.c
1885
&uprobe->arch.ixol, sizeof(uprobe->arch.ixol));
kernel/events/uprobes.c
2333
err = arch_uprobe_pre_xol(&uprobe->arch, regs);
kernel/events/uprobes.c
2556
utask->auprobe = &uprobe->arch;
kernel/events/uprobes.c
2767
if (arch_uprobe_ignore(&uprobe->arch, regs))
kernel/events/uprobes.c
2773
arch_uprobe_optimize(&uprobe->arch, bp_vaddr);
kernel/events/uprobes.c
2782
if (arch_uprobe_skip_sstep(&uprobe->arch, regs))
kernel/events/uprobes.c
2805
if (arch_uprobe_ignore(&uprobe->arch, regs))
kernel/events/uprobes.c
2821
err = arch_uprobe_post_xol(&uprobe->arch, regs);
kernel/events/uprobes.c
2823
arch_uprobe_abort_xol(&uprobe->arch, regs);
kernel/events/uprobes.c
512
uprobe = container_of(auprobe, struct uprobe, arch);
kernel/events/uprobes.c
88
struct arch_uprobe arch;
kernel/ptrace.c
1001
.arch = syscall_get_arch(child),
kernel/seccomp.c
255
sd->arch = syscall_get_arch(task);
kernel/seccomp.c
379
if (likely(sd->arch == SECCOMP_ARCH_NATIVE))
kernel/seccomp.c
383
if (likely(sd->arch == SECCOMP_ARCH_COMPAT))
kernel/seccomp.c
748
if (sd->arch == SECCOMP_ARCH_NATIVE)
kernel/seccomp.c
795
case offsetof(struct seccomp_data, arch):
kernel/seccomp.c
796
reg_value = sd->arch;
kernel/seccomp.c
849
size_t bitmap_size, int arch)
kernel/seccomp.c
869
sd.arch = arch;
lib/audit.c
32
int audit_classify_arch(int arch)
lib/audit.c
34
if (audit_is_compat(arch))
mm/rmap.c
718
arch_tlbbatch_flush(&tlb_ubc->arch);
mm/rmap.c
752
arch_tlbbatch_add_pending(&tlb_ubc->arch, mm, start, end);
samples/seccomp/dropper.c
28
static int install_filter(int arch, int nr, int error)
samples/seccomp/dropper.c
32
(offsetof(struct seccomp_data, arch))),
samples/seccomp/dropper.c
33
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, arch, 0, 3),
tools/bpf/bpf_jit_disasm.c
73
info.arch = bfd_get_arch(bfdf);
tools/bpf/bpf_jit_disasm.c
81
disassemble = disassembler(info.arch,
tools/bpf/bpftool/jit_disasm.c
240
static int init_context(disasm_ctx_t *ctx, const char *arch,
tools/bpf/bpftool/jit_disasm.c
283
if (arch) {
tools/bpf/bpftool/jit_disasm.c
284
const bfd_arch_info_type *inf = bfd_scan_arch(arch);
tools/bpf/bpftool/jit_disasm.c
289
p_err("No libbfd support for %s", arch);
tools/bpf/bpftool/jit_disasm.c
294
info->arch = bfd_get_arch(bfdf);
tools/bpf/bpftool/jit_disasm.c
305
ctx->disassemble = disassembler(info->arch,
tools/bpf/bpftool/jit_disasm.c
347
const char *arch, const char *disassembler_options,
tools/bpf/bpftool/jit_disasm.c
362
if (init_context(&ctx, arch, disassembler_options, image, len, func_ksym))
tools/bpf/bpftool/jit_disasm.c
81
init_context(disasm_ctx_t *ctx, const char *arch,
tools/bpf/bpftool/jit_disasm.c
88
if (arch)
tools/bpf/bpftool/jit_disasm.c
89
triple = LLVMNormalizeTargetTriple(arch);
tools/bpf/bpftool/main.h
191
const char *arch, const char *disassembler_options,
tools/bpf/bpftool/main.h
200
const char *arch, const char *disassembler_options,
tools/include/uapi/linux/kvm.h
271
struct kvm_debug_exit_arch arch;
tools/include/uapi/linux/kvm.h
637
struct kvm_guest_debug_arch arch;
tools/include/uapi/linux/seccomp.h
64
__u32 arch;
tools/objtool/disas.c
330
int arch, int mach32, int mach64,
tools/objtool/disas.c
336
dinfo->arch = arch;
tools/objtool/disas.c
383
dinfo->arch = bfd_arch_unknown;
tools/objtool/disas.c
387
if (err || dinfo->arch == bfd_arch_unknown || dinfo->mach == 0) {
tools/objtool/disas.c
397
dctx->disassembler = disassembler(dinfo->arch,
tools/objtool/include/objtool/disas.h
20
int arch, int mach32, int mach64,
tools/objtool/include/objtool/disas.h
47
int arch, int mach32, int mach64,
tools/perf/arch/common.c
148
const char *arch = perf_env__arch(env), *cross_env;
tools/perf/arch/common.c
156
if (!strcmp(perf_env__arch(NULL), arch))
tools/perf/arch/common.c
173
if (!strcmp(arch, "arc"))
tools/perf/arch/common.c
175
else if (!strcmp(arch, "arm"))
tools/perf/arch/common.c
177
else if (!strcmp(arch, "arm64"))
tools/perf/arch/common.c
179
else if (!strcmp(arch, "powerpc"))
tools/perf/arch/common.c
181
else if (!strcmp(arch, "riscv32"))
tools/perf/arch/common.c
183
else if (!strcmp(arch, "riscv64"))
tools/perf/arch/common.c
185
else if (!strcmp(arch, "sh"))
tools/perf/arch/common.c
187
else if (!strcmp(arch, "s390"))
tools/perf/arch/common.c
189
else if (!strcmp(arch, "sparc"))
tools/perf/arch/common.c
191
else if (!strcmp(arch, "x86"))
tools/perf/arch/common.c
193
else if (!strcmp(arch, "mips"))
tools/perf/arch/common.c
196
ui__error("binutils for %s not supported.\n", arch);
tools/perf/arch/common.c
205
name, arch, name);
tools/perf/arch/common.c
227
if (env->arch == NULL)
tools/perf/pmu-events/empty-pmu-events.c
2794
const char *arch;
tools/perf/pmu-events/empty-pmu-events.c
2806
.arch = "common",
tools/perf/pmu-events/empty-pmu-events.c
2818
.arch = "testarch",
tools/perf/pmu-events/empty-pmu-events.c
2830
.arch = 0,
tools/perf/pmu-events/empty-pmu-events.c
3181
if (!map->arch) {
tools/perf/pmu-events/empty-pmu-events.c
3212
while (strcmp("common", map->arch))
tools/perf/pmu-events/empty-pmu-events.c
3249
if (!map->arch)
tools/perf/pmu-events/empty-pmu-events.c
3273
if (!map->arch)
tools/perf/pmu-events/empty-pmu-events.c
3282
const struct pmu_events_table *find_core_events_table(const char *arch, const char *cpuid)
tools/perf/pmu-events/empty-pmu-events.c
3285
tables->arch;
tools/perf/pmu-events/empty-pmu-events.c
3287
if (!strcmp(tables->arch, arch) && !strcmp_cpuid_str(tables->cpuid, cpuid))
tools/perf/pmu-events/empty-pmu-events.c
3293
const struct pmu_metrics_table *find_core_metrics_table(const char *arch, const char *cpuid)
tools/perf/pmu-events/empty-pmu-events.c
3296
tables->arch;
tools/perf/pmu-events/empty-pmu-events.c
3298
if (!strcmp(tables->arch, arch) && !strcmp_cpuid_str(tables->cpuid, cpuid))
tools/perf/pmu-events/empty-pmu-events.c
3307
tables->arch;
tools/perf/pmu-events/empty-pmu-events.c
3321
tables->arch;
tools/perf/pmu-events/pmu-events.h
132
const struct pmu_events_table *find_core_events_table(const char *arch, const char *cpuid);
tools/perf/pmu-events/pmu-events.h
133
const struct pmu_metrics_table *find_core_metrics_table(const char *arch, const char *cpuid);
tools/perf/tests/topology.c
101
if (!env->cpu && strncmp(env->arch, "s390", 4) && strncmp(env->arch, "aarch64", 7))
tools/perf/tests/topology.c
110
if (!strncmp(env->arch, "ppc64le", 7)) {
tools/perf/ui/browsers/annotate.c
1189
err = symbol__annotate2(ms, evsel, &browser.arch);
tools/perf/ui/browsers/annotate.c
1201
err = thread__get_arch(ms->thread, &browser.arch);
tools/perf/ui/browsers/annotate.c
137
.arch = ab->arch,
tools/perf/ui/browsers/annotate.c
178
if (ins__is_fused(ab->arch, name, cursor->ins.name))
tools/perf/ui/browsers/annotate.c
26
struct arch;
tools/perf/ui/browsers/annotate.c
33
const struct arch *arch;
tools/perf/ui/browsers/annotate.c
470
symbol__annotate2(ms, evsel, &browser->arch);
tools/perf/ui/browsers/hists.c
3236
if (env->arch)
tools/perf/util/annotate-arch/annotate-arc.c
11
if (!arch)
tools/perf/util/annotate-arch/annotate-arc.c
14
arch->name = "arc";
tools/perf/util/annotate-arch/annotate-arc.c
15
arch->id = *id;
tools/perf/util/annotate-arch/annotate-arc.c
16
arch->objdump.comment_char = ';';
tools/perf/util/annotate-arch/annotate-arc.c
17
return arch;
tools/perf/util/annotate-arch/annotate-arc.c
6
const struct arch *arch__new_arc(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-arc.c
9
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-arm.c
11
struct arch arch;
tools/perf/util/annotate-arch/annotate-arm.c
16
static const struct ins_ops *arm__associate_instruction_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-arm.c
18
struct arch_arm *arm = container_of(arch, struct arch_arm, arch);
tools/perf/util/annotate-arch/annotate-arm.c
29
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-arm.c
33
const struct arch *arch__new_arm(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-arm.c
38
struct arch *arch;
tools/perf/util/annotate-arch/annotate-arm.c
43
arch = &arm->arch;
tools/perf/util/annotate-arch/annotate-arm.c
44
arch->name = "arm";
tools/perf/util/annotate-arch/annotate-arm.c
45
arch->id = *id;
tools/perf/util/annotate-arch/annotate-arm.c
46
arch->objdump.comment_char = ';';
tools/perf/util/annotate-arch/annotate-arm.c
47
arch->objdump.skip_functions_char = '+';
tools/perf/util/annotate-arch/annotate-arm.c
48
arch->associate_instruction_ops = arm__associate_instruction_ops;
tools/perf/util/annotate-arch/annotate-arm.c
60
return arch;
tools/perf/util/annotate-arch/annotate-arm64.c
101
arch = &arm->arch;
tools/perf/util/annotate-arch/annotate-arm64.c
102
arch->name = "arm64";
tools/perf/util/annotate-arch/annotate-arm64.c
103
arch->id = *id;
tools/perf/util/annotate-arch/annotate-arm64.c
104
arch->objdump.comment_char = '/';
tools/perf/util/annotate-arch/annotate-arm64.c
105
arch->objdump.skip_functions_char = '+';
tools/perf/util/annotate-arch/annotate-arm64.c
106
arch->associate_instruction_ops = arm64__associate_instruction_ops;
tools/perf/util/annotate-arch/annotate-arm64.c
119
return arch;
tools/perf/util/annotate-arch/annotate-arm64.c
12
struct arch arch;
tools/perf/util/annotate-arch/annotate-arm64.c
17
static int arm64_mov__parse(const struct arch *arch __maybe_unused,
tools/perf/util/annotate-arch/annotate-arm64.c
72
static const struct ins_ops *arm64__associate_instruction_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-arm64.c
74
struct arch_arm64 *arm = container_of(arch, struct arch_arm64, arch);
tools/perf/util/annotate-arch/annotate-arm64.c
87
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-arm64.c
91
const struct arch *arch__new_arm64(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-arm64.c
96
struct arch *arch;
tools/perf/util/annotate-arch/annotate-csky.c
39
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-csky.c
43
const struct arch *arch__new_csky(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-csky.c
46
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-csky.c
48
if (!arch)
tools/perf/util/annotate-arch/annotate-csky.c
51
arch->name = "csky";
tools/perf/util/annotate-arch/annotate-csky.c
52
arch->id = *id;
tools/perf/util/annotate-arch/annotate-csky.c
53
arch->objdump.comment_char = '/';
tools/perf/util/annotate-arch/annotate-csky.c
54
arch->associate_instruction_ops = csky__associate_ins_ops;
tools/perf/util/annotate-arch/annotate-csky.c
55
return arch;
tools/perf/util/annotate-arch/annotate-csky.c
8
static const struct ins_ops *csky__associate_ins_ops(struct arch *arch,
tools/perf/util/annotate-arch/annotate-loongarch.c
119
const struct ins_ops *loongarch__associate_ins_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-loongarch.c
138
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-loongarch.c
143
const struct arch *arch__new_loongarch(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-loongarch.c
146
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-loongarch.c
148
if (!arch)
tools/perf/util/annotate-arch/annotate-loongarch.c
151
arch->name = "loongarch";
tools/perf/util/annotate-arch/annotate-loongarch.c
152
arch->id = *id;
tools/perf/util/annotate-arch/annotate-loongarch.c
153
arch->associate_instruction_ops = loongarch__associate_ins_ops;
tools/perf/util/annotate-arch/annotate-loongarch.c
154
arch->objdump.comment_char = '#';
tools/perf/util/annotate-arch/annotate-loongarch.c
155
return arch;
tools/perf/util/annotate-arch/annotate-loongarch.c
17
static int loongarch_call__parse(const struct arch *arch, struct ins_operands *ops,
tools/perf/util/annotate-arch/annotate-loongarch.c
34
if (arch->objdump.skip_functions_char &&
tools/perf/util/annotate-arch/annotate-loongarch.c
35
strchr(name, arch->objdump.skip_functions_char))
tools/perf/util/annotate-arch/annotate-loongarch.c
68
static int loongarch_jump__parse(const struct arch *arch, struct ins_operands *ops,
tools/perf/util/annotate-arch/annotate-loongarch.c
81
ops->jump.raw_comment = strchr(ops->raw, arch->objdump.comment_char);
tools/perf/util/annotate-arch/annotate-mips.c
35
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-mips.c
40
const struct arch *arch__new_mips(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-mips.c
43
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-mips.c
45
if (!arch)
tools/perf/util/annotate-arch/annotate-mips.c
48
arch->name = "mips";
tools/perf/util/annotate-arch/annotate-mips.c
49
arch->id = *id;
tools/perf/util/annotate-arch/annotate-mips.c
50
arch->objdump.comment_char = '#';
tools/perf/util/annotate-arch/annotate-mips.c
51
arch->associate_instruction_ops = mips__associate_ins_ops;
tools/perf/util/annotate-arch/annotate-mips.c
52
return arch;
tools/perf/util/annotate-arch/annotate-mips.c
8
const struct ins_ops *mips__associate_ins_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-powerpc.c
140
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-powerpc.c
357
if (annotate_get_insn_location(dloc->arch, dl, &loc) < 0)
tools/perf/util/annotate-arch/annotate-powerpc.c
39
static int arithmetic__parse(const struct arch *arch __maybe_unused, struct ins_operands *ops,
tools/perf/util/annotate-arch/annotate-powerpc.c
393
const struct arch *arch__new_powerpc(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-powerpc.c
396
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-powerpc.c
398
if (!arch)
tools/perf/util/annotate-arch/annotate-powerpc.c
401
arch->name = "powerpc";
tools/perf/util/annotate-arch/annotate-powerpc.c
402
arch->id = *id;
tools/perf/util/annotate-arch/annotate-powerpc.c
403
arch->objdump.comment_char = '#';
tools/perf/util/annotate-arch/annotate-powerpc.c
405
arch->associate_instruction_ops = powerpc__associate_instruction_ops;
tools/perf/util/annotate-arch/annotate-powerpc.c
407
arch->update_insn_state = update_insn_state_powerpc;
tools/perf/util/annotate-arch/annotate-powerpc.c
409
return arch;
tools/perf/util/annotate-arch/annotate-powerpc.c
76
static int load_store__parse(const struct arch *arch __maybe_unused, struct ins_operands *ops,
tools/perf/util/annotate-arch/annotate-powerpc.c
96
static const struct ins_ops *powerpc__associate_instruction_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-riscv64.c
23
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-riscv64.c
28
const struct arch *arch__new_riscv64(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-riscv64.c
31
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-riscv64.c
33
if (!arch)
tools/perf/util/annotate-arch/annotate-riscv64.c
36
arch->name = "riscv";
tools/perf/util/annotate-arch/annotate-riscv64.c
37
arch->id = *id;
tools/perf/util/annotate-arch/annotate-riscv64.c
38
arch->objdump.comment_char = '#';
tools/perf/util/annotate-arch/annotate-riscv64.c
39
arch->associate_instruction_ops = riscv64__associate_ins_ops;
tools/perf/util/annotate-arch/annotate-riscv64.c
40
return arch;
tools/perf/util/annotate-arch/annotate-riscv64.c
8
const struct ins_ops *riscv64__associate_ins_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-s390.c
121
static const struct ins_ops *s390__associate_ins_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-s390.c
13
static int s390_call__parse(const struct arch *arch, struct ins_operands *ops,
tools/perf/util/annotate-arch/annotate-s390.c
147
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-s390.c
151
static int s390__cpuid_parse(struct arch *arch, const char *cpuid)
tools/perf/util/annotate-arch/annotate-s390.c
164
arch->family = family;
tools/perf/util/annotate-arch/annotate-s390.c
165
arch->model = 0;
tools/perf/util/annotate-arch/annotate-s390.c
172
const struct arch *arch__new_s390(const struct e_machine_and_e_flags *id, const char *cpuid)
tools/perf/util/annotate-arch/annotate-s390.c
174
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-s390.c
176
if (!arch)
tools/perf/util/annotate-arch/annotate-s390.c
179
arch->name = "s390";
tools/perf/util/annotate-arch/annotate-s390.c
180
arch->id = *id;
tools/perf/util/annotate-arch/annotate-s390.c
181
arch->associate_instruction_ops = s390__associate_ins_ops;
tools/perf/util/annotate-arch/annotate-s390.c
183
if (s390__cpuid_parse(arch, cpuid)) {
tools/perf/util/annotate-arch/annotate-s390.c
188
arch->objdump.comment_char = '#';
tools/perf/util/annotate-arch/annotate-s390.c
189
return arch;
tools/perf/util/annotate-arch/annotate-s390.c
33
if (arch->objdump.skip_functions_char &&
tools/perf/util/annotate-arch/annotate-s390.c
34
strchr(name, arch->objdump.skip_functions_char))
tools/perf/util/annotate-arch/annotate-s390.c
67
static int s390_mov__parse(const struct arch *arch __maybe_unused,
tools/perf/util/annotate-arch/annotate-sparc.c
124
static const struct ins_ops *sparc__associate_instruction_ops(struct arch *arch, const char *name)
tools/perf/util/annotate-arch/annotate-sparc.c
159
arch__associate_ins_ops(arch, name, ops);
tools/perf/util/annotate-arch/annotate-sparc.c
164
const struct arch *arch__new_sparc(const struct e_machine_and_e_flags *id,
tools/perf/util/annotate-arch/annotate-sparc.c
167
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-sparc.c
169
if (!arch)
tools/perf/util/annotate-arch/annotate-sparc.c
172
arch->name = "sparc";
tools/perf/util/annotate-arch/annotate-sparc.c
173
arch->id = *id;
tools/perf/util/annotate-arch/annotate-sparc.c
174
arch->associate_instruction_ops = sparc__associate_instruction_ops;
tools/perf/util/annotate-arch/annotate-sparc.c
175
arch->objdump.comment_char = '#';
tools/perf/util/annotate-arch/annotate-sparc.c
176
return arch;
tools/perf/util/annotate-arch/annotate-x86.c
134
static bool amd__ins_is_fused(const struct arch *arch, const char *ins1,
tools/perf/util/annotate-arch/annotate-x86.c
141
if (arch->family >= 0x15 && (strstarts(ins1, "test") ||
tools/perf/util/annotate-arch/annotate-x86.c
147
if (arch->family >= 0x19 && (strstarts(ins1, "add") ||
tools/perf/util/annotate-arch/annotate-x86.c
157
static bool intel__ins_is_fused(const struct arch *arch, const char *ins1,
tools/perf/util/annotate-arch/annotate-x86.c
160
if (arch->family != 6 || arch->model < 0x1e || strstr(ins2, "jmp"))
tools/perf/util/annotate-arch/annotate-x86.c
163
if (arch->model == 0x1e) {
tools/perf/util/annotate-arch/annotate-x86.c
185
static int x86__cpuid_parse(struct arch *arch, const char *cpuid)
tools/perf/util/annotate-arch/annotate-x86.c
195
arch->family = family;
tools/perf/util/annotate-arch/annotate-x86.c
196
arch->model = model;
tools/perf/util/annotate-arch/annotate-x86.c
197
arch->ins_is_fused = strstarts(cpuid, "AuthenticAMD") ?
tools/perf/util/annotate-arch/annotate-x86.c
220
if (annotate_get_insn_location(dloc->arch, dl, &loc) < 0)
tools/perf/util/annotate-arch/annotate-x86.c
780
const struct arch *arch__new_x86(const struct e_machine_and_e_flags *id, const char *cpuid)
tools/perf/util/annotate-arch/annotate-x86.c
782
struct arch *arch = zalloc(sizeof(*arch));
tools/perf/util/annotate-arch/annotate-x86.c
784
if (!arch)
tools/perf/util/annotate-arch/annotate-x86.c
787
arch->name = "x86";
tools/perf/util/annotate-arch/annotate-x86.c
788
arch->id = *id;
tools/perf/util/annotate-arch/annotate-x86.c
790
if (x86__cpuid_parse(arch, cpuid)) {
tools/perf/util/annotate-arch/annotate-x86.c
795
arch->instructions = x86__instructions;
tools/perf/util/annotate-arch/annotate-x86.c
796
arch->nr_instructions = ARRAY_SIZE(x86__instructions);
tools/perf/util/annotate-arch/annotate-x86.c
802
for (size_t i = 0; i < arch->nr_instructions - 1; i++) {
tools/perf/util/annotate-arch/annotate-x86.c
803
assert(strcmp(arch->instructions[i].name,
tools/perf/util/annotate-arch/annotate-x86.c
804
arch->instructions[i + 1].name) <= 0);
tools/perf/util/annotate-arch/annotate-x86.c
810
arch->sorted_instructions = true;
tools/perf/util/annotate-arch/annotate-x86.c
811
arch->objdump.comment_char = '#';
tools/perf/util/annotate-arch/annotate-x86.c
812
arch->objdump.register_char = '%';
tools/perf/util/annotate-arch/annotate-x86.c
813
arch->objdump.memory_ref_char = '(';
tools/perf/util/annotate-arch/annotate-x86.c
814
arch->objdump.imm_char = '$';
tools/perf/util/annotate-arch/annotate-x86.c
815
arch->insn_suffix = "bwlq";
tools/perf/util/annotate-arch/annotate-x86.c
817
arch->update_insn_state = update_insn_state_x86;
tools/perf/util/annotate-arch/annotate-x86.c
819
return arch;
tools/perf/util/annotate-data.c
1074
if (arch__is_x86(dloc->arch)) {
tools/perf/util/annotate-data.c
1314
arch__is_x86(dloc->arch)) {
tools/perf/util/annotate-data.c
1358
init_type_state(&state, dloc->arch);
tools/perf/util/annotate-data.c
1400
if ((arch__is_x86(dloc->arch)) || (arch__is_powerpc(dloc->arch)))
tools/perf/util/annotate-data.c
163
static void init_type_state(struct type_state *state, const struct arch *arch)
tools/perf/util/annotate-data.c
168
if (arch__is_x86(arch)) {
tools/perf/util/annotate-data.c
529
else if (arch__is_x86(dloc->arch) && reg == X86_REG_SP)
tools/perf/util/annotate-data.c
995
if (dloc->arch->update_insn_state)
tools/perf/util/annotate-data.c
996
dloc->arch->update_insn_state(state, dloc, cu_die, dl);
tools/perf/util/annotate-data.h
120
const struct arch *arch;
tools/perf/util/annotate.c
1000
arch = arch__find(e_machine, e_flags, machine->env ? machine->env->cpuid : NULL);
tools/perf/util/annotate.c
1001
if (arch == NULL) {
tools/perf/util/annotate.c
1006
*parch = arch;
tools/perf/util/annotate.c
1012
const struct arch **parch)
tools/perf/util/annotate.c
1019
const struct arch *arch = NULL;
tools/perf/util/annotate.c
1022
err = thread__get_arch(ms->thread, &arch);
tools/perf/util/annotate.c
1027
*parch = arch;
tools/perf/util/annotate.c
1032
args.arch = arch;
tools/perf/util/annotate.c
1270
thread__get_arch(ms->thread, &apd.arch);
tools/perf/util/annotate.c
1375
thread__get_arch(apd->he->ms.thread, &apd->arch);
tools/perf/util/annotate.c
1984
data_type = __hist_entry__get_data_type(apd->he, apd->arch, apd->dbg, dl, &offset);
tools/perf/util/annotate.c
2206
const struct arch **parch)
tools/perf/util/annotate.c
2453
static int arch__dwarf_regnum(const struct arch *arch, const char *str)
tools/perf/util/annotate.c
2459
p = strchr(str, arch->objdump.register_char);
tools/perf/util/annotate.c
2471
reg = get_dwarf_regnum(regname, arch->id.e_machine, arch->id.e_flags);
tools/perf/util/annotate.c
2482
static int extract_reg_offset(const struct arch *arch, const char *str,
tools/perf/util/annotate.c
2487
if (arch->objdump.register_char == 0)
tools/perf/util/annotate.c
2497
if (*str == arch->objdump.register_char) {
tools/perf/util/annotate.c
2498
if (arch__is_x86(arch)) {
tools/perf/util/annotate.c
2505
*str != arch->objdump.memory_ref_char)
tools/perf/util/annotate.c
2510
op_loc->reg1 = arch__dwarf_regnum(arch, p);
tools/perf/util/annotate.c
2516
op_loc->reg2 = arch__dwarf_regnum(arch, p + 1);
tools/perf/util/annotate.c
2545
int annotate_get_insn_location(const struct arch *arch, struct disasm_line *dl,
tools/perf/util/annotate.c
2578
if (!arch__is_powerpc(arch))
tools/perf/util/annotate.c
2587
if (arch__is_powerpc(arch)) {
tools/perf/util/annotate.c
2591
} else if (strchr(insn_str, arch->objdump.memory_ref_char)) {
tools/perf/util/annotate.c
2594
extract_reg_offset(arch, insn_str, op_loc);
tools/perf/util/annotate.c
2599
if (arch__is_x86(arch)) {
tools/perf/util/annotate.c
2611
if (*s == arch->objdump.register_char) {
tools/perf/util/annotate.c
2612
op_loc->reg1 = arch__dwarf_regnum(arch, s);
tools/perf/util/annotate.c
2614
else if (*s == arch->objdump.imm_char) {
tools/perf/util/annotate.c
2677
static bool is_stack_operation(const struct arch *arch, struct disasm_line *dl)
tools/perf/util/annotate.c
2679
if (arch__is_x86(arch)) {
tools/perf/util/annotate.c
2690
static bool is_stack_canary(const struct arch *arch, struct annotated_op_loc *loc)
tools/perf/util/annotate.c
2693
if (arch__is_x86(arch)) {
tools/perf/util/annotate.c
2706
static bool is_address_gen_insn(const struct arch *arch, struct disasm_line *dl)
tools/perf/util/annotate.c
2708
if (arch__is_x86(arch)) {
tools/perf/util/annotate.c
2795
__hist_entry__get_data_type(struct hist_entry *he, const struct arch *arch,
tools/perf/util/annotate.c
2812
if (annotate_get_insn_location(arch, dl, &loc) < 0) {
tools/perf/util/annotate.c
2818
if (is_stack_operation(arch, dl)) {
tools/perf/util/annotate.c
2824
if (is_address_gen_insn(arch, dl)) {
tools/perf/util/annotate.c
2832
.arch = arch,
tools/perf/util/annotate.c
2851
if (dso__kernel(map__dso(ms->map)) && arch__is_x86(arch) &&
tools/perf/util/annotate.c
2859
if (mem_type == NULL && is_stack_canary(arch, op_loc)) {
tools/perf/util/annotate.c
2899
const struct arch *arch;
tools/perf/util/annotate.c
2939
if (symbol__annotate(ms, evsel, &arch) < 0) {
tools/perf/util/annotate.c
2955
mem_type = __hist_entry__get_data_type(he, arch, di_cache.dbg, dl,
tools/perf/util/annotate.c
2971
if (prev_dl && ins__is_fused(arch, prev_dl->ins.name, dl->ins.name)) {
tools/perf/util/annotate.c
767
__hist_entry__get_data_type(struct hist_entry *he, const struct arch *arch,
tools/perf/util/annotate.c
859
data_type = __hist_entry__get_data_type(apd->he, apd->arch,
tools/perf/util/annotate.c
986
int thread__get_arch(struct thread *thread, const struct arch **parch)
tools/perf/util/annotate.c
988
const struct arch *arch;
tools/perf/util/annotate.h
205
const struct arch *arch;
tools/perf/util/annotate.h
444
const struct arch **parch);
tools/perf/util/annotate.h
447
const struct arch **parch);
tools/perf/util/annotate.h
549
int annotate_get_insn_location(const struct arch *arch, struct disasm_line *dl,
tools/perf/util/annotate.h
589
int thread__get_arch(struct thread *thread, const struct arch **parch);
tools/perf/util/arm-spe.c
355
simd_flags.arch |= SIMD_OP_FLAGS_ARCH_SVE;
tools/perf/util/capstone.c
143
enum cs_arch arch;
tools/perf/util/capstone.c
147
arch = CS_ARCH_X86;
tools/perf/util/capstone.c
150
arch = CS_ARCH_X86;
tools/perf/util/capstone.c
153
arch = CS_ARCH_ARM64;
tools/perf/util/capstone.c
156
arch = CS_ARCH_ARM;
tools/perf/util/capstone.c
159
arch = CS_ARCH_SYSZ;
tools/perf/util/capstone.c
165
if (perf_cs_open(arch, mode, cs_handle) != CS_ERR_OK) {
tools/perf/util/capstone.c
259
if (!arch__is_x86(args->arch))
tools/perf/util/capstone.c
38
static enum cs_err perf_cs_open(enum cs_arch arch, enum cs_mode mode, csh *handle)
tools/perf/util/capstone.c
41
return cs_open(arch, mode, handle);
tools/perf/util/capstone.c
44
static enum cs_err (*fn)(enum cs_arch arch, enum cs_mode mode, csh *handle);
tools/perf/util/capstone.c
54
return fn(arch, mode, handle);
tools/perf/util/data-convert-bt.c
1400
ADD("machine", env->arch);
tools/perf/util/data-convert-bt.c
1443
if (session->header.env.arch) {
tools/perf/util/data-convert-bt.c
1445
session->header.env.arch);
tools/perf/util/data-convert-json.c
301
output_json_key_string(out, true, 2, "arch", env->arch);
tools/perf/util/disasm.c
103
arch->nr_instructions++;
tools/perf/util/disasm.c
105
ins__sort(arch);
tools/perf/util/disasm.c
124
const struct arch *const *arch = archp;
tools/perf/util/disasm.c
126
return e_machine_and_eflags__cmp(key, &(*arch)->id);
tools/perf/util/disasm.c
131
const struct arch *const *aa = a;
tools/perf/util/disasm.c
132
const struct arch *const *ab = b;
tools/perf/util/disasm.c
137
const struct arch *arch__find(uint16_t e_machine, uint32_t e_flags, const char *cpuid)
tools/perf/util/disasm.c
139
static const struct arch *(*const arch_new_fn[])(const struct e_machine_and_e_flags *id,
tools/perf/util/disasm.c
156
static const struct arch **archs;
tools/perf/util/disasm.c
1590
if (arch__is_powerpc(args->arch)) {
tools/perf/util/disasm.c
162
const struct arch *result = NULL, **tmp;
tools/perf/util/disasm.c
195
bool arch__is_x86(const struct arch *arch)
tools/perf/util/disasm.c
197
return arch->id.e_machine == EM_386 || arch->id.e_machine == EM_X86_64;
tools/perf/util/disasm.c
200
bool arch__is_powerpc(const struct arch *arch)
tools/perf/util/disasm.c
202
return arch->id.e_machine == EM_PPC || arch->id.e_machine == EM_PPC64;
tools/perf/util/disasm.c
230
bool ins__is_fused(const struct arch *arch, const char *ins1, const char *ins2)
tools/perf/util/disasm.c
232
if (!arch || !arch->ins_is_fused)
tools/perf/util/disasm.c
235
return arch->ins_is_fused(arch, ins1, ins2);
tools/perf/util/disasm.c
238
static int call__parse(const struct arch *arch, struct ins_operands *ops, struct map_symbol *ms,
tools/perf/util/disasm.c
253
if (arch->objdump.skip_functions_char &&
tools/perf/util/disasm.c
254
strchr(name, arch->objdump.skip_functions_char))
tools/perf/util/disasm.c
337
static int jump__parse(const struct arch *arch, struct ins_operands *ops, struct map_symbol *ms,
tools/perf/util/disasm.c
348
ops->jump.raw_comment = strchr(ops->raw, arch->objdump.comment_char);
tools/perf/util/disasm.c
501
static int lock__parse(const struct arch *arch, struct ins_operands *ops, struct map_symbol *ms,
tools/perf/util/disasm.c
51
static void ins__sort(struct arch *arch);
tools/perf/util/disasm.c
511
ops->locked.ins.ops = ins__find(arch, ops->locked.ins.name, 0);
tools/perf/util/disasm.c
517
ops->locked.ins.ops->parse(arch, ops->locked.ops, ms, NULL) < 0)
tools/perf/util/disasm.c
568
static bool check_multi_regs(const struct arch *arch, const char *op)
tools/perf/util/disasm.c
572
if (arch->objdump.register_char == 0)
tools/perf/util/disasm.c
575
if (arch->objdump.memory_ref_char) {
tools/perf/util/disasm.c
576
op = strchr(op, arch->objdump.memory_ref_char);
tools/perf/util/disasm.c
581
while ((op = strchr(op, arch->objdump.register_char)) != NULL) {
tools/perf/util/disasm.c
589
static int mov__parse(const struct arch *arch, struct ins_operands *ops,
tools/perf/util/disasm.c
60
static int arch__grow_instructions(struct arch *arch)
tools/perf/util/disasm.c
618
ops->source.multi_regs = check_multi_regs(arch, ops->source.raw);
tools/perf/util/disasm.c
621
comment = strchr(s, arch->objdump.comment_char);
tools/perf/util/disasm.c
640
ops->target.multi_regs = check_multi_regs(arch, ops->target.raw);
tools/perf/util/disasm.c
65
if (arch->nr_instructions_allocated == 0 && arch->instructions)
tools/perf/util/disasm.c
669
static int dec__parse(const struct arch *arch __maybe_unused, struct ins_operands *ops,
tools/perf/util/disasm.c
68
new_nr_allocated = arch->nr_instructions_allocated + 128;
tools/perf/util/disasm.c
688
comment = strchr(s, arch->objdump.comment_char);
tools/perf/util/disasm.c
69
new_instructions = realloc((void *)arch->instructions,
tools/perf/util/disasm.c
75
arch->instructions = new_instructions;
tools/perf/util/disasm.c
754
static void ins__sort(struct arch *arch)
tools/perf/util/disasm.c
756
const int nmemb = arch->nr_instructions;
tools/perf/util/disasm.c
758
qsort((void *)arch->instructions, nmemb, sizeof(struct ins), ins__cmp);
tools/perf/util/disasm.c
76
arch->nr_instructions_allocated = new_nr_allocated;
tools/perf/util/disasm.c
761
static const struct ins_ops *__ins__find(const struct arch *arch, const char *name,
tools/perf/util/disasm.c
765
const int nmemb = arch->nr_instructions;
tools/perf/util/disasm.c
767
if (arch__is_powerpc(arch)) {
tools/perf/util/disasm.c
779
if (!arch->sorted_instructions) {
tools/perf/util/disasm.c
780
ins__sort((struct arch *)arch);
tools/perf/util/disasm.c
781
((struct arch *)arch)->sorted_instructions = true;
tools/perf/util/disasm.c
784
ins = bsearch(name, arch->instructions, nmemb, sizeof(struct ins), ins__key_cmp);
tools/perf/util/disasm.c
788
if (arch->insn_suffix) {
tools/perf/util/disasm.c
797
if (strchr(arch->insn_suffix, suffix) == NULL)
tools/perf/util/disasm.c
80
new_nr_allocated = arch->nr_instructions + 128;
tools/perf/util/disasm.c
803
ins = bsearch(tmp, arch->instructions, nmemb, sizeof(struct ins), ins__key_cmp);
tools/perf/util/disasm.c
808
const struct ins_ops *ins__find(const struct arch *arch, const char *name, struct disasm_line *dl)
tools/perf/util/disasm.c
810
const struct ins_ops *ops = __ins__find(arch, name, dl);
tools/perf/util/disasm.c
812
if (!ops && arch->associate_instruction_ops)
tools/perf/util/disasm.c
813
ops = arch->associate_instruction_ops((struct arch *)arch, name);
tools/perf/util/disasm.c
818
static void disasm_line__init_ins(struct disasm_line *dl, const struct arch *arch,
tools/perf/util/disasm.c
821
dl->ins.ops = ins__find(arch, dl->ins.name, dl);
tools/perf/util/disasm.c
826
if (dl->ins.ops->parse && dl->ins.ops->parse(arch, &dl->ops, ms, dl) < 0)
tools/perf/util/disasm.c
85
memcpy(new_instructions, arch->instructions, arch->nr_instructions * sizeof(struct ins));
tools/perf/util/disasm.c
89
int arch__associate_ins_ops(struct arch *arch, const char *name, const struct ins_ops *ops)
tools/perf/util/disasm.c
93
if (arch->nr_instructions == arch->nr_instructions_allocated &&
tools/perf/util/disasm.c
94
arch__grow_instructions(arch))
tools/perf/util/disasm.c
956
if (arch__is_powerpc(args->arch)) {
tools/perf/util/disasm.c
962
disasm_line__init_ins(dl, args->arch, args->ms);
tools/perf/util/disasm.c
97
ins = (struct ins *)&arch->instructions[arch->nr_instructions];
tools/perf/util/disasm.h
102
const struct arch *arch;
tools/perf/util/disasm.h
111
const struct arch *arch__find(uint16_t e_machine, uint32_t e_flags, const char *cpuid);
tools/perf/util/disasm.h
112
bool arch__is_x86(const struct arch *arch);
tools/perf/util/disasm.h
113
bool arch__is_powerpc(const struct arch *arch);
tools/perf/util/disasm.h
123
int arch__associate_ins_ops(struct arch *arch, const char *name, const struct ins_ops *ops);
tools/perf/util/disasm.h
125
const struct arch *arch__new_arc(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
126
const struct arch *arch__new_arm(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
127
const struct arch *arch__new_arm64(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
128
const struct arch *arch__new_csky(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
129
const struct arch *arch__new_loongarch(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
130
const struct arch *arch__new_mips(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
131
const struct arch *arch__new_powerpc(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
132
const struct arch *arch__new_riscv64(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
133
const struct arch *arch__new_s390(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
134
const struct arch *arch__new_sparc(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
135
const struct arch *arch__new_x86(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
137
const struct ins_ops *ins__find(const struct arch *arch, const char *name, struct disasm_line *dl);
tools/perf/util/disasm.h
141
bool ins__is_fused(const struct arch *arch, const char *ins1, const char *ins2);
tools/perf/util/disasm.h
44
bool (*ins_is_fused)(const struct arch *arch, const char *ins1,
tools/perf/util/disasm.h
46
const struct ins_ops *(*associate_instruction_ops)(struct arch *arch, const char *name);
tools/perf/util/disasm.h
93
int (*parse)(const struct arch *arch, struct ins_operands *ops, struct map_symbol *ms,
tools/perf/util/env.c
256
zfree(&env->arch);
tools/perf/util/env.c
312
const char *arch = perf_env__raw_arch(env);
tools/perf/util/env.c
314
if (!strncmp(arch, "x86_64", 6) || !strncmp(arch, "aarch64", 7) ||
tools/perf/util/env.c
315
!strncmp(arch, "arm64", 5) || !strncmp(arch, "mips64", 6) ||
tools/perf/util/env.c
316
!strncmp(arch, "parisc64", 8) || !strncmp(arch, "riscv64", 7) ||
tools/perf/util/env.c
317
!strncmp(arch, "s390x", 5) || !strncmp(arch, "sparc64", 7))
tools/perf/util/env.c
447
if (env->arch)
tools/perf/util/env.c
451
env->arch = strdup(uts.machine);
tools/perf/util/env.c
453
return env->arch ? 0 : -ENOMEM;
tools/perf/util/env.c
576
return env && !perf_env__read_arch(env) ? env->arch : "unknown";
tools/perf/util/env.c
595
static const char *normalize_arch(char *arch)
tools/perf/util/env.c
597
if (!strcmp(arch, "x86_64"))
tools/perf/util/env.c
599
if (arch[0] == 'i' && arch[2] == '8' && arch[3] == '6')
tools/perf/util/env.c
601
if (!strcmp(arch, "sun4u") || !strncmp(arch, "sparc", 5))
tools/perf/util/env.c
603
if (!strncmp(arch, "aarch64", 7) || !strncmp(arch, "arm64", 5))
tools/perf/util/env.c
605
if (!strncmp(arch, "arm", 3) || !strcmp(arch, "sa110"))
tools/perf/util/env.c
607
if (!strncmp(arch, "s390", 4))
tools/perf/util/env.c
609
if (!strncmp(arch, "parisc", 6))
tools/perf/util/env.c
611
if (!strncmp(arch, "powerpc", 7) || !strncmp(arch, "ppc", 3))
tools/perf/util/env.c
613
if (!strncmp(arch, "mips", 4))
tools/perf/util/env.c
615
if (!strncmp(arch, "sh", 2) && isdigit(arch[2]))
tools/perf/util/env.c
617
if (!strncmp(arch, "loongarch", 9))
tools/perf/util/env.c
620
return arch;
tools/perf/util/env.c
627
if (!env || !env->arch) { /* Assume local operation */
tools/perf/util/env.c
633
arch_name = env->arch;
tools/perf/util/env.h
76
char *arch;
tools/perf/util/header.c
1800
fprintf(fp, "# arch : %s\n", ff->ph->env.arch);
tools/perf/util/header.c
2632
FEAT_PROCESS_STR_FUN(arch, arch);
tools/perf/util/header.c
3736
FEAT_OPR(ARCH, arch, false),
tools/perf/util/libbfd.c
543
info.arch = bfd_get_arch(bfdf);
tools/perf/util/libbfd.c
574
disassemble = disassembler(info.arch,
tools/perf/util/llvm.c
149
if (arch__is_x86(args->arch)) {
tools/perf/util/llvm.c
158
args->arch->name);
tools/perf/util/lock-contention.c
107
const char *arch = perf_env__arch(machine->env);
tools/perf/util/lock-contention.c
128
if (!callstack || (!callstack[i] && (strcmp(arch, "powerpc") ||
tools/perf/util/machine.c
3148
bool machine__is(struct machine *machine, const char *arch)
tools/perf/util/machine.c
3150
return machine && !strcmp(perf_env__raw_arch(machine->env), arch);
tools/perf/util/machine.c
3153
bool machine__normalized_is(struct machine *machine, const char *arch)
tools/perf/util/machine.c
3155
return machine && !strcmp(perf_env__arch(machine->env), arch);
tools/perf/util/machine.h
230
bool machine__is(struct machine *machine, const char *arch);
tools/perf/util/machine.h
231
bool machine__normalized_is(struct machine *machine, const char *arch);
tools/perf/util/map.c
68
const char *arch;
tools/perf/util/map.c
83
arch = !strncmp(app_abi, "arm", 3) ? "arm" :
tools/perf/util/map.c
87
if (!arch)
tools/perf/util/map.c
92
+ strlen(arch);
tools/perf/util/map.c
98
ndk_length, ndk, app_length, app, arch, libname);
tools/perf/util/sample.h
73
u8 arch:1, /* architecture (isa) */
tools/perf/util/sort.c
186
if (left->simd_flags.arch != right->simd_flags.arch)
tools/perf/util/sort.c
187
return (int64_t) left->simd_flags.arch - right->simd_flags.arch;
tools/perf/util/sort.c
194
u64 arch = simd_flags->arch;
tools/perf/util/sort.c
196
if (arch & SIMD_OP_FLAGS_ARCH_SVE)
tools/perf/util/sort.c
207
if (!he->simd_flags.arch)
tools/perf/util/sort.c
2542
const char *arch = perf_env__arch(env);
tools/perf/util/sort.c
2544
if (!strcmp("x86", arch) || !strcmp("powerpc", arch)) {
tools/perf/util/sort.c
2555
const char *arch = perf_env__arch(env);
tools/perf/util/sort.c
2557
if (!strcmp("x86", arch)) {
tools/perf/util/sort.c
2562
} else if (!strcmp("powerpc", arch)) {
tools/perf/util/unwind-libunwind.c
17
const char *arch;
tools/perf/util/unwind-libunwind.c
36
if (!machine->env || !machine->env->arch)
tools/perf/util/unwind-libunwind.c
43
arch = perf_env__arch(machine->env);
tools/perf/util/unwind-libunwind.c
45
if (!strcmp(arch, "x86")) {
tools/perf/util/unwind-libunwind.c
48
} else if (!strcmp(arch, "arm64") || !strcmp(arch, "arm")) {
tools/perf/util/unwind-libunwind.c
54
pr_warning_once("unwind: target platform=%s is not supported\n", arch);
tools/testing/selftests/arm64/abi/ptrace.c
139
int slots, arch, ret;
tools/testing/selftests/arm64/abi/ptrace.c
151
arch = (state.dbg_info >> 8) & 0xf;
tools/testing/selftests/arm64/abi/ptrace.c
154
arch, slots);
tools/testing/selftests/arm64/abi/ptrace.c
157
ksft_test_result(arch, "%s_arch_set\n", type_name);
tools/testing/selftests/bpf/progs/bpf_misc.h
155
#define __arch(arch) __test_tag("test_arch=" arch)
tools/testing/selftests/bpf/test_loader.c
422
enum arch arch;
tools/testing/selftests/bpf/test_loader.c
575
arch = ARCH_X86_64;
tools/testing/selftests/bpf/test_loader.c
577
arch = ARCH_ARM64;
tools/testing/selftests/bpf/test_loader.c
579
arch = ARCH_RISCV64;
tools/testing/selftests/bpf/test_loader.c
581
arch = ARCH_S390X;
tools/testing/selftests/bpf/test_loader.c
587
arch_mask |= arch;
tools/testing/selftests/bpf/test_loader.c
588
collect_jit = get_current_arch() == arch;
tools/testing/selftests/kvm/include/kvm_util.h
126
struct kvm_vm_arch arch;
tools/testing/selftests/kvm/include/kvm_util.h
96
struct kvm_mmu_arch arch;
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
46
static inline bool __vm_arch_has_protected_memory(struct kvm_vm_arch *arch)
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
48
return arch->c_bit || arch->s_bit;
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
52
__vm_arch_has_protected_memory(&(vm)->arch)
tools/testing/selftests/kvm/include/x86/processor.h
1480
#define PTE_PRESENT_MASK(mmu) ((mmu)->arch.pte_masks.present)
tools/testing/selftests/kvm/include/x86/processor.h
1481
#define PTE_WRITABLE_MASK(mmu) ((mmu)->arch.pte_masks.writable)
tools/testing/selftests/kvm/include/x86/processor.h
1482
#define PTE_USER_MASK(mmu) ((mmu)->arch.pte_masks.user)
tools/testing/selftests/kvm/include/x86/processor.h
1483
#define PTE_READABLE_MASK(mmu) ((mmu)->arch.pte_masks.readable)
tools/testing/selftests/kvm/include/x86/processor.h
1484
#define PTE_EXECUTABLE_MASK(mmu) ((mmu)->arch.pte_masks.executable)
tools/testing/selftests/kvm/include/x86/processor.h
1485
#define PTE_ACCESSED_MASK(mmu) ((mmu)->arch.pte_masks.accessed)
tools/testing/selftests/kvm/include/x86/processor.h
1486
#define PTE_DIRTY_MASK(mmu) ((mmu)->arch.pte_masks.dirty)
tools/testing/selftests/kvm/include/x86/processor.h
1487
#define PTE_HUGE_MASK(mmu) ((mmu)->arch.pte_masks.huge)
tools/testing/selftests/kvm/include/x86/processor.h
1488
#define PTE_NX_MASK(mmu) ((mmu)->arch.pte_masks.nx)
tools/testing/selftests/kvm/include/x86/processor.h
1489
#define PTE_C_BIT_MASK(mmu) ((mmu)->arch.pte_masks.c)
tools/testing/selftests/kvm/include/x86/processor.h
1490
#define PTE_S_BIT_MASK(mmu) ((mmu)->arch.pte_masks.s)
tools/testing/selftests/kvm/include/x86/processor.h
1491
#define PTE_ALWAYS_SET_MASK(mmu) ((mmu)->arch.pte_masks.always_set)
tools/testing/selftests/kvm/include/x86/sev.h
89
.sev_fd = (vm)->arch.sev_fd, \
tools/testing/selftests/kvm/lib/arm64/processor.c
283
return vm_check_cap(vm, KVM_CAP_ARM_EL2) && vm->arch.has_gic;
tools/testing/selftests/kvm/lib/arm64/processor.c
709
vm->arch.gic_fd = __vgic_v3_setup(vm, nr_vcpus, 64);
tools/testing/selftests/kvm/lib/arm64/processor.c
710
vm->arch.has_gic = true;
tools/testing/selftests/kvm/lib/arm64/processor.c
716
if (vm->arch.has_gic)
tools/testing/selftests/kvm/lib/arm64/processor.c
717
__vgic_v3_init(vm->arch.gic_fd);
tools/testing/selftests/kvm/lib/arm64/processor.c
722
if (vm->arch.has_gic)
tools/testing/selftests/kvm/lib/arm64/processor.c
723
close(vm->arch.gic_fd);
tools/testing/selftests/kvm/lib/x86/processor.c
1321
vm->arch.sev_fd = open_sev_dev_path_or_exit();
tools/testing/selftests/kvm/lib/x86/processor.c
1322
vm->arch.c_bit = BIT_ULL(this_cpu_property(X86_PROPERTY_SEV_C_BIT));
tools/testing/selftests/kvm/lib/x86/processor.c
1323
vm->gpa_tag_mask = vm->arch.c_bit;
tools/testing/selftests/kvm/lib/x86/processor.c
1325
vm->arch.sev_fd = -1;
tools/testing/selftests/kvm/lib/x86/processor.c
171
mmu->arch.pte_masks = *pte_masks;
tools/testing/selftests/kvm/lib/x86/processor.c
193
.c = vm->arch.c_bit,
tools/testing/selftests/kvm/lib/x86/processor.c
194
.s = vm->arch.s_bit,
tools/testing/selftests/kvm/lib/x86/processor.c
366
TEST_ASSERT(!vm->arch.is_pt_protected,
tools/testing/selftests/kvm/lib/x86/processor.c
574
void *gdt = addr_gva2hva(vm, vm->arch.gdt);
tools/testing/selftests/kvm/lib/x86/processor.c
656
sregs.idt.base = vm->arch.idt;
tools/testing/selftests/kvm/lib/x86/processor.c
658
sregs.gdt.base = vm->arch.gdt;
tools/testing/selftests/kvm/lib/x86/processor.c
674
kvm_seg_set_tss_64bit(vm->arch.tss, &sregs.tr);
tools/testing/selftests/kvm/lib/x86/processor.c
698
(struct idt_entry *)addr_gva2hva(vm, vm->arch.idt);
tools/testing/selftests/kvm/lib/x86/processor.c
749
vm->arch.gdt = __vm_vaddr_alloc_page(vm, MEM_REGION_DATA);
tools/testing/selftests/kvm/lib/x86/processor.c
750
vm->arch.idt = __vm_vaddr_alloc_page(vm, MEM_REGION_DATA);
tools/testing/selftests/kvm/lib/x86/processor.c
752
vm->arch.tss = __vm_vaddr_alloc_page(vm, MEM_REGION_DATA);
tools/testing/selftests/kvm/lib/x86/processor.c
766
kvm_seg_set_tss_64bit(vm->arch.tss, &seg);
tools/testing/selftests/kvm/lib/x86/sev.c
103
vm->arch.is_pt_protected = true;
tools/testing/selftests/kvm/lib/x86/sev.c
151
vm->arch.is_pt_protected = true;
tools/testing/selftests/kvm/lib/x86/sev.c
51
TEST_ASSERT_EQ(vm->arch.sev_fd, -1);
tools/testing/selftests/kvm/lib/x86/sev.c
52
vm->arch.sev_fd = open_sev_dev_path_or_exit();
tools/testing/selftests/kvm/lib/x86/sev.c
64
TEST_ASSERT_EQ(vm->arch.sev_fd, -1);
tools/testing/selftests/kvm/lib/x86/sev.c
65
vm->arch.sev_fd = open_sev_dev_path_or_exit();
tools/testing/selftests/kvm/lib/x86/svm.c
75
pte_masks = vm->mmu.arch.pte_masks;
tools/testing/selftests/kvm/x86/debug_regs.c
112
run->debug.arch.exception == BP_VECTOR &&
tools/testing/selftests/kvm/x86/debug_regs.c
113
run->debug.arch.pc == CAST_TO_RIP(sw_bp),
tools/testing/selftests/kvm/x86/debug_regs.c
115
run->exit_reason, run->debug.arch.exception,
tools/testing/selftests/kvm/x86/debug_regs.c
116
run->debug.arch.pc, CAST_TO_RIP(sw_bp));
tools/testing/selftests/kvm/x86/debug_regs.c
123
debug.arch.debugreg[i] = CAST_TO_RIP(hw_bp);
tools/testing/selftests/kvm/x86/debug_regs.c
124
debug.arch.debugreg[7] = 0x400 | (1UL << (2*i+1));
tools/testing/selftests/kvm/x86/debug_regs.c
129
run->debug.arch.exception == DB_VECTOR &&
tools/testing/selftests/kvm/x86/debug_regs.c
130
run->debug.arch.pc == CAST_TO_RIP(hw_bp) &&
tools/testing/selftests/kvm/x86/debug_regs.c
131
run->debug.arch.dr6 == target_dr6,
tools/testing/selftests/kvm/x86/debug_regs.c
134
i, run->exit_reason, run->debug.arch.exception,
tools/testing/selftests/kvm/x86/debug_regs.c
135
run->debug.arch.pc, CAST_TO_RIP(hw_bp),
tools/testing/selftests/kvm/x86/debug_regs.c
136
run->debug.arch.dr6, target_dr6);
tools/testing/selftests/kvm/x86/debug_regs.c
145
debug.arch.debugreg[i] = CAST_TO_RIP(guest_value);
tools/testing/selftests/kvm/x86/debug_regs.c
146
debug.arch.debugreg[7] = 0x00000400 | (1UL << (2*i+1)) |
tools/testing/selftests/kvm/x86/debug_regs.c
152
run->debug.arch.exception == DB_VECTOR &&
tools/testing/selftests/kvm/x86/debug_regs.c
153
run->debug.arch.pc == CAST_TO_RIP(write_data) &&
tools/testing/selftests/kvm/x86/debug_regs.c
154
run->debug.arch.dr6 == target_dr6,
tools/testing/selftests/kvm/x86/debug_regs.c
157
i, run->exit_reason, run->debug.arch.exception,
tools/testing/selftests/kvm/x86/debug_regs.c
158
run->debug.arch.pc, CAST_TO_RIP(write_data),
tools/testing/selftests/kvm/x86/debug_regs.c
159
run->debug.arch.dr6, target_dr6);
tools/testing/selftests/kvm/x86/debug_regs.c
174
debug.arch.debugreg[7] = 0x00000400;
tools/testing/selftests/kvm/x86/debug_regs.c
178
run->debug.arch.exception == DB_VECTOR &&
tools/testing/selftests/kvm/x86/debug_regs.c
179
run->debug.arch.pc == target_rip &&
tools/testing/selftests/kvm/x86/debug_regs.c
180
run->debug.arch.dr6 == target_dr6,
tools/testing/selftests/kvm/x86/debug_regs.c
183
i, run->exit_reason, run->debug.arch.exception,
tools/testing/selftests/kvm/x86/debug_regs.c
184
run->debug.arch.pc, target_rip, run->debug.arch.dr6,
tools/testing/selftests/kvm/x86/debug_regs.c
191
debug.arch.debugreg[7] = 0x400 | DR7_GD;
tools/testing/selftests/kvm/x86/debug_regs.c
196
run->debug.arch.exception == DB_VECTOR &&
tools/testing/selftests/kvm/x86/debug_regs.c
197
run->debug.arch.pc == CAST_TO_RIP(bd_start) &&
tools/testing/selftests/kvm/x86/debug_regs.c
198
run->debug.arch.dr6 == target_dr6,
tools/testing/selftests/kvm/x86/debug_regs.c
201
run->exit_reason, run->debug.arch.exception,
tools/testing/selftests/kvm/x86/debug_regs.c
202
run->debug.arch.pc, target_rip, run->debug.arch.dr6,
tools/testing/selftests/kvm/x86/svm_nested_shutdown_test.c
53
vcpu_args_set(vcpu, 2, svm_gva, vm->arch.idt);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
166
idt = addr_gva2hva(vm, vm->arch.idt);
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
139
struct arch_vcpu_info arch;
tools/testing/selftests/ptrace/get_syscall_info.c
166
ASSERT_TRUE(info.arch) {
tools/testing/selftests/ptrace/get_syscall_info.c
194
ASSERT_TRUE(info.arch) {
tools/testing/selftests/ptrace/get_syscall_info.c
234
ASSERT_TRUE(info.arch) {
tools/testing/selftests/ptrace/set_syscall_info.c
112
ASSERT_TRUE(info->arch) {
tools/testing/selftests/ptrace/set_syscall_info.c
83
ASSERT_TRUE(info->arch) {
tools/testing/selftests/rseq/rseq-abi.h
148
} arch;
tools/testing/selftests/rseq/rseq-arm-bits.h
120
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm-bits.h
182
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm-bits.h
252
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm-bits.h
340
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm-bits.h
463
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm-bits.h
49
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm64-bits.h
102
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm64-bits.h
157
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm64-bits.h
213
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm64-bits.h
289
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm64-bits.h
356
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-arm64-bits.h
41
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-mips-bits.h
111
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-mips-bits.h
168
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-mips-bits.h
231
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-mips-bits.h
311
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-mips-bits.h
425
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-mips-bits.h
47
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-or1k-bits.h
142
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-or1k-bits.h
195
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-or1k-bits.h
254
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-or1k-bits.h
316
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-or1k-bits.h
35
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-or1k-bits.h
378
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-or1k-bits.h
92
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-ppc-bits.h
118
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-ppc-bits.h
181
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-ppc-bits.h
247
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-ppc-bits.h
333
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-ppc-bits.h
415
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-ppc-bits.h
47
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-riscv-bits.h
141
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-riscv-bits.h
194
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-riscv-bits.h
252
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-riscv-bits.h
314
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-riscv-bits.h
34
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-riscv-bits.h
376
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-riscv-bits.h
91
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-s390-bits.h
113
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-s390-bits.h
173
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-s390-bits.h
238
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-s390-bits.h
320
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-s390-bits.h
40
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq-s390-bits.h
432
[rseq_cs] "m" (rseq_get_abi()->rseq_cs.arch.ptr),
tools/testing/selftests/rseq/rseq.h
228
RSEQ_WRITE_ONCE(rseq_get_abi()->rseq_cs.arch.ptr, 0);
tools/testing/selftests/seccomp/seccomp_bpf.c
113
__u32 arch;
virt/kvm/async_pf.c
180
unsigned long hva, struct kvm_arch_async_pf *arch)
virt/kvm/async_pf.c
203
work->arch = *arch;
virt/kvm/kvm_main.c
1791
dest->arch = src->arch;
virt/kvm/kvm_main.c
1837
old->arch = invalid_slot->arch;
virt/kvm/kvm_main.c
4055
page = virt_to_page(vcpu->arch.pio_data);
virt/kvm/kvm_main.c
6497
offsetof(struct kvm_vcpu, arch),
virt/kvm/kvm_main.c
6499
- offsetof(struct kvm_vcpu, arch),