Symbol: guest
arch/arm64/kvm/hyp/include/hyp/switch.h
911
goto guest;
arch/arm64/kvm/hyp/include/hyp/switch.h
916
guest:
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1204
enum pkvm_page_state guest[2]; /* [ gfn, gfn + 1 ] */
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1271
WARN_ON(__guest_check_page_state_range(vm, ipa[0], size, selftest_state.guest[0]));
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1272
WARN_ON(__guest_check_page_state_range(vm, ipa[1], size, selftest_state.guest[1]));
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1302
selftest_state.guest[0] = selftest_state.guest[1] = PKVM_NOPAGE;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1370
selftest_state.guest[0] = PKVM_PAGE_SHARED_BORROWED;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1380
selftest_state.guest[1] = PKVM_PAGE_SHARED_BORROWED;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1384
selftest_state.guest[0] = PKVM_NOPAGE;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1387
selftest_state.guest[1] = PKVM_NOPAGE;
arch/mips/include/asm/cpu-features.h
665
#define cpu_guest_has_conf1 (cpu_data[0].guest.conf & (1 << 1))
arch/mips/include/asm/cpu-features.h
668
#define cpu_guest_has_conf2 (cpu_data[0].guest.conf & (1 << 2))
arch/mips/include/asm/cpu-features.h
671
#define cpu_guest_has_conf3 (cpu_data[0].guest.conf & (1 << 3))
arch/mips/include/asm/cpu-features.h
674
#define cpu_guest_has_conf4 (cpu_data[0].guest.conf & (1 << 4))
arch/mips/include/asm/cpu-features.h
677
#define cpu_guest_has_conf5 (cpu_data[0].guest.conf & (1 << 5))
arch/mips/include/asm/cpu-features.h
680
#define cpu_guest_has_conf6 (cpu_data[0].guest.conf & (1 << 6))
arch/mips/include/asm/cpu-features.h
683
#define cpu_guest_has_conf7 (cpu_data[0].guest.conf & (1 << 7))
arch/mips/include/asm/cpu-features.h
686
#define cpu_guest_has_fpu (cpu_data[0].guest.options & MIPS_CPU_FPU)
arch/mips/include/asm/cpu-features.h
689
#define cpu_guest_has_watch (cpu_data[0].guest.options & MIPS_CPU_WATCH)
arch/mips/include/asm/cpu-features.h
692
#define cpu_guest_has_contextconfig (cpu_data[0].guest.options & MIPS_CPU_CTXTC)
arch/mips/include/asm/cpu-features.h
695
#define cpu_guest_has_segments (cpu_data[0].guest.options & MIPS_CPU_SEGMENTS)
arch/mips/include/asm/cpu-features.h
698
#define cpu_guest_has_badinstr (cpu_data[0].guest.options & MIPS_CPU_BADINSTR)
arch/mips/include/asm/cpu-features.h
701
#define cpu_guest_has_badinstrp (cpu_data[0].guest.options & MIPS_CPU_BADINSTRP)
arch/mips/include/asm/cpu-features.h
704
#define cpu_guest_has_htw (cpu_data[0].guest.options & MIPS_CPU_HTW)
arch/mips/include/asm/cpu-features.h
707
#define cpu_guest_has_ldpte (cpu_data[0].guest.options & MIPS_CPU_LDPTE)
arch/mips/include/asm/cpu-features.h
710
#define cpu_guest_has_mvh (cpu_data[0].guest.options & MIPS_CPU_MVH)
arch/mips/include/asm/cpu-features.h
713
#define cpu_guest_has_msa (cpu_data[0].guest.ases & MIPS_ASE_MSA)
arch/mips/include/asm/cpu-features.h
716
#define cpu_guest_has_kscr(n) (cpu_data[0].guest.kscratch_mask & (1u << (n)))
arch/mips/include/asm/cpu-features.h
719
#define cpu_guest_has_rw_llb (cpu_has_mips_r6 || (cpu_data[0].guest.options & MIPS_CPU_RW_LLB))
arch/mips/include/asm/cpu-features.h
722
#define cpu_guest_has_perf (cpu_data[0].guest.options & MIPS_CPU_PERF)
arch/mips/include/asm/cpu-features.h
725
#define cpu_guest_has_maar (cpu_data[0].guest.options & MIPS_CPU_MAAR)
arch/mips/include/asm/cpu-features.h
728
#define cpu_guest_has_userlocal (cpu_data[0].guest.options & MIPS_CPU_ULRI)
arch/mips/include/asm/cpu-features.h
735
#define cpu_guest_has_dyn_fpu (cpu_data[0].guest.options_dyn & MIPS_CPU_FPU)
arch/mips/include/asm/cpu-features.h
738
#define cpu_guest_has_dyn_watch (cpu_data[0].guest.options_dyn & MIPS_CPU_WATCH)
arch/mips/include/asm/cpu-features.h
741
#define cpu_guest_has_dyn_contextconfig (cpu_data[0].guest.options_dyn & MIPS_CPU_CTXTC)
arch/mips/include/asm/cpu-features.h
744
#define cpu_guest_has_dyn_perf (cpu_data[0].guest.options_dyn & MIPS_CPU_PERF)
arch/mips/include/asm/cpu-features.h
747
#define cpu_guest_has_dyn_msa (cpu_data[0].guest.ases_dyn & MIPS_ASE_MSA)
arch/mips/include/asm/cpu-features.h
750
#define cpu_guest_has_dyn_maar (cpu_data[0].guest.options_dyn & MIPS_CPU_MAAR)
arch/mips/include/asm/cpu-info.h
102
struct guest_info guest;
arch/mips/kernel/cpu-probe.c
1675
c->guest.options |= MIPS_CPU_LDPTE;
arch/mips/kernel/cpu-probe.c
837
c->guest.conf |= BIT(1);
arch/mips/kernel/cpu-probe.c
850
c->guest.options |= MIPS_CPU_FPU;
arch/mips/kernel/cpu-probe.c
852
c->guest.options_dyn |= MIPS_CPU_FPU;
arch/mips/kernel/cpu-probe.c
855
c->guest.options |= MIPS_CPU_WATCH;
arch/mips/kernel/cpu-probe.c
857
c->guest.options_dyn |= MIPS_CPU_WATCH;
arch/mips/kernel/cpu-probe.c
860
c->guest.options |= MIPS_CPU_PERF;
arch/mips/kernel/cpu-probe.c
862
c->guest.options_dyn |= MIPS_CPU_PERF;
arch/mips/kernel/cpu-probe.c
865
c->guest.conf |= BIT(2);
arch/mips/kernel/cpu-probe.c
876
c->guest.conf |= BIT(3);
arch/mips/kernel/cpu-probe.c
889
c->guest.options |= MIPS_CPU_CTXTC;
arch/mips/kernel/cpu-probe.c
891
c->guest.options_dyn |= MIPS_CPU_CTXTC;
arch/mips/kernel/cpu-probe.c
894
c->guest.options |= MIPS_CPU_HTW;
arch/mips/kernel/cpu-probe.c
897
c->guest.options |= MIPS_CPU_ULRI;
arch/mips/kernel/cpu-probe.c
900
c->guest.options |= MIPS_CPU_SEGMENTS;
arch/mips/kernel/cpu-probe.c
903
c->guest.options |= MIPS_CPU_BADINSTR;
arch/mips/kernel/cpu-probe.c
905
c->guest.options |= MIPS_CPU_BADINSTRP;
arch/mips/kernel/cpu-probe.c
908
c->guest.ases |= MIPS_ASE_MSA;
arch/mips/kernel/cpu-probe.c
910
c->guest.ases_dyn |= MIPS_ASE_MSA;
arch/mips/kernel/cpu-probe.c
913
c->guest.conf |= BIT(4);
arch/mips/kernel/cpu-probe.c
924
c->guest.kscratch_mask = (config4 & MIPS_CONF4_KSCREXIST)
arch/mips/kernel/cpu-probe.c
928
c->guest.conf |= BIT(5);
arch/mips/kernel/cpu-probe.c
940
c->guest.options |= MIPS_CPU_MAAR;
arch/mips/kernel/cpu-probe.c
942
c->guest.options_dyn |= MIPS_CPU_MAAR;
arch/mips/kernel/cpu-probe.c
945
c->guest.options |= MIPS_CPU_RW_LLB;
arch/mips/kernel/cpu-probe.c
948
c->guest.options |= MIPS_CPU_MVH;
arch/mips/kernel/cpu-probe.c
951
c->guest.conf |= BIT(6);
arch/mips/kvm/tlb.c
340
for (entry = 0; entry < current_cpu_data.guest.tlbsize; entry++) {
arch/mips/kvm/vz.c
1801
ret += __arch_hweight8(cpu_data[0].guest.kscratch_mask);
arch/mips/kvm/vz.c
2902
current_cpu_data.guest.tlbsize = guest_mmu_size;
arch/mips/kvm/vz.c
2918
current_cpu_data.guest.tlbsize = guest_mmu_size + ftlb_size;
arch/mips/kvm/vz.c
2928
current_cpu_data.guest.tlbsize = guest_mmu_size + ftlb_size;
arch/mips/kvm/vz.c
3011
current_cpu_data.guest.tlbsize = 0;
arch/s390/kernel/vtime.c
119
u64 timer, clock, user, guest, system, hardirq, softirq;
arch/s390/kernel/vtime.c
144
guest = update_tsk_timer(&tsk->thread.guest_timer, lc->guest_timer);
arch/s390/kernel/vtime.c
148
lc->steal_timer += clock - user - guest - system - hardirq - softirq;
arch/s390/kernel/vtime.c
156
if (guest) {
arch/s390/kernel/vtime.c
157
account_guest_time(tsk, cputime_to_nsecs(guest));
arch/s390/kernel/vtime.c
158
tsk->utimescaled += cputime_to_nsecs(scale_vtime(guest));
arch/s390/kernel/vtime.c
168
return virt_timer_forward(user + guest + system + hardirq + softirq);
arch/x86/coco/tdx/tdx.c
1162
x86_platform.guest.enc_status_change_prepare = tdx_enc_status_change_prepare;
arch/x86/coco/tdx/tdx.c
1163
x86_platform.guest.enc_status_change_finish = tdx_enc_status_change_finish;
arch/x86/coco/tdx/tdx.c
1165
x86_platform.guest.enc_cache_flush_required = tdx_cache_flush_required;
arch/x86/coco/tdx/tdx.c
1166
x86_platform.guest.enc_tlb_flush_required = tdx_tlb_flush_required;
arch/x86/coco/tdx/tdx.c
1168
x86_platform.guest.enc_kexec_begin = tdx_kexec_begin;
arch/x86/coco/tdx/tdx.c
1169
x86_platform.guest.enc_kexec_finish = tdx_kexec_finish;
arch/x86/events/intel/core.c
5014
.guest = intel_ctrl & ~cpuc->intel_ctrl_host_mask & ~pebs_mask,
arch/x86/events/intel/core.c
5032
.guest = 0,
arch/x86/events/intel/core.c
5043
.guest = kvm_pmu->ds_area,
arch/x86/events/intel/core.c
5050
.guest = kvm_pmu->pebs_data_cfg,
arch/x86/events/intel/core.c
5058
.guest = pebs_mask & ~cpuc->intel_ctrl_host_mask & kvm_pmu->pebs_enable,
arch/x86/events/intel/core.c
5063
arr[pebs_enable].guest = 0;
arch/x86/events/intel/core.c
5066
arr[pebs_enable].guest &= ~kvm_pmu->host_cross_mapped_mask;
arch/x86/events/intel/core.c
5067
arr[global_ctrl].guest &= ~kvm_pmu->host_cross_mapped_mask;
arch/x86/events/intel/core.c
5069
arr[global_ctrl].guest |= arr[pebs_enable].guest;
arch/x86/events/intel/core.c
5085
arr[idx].host = arr[idx].guest = 0;
arch/x86/events/intel/core.c
5090
arr[idx].host = arr[idx].guest =
arch/x86/events/intel/core.c
5096
arr[idx].guest &= ~ARCH_PERFMON_EVENTSEL_ENABLE;
arch/x86/hyperv/ivm.c
891
x86_platform.guest.enc_cache_flush_required = hv_vtom_cache_flush_required;
arch/x86/hyperv/ivm.c
892
x86_platform.guest.enc_tlb_flush_required = hv_vtom_tlb_flush_required;
arch/x86/hyperv/ivm.c
893
x86_platform.guest.enc_status_change_prepare = hv_vtom_clear_present;
arch/x86/hyperv/ivm.c
894
x86_platform.guest.enc_status_change_finish = hv_vtom_set_host_visibility;
arch/x86/hyperv/ivm.c
895
x86_platform.guest.enc_kexec_begin = hv_vtom_kexec_begin;
arch/x86/hyperv/ivm.c
896
x86_platform.guest.enc_kexec_finish = hv_vtom_kexec_finish;
arch/x86/include/asm/perf_event.h
736
u64 host, guest;
arch/x86/include/asm/spec-ctrl.h
17
extern void x86_virt_spec_ctrl(u64 guest_virt_spec_ctrl, bool guest);
arch/x86/include/asm/x86_init.h
332
struct x86_guest guest;
arch/x86/kernel/crash.c
140
x86_platform.guest.enc_kexec_begin();
arch/x86/kernel/crash.c
141
x86_platform.guest.enc_kexec_finish();
arch/x86/kernel/fpu/xstate.c
1668
static int __xstate_request_perm(u64 permitted, u64 requested, bool guest)
arch/x86/kernel/fpu/xstate.c
1702
if (!guest) {
arch/x86/kernel/fpu/xstate.c
1708
perm = guest ? &fpu->guest_perm : &fpu->perm;
arch/x86/kernel/fpu/xstate.c
1724
static int xstate_request_perm(unsigned long idx, bool guest)
arch/x86/kernel/fpu/xstate.c
1745
permitted = xstate_get_group_perm(guest);
arch/x86/kernel/fpu/xstate.c
1751
permitted = xstate_get_group_perm(guest);
arch/x86/kernel/fpu/xstate.c
1754
if (guest && (permitted & FPU_GUEST_PERM_LOCKED))
arch/x86/kernel/fpu/xstate.c
1757
ret = __xstate_request_perm(permitted, requested, guest);
arch/x86/kernel/fpu/xstate.c
1812
static inline int xstate_request_perm(unsigned long idx, bool guest)
arch/x86/kernel/fpu/xstate.c
1846
bool guest = false;
arch/x86/kernel/fpu/xstate.c
1868
guest = true;
arch/x86/kernel/fpu/xstate.c
1875
return xstate_request_perm(idx, guest);
arch/x86/kernel/fpu/xstate.h
24
static inline u64 xstate_get_group_perm(bool guest)
arch/x86/kernel/fpu/xstate.h
30
perm = guest ? &fpu->guest_perm : &fpu->perm;
arch/x86/kernel/reboot.c
727
x86_platform.guest.enc_kexec_begin();
arch/x86/kernel/reboot.c
767
x86_platform.guest.enc_kexec_finish();
arch/x86/kernel/x86_init.c
162
.guest = {
arch/x86/kvm/vmx/nested.c
2335
vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val));
arch/x86/kvm/vmx/nested.c
2685
vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, vmx->msr_autoload.guest.nr);
arch/x86/kvm/vmx/nested.c
4923
for (i = 0; i < vmx->msr_autoload.guest.nr; ++i) {
arch/x86/kvm/vmx/nested.c
4924
if (vmx->msr_autoload.guest.val[i].index == MSR_EFER)
arch/x86/kvm/vmx/nested.c
4925
return vmx->msr_autoload.guest.val[i].value;
arch/x86/kvm/vmx/nested.c
5147
vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, vmx->msr_autoload.guest.nr);
arch/x86/kvm/vmx/nested.c
5424
vmx->pt_desc.guest.ctl = 0;
arch/x86/kvm/vmx/vmx.c
1080
vmx_remove_auto_msr(&m->guest, msr, VM_ENTRY_MSR_LOAD_COUNT);
arch/x86/kvm/vmx/vmx.c
1151
vmx_add_auto_msr(&m->guest, msr, guest_val, VM_ENTRY_MSR_LOAD_COUNT, kvm);
arch/x86/kvm/vmx/vmx.c
1250
!(vmx->pt_desc.guest.ctl & RTIT_CTL_TRACEEN);
arch/x86/kvm/vmx/vmx.c
1297
if (vmx->pt_desc.guest.ctl & RTIT_CTL_TRACEEN) {
arch/x86/kvm/vmx/vmx.c
1300
pt_load_msr(&vmx->pt_desc.guest, vmx->pt_desc.num_address_ranges);
arch/x86/kvm/vmx/vmx.c
1309
if (vmx->pt_desc.guest.ctl & RTIT_CTL_TRACEEN) {
arch/x86/kvm/vmx/vmx.c
1310
pt_save_msr(&vmx->pt_desc.guest, vmx->pt_desc.num_address_ranges);
arch/x86/kvm/vmx/vmx.c
1740
if ((vmx->pt_desc.guest.ctl & RTIT_CTL_TRACEEN) &&
arch/x86/kvm/vmx/vmx.c
1742
data != vmx->pt_desc.guest.ctl)
arch/x86/kvm/vmx/vmx.c
2217
msr_info->data = vmx->pt_desc.guest.ctl;
arch/x86/kvm/vmx/vmx.c
2222
msr_info->data = vmx->pt_desc.guest.status;
arch/x86/kvm/vmx/vmx.c
2229
msr_info->data = vmx->pt_desc.guest.cr3_match;
arch/x86/kvm/vmx/vmx.c
2238
msr_info->data = vmx->pt_desc.guest.output_base;
arch/x86/kvm/vmx/vmx.c
2247
msr_info->data = vmx->pt_desc.guest.output_mask;
arch/x86/kvm/vmx/vmx.c
2255
msr_info->data = vmx->pt_desc.guest.addr_b[index / 2];
arch/x86/kvm/vmx/vmx.c
2257
msr_info->data = vmx->pt_desc.guest.addr_a[index / 2];
arch/x86/kvm/vmx/vmx.c
2533
vmx->pt_desc.guest.ctl = data;
arch/x86/kvm/vmx/vmx.c
2541
vmx->pt_desc.guest.status = data;
arch/x86/kvm/vmx/vmx.c
2549
vmx->pt_desc.guest.cr3_match = data;
arch/x86/kvm/vmx/vmx.c
2561
vmx->pt_desc.guest.output_base = data;
arch/x86/kvm/vmx/vmx.c
2571
vmx->pt_desc.guest.output_mask = data;
arch/x86/kvm/vmx/vmx.c
2582
vmx->pt_desc.guest.addr_b[index / 2] = data;
arch/x86/kvm/vmx/vmx.c
2584
vmx->pt_desc.guest.addr_a[index / 2] = data;
arch/x86/kvm/vmx/vmx.c
4274
bool flag = !(vmx->pt_desc.guest.ctl & RTIT_CTL_TRACEEN);
arch/x86/kvm/vmx/vmx.c
5007
vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val));
arch/x86/kvm/vmx/vmx.c
5038
vmx->pt_desc.guest.output_mask = 0x7F;
arch/x86/kvm/vmx/vmx.c
6638
efer_slot = vmx_find_loadstore_msr_slot(&vmx->msr_autoload.guest, MSR_EFER);
arch/x86/kvm/vmx/vmx.c
6643
vmx->msr_autoload.guest.val[efer_slot].value);
arch/x86/kvm/vmx/vmx.c
6668
vmx_dump_msrs("guest autoload", &vmx->msr_autoload.guest);
arch/x86/kvm/vmx/vmx.c
7435
if (msrs[i].host == msrs[i].guest)
arch/x86/kvm/vmx/vmx.c
7438
add_atomic_switch_msr(vmx, msrs[i].msr, msrs[i].guest,
arch/x86/kvm/vmx/vmx.h
236
struct vmx_msrs guest;
arch/x86/kvm/vmx/vmx.h
67
struct pt_ctx guest;
arch/x86/mm/mem_encrypt_amd.c
489
x86_platform.guest.enc_status_change_prepare = amd_enc_status_change_prepare;
arch/x86/mm/mem_encrypt_amd.c
490
x86_platform.guest.enc_status_change_finish = amd_enc_status_change_finish;
arch/x86/mm/mem_encrypt_amd.c
491
x86_platform.guest.enc_tlb_flush_required = amd_enc_tlb_flush_required;
arch/x86/mm/mem_encrypt_amd.c
492
x86_platform.guest.enc_cache_flush_required = amd_enc_cache_flush_required;
arch/x86/mm/mem_encrypt_amd.c
493
x86_platform.guest.enc_kexec_begin = snp_kexec_begin;
arch/x86/mm/mem_encrypt_amd.c
494
x86_platform.guest.enc_kexec_finish = snp_kexec_finish;
arch/x86/mm/pat/set_memory.c
2385
if (x86_platform.guest.enc_tlb_flush_required(enc))
arch/x86/mm/pat/set_memory.c
2386
cpa_flush(&cpa, x86_platform.guest.enc_cache_flush_required());
arch/x86/mm/pat/set_memory.c
2389
ret = x86_platform.guest.enc_status_change_prepare(addr, numpages, enc);
arch/x86/mm/pat/set_memory.c
2408
ret = x86_platform.guest.enc_status_change_finish(addr, numpages, enc);
drivers/gpu/drm/i915/gvt/gtt.c
1980
struct intel_gvt_gtt_entry *e, unsigned long index, bool guest)
drivers/gpu/drm/i915/gvt/gtt.c
1990
if (!guest)
drivers/gpu/drm/i915/gvt/gtt.c
491
bool guest)
drivers/gpu/drm/i915/gvt/gtt.c
498
pte_ops->get_entry(guest ? mm->ppgtt_mm.guest_pdps :
drivers/gpu/drm/i915/gvt/gtt.c
518
bool guest)
drivers/gpu/drm/i915/gvt/gtt.c
522
pte_ops->set_entry(guest ? mm->ppgtt_mm.guest_pdps :
drivers/gpu/drm/i915/gvt/gtt.c
592
bool guest)
drivers/gpu/drm/i915/gvt/gtt.c
603
ret = ops->get_entry(page_table, e, index, guest,
drivers/gpu/drm/i915/gvt/gtt.c
609
update_entry_type_for_real(ops, e, guest ?
drivers/gpu/drm/i915/gvt/gtt.c
621
bool guest)
drivers/gpu/drm/i915/gvt/gtt.c
632
return ops->set_entry(page_table, e, index, guest,
drivers/gpu/drm/vmwgfx/device_include/svga3d_cmd.h
505
SVGAGuestImage guest;
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
358
if (cmd->dma.guest.ptr.offset % PAGE_SIZE ||
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
371
cmd->dma.guest.ptr.offset);
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
375
kmap_offset = cmd->dma.guest.ptr.offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
390
if (box->w == VMW_CURSOR_SNOOP_WIDTH && cmd->dma.guest.pitch == image_pitch) {
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
397
virtual + i * cmd->dma.guest.pitch,
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1519
&cmd->body.guest.ptr, &vmw_bo);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1525
if (unlikely(cmd->body.guest.ptr.offset > bo_size)) {
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1530
bo_size -= cmd->body.guest.ptr.offset;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
298
body->guest.ptr = *ptr;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
299
body->guest.ptr.offset += cur_offset->bo_offset;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
300
body->guest.pitch = vmw_surface_calculate_pitch(desc, cur_size);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
319
body->guest.pitch);
drivers/media/pci/zoran/zoran_device.c
123
int post_office_write(struct zoran *zr, unsigned int guest,
drivers/media/pci/zoran/zoran_device.c
129
ZR36057_POR_PO_DIR | ZR36057_POR_PO_TIME | ((guest & 7) << 20) |
drivers/media/pci/zoran/zoran_device.c
136
int post_office_read(struct zoran *zr, unsigned int guest, unsigned int reg)
drivers/media/pci/zoran/zoran_device.c
140
por = ZR36057_POR_PO_TIME | ((guest & 7) << 20) | ((reg & 7) << 16);
drivers/media/pci/zoran/zoran_device.h
20
int post_office_write(struct zoran *zr, unsigned int guest, unsigned int reg,
drivers/media/pci/zoran/zoran_device.h
22
int post_office_read(struct zoran *zr, unsigned int guest, unsigned int reg);
drivers/perf/apple_m1_cpu_pmu.c
411
bool guest = config_base & M1_PMU_CFG_COUNT_GUEST;
drivers/perf/apple_m1_cpu_pmu.c
417
__m1_pmu_configure_event_filter(index, user && guest, kernel && guest, false);
fs/proc/stat.c
113
guest += cpustat[CPUTIME_GUEST];
fs/proc/stat.c
135
seq_put_decimal_ull(p, " ", nsec_to_clock_t(guest));
fs/proc/stat.c
154
guest = cpustat[CPUTIME_GUEST];
fs/proc/stat.c
165
seq_put_decimal_ull(p, " ", nsec_to_clock_t(guest));
fs/proc/stat.c
86
u64 guest, guest_nice;
fs/proc/stat.c
94
guest = guest_nice = 0;
kernel/events/core.c
1011
perf_cgroup_set_timestamp(struct perf_cpu_context *cpuctx, bool guest)
kernel/events/core.c
1031
if (guest) {
kernel/events/core.c
1240
perf_cgroup_set_timestamp(struct perf_cpu_context *cpuctx, bool guest)
tools/perf/util/evsel.c
730
MOD_PRINT(guest, 'G');
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2678
goto guest;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2694
guest: /* Assuming Guest */
tools/perf/util/parse-events.c
1831
if (mod.guest) {
tools/perf/util/parse-events.h
210
bool guest : 1; /* 'G' */
tools/testing/selftests/kvm/s390/memop.c
524
static void choose_block(bool guest, int i, int *size, int *offset)
tools/testing/selftests/kvm/s390/memop.c
529
if (guest) {
tools/testing/selftests/kvm/s390/memop.c
542
static __uint128_t permutate_bits(bool guest, int i, int size, __uint128_t old)
tools/testing/selftests/kvm/s390/memop.c
550
if (guest)
tools/testing/selftests/kvm/x86/sev_smoke_test.c
221
static void test_sev_smoke(void *guest, uint32_t type, uint64_t policy)
tools/testing/selftests/kvm/x86/sev_smoke_test.c
226
test_sev(guest, type, policy | SNP_POLICY_DBG);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
228
test_sev(guest, type, policy | SEV_POLICY_NO_DBG);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
229
test_sev(guest, type, policy);
tools/virtio/ringtest/main.c
295
pthread_t host, guest;
tools/virtio/ringtest/main.c
383
ret = pthread_create(&guest, NULL, start_guest, guest_arg);
tools/virtio/ringtest/main.c
386
ret = pthread_join(guest, &tret);
tools/virtio/ringtest/ring.c
103
guest.num_free = ring_size;
tools/virtio/ringtest/ring.c
116
if (!guest.num_free)
tools/virtio/ringtest/ring.c
119
guest.num_free--;
tools/virtio/ringtest/ring.c
120
head = (ring_size - 1) & (guest.avail_idx++);
tools/virtio/ringtest/ring.c
145
unsigned head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/ring.c
159
guest.num_free++;
tools/virtio/ringtest/ring.c
160
guest.last_used_idx++;
tools/virtio/ringtest/ring.c
166
unsigned head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/ring.c
180
event->call_index = guest.last_used_idx;
tools/virtio/ringtest/ring.c
195
guest.avail_idx,
tools/virtio/ringtest/ring.c
196
guest.kicked_avail_idx);
tools/virtio/ringtest/ring.c
198
guest.kicked_avail_idx = guest.avail_idx;
tools/virtio/ringtest/ring.c
65
} guest;
tools/virtio/ringtest/ring.c
92
guest.avail_idx = 0;
tools/virtio/ringtest/ring.c
93
guest.kicked_avail_idx = -1;
tools/virtio/ringtest/ring.c
94
guest.last_used_idx = 0;
tools/virtio/ringtest/virtio_ring_0_9.c
107
if (!guest.num_free)
tools/virtio/ringtest/virtio_ring_0_9.c
111
head = (ring_size - 1) & (guest.avail_idx++);
tools/virtio/ringtest/virtio_ring_0_9.c
113
head = guest.free_head;
tools/virtio/ringtest/virtio_ring_0_9.c
115
guest.num_free--;
tools/virtio/ringtest/virtio_ring_0_9.c
127
guest.free_head = desc[head].next;
tools/virtio/ringtest/virtio_ring_0_9.c
135
avail = guest.avail_idx++;
tools/virtio/ringtest/virtio_ring_0_9.c
142
avail = (ring_size - 1) & (guest.avail_idx++);
tools/virtio/ringtest/virtio_ring_0_9.c
148
ring.avail->idx = guest.avail_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
159
head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
161
if ((index ^ guest.last_used_idx ^ 0x8000) & ~(ring_size - 1))
tools/virtio/ringtest/virtio_ring_0_9.c
167
if (ring.used->idx == guest.last_used_idx)
tools/virtio/ringtest/virtio_ring_0_9.c
172
head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
175
head = (ring_size - 1) & guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
189
ring.desc[index].next = guest.free_head;
tools/virtio/ringtest/virtio_ring_0_9.c
190
guest.free_head = index;
tools/virtio/ringtest/virtio_ring_0_9.c
192
guest.num_free++;
tools/virtio/ringtest/virtio_ring_0_9.c
193
guest.last_used_idx++;
tools/virtio/ringtest/virtio_ring_0_9.c
199
unsigned short last_used_idx = guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
219
vring_used_event(&ring) = guest.last_used_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
234
guest.avail_idx,
tools/virtio/ringtest/virtio_ring_0_9.c
235
guest.kicked_avail_idx);
tools/virtio/ringtest/virtio_ring_0_9.c
237
guest.kicked_avail_idx = guest.avail_idx;
tools/virtio/ringtest/virtio_ring_0_9.c
52
} guest;
tools/virtio/ringtest/virtio_ring_0_9.c
78
guest.avail_idx = 0;
tools/virtio/ringtest/virtio_ring_0_9.c
79
guest.kicked_avail_idx = -1;
tools/virtio/ringtest/virtio_ring_0_9.c
80
guest.last_used_idx = 0;
tools/virtio/ringtest/virtio_ring_0_9.c
83
guest.free_head = 0;
tools/virtio/ringtest/virtio_ring_0_9.c
89
guest.num_free = ring_size;