Symbol: vmcb
sys/amd64/vmm/amd/svm.c
1352
struct vmcb *vmcb;
sys/amd64/vmm/amd/svm.c
1362
vmcb = svm_get_vmcb(vcpu);
sys/amd64/vmm/amd/svm.c
1363
state = &vmcb->state;
sys/amd64/vmm/amd/svm.c
1364
ctrl = &vmcb->ctrl;
sys/amd64/vmm/amd/svm.c
1624
svm_handle_inst_emul(vmcb, info2, vmexit);
sys/amd64/vmm/amd/svm.c
2278
free(vcpu->vmcb, M_SVM);
sys/amd64/vmm/amd/svm.c
321
svm_get_cs_info(struct vmcb *vmcb, struct vm_guest_paging *paging, int *cs_d,
sys/amd64/vmm/amd/svm.c
327
error = vmcb_seg(vmcb, VM_REG_GUEST_CS, &seg);
sys/amd64/vmm/amd/svm.c
655
vcpu->vmcb = malloc_aligned(sizeof(struct vmcb), PAGE_SIZE, M_SVM,
sys/amd64/vmm/amd/svm.c
659
vcpu->vmcb_pa = vtophys(vcpu->vmcb);
sys/amd64/vmm/amd/svm.c
692
svm_vcpu_mode(struct vmcb *vmcb)
sys/amd64/vmm/amd/svm.c
698
state = &vmcb->state;
sys/amd64/vmm/amd/svm.c
701
error = vmcb_seg(vmcb, VM_REG_GUEST_CS, &seg);
sys/amd64/vmm/amd/svm.c
785
svm_get_cs_info(vcpu->vmcb, &vis->paging, &vis->cs_d,
sys/amd64/vmm/amd/svm.c
813
svm_paging_info(struct vmcb *vmcb, struct vm_guest_paging *paging)
sys/amd64/vmm/amd/svm.c
817
state = &vmcb->state;
sys/amd64/vmm/amd/svm.c
820
paging->cpu_mode = svm_vcpu_mode(vmcb);
sys/amd64/vmm/amd/svm.c
904
svm_handle_inst_emul(struct vmcb *vmcb, uint64_t gpa, struct vm_exit *vmexit)
sys/amd64/vmm/amd/svm.c
911
ctrl = &vmcb->ctrl;
sys/amd64/vmm/amd/svm.c
917
svm_paging_info(vmcb, paging);
sys/amd64/vmm/amd/svm.c
919
svm_get_cs_info(vmcb, paging, &vmexit->u.inst_emul.cs_d,
sys/amd64/vmm/amd/svm_softc.h
103
return (&vcpu->vmcb->state);
sys/amd64/vmm/amd/svm_softc.h
110
return (&vcpu->vmcb->ctrl);
sys/amd64/vmm/amd/svm_softc.h
53
struct vmcb *vmcb; /* hardware saved vcpu context */
sys/amd64/vmm/amd/svm_softc.h
92
static __inline struct vmcb *
sys/amd64/vmm/amd/svm_softc.h
96
return (vcpu->vmcb);
sys/amd64/vmm/amd/vmcb.c
120
struct vmcb *vmcb;
sys/amd64/vmm/amd/vmcb.c
124
vmcb = svm_get_vmcb(vcpu);
sys/amd64/vmm/amd/vmcb.c
128
if ((off + bytes) >= sizeof (struct vmcb))
sys/amd64/vmm/amd/vmcb.c
131
ptr = (char *)vmcb;
sys/amd64/vmm/amd/vmcb.c
164
struct vmcb *vmcb;
sys/amd64/vmm/amd/vmcb.c
169
vmcb = svm_get_vmcb(vcpu);
sys/amd64/vmm/amd/vmcb.c
170
state = &vmcb->state;
sys/amd64/vmm/amd/vmcb.c
229
seg = vmcb_segptr(vmcb, ident);
sys/amd64/vmm/amd/vmcb.c
237
seg = vmcb_segptr(vmcb, ident == VM_REG_GUEST_FS_BASE ?
sys/amd64/vmm/amd/vmcb.c
270
struct vmcb *vmcb;
sys/amd64/vmm/amd/vmcb.c
275
vmcb = svm_get_vmcb(vcpu);
sys/amd64/vmm/amd/vmcb.c
276
state = &vmcb->state;
sys/amd64/vmm/amd/vmcb.c
345
seg = vmcb_segptr(vmcb, ident);
sys/amd64/vmm/amd/vmcb.c
367
vmcb_seg(struct vmcb *vmcb, int ident, struct vmcb_segment *seg2)
sys/amd64/vmm/amd/vmcb.c
371
seg = vmcb_segptr(vmcb, ident);
sys/amd64/vmm/amd/vmcb.c
383
struct vmcb *vmcb;
sys/amd64/vmm/amd/vmcb.c
387
vmcb = svm_get_vmcb(vcpu);
sys/amd64/vmm/amd/vmcb.c
389
seg = vmcb_segptr(vmcb, reg);
sys/amd64/vmm/amd/vmcb.c
433
struct vmcb *vmcb;
sys/amd64/vmm/amd/vmcb.c
436
vmcb = svm_get_vmcb(vcpu);
sys/amd64/vmm/amd/vmcb.c
437
seg = vmcb_segptr(vmcb, reg);
sys/amd64/vmm/amd/vmcb.c
61
vmcb_segptr(struct vmcb *vmcb, int type)
sys/amd64/vmm/amd/vmcb.c
66
state = &vmcb->state;
sys/amd64/vmm/amd/vmcb.h
352
CTASSERT(sizeof(struct vmcb) == PAGE_SIZE);
sys/amd64/vmm/amd/vmcb.h
353
CTASSERT(offsetof(struct vmcb, state) == 0x400);
sys/amd64/vmm/amd/vmcb.h
359
int vmcb_seg(struct vmcb *vmcb, int ident, struct vmcb_segment *seg);