Symbol: vmm_softc
sys/arch/amd64/amd64/vmm_machdep.c
2794
if (vmm_softc->mode == VMM_MODE_EPT)
sys/arch/amd64/amd64/vmm_machdep.c
2796
else if (vmm_softc->mode == VMM_MODE_RVI)
sys/arch/amd64/amd64/vmm_machdep.c
2799
panic("%s: unknown vmm mode: %d", __func__, vmm_softc->mode);
sys/arch/amd64/amd64/vmm_machdep.c
284
struct vmm_softc *sc = (struct vmm_softc *)self;
sys/arch/amd64/amd64/vmm_machdep.c
2958
vcpu->vc_virt_mode = vmm_softc->mode;
sys/arch/amd64/amd64/vmm_machdep.c
2969
if (vmm_softc->mode == VMM_MODE_EPT)
sys/arch/amd64/amd64/vmm_machdep.c
2971
else if (vmm_softc->mode == VMM_MODE_RVI)
sys/arch/amd64/amd64/vmm_machdep.c
2974
panic("%s: unknown vmm mode: %d", __func__, vmm_softc->mode);
sys/arch/amd64/amd64/vmm_machdep.c
3067
if (vmm_softc->mode == VMM_MODE_EPT)
sys/arch/amd64/amd64/vmm_machdep.c
3069
else if (vmm_softc->mode == VMM_MODE_RVI)
sys/arch/amd64/amd64/vmm_machdep.c
3072
panic("%s: unknown vmm mode: %d", __func__, vmm_softc->mode);
sys/arch/amd64/amd64/vmm_machdep.c
3517
if (vmm_softc->mode == VMM_MODE_EPT) {
sys/arch/amd64/amd64/vmm_machdep.c
3520
} else if (vmm_softc->mode == VMM_MODE_RVI) {
sys/arch/amd64/amd64/vmm_machdep.c
356
if ((err = rw_enter(&vmm_softc->vm_lock, RW_WRITE | RW_NOSLEEP)))
sys/arch/amd64/amd64/vmm_machdep.c
360
SLIST_FOREACH(vm, &vmm_softc->vm_list, vm_link) {
sys/arch/amd64/amd64/vmm_machdep.c
3917
if (vmm_softc->sc_md.pkru_enabled)
sys/arch/amd64/amd64/vmm_machdep.c
3926
if (vmm_softc->sc_md.pkru_enabled) {
sys/arch/amd64/amd64/vmm_machdep.c
406
rw_exit_write(&vmm_softc->vm_lock);
sys/arch/amd64/amd64/vmm_machdep.c
425
if (vmm_softc->mode == VMM_MODE_EPT)
sys/arch/amd64/amd64/vmm_machdep.c
435
if (vmm_softc->vm_ct > 0)
sys/arch/amd64/amd64/vmm_machdep.c
5240
if (vmm_softc->mode == VMM_MODE_RVI) {
sys/arch/amd64/amd64/vmm_machdep.c
5243
} else if (vmm_softc->mode == VMM_MODE_EPT) {
sys/arch/amd64/amd64/vmm_machdep.c
5271
if (vmm_softc->mode == VMM_MODE_RVI) {
sys/arch/amd64/amd64/vmm_machdep.c
5277
} else if (vmm_softc->mode == VMM_MODE_EPT) {
sys/arch/amd64/amd64/vmm_machdep.c
5968
if (vmm_softc->sc_md.pkru_enabled)
sys/arch/amd64/amd64/vmm_machdep.c
612
if (vmm_softc->mode == VMM_MODE_EPT)
sys/arch/amd64/amd64/vmm_machdep.c
616
else if (vmm_softc->mode == VMM_MODE_RVI)
sys/arch/amd64/amd64/vmm_machdep.c
6223
if (vmm_softc->sc_md.pkru_enabled) {
sys/arch/amd64/amd64/vmm_machdep.c
6235
if (vmm_softc->sc_md.pkru_enabled) {
sys/arch/amd64/amd64/vmm_machdep.c
6283
if (vmm_softc->mode == VMM_MODE_EPT) {
sys/arch/amd64/amd64/vmm_machdep.c
6420
if (vmm_softc->sc_md.pkru_enabled)
sys/arch/amd64/amd64/vmm_machdep.c
648
rw_enter_write(&vmm_softc->sc_slock);
sys/arch/amd64/amd64/vmm_machdep.c
6612
if (vmm_softc->mode == VMM_MODE_RVI) {
sys/arch/amd64/amd64/vmm_machdep.c
6799
if (vmm_softc->sc_md.pkru_enabled)
sys/arch/amd64/amd64/vmm_machdep.c
6814
if (vmm_softc->sc_md.pkru_enabled) {
sys/arch/amd64/amd64/vmm_machdep.c
686
rw_exit_write(&vmm_softc->sc_slock);
sys/arch/amd64/amd64/vmm_machdep.c
6918
struct vmm_softc *sc = vmm_softc;
sys/arch/amd64/amd64/vmm_machdep.c
6920
rw_enter_write(&vmm_softc->vpid_lock);
sys/arch/amd64/amd64/vmm_machdep.c
6934
rw_exit_write(&vmm_softc->vpid_lock);
sys/arch/amd64/amd64/vmm_machdep.c
6943
rw_exit_write(&vmm_softc->vpid_lock);
sys/arch/amd64/amd64/vmm_machdep.c
6971
struct vmm_softc *sc = vmm_softc;
sys/arch/amd64/amd64/vmm_machdep.c
6973
rw_enter_write(&vmm_softc->vpid_lock);
sys/arch/amd64/amd64/vmm_machdep.c
6979
rw_exit_write(&vmm_softc->vpid_lock);
sys/arch/amd64/amd64/vmm_machdep.c
708
rw_enter_write(&vmm_softc->sc_slock);
sys/arch/amd64/amd64/vmm_machdep.c
746
rw_exit_write(&vmm_softc->sc_slock);
sys/arch/amd64/amd64/vmm_machdep.c
919
switch (vmm_softc->mode) {
sys/arch/amd64/amd64/vmm_machdep.c
927
printf("%s: invalid vmm mode %d\n", __func__, vmm_softc->mode);
sys/dev/vmm/vmm.c
105
rw_enter_write(&vmm_softc->sc_slock);
sys/dev/vmm/vmm.c
106
KASSERT(vmm_softc->sc_status == VMM_SUSPENDED);
sys/dev/vmm/vmm.c
107
refcnt_init(&vmm_softc->sc_refcnt);
sys/dev/vmm/vmm.c
108
vmm_softc->sc_status = VMM_ACTIVE;
sys/dev/vmm/vmm.c
109
rw_exit_write(&vmm_softc->sc_slock);
sys/dev/vmm/vmm.c
112
wakeup(&vmm_softc->sc_status);
sys/dev/vmm/vmm.c
136
if (vmm_softc == NULL)
sys/dev/vmm/vmm.c
140
if (vmm_softc->mode == VMM_MODE_UNKNOWN)
sys/dev/vmm/vmm.c
181
rw_enter_read(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
182
SLIST_FOREACH(vm, &vmm_softc->vm_list, vm_link) {
sys/dev/vmm/vmm.c
202
rw_exit_read(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
222
ret = rw_enter(&vmm_softc->sc_slock, RW_READ | RW_INTR);
sys/dev/vmm/vmm.c
225
while (vmm_softc->sc_status != VMM_ACTIVE) {
sys/dev/vmm/vmm.c
226
ret = rwsleep_nsec(&vmm_softc->sc_status, &vmm_softc->sc_slock,
sys/dev/vmm/vmm.c
229
rw_exit(&vmm_softc->sc_slock);
sys/dev/vmm/vmm.c
233
refcnt_take(&vmm_softc->sc_refcnt);
sys/dev/vmm/vmm.c
234
rw_exit(&vmm_softc->sc_slock);
sys/dev/vmm/vmm.c
276
refcnt_rele_wake(&vmm_softc->sc_refcnt);
sys/dev/vmm/vmm.c
35
struct vmm_softc *vmm_softc;
sys/dev/vmm/vmm.c
379
rw_enter_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
380
if (vmm_softc->vcpu_ct + vcp->vcp_ncpus > vmm_softc->vcpu_max) {
sys/dev/vmm/vmm.c
382
vmm_softc->vcpu_max);
sys/dev/vmm/vmm.c
383
rw_exit_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
386
vmm_softc->vcpu_ct += vcp->vcp_ncpus;
sys/dev/vmm/vmm.c
387
vmm_softc->vm_ct++;
sys/dev/vmm/vmm.c
388
rw_exit_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
44
sizeof(struct vmm_softc), vmm_probe, vmm_attach, NULL, vmm_activate
sys/dev/vmm/vmm.c
474
rw_enter_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
475
vmm_softc->vm_idx++;
sys/dev/vmm/vmm.c
476
vm->vm_id = vmm_softc->vm_idx;
sys/dev/vmm/vmm.c
480
SLIST_INSERT_HEAD(&vmm_softc->vm_list, vm, vm_link);
sys/dev/vmm/vmm.c
481
rw_exit_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
491
rw_enter_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
492
vmm_softc->vm_ct--;
sys/dev/vmm/vmm.c
493
vmm_softc->vcpu_ct -= vcp->vcp_ncpus;
sys/dev/vmm/vmm.c
494
if (vmm_softc->vm_ct < 1)
sys/dev/vmm/vmm.c
496
rw_exit_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
60
struct vmm_softc *sc = (struct vmm_softc *)self;
sys/dev/vmm/vmm.c
644
rw_enter_read(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
645
vm_ct = vmm_softc->vm_ct;
sys/dev/vmm/vmm.c
646
rw_exit_read(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
663
rw_enter_read(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
664
SLIST_FOREACH(vm, &vmm_softc->vm_list, vm_link) {
sys/dev/vmm/vmm.c
690
rw_exit_read(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
733
rw_enter_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
734
SLIST_REMOVE(&vmm_softc->vm_list, vm, vm, vm_link);
sys/dev/vmm/vmm.c
735
rw_exit_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
751
rw_enter_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
752
vmm_softc->vm_ct--;
sys/dev/vmm/vmm.c
753
vmm_softc->vcpu_ct -= nvcpu;
sys/dev/vmm/vmm.c
754
if (vmm_softc->vm_ct < 1)
sys/dev/vmm/vmm.c
756
rw_exit_write(&vmm_softc->vm_lock);
sys/dev/vmm/vmm.c
81
vmm_softc = sc;
sys/dev/vmm/vmm.c
91
rw_enter_write(&vmm_softc->sc_slock);
sys/dev/vmm/vmm.c
92
KASSERT(vmm_softc->sc_status == VMM_ACTIVE);
sys/dev/vmm/vmm.c
93
vmm_softc->sc_status = VMM_SUSPENDED;
sys/dev/vmm/vmm.c
94
rw_exit_write(&vmm_softc->sc_slock);
sys/dev/vmm/vmm.c
97
refcnt_finalize(&vmm_softc->sc_refcnt, "vmmsusp");
sys/dev/vmm/vmm.h
241
extern struct vmm_softc *vmm_softc;