Symbol: vrp
sys/arch/amd64/amd64/vmm_machdep.c
3343
vm_run(struct vm_run_params *vrp)
sys/arch/amd64/amd64/vmm_machdep.c
3353
ret = vm_find(vrp->vrp_vm_id, &vm);
sys/arch/amd64/amd64/vmm_machdep.c
3357
vcpu = vm_find_vcpu(vm, vrp->vrp_vcpu_id);
sys/arch/amd64/amd64/vmm_machdep.c
3381
ret = copyin(vrp->vrp_exit, &vcpu->vc_exit, sizeof(struct vm_exit));
sys/arch/amd64/amd64/vmm_machdep.c
3385
vcpu->vc_inject.vie_type = vrp->vrp_inject.vie_type;
sys/arch/amd64/amd64/vmm_machdep.c
3386
vcpu->vc_inject.vie_vector = vrp->vrp_inject.vie_vector;
sys/arch/amd64/amd64/vmm_machdep.c
3387
vcpu->vc_inject.vie_errorcode = vrp->vrp_inject.vie_errorcode;
sys/arch/amd64/amd64/vmm_machdep.c
3392
vcpu_rv = vcpu_run_vmx(vcpu, vrp);
sys/arch/amd64/amd64/vmm_machdep.c
3394
vcpu_rv = vcpu_run_svm(vcpu, vrp);
sys/arch/amd64/amd64/vmm_machdep.c
3400
vrp->vrp_exit_reason = (vcpu_rv == 0) ? VM_EXIT_NONE
sys/arch/amd64/amd64/vmm_machdep.c
3402
vrp->vrp_irqready = vcpu->vc_irqready;
sys/arch/amd64/amd64/vmm_machdep.c
3404
ret = copyout(&vcpu->vc_exit, vrp->vrp_exit,
sys/arch/amd64/amd64/vmm_machdep.c
3408
vrp->vrp_exit_reason = VM_EXIT_TERMINATED;
sys/arch/amd64/amd64/vmm_machdep.c
3650
vcpu_run_vmx(struct vcpu *vcpu, struct vm_run_params *vrp)
sys/arch/amd64/amd64/vmm_machdep.c
3676
if (vrp->vrp_intr_pending)
sys/arch/amd64/amd64/vmm_machdep.c
3904
TRACEPOINT(vmm, guest_enter, vcpu, vrp);
sys/arch/amd64/amd64/vmm_machdep.c
3967
TRACEPOINT(vmm, guest_exit, vcpu, vrp, exit_reason);
sys/arch/amd64/amd64/vmm_machdep.c
6638
vcpu_run_svm(struct vcpu *vcpu, struct vm_run_params *vrp)
sys/arch/amd64/amd64/vmm_machdep.c
6647
if (vrp->vrp_intr_pending)
sys/arch/amd64/amd64/vmm_machdep.c
6778
TRACEPOINT(vmm, guest_enter, vcpu, vrp);
sys/arch/amd64/amd64/vmm_machdep.c
6840
TRACEPOINT(vmm, guest_exit, vcpu, vrp, exit_reason);
sys/dev/vmm/vmm.c
777
vm_resetcpu(struct vm_resetcpu_params *vrp)
sys/dev/vmm/vmm.c
784
error = vm_find(vrp->vrp_vm_id, &vm);
sys/dev/vmm/vmm.c
789
vrp->vrp_vm_id);
sys/dev/vmm/vmm.c
793
vcpu = vm_find_vcpu(vm, vrp->vrp_vcpu_id);
sys/dev/vmm/vmm.c
797
vrp->vrp_vcpu_id, vrp->vrp_vm_id);
sys/dev/vmm/vmm.c
806
if (vcpu_reset_regs(vcpu, &vrp->vrp_init_state)) {
sys/net/if_veb.c
2568
struct veb_rule vr, *vrp;
sys/net/if_veb.c
2644
vrp = pool_get(&veb_rule_pool, PR_WAITOK|PR_LIMITFAIL|PR_ZERO);
sys/net/if_veb.c
2645
if (vrp == NULL) {
sys/net/if_veb.c
2650
*vrp = vr;
sys/net/if_veb.c
2657
TAILQ_INSERT_TAIL(&p->p_vrl, vrp, vr_entry);
sys/net/if_veb.c
2661
vrp, vr_lentry[0]);
sys/net/if_veb.c
2665
vrp, vr_lentry[1]);
sys/net/if_veb.c
2674
pool_put(&veb_rule_pool, vrp);
usr.sbin/rpki-client/ccr.c
1109
struct vrp *prev, *vrp = NULL;
usr.sbin/rpki-client/ccr.c
1122
if ((vrp = calloc(1, sizeof(*vrp))) == NULL)
usr.sbin/rpki-client/ccr.c
1125
vrp->asid = asid;
usr.sbin/rpki-client/ccr.c
1126
vrp->afi = afi;
usr.sbin/rpki-client/ccr.c
1128
if (!ip_addr_parse(r->address, afi, fn, &vrp->addr)) {
usr.sbin/rpki-client/ccr.c
1133
maxlen = vrp->addr.prefixlen;
usr.sbin/rpki-client/ccr.c
1139
if (vrp->addr.prefixlen > maxlen) {
usr.sbin/rpki-client/ccr.c
1147
vrp->maxlength = maxlen;
usr.sbin/rpki-client/ccr.c
1151
if (ccr_vrp_cmp(vrp, prev) <= 0) {
usr.sbin/rpki-client/ccr.c
1157
if ((RB_INSERT(ccr_vrp_tree, &ccr->vrps, vrp)) != NULL) {
usr.sbin/rpki-client/ccr.c
1162
prev = vrp;
usr.sbin/rpki-client/ccr.c
1163
vrp = NULL;
usr.sbin/rpki-client/ccr.c
1168
free(vrp);
usr.sbin/rpki-client/ccr.c
382
append_cached_vrp(STACK_OF(ROAIPAddress) *addresses, struct vrp *vrp)
usr.sbin/rpki-client/ccr.c
391
num_bytes = (vrp->addr.prefixlen + 7) / 8;
usr.sbin/rpki-client/ccr.c
392
num_bits = vrp->addr.prefixlen % 8;
usr.sbin/rpki-client/ccr.c
398
if (!ASN1_BIT_STRING_set(ripa->address, vrp->addr.addr, num_bytes))
usr.sbin/rpki-client/ccr.c
406
if (vrp->maxlength > vrp->addr.prefixlen) {
usr.sbin/rpki-client/ccr.c
410
if (!ASN1_INTEGER_set_uint64(ripa->maxLength, vrp->maxlength))
usr.sbin/rpki-client/ccr.c
423
struct vrp *prev, *vrp;
usr.sbin/rpki-client/ccr.c
432
RB_FOREACH(vrp, ccr_vrp_tree, &ccr->vrps) {
usr.sbin/rpki-client/ccr.c
433
if (prev == NULL || prev->asid != vrp->asid) {
usr.sbin/rpki-client/ccr.c
437
if (!ASN1_INTEGER_set_uint64(rp->asID, vrp->asid))
usr.sbin/rpki-client/ccr.c
444
if (prev == NULL || prev->asid != vrp->asid ||
usr.sbin/rpki-client/ccr.c
445
prev->afi != vrp->afi) {
usr.sbin/rpki-client/ccr.c
450
assert(vrp->afi == AFI_IPV4 || vrp->afi == AFI_IPV6);
usr.sbin/rpki-client/ccr.c
453
afibuf[1] = vrp->afi;
usr.sbin/rpki-client/ccr.c
463
append_cached_vrp(ripaf->addresses, vrp);
usr.sbin/rpki-client/ccr.c
464
prev = vrp;
usr.sbin/rpki-client/ccr.c
784
struct vrp *vrp;
usr.sbin/rpki-client/ccr.c
788
if ((vrp = calloc(1, sizeof(*vrp))) == NULL)
usr.sbin/rpki-client/ccr.c
791
vrp->asid = roa->asid;
usr.sbin/rpki-client/ccr.c
792
vrp->afi = roa->ips[i].afi;
usr.sbin/rpki-client/ccr.c
793
vrp->addr = roa->ips[i].addr;
usr.sbin/rpki-client/ccr.c
794
vrp->maxlength = roa->ips[i].maxlength;
usr.sbin/rpki-client/ccr.c
796
if (RB_INSERT(ccr_vrp_tree, tree, vrp) != NULL)
usr.sbin/rpki-client/ccr.c
797
free(vrp);
usr.sbin/rpki-client/ccr.c
805
ccr_vrp_cmp(const struct vrp *a, const struct vrp *b)
usr.sbin/rpki-client/ccr.c
848
RB_GENERATE(ccr_vrp_tree, vrp, entry, ccr_vrp_cmp);
usr.sbin/rpki-client/ccr.c
892
struct vrp *vrp, *tmp_vrp;
usr.sbin/rpki-client/ccr.c
894
RB_FOREACH_SAFE(vrp, ccr_vrp_tree, vrps, tmp_vrp) {
usr.sbin/rpki-client/ccr.c
895
RB_REMOVE(ccr_vrp_tree, vrps, vrp);
usr.sbin/rpki-client/ccr.c
896
free(vrp);
usr.sbin/rpki-client/extern.h
383
RB_ENTRY(vrp) entry;
usr.sbin/rpki-client/extern.h
395
RB_HEAD(vrp_tree, vrp);
usr.sbin/rpki-client/extern.h
396
RB_PROTOTYPE(vrp_tree, vrp, entry, vrpcmp);
usr.sbin/rpki-client/extern.h
456
RB_HEAD(ccr_vrp_tree, vrp);
usr.sbin/rpki-client/extern.h
457
RB_PROTOTYPE(ccr_vrp_tree, vrp, entry, ccr_vrp_cmp);
usr.sbin/rpki-client/output-bgpd.c
25
struct vrp *vrp;
usr.sbin/rpki-client/output-bgpd.c
35
RB_FOREACH(vrp, vrp_tree, &vd->vrps) {
usr.sbin/rpki-client/output-bgpd.c
38
ip_addr_print(&vrp->addr, vrp->afi, ipbuf, sizeof(ipbuf));
usr.sbin/rpki-client/output-bgpd.c
39
if (vrp->maxlength > vrp->addr.prefixlen) {
usr.sbin/rpki-client/output-bgpd.c
41
"maxlen %u ", vrp->maxlength);
usr.sbin/rpki-client/output-bgpd.c
47
ipbuf, maxlenbuf, vrp->asid, (long long)vrp->expires) < 0)
usr.sbin/rpki-client/output-bird.c
26
struct vrp *v;
usr.sbin/rpki-client/output-csv.c
25
struct vrp *v;
usr.sbin/rpki-client/output-json.c
155
struct vrp *v;
usr.sbin/rpki-client/output.c
94
struct vrp *v, *tv;
usr.sbin/rpki-client/print.c
887
struct vrp *vrp;
usr.sbin/rpki-client/print.c
898
RB_FOREACH(vrp, ccr_vrp_tree, &ccr->vrps) {
usr.sbin/rpki-client/print.c
899
ip_addr_print(&vrp->addr, vrp->afi, buf, sizeof(buf));
usr.sbin/rpki-client/print.c
904
json_do_int("asn", vrp->asid);
usr.sbin/rpki-client/print.c
905
if (vrp->maxlength)
usr.sbin/rpki-client/print.c
906
json_do_int("maxlen", vrp->maxlength);
usr.sbin/rpki-client/print.c
910
if (vrp->maxlength)
usr.sbin/rpki-client/print.c
911
printf("-%hhu", vrp->maxlength);
usr.sbin/rpki-client/print.c
912
printf(" AS %u\n", vrp->asid);
usr.sbin/rpki-client/roa.c
337
struct vrp *v, *found;
usr.sbin/rpki-client/roa.c
378
vrpcmp(struct vrp *a, struct vrp *b)
usr.sbin/rpki-client/roa.c
418
RB_GENERATE(vrp_tree, vrp, entry, vrpcmp);
usr.sbin/vmd/arm64_vm.c
119
vcpu_exit(struct vm_run_params *vrp)
usr.sbin/vmd/arm64_vm.c
127
vcpu_exit_pci(struct vm_run_params *vrp)
usr.sbin/vmd/fw_cfg.c
214
vcpu_exit_fw_cfg(struct vm_run_params *vrp)
usr.sbin/vmd/fw_cfg.c
217
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/fw_cfg.c
253
vcpu_exit_fw_cfg_dma(struct vm_run_params *vrp)
usr.sbin/vmd/fw_cfg.c
257
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/i8253.c
165
vcpu_exit_i8253_misc(struct vm_run_params *vrp)
usr.sbin/vmd/i8253.c
167
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/i8253.c
221
vcpu_exit_i8253(struct vm_run_params *vrp)
usr.sbin/vmd/i8253.c
227
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/i8259.c
666
vcpu_exit_i8259(struct vm_run_params *vrp)
usr.sbin/vmd/i8259.c
668
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/i8259.c
724
vcpu_exit_elcr(struct vm_run_params *vrp)
usr.sbin/vmd/i8259.c
726
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/mc146818.c
277
vcpu_exit_mc146818(struct vm_run_params *vrp)
usr.sbin/vmd/mc146818.c
279
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/mc146818.h
19
uint8_t vcpu_exit_mc146818(struct vm_run_params *vrp);
usr.sbin/vmd/ns8250.c
585
vcpu_exit_com(struct vm_run_params *vrp)
usr.sbin/vmd/ns8250.c
588
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/ns8250.c
615
intr = vcpu_process_com_data(vei, vrp->vrp_vm_id,
usr.sbin/vmd/ns8250.c
616
vrp->vrp_vcpu_id);
usr.sbin/vmd/pci.c
268
pci_handle_address_reg(struct vm_run_params *vrp)
usr.sbin/vmd/pci.c
270
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/pci.c
290
pci_handle_io(struct vm_run_params *vrp)
usr.sbin/vmd/pci.c
297
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/pci.c
336
pci_handle_data_reg(struct vm_run_params *vrp)
usr.sbin/vmd/pci.c
338
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/vm.c
501
struct vm_resetcpu_params vrp;
usr.sbin/vmd/vm.c
503
memset(&vrp, 0, sizeof(vrp));
usr.sbin/vmd/vm.c
504
vrp.vrp_vm_id = vmid;
usr.sbin/vmd/vm.c
505
vrp.vrp_vcpu_id = vcpu_id;
usr.sbin/vmd/vm.c
506
memcpy(&vrp.vrp_init_state, vrs, sizeof(struct vcpu_reg_state));
usr.sbin/vmd/vm.c
510
if (ioctl(env->vmd_fd, VMM_IOC_RESETCPU, &vrp) == -1)
usr.sbin/vmd/vm.c
596
struct vm_run_params **vrp;
usr.sbin/vmd/vm.c
610
vrp = calloc(vmc->vmc_ncpus, sizeof(struct vm_run_params *));
usr.sbin/vmd/vm.c
611
if (vrp == NULL) {
usr.sbin/vmd/vm.c
632
vrp[i] = malloc(sizeof(struct vm_run_params));
usr.sbin/vmd/vm.c
633
if (vrp[i] == NULL) {
usr.sbin/vmd/vm.c
638
vrp[i]->vrp_exit = malloc(sizeof(struct vm_exit));
usr.sbin/vmd/vm.c
639
if (vrp[i]->vrp_exit == NULL) {
usr.sbin/vmd/vm.c
644
vrp[i]->vrp_vm_id = vm->vm_vmmid;
usr.sbin/vmd/vm.c
645
vrp[i]->vrp_vcpu_id = i;
usr.sbin/vmd/vm.c
699
ret = pthread_create(&tid[i], NULL, vcpu_run_loop, vrp[i]);
usr.sbin/vmd/vm.c
810
struct vm_run_params *vrp = (struct vm_run_params *)arg;
usr.sbin/vmd/vm.c
812
uint32_t n = vrp->vrp_vcpu_id;
usr.sbin/vmd/vm.c
888
if (vrp->vrp_irqready && intr_pending(current_vm)) {
usr.sbin/vmd/vm.c
889
vrp->vrp_inject.vie_vector = intr_ack(current_vm);
usr.sbin/vmd/vm.c
890
vrp->vrp_inject.vie_type = VCPU_INJECT_INTR;
usr.sbin/vmd/vm.c
892
vrp->vrp_inject.vie_type = VCPU_INJECT_NONE;
usr.sbin/vmd/vm.c
895
vrp->vrp_intr_pending = intr_pending(current_vm);
usr.sbin/vmd/vm.c
897
if (ioctl(env->vmd_fd, VMM_IOC_RUN, vrp) == -1) {
usr.sbin/vmd/vm.c
906
if (vrp->vrp_exit_reason == VM_EXIT_TERMINATED) {
usr.sbin/vmd/vm.c
911
if (vrp->vrp_exit_reason != VM_EXIT_NONE) {
usr.sbin/vmd/vm.c
916
ret = vcpu_exit(vrp);
usr.sbin/vmd/x86_vm.c
453
vcpu_exit_inout(struct vm_run_params *vrp)
usr.sbin/vmd/x86_vm.c
455
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/x86_vm.c
478
intr = ioports_map[vei->vei.vei_port](vrp);
usr.sbin/vmd/x86_vm.c
485
vcpu_assert_irq(vrp->vrp_vm_id, vrp->vrp_vcpu_id, intr);
usr.sbin/vmd/x86_vm.c
508
vcpu_exit(struct vm_run_params *vrp)
usr.sbin/vmd/x86_vm.c
512
switch (vrp->vrp_exit_reason) {
usr.sbin/vmd/x86_vm.c
530
ret = vcpu_exit_eptviolation(vrp);
usr.sbin/vmd/x86_vm.c
536
vcpu_exit_inout(vrp);
usr.sbin/vmd/x86_vm.c
540
vcpu_halt(vrp->vrp_vcpu_id);
usr.sbin/vmd/x86_vm.c
547
log_debug("unknown exit reason 0x%x", vrp->vrp_exit_reason);
usr.sbin/vmd/x86_vm.c
566
vcpu_exit_eptviolation(struct vm_run_params *vrp)
usr.sbin/vmd/x86_vm.c
568
struct vm_exit *ve = vrp->vrp_exit;
usr.sbin/vmd/x86_vm.c
651
vcpu_exit_pci(struct vm_run_params *vrp)
usr.sbin/vmd/x86_vm.c
653
struct vm_exit *vei = vrp->vrp_exit;
usr.sbin/vmd/x86_vm.c
660
pci_handle_address_reg(vrp);
usr.sbin/vmd/x86_vm.c
666
pci_handle_data_reg(vrp);
usr.sbin/vmd/x86_vm.c
669
intr = pci_handle_io(vrp);