km_alloc
pcbva = (vaddr_t)km_alloc(USPACE, &kv_any, &kp_contig, &kd_waitok);
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
vaddr_t va = (vaddr_t)km_alloc(endpa - pgpa, &kv_any, &kp_none,
va = (vaddr_t)km_alloc(end - pa, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(sz, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(map_size, &kv_any, &kp_none, &kd_nowait);
*nva = (vaddr_t)km_alloc(2 * PAGE_SIZE, &kv_any, &kp_none,
cif = km_alloc(sizeof *cif, &kv_any, &kp_zero, &kd_waitok);
kstack = (vaddr_t)km_alloc(USPACE, &kv_any, &kp_dirty, &kd_nowait);
addr = km_alloc(size, &kv_any, &kp_kexec, &kd_nowait);
kstart = (vaddr_t)km_alloc(ksz, &kv_any, &kp_none, &kd_waitok);
vaddr_t va = (vaddr_t)km_alloc(endpa - pgpa, &kv_any, &kp_none,
new_pd_va = km_alloc(PAGE_SIZE, &kv_page, &kp_zero, &kd_nowait);
pml4va = km_alloc(PAGE_SIZE, &kv_page, &kp_zero, &kd_nowait);
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
vcpu->vc_control_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page, &kp_zero,
vcpu->vc_msr_bitmap_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page, &kp_zero,
vcpu->vc_vmx_msr_exit_load_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page,
vcpu->vc_vmx_msr_exit_save_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page,
vcpu->vc_vmx_msr_entry_load_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page,
vcpu->vc_control_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page, &kp_zero,
vcpu->vc_msr_bitmap_va = (vaddr_t)km_alloc(2 * PAGE_SIZE, &kv_any,
vcpu->vc_svm_hsa_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page,
vcpu->vc_svm_ioio_va = (vaddr_t)km_alloc(3 * PAGE_SIZE, &kv_any,
vcpu->vc_svm_vmsa_va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_page,
l1tf_flush_region = km_alloc(VMX_L1D_FLUSH_SIZE,
cr3_host_virt = (vaddr_t)km_alloc(PAGE_SIZE, &kv_any, &kp_none,
sys_bios_data = (vaddr_t)km_alloc(PAGE_SIZE, &kv_any, &kp_none,
sys_image = (vaddr_t)km_alloc(1024 * 1024, &kv_any, &kp_none,
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
kstack = km_alloc(USPACE, &kv_any, &kp_zero, &kd_waitok);
kstack = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_waitok);
kstack = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_waitok);
kstack = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_waitok);
return (km_alloc(pp->pr_pgsize,
va = (vaddr_t)km_alloc(L1_TABLE_SIZE, &kv_any, &kp_none,
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
va = (vaddr_t)km_alloc(endpa - startpa, &kv_any, &kp_none, &kd_nowait);
vaddr_t va = (vaddr_t)km_alloc(endpa - pgpa, &kv_any, &kp_none,
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(endpa - startpa, &kv_any, &kp_none, &kd_nowait);
kstack = km_alloc(USPACE, &kv_any, &kp_zero, &kd_waitok);
kvp = km_alloc(sizeof(struct pmapvp0), &kv_kvp, &kp_zero,
pted = km_alloc(PAGE_SIZE, &kv_kvp, &kp_zero,
return km_alloc(pp->pr_pgsize, &kv_any, &kp_dirty, &kd);
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
l1 = km_alloc(nl1 * DART_PAGE_SIZE, &kv_any, &kp_none, &kd_waitok);
state->crashlog = km_alloc(size * PAGE_SIZE,
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, &kd_nowait);
kva = (vaddr_t)km_alloc(size, &kv_physwait, &kp_none, &kd_prefer);
vaddr_t va = (vaddr_t)km_alloc(endpa - pgpa, &kv_any, &kp_none,
eva = (vaddr_t)km_alloc(end - pa, &kv_any,
sva = va = (vaddr_t)km_alloc(endpa, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
*nva = (vaddr_t)km_alloc(2 * PAGE_SIZE, &kv_any, &kp_none,
cif = km_alloc(sizeof *cif, &kv_any, &kp_zero, &kd_waitok);
kstack = (vaddr_t)km_alloc(USPACE, &kv_any, &kp_dirty, &kd_nowait);
va = (vaddr_t)km_alloc(NBPG*2, &kv_any, &kp_zero, &kd_waitok);
va = (vaddr_t)km_alloc(map_size, &kv_any, &kp_none, &kd_nowait);
vaddr_t va = (vaddr_t)km_alloc(endpa - pgpa, &kv_any, &kp_none,
return (km_alloc(pp->pr_pgsize,
pmap->pm_pdir = (vaddr_t)km_alloc(NBPG, &kv_any, &kp_dirty, &kd_waitok);
va = (vaddr_t)km_alloc(NBPG, &kv_any, &kp_zero, &kd_waitok);
va = (vaddr_t)km_alloc(4 * NBPG, &kv_any, &kp_zero, &kd_waitok);
pmap->pm_pdir = (vaddr_t)km_alloc(4 * NBPG, &kv_any, &kp_dirty,
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, &kd_nowait);
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
sys_bios_data = (vaddr_t)km_alloc(PAGE_SIZE, &kv_any, &kp_none,
sys_image = (vaddr_t)km_alloc(1024 * 1024, &kv_any, &kp_none,
va = (vaddr_t)km_alloc(endpa - pa, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
hatch_stacks[cpu] = (vaddr_t)km_alloc(USPACE, &kv_any,
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
pa = (paddr_t)km_alloc(round_page(sz), &kv_any, &kp_contig, &kd_nowait);
kva = (vaddr_t)km_alloc(size, &kv_physwait, &kp_none, &kd_prefer);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
vaddr = (vaddr_t)km_alloc(len, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, &kd_nowait);
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
addr = km_alloc(size, &kv_any, &kp_kexec, &kd_nowait);
symaddr = km_alloc(symsize, &kv_any, &kp_kexec, &kd_nowait);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(endpa - startpa, &kv_any, &kp_none, &kd_nowait);
ci->ci_initstack_end = km_alloc(PAGE_SIZE, &kv_any, &kp_zero,
va = (vaddr_t)km_alloc(epa - pa, &kv_any, &kp_none, &kd_waitok);
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(endpa - startpa, &kv_any, &kp_none, &kd_nowait);
kstack = km_alloc(USPACE, &kv_any, &kp_zero, &kd_waitok);
kvp = km_alloc(sizeof(struct pmapvp1), &kv_kvp, &kp_zero,
pted = km_alloc(PAGE_SIZE, &kv_kvp, &kp_zero,
return km_alloc(pp->pr_pgsize, &kv_any, &kp_dirty, &kd);
taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
kva = (vaddr_t)km_alloc(size, &kv_physwait, &kp_none, &kd_prefer);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(sz, &kv_cpu_info, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, kd);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(size, &kv_any, &kp_none, &kd_nowait);
kva = (vaddr_t)km_alloc(size, &kv_physwait, &kp_none, &kd_prefer);
va = km_alloc(VTD_PAGE_SIZE, &kv_page, &kp_zero, kd);
sc->sc_mem_region[i] = km_alloc(sc->sc_mem_size[i],
rd->rdata_desc = (caddr_t)km_alloc(size, &kv_intrsafe, &bwi_pa_mode,
sc->sc_bounce_tx_data[i] = (caddr_t)km_alloc(
sc->sc_bounce_rx_data = (caddr_t)km_alloc(BWI_RX_NDESC * MCLBYTES,
if (!(rom->rom_code = km_alloc(round_page(size), &kv_any,
buf = km_alloc(size, &kv_any, &kp_zero, &kd_waitok);
(vaddr_t)km_alloc(PAGE_SIZE, &kv_any, &kp_none, &kd_waitok);
va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_physwait, &kp_none, &kd_waitok);
va = (vaddr_t)km_alloc(PAGE_SIZE * npages, &kv_any, &kp_none,
va = (vaddr_t)km_alloc(PAGE_SIZE * npfn, &kv_any, &kp_none,
void *addr = km_alloc(PAGE_SIZE, &kv_page, &kp_dirty, &kd_nowait);
ttm->segs = km_alloc(round_page(ttm->num_pages *
ch->ch_ring = km_alloc(2 * buflen, &kv_any, &kp_zero, cold ?
sc->sc_simp[cpu] = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_nowait);
sc->sc_siep[cpu] = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_nowait);
sc->sc_events = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_nowait);
sc->sc_monitor[0] = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_nowait);
sc->sc_monitor[1] = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_nowait);
sc->sc_rx_ring = km_alloc(sc->sc_rx_size, &kv_any, &kp_zero,
sc->sc_page = km_alloc(PAGE_SIZE, &kv_any, &kp_none, &kd_nowait);
va = km_alloc(PAGE_SIZE, &kv_any, &kp_zero, &kd_nowait);
xs->xs_ring = km_alloc(PAGE_SIZE, &kv_any, &kp_none, &kd_nowait);
rwva = (vaddr_t)km_alloc(PAGE_SIZE, &kv_any, &kp_none,
argp = km_alloc(NCARGS, &kv_exec, &kp_pageable, &kd_waitok);
va = (caddr_t)km_alloc(ptoa(npg), &kv_intrsafe, &kp_dirty, kdp);
kmemusage = km_alloc(round_page(nkmempages * sizeof(struct kmemusage)),
entropy = km_alloc(PAGE_SIZE, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(PAGE_SIZE, &kv_any, &kp_none, &kd_avoidalias);
chunktable = (vaddr_t)km_alloc(HIBERNATE_CHUNK_TABLE_SIZE, &kv_any,
tempva = (vaddr_t)km_alloc(MAXPHYS + PAGE_SIZE, &kv_any, &kp_none,
hibernate_fchunk_area = (vaddr_t)km_alloc(24 * PAGE_SIZE, &kv_any,
hibernate_temp_page = (vaddr_t)km_alloc(PAGE_SIZE, &kv_any,
va = (vaddr_t)km_alloc(sz, &kv_any, &kp_pig, &kd_nowait);
va = (vaddr_t)km_alloc(sz, &kv_any, &kp_pig, &kd_nowait);
*va = (vaddr_t)km_alloc(sz, &kv_any, &kp_piglet, &kd_nowait);
fn = km_alloc;
return (km_alloc(pp->pr_pgsize, &kv_page, pp->pr_crange, &kd));
v = km_alloc(pp->pr_pgsize, &kv, pp->pr_crange, &kd);
v = km_alloc(pp->pr_pgsize, &kv, pp->pr_crange, &kd);
cp = km_alloc(round_page(size), &kv_any, &kp_zero, &kd_nowait);
buffer = km_alloc(size, &kv_any, &kp_pageable, &kd_waitok);
void *km_alloc(size_t, const struct kmem_va_mode *,
va = (vaddr_t)km_alloc(sz, &kv_any, &kp_none, &kd_nowait);
va = (vaddr_t)km_alloc(USPACE, &kv_uarea, &kp_zero, &kd_waitok);
ne = km_alloc(PAGE_SIZE, &kv_page, &kp_dirty,
pgs = km_alloc(round_page(npages * sizeof(*pgs)),
pseg->start = (vaddr_t)km_alloc(MAX_PAGER_SEGS * MAXBSIZE,