M_FIRSTFIT
PAGE_SIZE, PAGE_SIZE, M_FIRSTFIT | M_WAITOK);
if (vmem_alloc(sc->vmem_epc, PAGE_SIZE, M_FIRSTFIT | M_NOWAIT,
1, 0, M_FIRSTFIT | M_WAITOK);
if (vmem_alloc(sc->irq_alloc, count, M_FIRSTFIT | M_NOWAIT,
gicv3_get_nirqs(dev), 1, 0, M_FIRSTFIT | M_WAITOK);
if (vmem_alloc(sc->sc_irq_alloc, nvecs, M_FIRSTFIT | M_NOWAIT,
M_FIRSTFIT | M_NOWAIT, &mem->paddr)) {
if (vmem_alloc(sc->key_map, TLS_KEY_CONTEXT_SZ, M_NOWAIT | M_FIRSTFIT,
M_FIRSTFIT|M_NOWAIT, &addr);
1, 0, M_FIRSTFIT| M_NOWAIT);
sc->vres.key.size, 32, 0, M_FIRSTFIT | M_WAITOK);
sc->vres.stag.size >> 5, 1, 0, M_FIRSTFIT | M_WAITOK);
4, 0, 0, VMEM_ADDR_MIN, VMEM_ADDR_MAX, M_FIRSTFIT | M_NOWAIT,
if (vmem_alloc(sc->stag_arena, size, M_FIRSTFIT | M_NOWAIT,
sc->vres.pbl.size, 1, 0, M_FIRSTFIT | M_WAITOK);
if (vmem_alloc(pr->pr_arena, PPOD_SZ(nppods), M_NOWAIT | M_FIRSTFIT,
M_FIRSTFIT | M_NOWAIT);
M_FIRSTFIT | M_NOWAIT, &addr)) {
PAGE_SIZE, M_FIRSTFIT | M_WAITOK);
0, 0, VMEM_ADDR_MIN, VMEM_ADDR_MAX, flags | M_FIRSTFIT, &res);
#define VMEM_FITMASK (M_BESTFIT | M_FIRSTFIT | M_NEXTFIT)
M_BESTFIT | M_FIRSTFIT | M_NEXTFIT)
#define VMEM_FLAGS (M_NOWAIT | M_BESTFIT | M_FIRSTFIT | M_NEXTFIT)
MPASS(strat == M_BESTFIT || strat == M_FIRSTFIT || strat == M_NEXTFIT);
MPASS(strat == M_BESTFIT || strat == M_FIRSTFIT || strat == M_NEXTFIT);
if (strat == M_FIRSTFIT)
if (strat == M_FIRSTFIT) {
if (strat == M_FIRSTFIT && ORDER2SIZE(idx) != qsize) {
vmem_alloc(asid_arena, 1, M_FIRSTFIT|M_WAITOK, &pid);
M_WAITOK | M_FIRSTFIT);
vmem_alloc(async_token_pool, 1, M_FIRSTFIT | M_WAITOK, &token);
M_FIRSTFIT | M_NOWAIT);
error = vmem_alloc(ctx->irtids, count, M_FIRSTFIT | M_NOWAIT,
error = vmem_alloc(unit->irtids, 1, M_FIRSTFIT | M_NOWAIT, &vmem_res);
M_FIRSTFIT | M_NOWAIT);
error = vmem_alloc(unit->irtids, count, M_FIRSTFIT | M_NOWAIT,