spgcnt_t
nv = atomic_add_long_nv(&swapfs_pending_delete, (spgcnt_t)delta_swaps);
nv = atomic_add_long_nv(&swapfs_pending_delete, -(spgcnt_t)delta_swaps);
spgcnt_t *nreloc,
uint_t *pszc, spgcnt_t *nreloc, struct seg *seg, caddr_t addr,
spgcnt_t i;
MAX((spgcnt_t)(availrmem - swapfs_minfree), 0);
spgcnt_t prefltmem, availm, npages;
npages = (spgcnt_t)btopr(len);
spgcnt_t rmem = availrmem - tune.t_minarmem;
spgcnt_t fmem = freemem - minfree;
spgcnt_t pmem = availrmem - tune.t_minarmem;
spgcnt_t vmem = btop(vmem_size(heap_arena, VMEM_FREE));
spgcnt_t wanted;
spgcnt_t npgs;
page_needfree(-(spgcnt_t)free_get);
spgcnt_t freemem_left;
spgcnt_t pgcnt;
spgcnt_t availm, npages;
npages = (spgcnt_t)btopr(len);
spgcnt_t avail;
if (avail > (spgcnt_t)needs) {
spgcnt_t vavail;
spgcnt_t faststmp, slowstmp, result;
spgcnt_t a_softlockcnt; /* number of pages locked in pcache */
(k_anoninfo.ani_max + MAX((spgcnt_t)(availrmem - swapfs_minfree), 0))
MAX((spgcnt_t)(availrmem - swapfs_minfree), 0))
spgcnt_t *, struct seg *, caddr_t,
spgcnt_t *, int);
void page_needfree(spgcnt_t);
void page_create_putback(spgcnt_t npages);
int page_relocate(page_t **, page_t **, int, int, spgcnt_t *, struct lgrp *);
int do_page_relocate(page_t **, page_t **, int, spgcnt_t *, struct lgrp *);
spgcnt_t pg;
spgcnt_t pg;
spgcnt_t pgno = seg_page(seg, addr+len) - seg_page(seg, addr) + 1;
spgcnt_t an_idx;
spgcnt_t shm_softlockcnt; /* # outstanding lock operations */
spgcnt_t shm_lckpgs; /* # of locked pages per attached seg */
spgcnt_t npgs;
spgcnt_t nreloc;
spgcnt_t nreloc;
spgcnt_t pages = btop(len);
spgcnt_t softlockcnt; /* # of pages SOFTLOCKED in seg */
spgcnt_t softlockcnt_sbase; /* # of softlocks for seg start addr */
spgcnt_t softlockcnt_send; /* # of softlocks for seg end addr */
spgcnt_t npages;
spgcnt_t npages;
spgcnt_t npages;
spgcnt_t npages;
spgcnt_t npages;
spgcnt_t npages = btopr(size);
spgcnt_t nreloc = 0;
spgcnt_t npgs = nreloc;
spgcnt_t nreloc = 0;
spgcnt_t nreloc = 0;
spgcnt_t adjusted_swap = npages - unlocked_mem_swap;
spgcnt_t mem_free_pages = 0;
mem_free_pages = MIN((spgcnt_t)(k_anoninfo.ani_mem_resv -
page_needfree(spgcnt_t npages)
page_create_putback(spgcnt_t npages)
page_needfree(-(spgcnt_t)npages);
page_needfree(-(spgcnt_t)npages);
spgcnt_t *nrelocp,
spgcnt_t npgs, i;
spgcnt_t dofree = 0;
nv = atomic_add_long_nv(&pending_delete, (spgcnt_t)delta_pages);
spgcnt_t *nrelocp,
spgcnt_t ret;
nv = atomic_add_long_nv(&pending_delete, -(spgcnt_t)delta_pages);
spgcnt_t curnpgs = page_get_pagecnt(pp->p_szc);
spgcnt_t pgcnt, npgs;
spgcnt_t npgs;
spgcnt_t i, npgs;
spgcnt_t page_cnt;
spgcnt_t i;
page_needfree(-(spgcnt_t)deficit);
spgcnt_t count;
spgcnt_t *nrelocp,
spgcnt_t npgs = page_get_pagecnt(pp->p_szc);
spgcnt_t pgcnt, npgs, i;
spgcnt_t npgs, pg_cnt;
spgcnt_t pagesavail;
if (pagesavail < (spgcnt_t)btopr(vp_len)) {
spgcnt_t fmem = MAX((spgcnt_t)(freemem - needfree), 0);
spgcnt_t pages;
MAX((spgcnt_t)(availrmem - swapfs_minfree), 0)) < pages) {
(spgcnt_t)sip->si_nfpgs > 0) {
ASSERT((spgcnt_t)sip->si_nfpgs >= 0);
spgcnt_t avail;
avail = MAX((spgcnt_t)(availrmem - swapfs_minfree), 0);
spgcnt_t avail;
avail = MAX((spgcnt_t)(availrmem - swapfs_minfree), 0);
spgcnt_t _cnt = (spgcnt_t)(cnt); \
spgcnt_t scnt = (spgcnt_t)(cnt);
static spgcnt_t
static spgcnt_t
spgcnt_t rv;
bln_stats.bln_hard_limit = (spgcnt_t)HYPERVISOR_memory_op(
spgcnt_t bln_hv_pages; /* number of total pages given to hypervisor */
spgcnt_t bln_hard_limit; /* domain's max-mem limit */
hat_page_relocate(page_t **target, page_t **replacement, spgcnt_t *nrelocp)
spgcnt_t npages, i;
extern int hat_page_relocate(page_t **, page_t **, spgcnt_t *);
spgcnt_t _cnt = (spgcnt_t)(cnt); \
spgcnt_t _cnt = (spgcnt_t)(cnt); \
spgcnt_t excess;
spgcnt_t excess;
spgcnt_t npgs;
spgcnt_t i = 0;