atomic_load_long
mygen = atomic_load_long(&invl_gen->gen);
while (*m_gen > atomic_load_long(&pmap_invl_gen_head.gen)) {
atomic_load_long(&pmap_invl_gen_head.gen)) {
if (*m_gen > atomic_load_long(invl_gen))
eptgen = atomic_load_long(&pmap->pm_eptgen);
if (atomic_load_long(&pmap->pm_eptgen) == vmx->eptgen[curcpu]) {
eptgen = atomic_load_long(&pmap->pm_eptgen);
rxbytes = atomic_load_long(&rxr->rx_bytes);
txbytes = atomic_load_long(&txr->tx_bytes);
txpackets = atomic_load_long(&txr->tx_packets);
rxpackets = atomic_load_long(&rxr->rx_packets);
rxbytes = atomic_load_long(&rxr->rx_bytes);
txbytes = atomic_load_long(&txr->tx_bytes);
txpackets = atomic_load_long(&txr->tx_packets);
rxpackets = atomic_load_long(&rxr->rx_packets);
return (tm->tm_pages_max > atomic_load_long(&tm->tm_pages_used));
reserved = atomic_load_long(&tmpfs_pages_reserved);
global = atomic_load_long(&ticksl);
aio_complete(job, atomic_load_long(&job->nbytes), 0);
lnumneg = atomic_load_long(&numneg);
lnumcache = atomic_load_long(&numcache);
rfreevnodes = atomic_load_long(&freevnodes);
rfreevnodes_old = atomic_load_long(&freevnodes_old);
rnumvnodes = atomic_load_long(&numvnodes);
rnumvnodes = atomic_load_long(&numvnodes);
rnumvnodes = atomic_load_long(&numvnodes);
rnumvnodes = atomic_load_long(&numvnodes);
if (atomic_load_long(&numvnodes) + 1 > desiredvnodes &&
return (atomic_load_long(&fp->f_offset));
return (atomic_load_long(&fp->f_offset));
return (atomic_load_long(&uma_kmem_total));