vmem
vbe_print_memory(unsigned vmem)
vmem /= 1024;
if (vmem >= 10240000) {
vmem /= 1048576;
} else if (vmem >= 10000) {
vmem /= 1024;
printf("Total memory: %u%cB\n", vmem, unit);
struct vmem *vmem_epc;
if (vmem_alloc(sc->vmem, size,
vmem_free(sc->vmem, mem->paddr, mem->size);
vmem_t *vmem;
vmem = vmem_create("stratix10 vmem", 0, 0, PAGE_SIZE,
if (vmem == NULL)
vmem_add(vmem, addr, size, 0);
sc->vmem = vmem;
vmem_t *vmem;
void **vmem;
if (rmem->vmem_size && *rmem->vmem) {
free(*rmem->vmem, M_DEVBUF);
*rmem->vmem = NULL;
*rmem->vmem = malloc(rmem->vmem_size, M_DEVBUF, M_NOWAIT | M_ZERO);
if (!(*rmem->vmem))
void *vmem;
vmem = ice_calloc(hw, words, sizeof(u16));
if (!vmem)
data_local = (_FORCE_ __le16 *)vmem;
ice_free(hw, vmem);
void *vmem;
vmem = ice_calloc(hw, ICE_SR_SECTOR_SIZE_IN_WORDS, sizeof(u16));
if (!vmem)
data = (u16 *)vmem;
ice_free(hw, vmem);
struct irdma_virt_mem *vmem =
vmem->size = sizeof(struct irdma_hmc_pd_entry) * 512;
vmem->va = kzalloc(vmem->size, GFP_KERNEL);
if (!vmem->va) {
sd_entry->u.pd_table.pd_entry = vmem->va;
kfree(vmem->va);
struct irdma_virt_mem *vmem;
vmem = &vsi->ilq_mem;
vmem = &vsi->ieq_mem;
vmem->size = pudasize + sqwridsize + rqwridsize;
vmem->va = kzalloc(vmem->size, GFP_KERNEL);
if (!vmem->va)
rsrc = vmem->va;
vsi->ilq = vmem->va;
vsi->ieq = vmem->va;
((u8 *)vmem->va + pudasize);
rsrc->rq_wrid_array = (u64 *)((u8 *)vmem->va + pudasize + sqwridsize);
struct irdma_virt_mem *vmem;
vmem = &vsi->ilq_mem;
vmem = &vsi->ieq_mem;
void *vmem;
vmem = ixgbe_calloc(hw, words, sizeof(u16));
if (!vmem)
data_local = (__le16 *)vmem;
ixgbe_free(hw, vmem);
void *vmem;
vmem = ixgbe_calloc(hw, IXGBE_SR_SECTOR_SIZE_IN_WORDS, sizeof(u16));
if (!vmem)
data = (u16 *)vmem;
ixgbe_free(hw, vmem);
struct i40e_virt_mem vmem;
ret_code = i40e_allocate_virt_mem(hw, &vmem,
data = (u16 *)vmem.va;
i40e_free_virt_mem(hw, &vmem);
xdma_handle_mem_node(vmem_t *vmem, phandle_t memory)
vmem_add(vmem, mem_start, mem_size, 0);
vmem_t *vmem;
vmem = vmem_create("xDMA vmem", 0, 0, PAGE_SIZE,
if (vmem == NULL)
if (xdma_handle_mem_node(vmem, mem_node) != 0) {
vmem_destroy(vmem);
return (vmem);
xdma_put_memory(vmem_t *vmem)
vmem_destroy(vmem);
xchan_set_memory(xdma_channel_t *xchan, vmem_t *vmem)
xchan->vmem = vmem;
vmem_t *vmem; /* VA space */
vmem_t *vmem;
void xdma_put_memory(vmem_t *vmem);
int xdma_handle_mem_node(vmem_t *vmem, phandle_t memory);
void xchan_set_memory(xdma_channel_t *xchan, vmem_t *vmem);
vmem_t *vmem; /* Bounce memory. */
if (vmem_alloc(xio->vmem, size,
xio->vmem = vmem_create("xDMA vmem", 0, 0, PAGE_SIZE,
if (xio->vmem == NULL)
if (xdma_handle_mem_node(xio->vmem, mem_node) != 0) {
vmem_destroy(xio->vmem);
vmem_destroy(xio->vmem);
vmem_free(xio->vmem, va, PAGE_SIZE);
if (xchan->vmem == NULL)
if (vmem_alloc(xchan->vmem, size,
vmem_free(xchan->vmem, xr->buf.paddr, size);
DB_SHOW_COMMAND(vmem, vmem_summ)
LIST_ENTRY(vmem) vm_alllist;
DB_SHOW_ALL_COMMAND(vmem, vmem_summall)
static LIST_HEAD(, vmem) vmem_list = LIST_HEAD_INITIALIZER(vmem_list);
static struct vmem kernel_arena_storage;
static struct vmem buffer_arena_storage;
static struct vmem transient_arena_storage;
static struct vmem memguard_arena_storage;
sizeof(struct vmem), NULL, NULL, NULL, NULL,
bufkva_reclaim(vmem_t *vmem, int flags)
window->map->vmem = vmem_create("IOMMU mappings", PAGE_SIZE,
error = vmem_xalloc(window->map->vmem, allocsize,
vmem_xfree(window->map->vmem, pageround, roundedsize);
vmem_t *vmem;
vmem_t *vmem;
typedef struct vmem vmem_t;
struct vmem;
void memguard_init(struct vmem *);
struct vmem;
vmem_t *vmem;
vmem = vm_dom[domain].vmd_kernel_arena;
if (vmem_alloc(vmem, asize, M_BESTFIT | flags, &addr))
vmem_free(vmem, addr, asize);
vmem_t *vmem;
vmem = vm_dom[domain].vmd_kernel_arena;
if (vmem_alloc(vmem, asize, flags | M_BESTFIT, &addr))
vmem_free(vmem, addr, asize);
static struct vmem *
struct vmem *arena;
struct vmem *arena;
extern struct vmem *kernel_arena;
extern struct vmem *buffer_arena;
extern struct vmem *transient_arena;
extern struct vmem *memguard_arena;
struct vmem *vmd_kernel_arena; /* (c) per-domain kva R/W arena. */
struct vmem *vmd_kernel_rwx_arena; /* (c) per-domain kva R/W/X arena. */
struct vmem *vmd_kernel_nofree_arena; /* (c) per-domain kva NOFREE arena. */