arch/arm/mm/init.c
135
if (memblock_overlaps_region(&memblock.memory,
arch/arm/mm/mmu.c
1318
end = memblock.memory.regions[0].base + memblock.memory.regions[0].size;
arch/arm/mm/pmsa-v8.c
248
add_range(mem, ARRAY_SIZE(mem), 0, memblock.memory.regions[0].base,
arch/arm/mm/pmsa-v8.c
249
memblock.memory.regions[0].base + memblock.memory.regions[0].size);
arch/arm64/kernel/setup.c
224
num_standard_resources = memblock.memory.cnt;
arch/loongarch/kernel/mem.c
60
memblock_set_node(0, PHYS_ADDR_MAX, &memblock.memory, 0);
arch/loongarch/kernel/setup.c
435
num_standard_resources = memblock.memory.cnt;
arch/m68k/math-emu/fp_decode.h
355
3: | addressing mode: address register/programm counter memory indirect
arch/microblaze/mm/init.c
131
memblock.memory.regions[0].size = memory_size;
arch/microblaze/mm/init.c
169
if ((u32) memblock.memory.regions[0].size < 0x400000) {
arch/microblaze/mm/init.c
174
if ((u32) memblock.memory.regions[0].size < kernel_tlb) {
arch/microblaze/mm/init.c
180
memory_start = (u32) memblock.memory.regions[0].base;
arch/microblaze/mm/init.c
181
lowmem_size = memory_size = (u32) memblock.memory.regions[0].size;
arch/mips/cavium-octeon/setup.c
1007
memory = cvmx_bootmem_phy_alloc(mem_alloc_size,
arch/mips/cavium-octeon/setup.c
1011
if (memory >= 0) {
arch/mips/cavium-octeon/setup.c
1024
&memory, &size);
arch/mips/cavium-octeon/setup.c
1027
&memory, &size);
arch/mips/cavium-octeon/setup.c
1029
end = memory + mem_alloc_size;
arch/mips/cavium-octeon/setup.c
1036
if (memory < crashk_base && end > crashk_end) {
arch/mips/cavium-octeon/setup.c
1038
memblock_add(memory, crashk_base - memory);
arch/mips/cavium-octeon/setup.c
1039
total += crashk_base - memory;
arch/mips/cavium-octeon/setup.c
1045
if (memory >= crashk_base && end <= crashk_end)
arch/mips/cavium-octeon/setup.c
1052
if (memory > crashk_base && memory < crashk_end &&
arch/mips/cavium-octeon/setup.c
1058
mem_alloc_size -= crashk_end - memory;
arch/mips/cavium-octeon/setup.c
1059
memory = crashk_end;
arch/mips/cavium-octeon/setup.c
1060
} else if (memory < crashk_base && end > crashk_base &&
arch/mips/cavium-octeon/setup.c
1068
memblock_add(memory, mem_alloc_size);
arch/mips/cavium-octeon/setup.c
973
int64_t memory;
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
21
# addresses, and need to have the appropriate memory region set
arch/mips/include/asm/mach-ip27/kernel-entry-init.h
36
li t0, 0x1c000 # Offset of text into node memory
arch/mips/kernel/setup.c
650
memblock_set_node(0, PHYS_ADDR_MAX, &memblock.memory, 0);
arch/powerpc/boot/devtree.c
20
void *root, *memory;
arch/powerpc/boot/devtree.c
47
memory = finddevice("/memory");
arch/powerpc/boot/devtree.c
48
if (! memory) {
arch/powerpc/boot/devtree.c
49
memory = create_node(NULL, "memory");
arch/powerpc/boot/devtree.c
50
setprop_str(memory, "device_type", "memory");
arch/powerpc/boot/devtree.c
58
setprop(memory, "reg", memreg, (naddr + nsize)*sizeof(u32));
arch/powerpc/boot/oflib.c
111
static ihandle memory;
arch/powerpc/boot/oflib.c
140
memory = of_call_prom("open", 1, 1, "/memory");
arch/powerpc/boot/oflib.c
141
if (memory == PROM_ERROR) {
arch/powerpc/boot/oflib.c
142
memory = of_call_prom("open", 1, 1, "/memory@0");
arch/powerpc/boot/oflib.c
143
if (memory == PROM_ERROR) {
arch/powerpc/boot/oflib.c
163
ret = of_call_prom_ret("call-method", 5, 2, &result, "claim", memory,
arch/powerpc/boot/treeboot-iss4xx.c
37
void *memory;
arch/powerpc/boot/treeboot-iss4xx.c
40
memory = finddevice("/memory");
arch/powerpc/boot/treeboot-iss4xx.c
41
if (!memory)
arch/powerpc/boot/treeboot-iss4xx.c
44
getprop(memory, "reg", reg, sizeof(reg));
arch/powerpc/kernel/fadump.c
1656
elf_phdr_cnt = 2 + fw_dump.boot_mem_regs_cnt + memblock_num_regions(memory);
arch/powerpc/kernel/prom_init.c
121
ihandle memory;
arch/powerpc/kernel/prom_init.c
2262
prom.memory = call_prom("open", 1, 1, ADDR("/memory"));
arch/powerpc/kernel/prom_init.c
2266
if (!IHANDLE_VALID(prom.memory) || !IHANDLE_VALID(prom.mmumap))
arch/powerpc/kernel/prom_init.c
605
ADDR("claim"), prom.memory,
arch/powerpc/kernel/prom_init.c
614
prom.memory, size, virt);
arch/powerpc/mm/drmem.c
103
static int drmem_update_dt_v2(struct device_node *memory,
arch/powerpc/mm/drmem.c
168
of_update_property(memory, new_prop);
arch/powerpc/mm/drmem.c
174
struct device_node *memory;
arch/powerpc/mm/drmem.c
178
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
arch/powerpc/mm/drmem.c
179
if (!memory)
arch/powerpc/mm/drmem.c
187
prop = of_find_property(memory, "ibm,dynamic-memory", NULL);
arch/powerpc/mm/drmem.c
189
rc = drmem_update_dt_v1(memory, prop);
arch/powerpc/mm/drmem.c
191
prop = of_find_property(memory, "ibm,dynamic-memory-v2", NULL);
arch/powerpc/mm/drmem.c
193
rc = drmem_update_dt_v2(memory, prop);
arch/powerpc/mm/drmem.c
197
of_node_put(memory);
arch/powerpc/mm/drmem.c
64
static int drmem_update_dt_v1(struct device_node *memory,
arch/powerpc/mm/drmem.c
90
of_update_property(memory, new_prop);
arch/powerpc/mm/mem.c
186
memblock_set_node(0, PHYS_ADDR_MAX, &memblock.memory, 0);
arch/powerpc/mm/numa.c
1005
memblock_set_node(start, size, &memblock.memory, nid);
arch/powerpc/mm/numa.c
1028
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
arch/powerpc/mm/numa.c
1029
if (memory) {
arch/powerpc/mm/numa.c
1030
walk_drmem_lmbs(memory, NULL, numa_setup_drmem_lmb);
arch/powerpc/mm/numa.c
1031
of_node_put(memory);
arch/powerpc/mm/numa.c
1052
&memblock.memory, nid);
arch/powerpc/mm/numa.c
1282
struct device_node *memory;
arch/powerpc/mm/numa.c
1285
for_each_node_by_type(memory, "memory") {
arch/powerpc/mm/numa.c
1291
if (of_address_to_resource(memory, i++, &res))
arch/powerpc/mm/numa.c
1297
nid = of_node_to_nid_single(memory);
arch/powerpc/mm/numa.c
1305
of_node_put(memory);
arch/powerpc/mm/numa.c
1317
struct device_node *memory = NULL;
arch/powerpc/mm/numa.c
1323
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
arch/powerpc/mm/numa.c
1324
if (memory) {
arch/powerpc/mm/numa.c
1326
of_node_put(memory);
arch/powerpc/mm/numa.c
1339
struct device_node *memory = NULL;
arch/powerpc/mm/numa.c
1351
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
arch/powerpc/mm/numa.c
1352
if (memory) {
arch/powerpc/mm/numa.c
1353
of_node_put(memory);
arch/powerpc/mm/numa.c
515
struct device_node *memory = NULL;
arch/powerpc/mm/numa.c
517
memory = of_find_node_by_type(memory, "memory");
arch/powerpc/mm/numa.c
518
if (!memory)
arch/powerpc/mm/numa.c
521
*n_addr_cells = of_n_addr_cells(memory);
arch/powerpc/mm/numa.c
522
*n_size_cells = of_n_size_cells(memory);
arch/powerpc/mm/numa.c
523
of_node_put(memory);
arch/powerpc/mm/numa.c
555
struct device_node *memory;
arch/powerpc/mm/numa.c
559
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
arch/powerpc/mm/numa.c
560
if (!memory)
arch/powerpc/mm/numa.c
563
prop = of_get_property(memory, "ibm,associativity-lookup-arrays", &len);
arch/powerpc/mm/numa.c
565
of_node_put(memory);
arch/powerpc/mm/numa.c
572
of_node_put(memory);
arch/powerpc/mm/numa.c
889
memblock_set_node(base, sz, &memblock.memory, nid);
arch/powerpc/mm/numa.c
897
struct device_node *memory, *pci;
arch/powerpc/mm/numa.c
966
for_each_node_by_type(memory, "memory") {
arch/powerpc/mm/numa.c
974
memcell_buf = of_get_property(memory,
arch/powerpc/mm/numa.c
977
memcell_buf = of_get_property(memory, "reg", &len);
arch/powerpc/mm/numa.c
993
associativity = of_get_associativity(memory);
arch/riscv/kernel/setup.c
152
num_resources = memblock.memory.cnt + memblock.reserved.cnt + 1;
arch/s390/kernel/setup.c
613
if (memblock.memory.regions[0].size < crash_size) {
arch/s390/kernel/setup.c
659
(unsigned long)memblock.memory.total_size >> 20);
arch/s390/kernel/setup.c
717
memblock_set_node(0, ULONG_MAX, &memblock.memory, 0);
arch/sh/kernel/setup.c
236
&memblock.memory, nid);
arch/sparc/mm/init_64.c
1203
prev_max = memblock.memory.max;
arch/sparc/mm/init_64.c
1217
&memblock.memory, nid);
arch/sparc/mm/init_64.c
1218
if (memblock.memory.max != prev_max)
arch/sparc/mm/init_64.c
1591
memblock_set_node(0, PHYS_ADDR_MAX, &memblock.memory, 0);
arch/x86/mm/init_32.c
628
memblock_set_node(0, PHYS_ADDR_MAX, &memblock.memory, 0);
arch/x86/mm/init_64.c
825
memblock_set_node(0, PHYS_ADDR_MAX, &memblock.memory, 0);
drivers/base/memory.c
692
static int __add_memory_block(struct memory_block *memory)
drivers/base/memory.c
696
memory->dev.bus = &memory_subsys;
drivers/base/memory.c
697
memory->dev.id = memory->start_section_nr / sections_per_block;
drivers/base/memory.c
698
memory->dev.release = memory_block_release;
drivers/base/memory.c
699
memory->dev.groups = memory_memblk_attr_groups;
drivers/base/memory.c
700
memory->dev.offline = memory->state == MEM_OFFLINE;
drivers/base/memory.c
702
ret = device_register(&memory->dev);
drivers/base/memory.c
704
put_device(&memory->dev);
drivers/base/memory.c
707
ret = xa_err(xa_store(&memory_blocks, memory->dev.id, memory,
drivers/base/memory.c
710
device_unregister(&memory->dev);
drivers/base/memory.c
836
static void remove_memory_block(struct memory_block *memory)
drivers/base/memory.c
838
if (WARN_ON_ONCE(memory->dev.bus != &memory_subsys))
drivers/base/memory.c
841
WARN_ON(xa_erase(&memory_blocks, memory->dev.id) == NULL);
drivers/base/memory.c
843
if (memory->group) {
drivers/base/memory.c
844
list_del(&memory->group_next);
drivers/base/memory.c
845
memory->group = NULL;
drivers/base/memory.c
849
put_device(&memory->dev);
drivers/base/memory.c
850
device_unregister(&memory->dev);
drivers/char/agp/agp.h
198
struct agp_memory *memory, size_t page_count);
drivers/char/agp/agp.h
200
void agp_generic_destroy_pages(struct agp_memory *memory);
drivers/char/agp/backend.c
117
long memory, index, result;
drivers/char/agp/backend.c
120
memory = totalram_pages() >> (20 - PAGE_SHIFT);
drivers/char/agp/backend.c
122
memory = totalram_pages() << (PAGE_SHIFT - 20);
drivers/char/agp/backend.c
126
while ((memory > maxes_table[index].mem) && (index < 8))
drivers/char/agp/backend.c
130
( (memory - maxes_table[index - 1].mem) *
drivers/gpu/drm/amd/pm/powerplay/hwmgr/processpptables.c
731
ps->memory.dllOff = (0 != tmp);
drivers/gpu/drm/amd/pm/powerplay/hwmgr/processpptables.c
733
ps->memory.m3arb = (le32_to_cpu(pnon_clock_info->ulCapsAndSettings) &
drivers/gpu/drm/amd/pm/powerplay/inc/power_state.h
160
struct PP_StateMemroyBlock memory;
drivers/gpu/drm/amd/pm/swsmu/inc/amdgpu_smu.h
202
struct smu_state_memory_block memory;
drivers/gpu/drm/nouveau/include/nvif/if000c.h
62
__u64 memory;
drivers/gpu/drm/nouveau/include/nvif/if000c.h
89
__u64 memory;
drivers/gpu/drm/nouveau/include/nvkm/core/firmware.h
24
struct nvkm_memory memory;
drivers/gpu/drm/nouveau/include/nvkm/core/gpuobj.h
16
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/perf.h
13
u32 memory;
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
22
struct nvkm_memory *memory; /* Memory currently mapped into VMA. */
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
91
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nouveau_svm.c
456
struct nvif_object *memory = &buffer->object;
drivers/gpu/drm/nouveau/nouveau_svm.c
457
const u32 instlo = nvif_rd32(memory, offset + 0x00);
drivers/gpu/drm/nouveau/nouveau_svm.c
458
const u32 insthi = nvif_rd32(memory, offset + 0x04);
drivers/gpu/drm/nouveau/nouveau_svm.c
459
const u32 addrlo = nvif_rd32(memory, offset + 0x08);
drivers/gpu/drm/nouveau/nouveau_svm.c
460
const u32 addrhi = nvif_rd32(memory, offset + 0x0c);
drivers/gpu/drm/nouveau/nouveau_svm.c
461
const u32 timelo = nvif_rd32(memory, offset + 0x10);
drivers/gpu/drm/nouveau/nouveau_svm.c
462
const u32 timehi = nvif_rd32(memory, offset + 0x14);
drivers/gpu/drm/nouveau/nouveau_svm.c
463
const u32 engine = nvif_rd32(memory, offset + 0x18);
drivers/gpu/drm/nouveau/nouveau_svm.c
464
const u32 info = nvif_rd32(memory, offset + 0x1c);
drivers/gpu/drm/nouveau/nouveau_svm.c
475
nvif_mask(memory, offset + 0x1c, 0x80000000, 0x00000000);
drivers/gpu/drm/nouveau/nvif/vmm.c
148
.memory = nvif_handle(&mem->object),
drivers/gpu/drm/nouveau/nvif/vmm.c
53
args->memory = nvif_handle(&mem->object);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
113
#define nvkm_firmware_mem(p) container_of((p), struct nvkm_firmware, mem.memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
116
nvkm_firmware_mem_sgl(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
118
struct nvkm_firmware *fw = nvkm_firmware_mem(memory);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
132
nvkm_firmware_mem_map(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm,
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
135
struct nvkm_firmware *fw = nvkm_firmware_mem(memory);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
137
.memory = &fw->mem.memory,
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
139
.sgl = nvkm_firmware_mem_sgl(memory),
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
149
nvkm_firmware_mem_size(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
151
struct scatterlist *sgl = nvkm_firmware_mem_sgl(memory);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
157
nvkm_firmware_mem_addr(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
159
BUG_ON(nvkm_firmware_mem(memory)->func->type != NVKM_FIRMWARE_IMG_DMA);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
160
return nvkm_firmware_mem(memory)->phys;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
164
nvkm_firmware_mem_page(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
170
nvkm_firmware_mem_target(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
172
if (nvkm_firmware_mem(memory)->device->func->tegra)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
179
nvkm_firmware_mem_dtor(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
197
struct nvkm_memory *memory = &fw->mem.memory;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
207
nvkm_memory_unref(&memory);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
212
nvkm_memory_unref(&memory);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
307
nvkm_memory_ctor(&nvkm_firmware_mem, &fw->mem.memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
202
abs(align), zero, &gpuobj->memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
207
gpuobj->addr = nvkm_memory_addr(gpuobj->memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
208
gpuobj->size = nvkm_memory_size(gpuobj->memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
222
nvkm_memory_unref(&gpuobj->memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
250
nvkm_gpuobj_wrap(struct nvkm_memory *memory, struct nvkm_gpuobj **pgpuobj)
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
255
(*pgpuobj)->addr = nvkm_memory_addr(memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
256
(*pgpuobj)->size = nvkm_memory_size(memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
50
return nvkm_memory_map(gpuobj->memory, offset, vmm, vma, argv, argc);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
56
return nvkm_ro32(gpuobj->memory, offset);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
62
nvkm_wo32(gpuobj->memory, offset, data);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
70
nvkm_done(gpuobj->memory);
drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c
92
gpuobj->map = nvkm_kmap(gpuobj->memory);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
101
struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/memory.c
103
memory->func = func;
drivers/gpu/drm/nouveau/nvkm/core/memory.c
104
kref_init(&memory->kref);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
110
struct nvkm_memory *memory = container_of(kref, typeof(*memory), kref);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
111
if (!WARN_ON(!memory->func)) {
drivers/gpu/drm/nouveau/nvkm/core/memory.c
112
if (memory->func->dtor)
drivers/gpu/drm/nouveau/nvkm/core/memory.c
113
memory = memory->func->dtor(memory);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
114
kfree(memory);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
121
struct nvkm_memory *memory = *pmemory;
drivers/gpu/drm/nouveau/nvkm/core/memory.c
122
if (memory) {
drivers/gpu/drm/nouveau/nvkm/core/memory.c
123
kref_put(&memory->kref, nvkm_memory_del);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
129
nvkm_memory_ref(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/core/memory.c
131
if (memory)
drivers/gpu/drm/nouveau/nvkm/core/memory.c
132
kref_get(&memory->kref);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
133
return memory;
drivers/gpu/drm/nouveau/nvkm/core/memory.c
142
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/core/memory.c
159
ret = nvkm_instobj_new(imem, size, align, zero, preserve, &memory);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
163
*pmemory = memory;
drivers/gpu/drm/nouveau/nvkm/core/memory.c
30
nvkm_memory_tags_put(struct nvkm_memory *memory, struct nvkm_device *device,
drivers/gpu/drm/nouveau/nvkm/core/memory.c
39
kfree(memory->tags);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
40
memory->tags = NULL;
drivers/gpu/drm/nouveau/nvkm/core/memory.c
48
nvkm_memory_tags_get(struct nvkm_memory *memory, struct nvkm_device *device,
drivers/gpu/drm/nouveau/nvkm/core/memory.c
56
if ((tags = memory->tags)) {
drivers/gpu/drm/nouveau/nvkm/core/memory.c
94
*ptags = memory->tags = tags;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
160
nvkm_memory_unref(&chan->memory);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.h
21
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
1186
switch (nvkm_memory_target(disp->inst->memory)) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
686
chan->memory = nvkm_umem_search(chan->object.client, object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
687
if (IS_ERR(chan->memory))
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
688
return PTR_ERR(chan->memory);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
690
if (nvkm_memory_size(chan->memory) < 0x1000)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
693
switch (nvkm_memory_target(chan->memory)) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
701
chan->push |= nvkm_memory_addr(chan->memory) >> 8;
drivers/gpu/drm/nouveau/nvkm/engine/disp/tu102.c
170
switch (nvkm_memory_target(disp->inst->memory)) {
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv04.c
53
device->mmu->vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
285
nvkm_vmm_part(chan->vmm, chan->inst->memory);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
421
ret = nvkm_vmm_join(vmm, chan->inst->memory);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ga100.c
320
ga100_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ga100.c
323
u64 addr = nvkm_memory_addr(memory) + start;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
400
gf100_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
403
u64 addr = nvkm_memory_addr(memory) + start;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
406
switch (nvkm_memory_target(memory)) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
419
gf100_runl_insert_chan(struct nvkm_chan *chan, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
421
nvkm_wo32(memory, offset + 0, chan->id);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
422
nvkm_wo32(memory, offset + 4, 0x00000004);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
430
gk104_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
434
u64 addr = nvkm_memory_addr(memory) + start;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
437
switch (nvkm_memory_target(memory)) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
452
gk104_runl_insert_chan(struct nvkm_chan *chan, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
454
nvkm_wo32(memory, offset + 0, chan->id);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
455
nvkm_wo32(memory, offset + 4, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c
72
gk110_runl_insert_cgrp(struct nvkm_cgrp *cgrp, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c
74
nvkm_wo32(memory, offset + 0, (cgrp->chan_nr << 26) | (128 << 18) |
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c
76
nvkm_wo32(memory, offset + 4, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c
46
gm107_runl_insert_chan(struct nvkm_chan *chan, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c
48
nvkm_wo32(memory, offset + 0, chan->id);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c
49
nvkm_wo32(memory, offset + 4, chan->inst->addr >> 12);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c
33
gp100_runl_insert_chan(struct nvkm_chan *chan, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c
35
nvkm_wo32(memory, offset + 0, chan->id | chan->runq << 14);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c
36
nvkm_wo32(memory, offset + 4, chan->inst->addr >> 12);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
118
const u64 bar2 = cctx ? nvkm_memory_bar2(cctx->vctx->inst->memory) : 0ULL;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
130
if (nvkm_memory_bar2(vctx->inst->memory) == ~0ULL)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
182
gv100_runl_insert_chan(struct nvkm_chan *chan, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
187
nvkm_wo32(memory, offset + 0x0, lower_32_bits(user) | chan->runq << 1);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
188
nvkm_wo32(memory, offset + 0x4, upper_32_bits(user));
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
189
nvkm_wo32(memory, offset + 0x8, lower_32_bits(inst) | chan->id);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
190
nvkm_wo32(memory, offset + 0xc, upper_32_bits(inst));
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
194
gv100_runl_insert_cgrp(struct nvkm_cgrp *cgrp, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
196
nvkm_wo32(memory, offset + 0x0, (128 << 24) | (3 << 16) | 0x00000001);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
197
nvkm_wo32(memory, offset + 0x4, cgrp->chan_nr);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
198
nvkm_wo32(memory, offset + 0x8, cgrp->id);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
199
nvkm_wo32(memory, offset + 0xc, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
238
nv50_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
241
u64 addr = nvkm_memory_addr(memory) + start;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
248
nv50_runl_insert_chan(struct nvkm_chan *chan, struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
250
nvkm_wo32(memory, offset, chan->id);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
287
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
295
memory = nv50_runl_alloc(runl, &start);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
296
if (IS_ERR(memory))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
297
return PTR_ERR(memory);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
302
nvkm_kmap(memory);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
306
runl->func->insert_cgrp(cgrp, memory, offset);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
312
runl->func->insert_chan(chan, memory, offset);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
316
nvkm_done(memory);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
324
runl->func->commit(runl, memory, start, count);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c
71
tu102_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c
74
u64 addr = nvkm_memory_addr(memory) + start;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
402
switch (nvkm_memory_target(chan->inst->memory)) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
413
args->v0.inst = nvkm_memory_addr(chan->inst->memory);
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
138
ret = nvkm_vmm_get(fw->vmm, 12, nvkm_memory_size(&fw->fw.mem.memory), &fw->vma);
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
144
ret = nvkm_memory_map(&fw->fw.mem.memory, 0, fw->vmm, fw->vma, NULL, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
147
ret = nvkm_vmm_join(bar->bar2_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
183
ret = nvkm_vmm_join(bar->bar1_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
209
nvkm_vmm_part(bar->bar1_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
212
nvkm_vmm_part(bar->bar2_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c
106
info->memory = nvbios_rd32(bios, perf + 0x05) * 20;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c
122
info->memory = nvbios_rd16(bios, perf + 0x0b) * 1000;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c
125
info->memory = nvbios_rd16(bios, perf + 0x0b) * 2000;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c
134
info->memory = nvbios_rd16(bios, perf + 0x0c) * 1000;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c
144
info->memory = nvbios_rd16(bios, perf + 0x0c) * 1000;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c
433
cstate->domain[nv_clk_src_mem] = perfE.memory;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
133
nvkm_memory_ctor(&nvkm_vram, &vram->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
136
*pmemory = &vram->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
168
nvkm_memory_ctor(&nvkm_vram, &vram->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
171
*pmemory = &vram->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
24
#define nvkm_vram(p) container_of((p), struct nvkm_vram, memory)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
32
struct nvkm_memory memory;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
39
nvkm_vram_kmap(struct nvkm_memory *memory, struct nvkm_memory **pmemory)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
41
return nvkm_instobj_wrap(nvkm_vram(memory)->ram->fb->subdev.device, memory, pmemory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
45
nvkm_vram_map(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
48
struct nvkm_vram *vram = nvkm_vram(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
50
.memory = &vram->memory,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
59
nvkm_vram_size(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
61
return (u64)nvkm_mm_size(nvkm_vram(memory)->mn) << NVKM_RAM_MM_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
65
nvkm_vram_addr(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
67
struct nvkm_vram *vram = nvkm_vram(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
74
nvkm_vram_page(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
76
return nvkm_vram(memory)->page;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
80
nvkm_vram_target(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
86
nvkm_vram_dtor(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
88
struct nvkm_vram *vram = nvkm_vram(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
302
true, true, &ram->ltrain.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
306
addr = nvkm_memory_addr(ram->ltrain.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
340
nvkm_memory_unref(&ram->ltrain.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
90
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
248
} while (perfE.memory < freq);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
140
nvkm_memory_unref(&vmm->pd->pt[0]->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
142
ret = nvkm_memory_kmap(pd3, &vmm->pd->pt[0]->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
147
vmm->pd->pt[0]->addr = nvkm_memory_addr(vmm->pd->pt[0]->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
86
struct nvkm_memory *pdb = vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
102
ctrl->physicalAddr = nvkm_memory_addr(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
103
ctrl->limit = nvkm_memory_size(memory) - 1;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1521
if (WARN_ON(nvkm_memory_target(disp->inst->memory) != NVKM_MEM_TARGET_VRAM))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1530
ctrl->instMemPhysAddr = nvkm_memory_addr(disp->inst->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1531
ctrl->instMemSize = nvkm_memory_size(disp->inst->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
190
ret = rmapi->disp->chan.set_pushbuf(chan->disp, chan->object.oclass, chan->head, chan->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
199
r535_dmac_push(struct nvkm_disp_chan *chan, u64 memory)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
201
chan->memory = nvkm_umem_search(chan->object.client, memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
202
if (IS_ERR(chan->memory))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
203
return PTR_ERR(chan->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
73
r535_disp_chan_set_pushbuf(struct nvkm_disp *disp, s32 oclass, int inst, struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
84
if (memory) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
85
switch (nvkm_memory_target(memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
198
fbsr_inst(struct fbsr *fbsr, const char *type, struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
200
return fbsr_vram(fbsr, type, nvkm_memory_addr(memory), nvkm_memory_size(memory));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
228
if (!fbsr_inst(&fbsr, "inst", &iobj->memory))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
234
if (!fbsr_inst(&fbsr, "boot", &iobj->memory))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
33
r570_disp_chan_set_pushbuf(struct nvkm_disp *disp, s32 oclass, int inst, struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
44
if (memory) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
45
switch (nvkm_memory_target(memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
65
ctrl->physicalAddr = nvkm_memory_addr(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
66
ctrl->limit = nvkm_memory_size(memory) - 1;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
102
ret = imem->func->memory_wrap(imem, memory, pmemory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
106
container_of(*pmemory, struct nvkm_instobj, memory)->preserve = true;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
115
struct nvkm_memory *memory = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
119
ret = imem->func->memory_new(imem, size, align, zero, &memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
126
zero, nvkm_memory_addr(memory), nvkm_memory_size(memory));
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
129
void __iomem *map = nvkm_kmap(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
132
nvkm_wo32(memory, offset, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
136
nvkm_done(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
139
container_of(memory, struct nvkm_instobj, memory)->preserve = preserve;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
142
nvkm_memory_unref(&memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
143
*pmemory = memory;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
34
struct nvkm_memory *memory = &iobj->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
35
const u64 size = nvkm_memory_size(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
39
if (!(map = nvkm_kmap(memory))) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
41
nvkm_wo32(memory, i, iobj->suspend[i / 4]);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
45
nvkm_done(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
54
struct nvkm_memory *memory = &iobj->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
55
const u64 size = nvkm_memory_size(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
63
if (!(map = nvkm_kmap(memory))) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
65
iobj->suspend[i / 4] = nvkm_ro32(memory, i);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
69
nvkm_done(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
85
nvkm_memory_ctor(func, &iobj->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/base.c
94
struct nvkm_memory *memory, struct nvkm_memory **pmemory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
116
gk20a_instobj_target(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
122
gk20a_instobj_page(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
128
gk20a_instobj_addr(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
130
return (u64)gk20a_instobj(memory)->mn->offset << 12;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
134
gk20a_instobj_size(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
136
return (u64)gk20a_instobj(memory)->mn->length << 12;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
151
imem->vaddr_use -= nvkm_memory_size(&obj->base.base.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
174
gk20a_instobj_acquire_dma(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
176
struct gk20a_instobj *node = gk20a_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
186
gk20a_instobj_acquire_iommu(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
188
struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
191
const u64 size = nvkm_memory_size(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
229
gk20a_instobj_release_dma(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
231
struct gk20a_instobj *node = gk20a_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
241
gk20a_instobj_release_iommu(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
243
struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
265
gk20a_instobj_rd32(struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
267
struct gk20a_instobj *node = gk20a_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
273
gk20a_instobj_wr32(struct nvkm_memory *memory, u64 offset, u32 data)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
275
struct gk20a_instobj *node = gk20a_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
281
gk20a_instobj_map(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
284
struct gk20a_instobj *node = gk20a_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
286
.memory = &node->base.memory,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
295
gk20a_instobj_dtor_dma(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
297
struct gk20a_instobj_dma *node = gk20a_instobj_dma(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
312
gk20a_instobj_dtor_iommu(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
314
struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
394
nvkm_memory_ctor(&gk20a_instobj_func_dma, &node->base.base.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
395
node->base.base.memory.ptrs = &gk20a_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
441
nvkm_memory_ctor(&gk20a_instobj_func_iommu, &node->base.base.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
442
node->base.base.memory.ptrs = &gk20a_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
536
*pmemory = node ? &node->base.memory : NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
59
#define gk20a_instobj(p) container_of((p), struct gk20a_instobj, base.memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
100
nv04_instobj_dtor(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
102
struct nv04_instobj *iobj = nv04_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
130
*pmemory = &iobj->base.memory;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
133
iobj->base.memory.ptrs = &nv04_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
38
#define nv04_instobj(p) container_of((p), struct nv04_instobj, base.memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
47
nv04_instobj_wr32(struct nvkm_memory *memory, u64 offset, u32 data)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
49
struct nv04_instobj *iobj = nv04_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
55
nv04_instobj_rd32(struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
57
struct nv04_instobj *iobj = nv04_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
69
nv04_instobj_release(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
74
nv04_instobj_acquire(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
76
struct nv04_instobj *iobj = nv04_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
82
nv04_instobj_size(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
84
return nv04_instobj(memory)->node->length;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
88
nv04_instobj_addr(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
90
return nv04_instobj(memory)->node->offset;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
94
nv04_instobj_target(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
101
struct nv40_instobj *iobj = nv40_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
129
*pmemory = &iobj->base.memory;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
132
iobj->base.memory.ptrs = &nv40_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
39
#define nv40_instobj(p) container_of((p), struct nv40_instobj, base.memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
48
nv40_instobj_wr32(struct nvkm_memory *memory, u64 offset, u32 data)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
50
struct nv40_instobj *iobj = nv40_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
55
nv40_instobj_rd32(struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
57
struct nv40_instobj *iobj = nv40_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
68
nv40_instobj_release(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
74
nv40_instobj_acquire(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
76
struct nv40_instobj *iobj = nv40_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
81
nv40_instobj_size(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
83
return nv40_instobj(memory)->node->length;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
87
nv40_instobj_addr(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
89
return nv40_instobj(memory)->node->offset;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
93
nv40_instobj_target(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
99
nv40_instobj_dtor(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
103
nv50_instobj_wr32(struct nvkm_memory *memory, u64 offset, u32 data)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
105
iowrite32_native(data, nv50_instobj(memory)->map + offset);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
109
nv50_instobj_rd32(struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
111
return ioread32_native(nv50_instobj(memory)->map + offset);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
125
struct nvkm_memory *memory = &iobj->base.memory;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
129
u64 size = nvkm_memory_size(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
146
nvkm_memory_addr(&eobj->base.memory),
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
147
nvkm_memory_size(&eobj->base.memory),
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
163
ret = nvkm_memory_map(memory, 0, vmm, bar, NULL, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
184
nv50_instobj_map(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
187
memory = nv50_instobj(memory)->ram;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
188
return nvkm_memory_map(memory, offset, vmm, vma, argv, argc);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
192
nv50_instobj_release(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
194
struct nv50_instobj *iobj = nv50_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
211
iobj->base.memory.ptrs = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
217
nv50_instobj_acquire(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
219
struct nv50_instobj *iobj = nv50_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
253
iobj->base.memory.ptrs = &nv50_instobj_fast;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
255
iobj->base.memory.ptrs = &nv50_instobj_slow;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
266
nv50_instobj_boot(struct nvkm_memory *memory, struct nvkm_vmm *vmm)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
268
struct nv50_instobj *iobj = nv50_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
286
nv50_instobj_size(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
288
return nvkm_memory_size(nv50_instobj(memory)->ram);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
292
nv50_instobj_addr(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
294
return nvkm_memory_addr(nv50_instobj(memory)->ram);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
298
nv50_instobj_bar2(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
300
struct nv50_instobj *iobj = nv50_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
302
if (nv50_instobj_acquire(&iobj->base.memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
306
nv50_instobj_release(&iobj->base.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
311
nv50_instobj_target(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
313
return nvkm_memory_target(nv50_instobj(memory)->ram);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
317
nv50_instobj_dtor(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
319
struct nv50_instobj *iobj = nv50_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
358
struct nvkm_memory *memory, struct nvkm_memory **pmemory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
365
*pmemory = &iobj->base.memory;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
372
iobj->ram = nvkm_memory_ref(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
44
#define nv50_instobj(p) container_of((p), struct nv50_instobj, base.memory)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
57
nv50_instobj_wr32_slow(struct nvkm_memory *memory, u64 offset, u32 data)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
59
struct nv50_instobj *iobj = nv50_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
76
nv50_instobj_rd32_slow(struct nvkm_memory *memory, u64 offset)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
78
struct nv50_instobj *iobj = nv50_instobj(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/priv.h
43
struct nvkm_memory memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
106
pt->memory = pt->ptp->pt->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
159
nvkm_memory_unref(&pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
193
nvkm_fo64(pt->memory, 0, 0, size >> 3);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
208
size, align, zero, &pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
215
pt->addr = nvkm_memory_addr(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
226
nvkm_memory_unref(&pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
90
slot = nvkm_memory_size(ptp->pt->memory) >> ptp->shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
110
nvkm_mem_map_sgl(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
113
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
115
.memory = &mem->memory,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
133
nvkm_mem_map_host(struct nvkm_memory *memory, void **pmap)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
135
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
170
*pmemory = &mem->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
174
nvkm_memory_ctor(&nvkm_mem_dma, &mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
177
nvkm_memory_ctor(&nvkm_mem_sgl, &mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
191
nvkm_memory_ctor(&nvkm_mem_dma, &mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
22
#define nvkm_mem(p) container_of((p), struct nvkm_mem, memory)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
227
struct nvkm_memory *memory = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
232
argv, argc, &memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
235
argv, argc, &memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
239
nvkm_memory_unref(&memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
240
*pmemory = memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
31
struct nvkm_memory memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
43
nvkm_mem_target(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
45
return nvkm_mem(memory)->target;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
49
nvkm_mem_page(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
55
nvkm_mem_addr(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
57
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
64
nvkm_mem_size(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
66
return nvkm_mem(memory)->pages << PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
70
nvkm_mem_map_dma(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
73
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
75
.memory = &mem->memory,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
83
nvkm_mem_dtor(struct nvkm_memory *memory)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
85
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c
34
gf100_mem_map(struct nvkm_mmu *mmu, struct nvkm_memory *memory, void *argv,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c
54
ret = nvkm_vmm_get(bar, nvkm_memory_page(memory),
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c
55
nvkm_memory_size(memory), pvma);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c
59
ret = nvkm_memory_map(memory, 0, bar, *pvma, &uvmm, sizeof(uvmm));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c
31
nv04_mem_map(struct nvkm_mmu *mmu, struct nvkm_memory *memory, void *argv,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c
38
const u64 addr = nvkm_memory_addr(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c
45
*psize = nvkm_memory_size(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c
34
nv50_mem_map(struct nvkm_mmu *mmu, struct nvkm_memory *memory, void *argv,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c
44
u64 size = nvkm_memory_size(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c
62
return nvkm_memory_map(memory, 0, bar, *pvma, &uvmm, sizeof(uvmm));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv44.c
35
struct nvkm_memory *pt = mmu->vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/priv.h
63
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
103
*length = nvkm_memory_size(umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
109
int ret = mmu->func->mem.umap(mmu, umem->memory, argv, argc,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
130
nvkm_memory_unref(&umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
178
&umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
186
args->v0.page = nvkm_memory_page(umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
187
args->v0.addr = nvkm_memory_addr(umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
188
args->v0.size = nvkm_memory_size(umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
38
struct nvkm_memory *memory = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
48
memory = nvkm_memory_ref(umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
56
memory = nvkm_memory_ref(umem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
59
return memory ? memory : ERR_PTR(-ENOENT);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
98
int ret = nvkm_mem_map_host(umem->memory, &umem->map);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.h
14
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
138
if (ret = -EINVAL, !vma->memory) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
160
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
166
handle = args->v0.memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
174
memory = nvkm_umem_search(client, handle);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
175
if (IS_ERR(memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
176
VMM_DEBUG(vmm, "memory %016llx %ld\n", handle, PTR_ERR(memory));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
177
return PTR_ERR(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
191
if (ret = -EINVAL, vma->mapped && !vma->memory) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
197
if (addr + size > vma->addr + vma->size || vma->memory ||
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
201
!!vma->memory, vma->refd, vma->mapref,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
215
ret = nvkm_memory_map(memory, offset, vmm, vma, argv, argc);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
218
nvkm_memory_unref(&memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
227
nvkm_memory_unref(&memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
401
struct nvkm_memory *memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
404
u64 handle = args->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
417
memory = nvkm_umem_search(client, args->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
418
if (IS_ERR(memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
419
VMM_DEBUG(vmm, "memory %016llx %ld\n", handle, PTR_ERR(memory));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
420
return PTR_ERR(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
423
ret = nvkm_memory_map(memory, args->offset, vmm, &vma, argv, argc);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
425
nvkm_memory_unref(&vma.memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
426
nvkm_memory_unref(&memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1039
vma->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1242
if (prev->memory || prev->mapped != map)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1248
next->memory || next->mapped != map)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1269
if (!vma->mapped || vma->memory)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1341
if (!vma->mapref || vma->memory) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1422
nvkm_memory_tags_put(vma->memory, vmm->mmu->subdev.device, &vma->tags);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1423
nvkm_memory_unref(&vma->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1451
if (vma->memory) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1462
switch (nvkm_memory_target(map->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1484
nvkm_memory_page(map->memory) < map->page->shift) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1487
nvkm_memory_page(map->memory));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1516
if (unlikely(nvkm_memory_size(map->memory) < map->offset + vma->size)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1518
nvkm_memory_size(map->memory),
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1586
nvkm_memory_tags_put(vma->memory, vmm->mmu->subdev.device, &vma->tags);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1587
nvkm_memory_unref(&vma->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1588
vma->memory = nvkm_memory_ref(map->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1640
const bool mem = next->memory != NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1649
(next->memory != NULL) == mem &&
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1939
nvkm_memory_boot(it->pt[0]->pt[type]->memory, it->vmm);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
327
nvkm_kmap((PT)->memory); \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
346
nvkm_done((PT)->memory); \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
362
#define VMM_FO(m,o,d,c,b) nvkm_fo##b((m)->memory, (o), (d), (c))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
363
#define VMM_WO(m,o,d,c,b) nvkm_wo##b((m)->memory, (o), (d))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
383
nvkm_wo64((m)->memory, (m)->base + _pteo + 0, (lo)); \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
384
nvkm_wo64((m)->memory, (m)->base + _pteo + 8, (hi)); \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
391
nvkm_kmap((m)->memory); \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
393
nvkm_done((m)->memory); \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
114
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
128
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
141
nvkm_kmap(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
143
nvkm_done(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
204
switch (nvkm_memory_target(pd->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
240
const enum nvkm_memory_target target = nvkm_memory_target(map->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
248
struct nvkm_memory *memory = map->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
284
u32 tags = ALIGN(nvkm_memory_size(memory), 1 << 17) >> comp;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
291
ret = nvkm_memory_tags_get(memory, device, tags,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
346
switch (nvkm_memory_target(pd->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
70
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
76
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
115
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
149
nvkm_kmap(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
151
nvkm_done(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
188
nvkm_kmap(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
190
nvkm_done(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
226
const enum nvkm_memory_target target = nvkm_memory_target(map->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
38
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
44
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
109
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
159
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
164
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
239
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
265
nvkm_kmap(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
267
nvkm_done(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
292
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
294
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
295
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
304
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
313
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
315
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
316
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
325
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
336
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
366
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
38
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
390
nvkm_kmap(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
392
nvkm_done(pd->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
40
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
41
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
426
const enum nvkm_memory_target target = nvkm_memory_target(map->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
49
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
57
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
59
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
60
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
68
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
79
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
135
mem = vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
50
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
53
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
49
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
54
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
116
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
140
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
150
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
168
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
33
tmp[0] = nvkm_ro32(pt->memory, pteo + 0x0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
34
tmp[1] = nvkm_ro32(pt->memory, pteo + 0x4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
35
tmp[2] = nvkm_ro32(pt->memory, pteo + 0x8);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
36
tmp[3] = nvkm_ro32(pt->memory, pteo + 0xc);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
114
switch (nvkm_memory_size(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
129
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
236
struct nvkm_memory *memory = map->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
260
switch (nvkm_memory_target(memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
293
u32 tags = (nvkm_memory_size(memory) >> 16) * comp;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
300
ret = nvkm_memory_tags_get(memory, device, tags, NULL,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
70
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
76
nvkm_done(pt->memory);
drivers/gpu/drm/radeon/radeon.h
1095
struct list_head memory;
drivers/gpu/drm/radeon/radeon_agp.c
154
INIT_LIST_HEAD(&head->memory);
drivers/gpu/drm/xe/xe_lrc.c
2282
snapshot->tail.memory = xe_lrc_ring_tail(lrc);
drivers/gpu/drm/xe/xe_lrc.c
2341
snapshot->tail.internal, snapshot->tail.memory);
drivers/gpu/drm/xe/xe_lrc.h
36
u32 memory;
drivers/infiniband/hw/mlx5/odp.c
1541
pfault->memory.va - pfault->memory.prefetch_before_byte_count;
drivers/infiniband/hw/mlx5/odp.c
1542
size_t prefetch_size = pfault->memory.prefetch_before_byte_count +
drivers/infiniband/hw/mlx5/odp.c
1543
pfault->memory.fault_byte_count +
drivers/infiniband/hw/mlx5/odp.c
1544
pfault->memory.prefetch_after_byte_count;
drivers/infiniband/hw/mlx5/odp.c
1549
mmkey = find_odp_mkey(dev, pfault->memory.mkey);
drivers/infiniband/hw/mlx5/odp.c
1569
ret = pagefault_mr(mr, pfault->memory.va,
drivers/infiniband/hw/mlx5/odp.c
1570
pfault->memory.fault_byte_count, NULL, 0,
drivers/infiniband/hw/mlx5/odp.c
1579
if (pfault->memory.flags & MLX5_MEMORY_PAGE_FAULT_FLAGS_LAST)
drivers/infiniband/hw/mlx5/odp.c
1585
pfault->memory.flags & MLX5_MEMORY_PAGE_FAULT_FLAGS_LAST ?
drivers/infiniband/hw/mlx5/odp.c
1588
pfault->token, pfault->memory.mkey, pfault->memory.va,
drivers/infiniband/hw/mlx5/odp.c
1589
pfault->memory.fault_byte_count);
drivers/infiniband/hw/mlx5/odp.c
1600
pfault->token, pfault->memory.mkey, pfault->memory.va,
drivers/infiniband/hw/mlx5/odp.c
1601
pfault->memory.fault_byte_count, ret);
drivers/infiniband/hw/mlx5/odp.c
1711
be32_to_cpu(pf_eqe->memory.token31_0) |
drivers/infiniband/hw/mlx5/odp.c
1712
((u64)be16_to_cpu(pf_eqe->memory.token47_32)
drivers/infiniband/hw/mlx5/odp.c
1714
pfault->memory.va = be64_to_cpu(pf_eqe->memory.va);
drivers/infiniband/hw/mlx5/odp.c
1715
pfault->memory.mkey = be32_to_cpu(pf_eqe->memory.mkey);
drivers/infiniband/hw/mlx5/odp.c
1716
pfault->memory.fault_byte_count = (be32_to_cpu(
drivers/infiniband/hw/mlx5/odp.c
1717
pf_eqe->memory.demand_fault_pages) >> 12) *
drivers/infiniband/hw/mlx5/odp.c
1719
pfault->memory.prefetch_before_byte_count =
drivers/infiniband/hw/mlx5/odp.c
1721
pf_eqe->memory.pre_demand_fault_pages) *
drivers/infiniband/hw/mlx5/odp.c
1723
pfault->memory.prefetch_after_byte_count =
drivers/infiniband/hw/mlx5/odp.c
1725
pf_eqe->memory.post_demand_fault_pages) *
drivers/infiniband/hw/mlx5/odp.c
1727
pfault->memory.flags = pf_eqe->memory.flags;
drivers/infiniband/hw/mlx5/odp.c
1732
pfault->memory.mkey,
drivers/infiniband/hw/mlx5/odp.c
1733
pfault->memory.fault_byte_count,
drivers/infiniband/hw/mlx5/odp.c
1734
pfault->memory.va, pfault->memory.flags);
drivers/infiniband/hw/mlx5/odp.c
1738
pfault->memory.prefetch_before_byte_count,
drivers/infiniband/hw/mlx5/odp.c
1739
pfault->memory.prefetch_after_byte_count);
drivers/infiniband/hw/mlx5/odp.c
87
} memory;
drivers/iommu/iommufd/selftest.c
1959
void *memory;
drivers/iommu/iommufd/selftest.c
1992
kfree(priv->memory);
drivers/iommu/iommufd/selftest.c
2017
phys->paddr = virt_to_phys(priv->memory);
drivers/iommu/iommufd/selftest.c
2040
priv->memory = kzalloc(len, GFP_KERNEL);
drivers/iommu/iommufd/selftest.c
2041
if (!priv->memory) {
drivers/iommu/iommufd/selftest.c
2060
kfree(priv->memory);
drivers/md/dm-vdo/indexer/chapter-index.c
123
u8 *memory, u32 first_list, bool last_page,
drivers/md/dm-vdo/indexer/chapter-index.c
139
result = uds_pack_delta_index_page(delta_index, nonce, memory,
drivers/md/dm-vdo/indexer/chapter-index.h
46
u8 *memory, u32 first_list,
drivers/md/dm-vdo/indexer/delta-index.c
1110
zone->memory + get_delta_list_byte_start(delta_list),
drivers/md/dm-vdo/indexer/delta-index.c
1264
delta_list->start = get_immutable_start(delta_zone->memory, list_number);
drivers/md/dm-vdo/indexer/delta-index.c
1265
end_offset = get_immutable_start(delta_zone->memory, list_number + 1);
drivers/md/dm-vdo/indexer/delta-index.c
1282
uds_prefetch_range(&delta_zone->memory[delta_list->start / BITS_PER_BYTE],
drivers/md/dm-vdo/indexer/delta-index.c
1313
const u8 *memory = delta_zone->memory;
drivers/md/dm-vdo/indexer/delta-index.c
1315
const u8 *addr = memory + delta_offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
1416
const u8 *addr = entry->delta_zone->memory + offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
1427
u8 *addr = entry->delta_zone->memory + offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
1503
return get_field(delta_entry->delta_zone->memory,
drivers/md/dm-vdo/indexer/delta-index.c
1536
set_field(value, delta_entry->delta_zone->memory,
drivers/md/dm-vdo/indexer/delta-index.c
1587
u8 *memory;
drivers/md/dm-vdo/indexer/delta-index.c
1653
memory = delta_zone->memory;
drivers/md/dm-vdo/indexer/delta-index.c
1654
move_bits(memory, source, memory, destination, count);
drivers/md/dm-vdo/indexer/delta-index.c
1665
u8 *memory = delta_zone->memory;
drivers/md/dm-vdo/indexer/delta-index.c
1669
set_field(delta_entry->delta, memory, offset, delta_zone->min_bits);
drivers/md/dm-vdo/indexer/delta-index.c
1676
set_field(t1, memory, offset, delta_zone->min_bits);
drivers/md/dm-vdo/indexer/delta-index.c
1677
set_zero(memory, offset + delta_zone->min_bits, t2);
drivers/md/dm-vdo/indexer/delta-index.c
1678
set_field(1, memory, offset + delta_zone->min_bits + t2, 1);
drivers/md/dm-vdo/indexer/delta-index.c
1684
u8 *memory = delta_entry->delta_zone->memory;
drivers/md/dm-vdo/indexer/delta-index.c
1687
set_field(value, memory, offset, delta_entry->value_bits);
drivers/md/dm-vdo/indexer/delta-index.c
171
memmove(delta_zone->memory + destination,
drivers/md/dm-vdo/indexer/delta-index.c
172
delta_zone->memory + source,
drivers/md/dm-vdo/indexer/delta-index.c
1807
u8 *memory = delta_entry->delta_zone->memory;
drivers/md/dm-vdo/indexer/delta-index.c
1843
move_bits(memory, source, memory, destination, count);
drivers/md/dm-vdo/indexer/delta-index.c
233
memset(zone->memory + (list_bits / BITS_PER_BYTE), ~0,
drivers/md/dm-vdo/indexer/delta-index.c
301
vdo_free(vdo_forget(delta_index->delta_zones[z].memory));
drivers/md/dm-vdo/indexer/delta-index.c
314
result = vdo_allocate(size, u8, "delta list", &delta_zone->memory);
drivers/md/dm-vdo/indexer/delta-index.c
404
static inline u32 get_field(const u8 *memory, u64 offset, u8 size)
drivers/md/dm-vdo/indexer/delta-index.c
406
const void *addr = memory + offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
412
static inline void set_field(u32 value, u8 *memory, u64 offset, u8 size)
drivers/md/dm-vdo/indexer/delta-index.c
414
void *addr = memory + offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
431
static inline u32 get_immutable_start(const u8 *memory, u32 list_number)
drivers/md/dm-vdo/indexer/delta-index.c
433
return get_field(memory, get_immutable_header_offset(list_number),
drivers/md/dm-vdo/indexer/delta-index.c
438
static inline void set_immutable_start(u8 *memory, u32 list_number, u32 start)
drivers/md/dm-vdo/indexer/delta-index.c
440
set_field(start, memory, get_immutable_header_offset(list_number),
drivers/md/dm-vdo/indexer/delta-index.c
445
u8 *memory, size_t memory_size)
drivers/md/dm-vdo/indexer/delta-index.c
465
if (get_immutable_start(memory, 0) != get_immutable_header_offset(list_count + 1))
drivers/md/dm-vdo/indexer/delta-index.c
470
if (get_immutable_start(memory, i) > get_immutable_start(memory, i + 1))
drivers/md/dm-vdo/indexer/delta-index.c
478
if (get_immutable_start(memory, list_count) >
drivers/md/dm-vdo/indexer/delta-index.c
484
if (memory[memory_size - POST_FIELD_GUARD_BYTES + i] != (u8) ~0)
drivers/md/dm-vdo/indexer/delta-index.c
495
u8 *memory, size_t memory_size)
drivers/md/dm-vdo/indexer/delta-index.c
501
struct delta_page_header *header = (struct delta_page_header *) memory;
drivers/md/dm-vdo/indexer/delta-index.c
513
if (!verify_delta_index_page(nonce, list_count, expected_nonce, memory,
drivers/md/dm-vdo/indexer/delta-index.c
520
if (!verify_delta_index_page(nonce, list_count, expected_nonce, memory,
drivers/md/dm-vdo/indexer/delta-index.c
543
delta_zone->memory = memory;
drivers/md/dm-vdo/indexer/delta-index.c
562
static inline u64 get_big_field(const u8 *memory, u64 offset, u8 size)
drivers/md/dm-vdo/indexer/delta-index.c
564
const void *addr = memory + offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
570
static inline void set_big_field(u64 value, u8 *memory, u64 offset, u8 size)
drivers/md/dm-vdo/indexer/delta-index.c
572
void *addr = memory + offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
582
static inline void set_zero(u8 *memory, u64 offset, u32 size)
drivers/md/dm-vdo/indexer/delta-index.c
585
u8 *addr = memory + offset / BITS_PER_BYTE;
drivers/md/dm-vdo/indexer/delta-index.c
708
u8 *memory, size_t memory_size, u64 virtual_chapter_number,
drivers/md/dm-vdo/indexer/delta-index.c
752
header = (struct delta_page_header *) memory;
drivers/md/dm-vdo/indexer/delta-index.c
761
set_immutable_start(memory, 0, offset);
drivers/md/dm-vdo/indexer/delta-index.c
764
set_immutable_start(memory, i + 1, offset);
drivers/md/dm-vdo/indexer/delta-index.c
769
move_bits(delta_zone->memory, delta_lists[i].start, memory,
drivers/md/dm-vdo/indexer/delta-index.c
770
get_immutable_start(memory, i), delta_lists[i].size);
drivers/md/dm-vdo/indexer/delta-index.c
774
memset(memory + memory_size - POST_FIELD_GUARD_BYTES, ~0,
drivers/md/dm-vdo/indexer/delta-index.c
984
move_bits(data, save_info->bit_offset, delta_zone->memory, delta_list->start,
drivers/md/dm-vdo/indexer/delta-index.h
208
u32 payload_bits, u8 *memory,
drivers/md/dm-vdo/indexer/delta-index.h
216
u64 header_nonce, u8 *memory,
drivers/md/dm-vdo/indexer/delta-index.h
43
u8 *memory;
drivers/md/raid5.c
7462
int raid_disk, memory, max_disks;
drivers/md/raid5.c
7666
memory = conf->min_nr_stripes * (sizeof(struct stripe_head) +
drivers/md/raid5.c
7671
mdname(mddev), memory);
drivers/md/raid5.c
7675
pr_debug("md/raid:%s: allocated %dkB\n", mdname(mddev), memory);
drivers/media/common/videobuf2/videobuf2-core.c
1025
q->memory = VB2_MEMORY_UNKNOWN;
drivers/media/common/videobuf2/videobuf2-core.c
1032
int vb2_core_create_bufs(struct vb2_queue *q, enum vb2_memory memory,
drivers/media/common/videobuf2/videobuf2-core.c
1062
q->memory = memory;
drivers/media/common/videobuf2/videobuf2-core.c
1069
if (q->memory != memory) {
drivers/media/common/videobuf2/videobuf2-core.c
1094
allocated_buffers = __vb2_queue_alloc(q, memory, num_buffers,
drivers/media/common/videobuf2/videobuf2-core.c
1150
q->memory = VB2_MEMORY_UNKNOWN;
drivers/media/common/videobuf2/videobuf2-core.c
1568
switch (q->memory) {
drivers/media/common/videobuf2/videobuf2-core.c
2399
if (q->memory != VB2_MEMORY_MMAP) {
drivers/media/common/videobuf2/videobuf2-core.c
2429
if (q->memory != VB2_MEMORY_MMAP) {
drivers/media/common/videobuf2/videobuf2-core.c
2654
q->memory = VB2_MEMORY_UNKNOWN;
drivers/media/common/videobuf2/videobuf2-core.c
2820
unsigned int memory;
drivers/media/common/videobuf2/videobuf2-core.c
2876
fileio->memory = VB2_MEMORY_MMAP;
drivers/media/common/videobuf2/videobuf2-core.c
2879
ret = vb2_core_reqbufs(q, fileio->memory, 0, &fileio->count);
drivers/media/common/videobuf2/videobuf2-core.c
2957
vb2_core_reqbufs(q, fileio->memory, 0, &fileio->count);
drivers/media/common/videobuf2/videobuf2-core.c
2977
vb2_core_reqbufs(q, fileio->memory, 0, &fileio->count);
drivers/media/common/videobuf2/videobuf2-core.c
415
if (q->memory == VB2_MEMORY_DMABUF) {
drivers/media/common/videobuf2/videobuf2-core.c
465
static int __vb2_queue_alloc(struct vb2_queue *q, enum vb2_memory memory,
drivers/media/common/videobuf2/videobuf2-core.c
511
vb->memory = memory;
drivers/media/common/videobuf2/videobuf2-core.c
522
if (memory == VB2_MEMORY_MMAP) {
drivers/media/common/videobuf2/videobuf2-core.c
570
if (q->memory == VB2_MEMORY_MMAP)
drivers/media/common/videobuf2/videobuf2-core.c
572
else if (q->memory == VB2_MEMORY_DMABUF)
drivers/media/common/videobuf2/videobuf2-core.c
689
q->memory = VB2_MEMORY_UNKNOWN;
drivers/media/common/videobuf2/videobuf2-core.c
778
enum vb2_memory memory, unsigned int type)
drivers/media/common/videobuf2/videobuf2-core.c
780
if (memory != VB2_MEMORY_MMAP && memory != VB2_MEMORY_USERPTR &&
drivers/media/common/videobuf2/videobuf2-core.c
781
memory != VB2_MEMORY_DMABUF) {
drivers/media/common/videobuf2/videobuf2-core.c
795
if (memory == VB2_MEMORY_MMAP && __verify_mmap_ops(q)) {
drivers/media/common/videobuf2/videobuf2-core.c
800
if (memory == VB2_MEMORY_USERPTR && __verify_userptr_ops(q)) {
drivers/media/common/videobuf2/videobuf2-core.c
805
if (memory == VB2_MEMORY_DMABUF && __verify_dmabuf_ops(q)) {
drivers/media/common/videobuf2/videobuf2-core.c
867
int vb2_core_reqbufs(struct vb2_queue *q, enum vb2_memory memory,
drivers/media/common/videobuf2/videobuf2-core.c
888
(q->memory != VB2_MEMORY_UNKNOWN && q->memory != memory) ||
drivers/media/common/videobuf2/videobuf2-core.c
895
if (debug && q->memory == VB2_MEMORY_MMAP &&
drivers/media/common/videobuf2/videobuf2-core.c
929
q->memory = memory;
drivers/media/common/videobuf2/videobuf2-core.c
958
__vb2_queue_alloc(q, memory, num_buffers, num_planes, plane_sizes, &first_index);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1026
int res = vb2_verify_memory_type(vdev->queue, p->memory, p->type);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1029
vb2_set_flags_and_caps(vdev->queue, p->memory, &flags,
drivers/media/common/videobuf2/videobuf2-v4l2.c
1036
res = vb2_core_reqbufs(vdev->queue, p->memory, p->flags, &p->count);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1049
int res = vb2_verify_memory_type(vdev->queue, p->memory, p->format.type);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1052
vb2_set_flags_and_caps(vdev->queue, p->memory, &p->flags,
drivers/media/common/videobuf2/videobuf2-v4l2.c
106
length = (b->memory == VB2_MEMORY_USERPTR ||
drivers/media/common/videobuf2/videobuf2-v4l2.c
107
b->memory == VB2_MEMORY_DMABUF)
drivers/media/common/videobuf2/videobuf2-v4l2.c
121
length = (b->memory == VB2_MEMORY_USERPTR)
drivers/media/common/videobuf2/videobuf2-v4l2.c
207
switch (b->memory) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
282
switch (b->memory) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
380
if (b->memory != q->memory) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
489
b->memory = vb->memory;
drivers/media/common/videobuf2/videobuf2-v4l2.c
512
if (q->memory == VB2_MEMORY_MMAP)
drivers/media/common/videobuf2/videobuf2-v4l2.c
514
else if (q->memory == VB2_MEMORY_USERPTR)
drivers/media/common/videobuf2/videobuf2-v4l2.c
516
else if (q->memory == VB2_MEMORY_DMABUF)
drivers/media/common/videobuf2/videobuf2-v4l2.c
528
if (q->memory == VB2_MEMORY_MMAP)
drivers/media/common/videobuf2/videobuf2-v4l2.c
530
else if (q->memory == VB2_MEMORY_USERPTR)
drivers/media/common/videobuf2/videobuf2-v4l2.c
532
else if (q->memory == VB2_MEMORY_DMABUF)
drivers/media/common/videobuf2/videobuf2-v4l2.c
597
if (vb->vb2_queue->memory != VB2_MEMORY_MMAP) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
674
static void vb2_set_flags_and_caps(struct vb2_queue *q, u32 memory,
drivers/media/common/videobuf2/videobuf2-v4l2.c
677
if (!q->allow_cache_hints || memory != V4L2_MEMORY_MMAP) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
709
int ret = vb2_verify_memory_type(q, req->memory, req->type);
drivers/media/common/videobuf2/videobuf2-v4l2.c
712
vb2_set_flags_and_caps(q, req->memory, &flags,
drivers/media/common/videobuf2/videobuf2-v4l2.c
715
return ret ? ret : vb2_core_reqbufs(q, req->memory,
drivers/media/common/videobuf2/videobuf2-v4l2.c
751
int ret = vb2_verify_memory_type(q, create->memory, f->type);
drivers/media/common/videobuf2/videobuf2-v4l2.c
755
vb2_set_flags_and_caps(q, create->memory, &create->flags,
drivers/media/common/videobuf2/videobuf2-v4l2.c
801
return vb2_core_create_bufs(q, create->memory,
drivers/media/platform/nvidia/tegra-vde/v4l2.c
176
if (vq->memory == VB2_MEMORY_DMABUF) {
drivers/media/platform/samsung/exynos4-is/fimc-is.c
242
buf = is->memory.vaddr + is->setfile.base;
drivers/media/platform/samsung/exynos4-is/fimc-is.c
247
pr_debug("mem vaddr: %p, setfile buf: %p\n", is->memory.vaddr, buf);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
272
mcuctl_write(is->memory.addr, is, MCUCTL_REG_BBOAR);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
318
memcpy(is->memory.vaddr, is->fw.f_w->data, is->fw.f_w->size);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
338
is->memory.vaddr = dma_alloc_coherent(dev, FIMC_IS_CPU_MEM_SIZE,
drivers/media/platform/samsung/exynos4-is/fimc-is.c
339
&is->memory.addr, GFP_KERNEL);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
340
if (is->memory.vaddr == NULL)
drivers/media/platform/samsung/exynos4-is/fimc-is.c
343
is->memory.size = FIMC_IS_CPU_MEM_SIZE;
drivers/media/platform/samsung/exynos4-is/fimc-is.c
345
dev_info(dev, "FIMC-IS CPU memory base: %pad\n", &is->memory.addr);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
347
if (((u32)is->memory.addr) & FIMC_IS_FW_ADDR_MASK) {
drivers/media/platform/samsung/exynos4-is/fimc-is.c
349
(u32)is->memory.addr);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
350
dma_free_coherent(dev, is->memory.size, is->memory.vaddr,
drivers/media/platform/samsung/exynos4-is/fimc-is.c
351
is->memory.addr);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
355
is->is_p_region = (struct is_region *)(is->memory.vaddr +
drivers/media/platform/samsung/exynos4-is/fimc-is.c
358
is->is_dma_p_region = is->memory.addr +
drivers/media/platform/samsung/exynos4-is/fimc-is.c
361
is->is_shared_region = (struct is_share_region *)(is->memory.vaddr +
drivers/media/platform/samsung/exynos4-is/fimc-is.c
370
if (is->memory.vaddr == NULL)
drivers/media/platform/samsung/exynos4-is/fimc-is.c
373
dma_free_coherent(dev, is->memory.size, is->memory.vaddr,
drivers/media/platform/samsung/exynos4-is/fimc-is.c
374
is->memory.addr);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
403
memcpy(is->memory.vaddr, fw->data, fw->size);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
407
buf = (void *)(is->memory.vaddr + fw->size - FIMC_IS_FW_DESC_LEN);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
411
buf = (void *)(is->memory.vaddr + fw->size - FIMC_IS_FW_VER_LEN);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
419
dev_dbg(dev, "FW size: %zu, DMA addr: %pad\n", fw->size, &is->memory.addr);
drivers/media/platform/samsung/exynos4-is/fimc-is.c
702
&is->memory.addr + FIMC_IS_SHARED_REGION_OFFSET,
drivers/media/platform/samsung/exynos4-is/fimc-is.c
740
const u8 *buf = is->memory.vaddr + FIMC_IS_DEBUG_REGION_OFFSET;
drivers/media/platform/samsung/exynos4-is/fimc-is.c
742
if (is->memory.vaddr == NULL) {
drivers/media/platform/samsung/exynos4-is/fimc-is.h
266
struct fimc_is_memory memory;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_dec.c
603
if (reqbufs->memory != V4L2_MEMORY_MMAP) {
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_dec.c
626
if (buf->memory != V4L2_MEMORY_MMAP) {
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1538
if ((reqbufs->memory != V4L2_MEMORY_MMAP) &&
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1539
(reqbufs->memory != V4L2_MEMORY_USERPTR) &&
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1540
(reqbufs->memory != V4L2_MEMORY_DMABUF))
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1608
if ((buf->memory != V4L2_MEMORY_MMAP) &&
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1609
(buf->memory != V4L2_MEMORY_USERPTR) &&
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1610
(buf->memory != V4L2_MEMORY_DMABUF))
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
2572
if (q->memory != V4L2_MEMORY_DMABUF) {
drivers/media/platform/ti/omap/omap_vout_vrfb.c
42
if (vout->vq.memory == V4L2_MEMORY_MMAP && i >= startindex)
drivers/media/usb/gspca/gspca.h
208
char memory; /* memory type (V4L2_MEMORY_xxx) */
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
1056
err = get_v4l2_plane32(p64, p32, b64->memory);
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
1126
err = put_v4l2_plane32(p64, p32, b64->memory);
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
127
__u32 memory; /* enum v4l2_memory */
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
297
__u32 memory; /* enum v4l2_memory */
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
321
__u32 memory; /* enum v4l2_memory */
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
336
enum v4l2_memory memory)
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
344
switch (memory) {
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
370
enum v4l2_memory memory)
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
381
switch (memory) {
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
419
.memory = vb32.memory,
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
425
switch (vb->memory) {
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
464
.memory = vb32.memory,
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
469
switch (vb->memory) {
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
506
.memory = vb->memory,
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
512
switch (vb->memory) {
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
551
.memory = vb->memory,
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
556
switch (vb->memory) {
drivers/media/v4l2-core/v4l2-ioctl.c
3330
.memory = vb32.memory,
drivers/media/v4l2-core/v4l2-ioctl.c
3403
vb32.memory = vb->memory;
drivers/media/v4l2-core/v4l2-ioctl.c
435
prt_names(p->memory, v4l2_memory_names));
drivers/media/v4l2-core/v4l2-ioctl.c
453
p->sequence, prt_names(p->memory, v4l2_memory_names));
drivers/media/v4l2-core/v4l2-ioctl.c
488
p->index, p->count, prt_names(p->memory, v4l2_memory_names),
drivers/media/v4l2-core/v4l2-mem2mem.c
593
if (buf->memory == V4L2_MEMORY_MMAP && V4L2_TYPE_IS_CAPTURE(vq->type)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3467
static const char * const memory[] = { "EDC0:", "EDC1:", "MC:",
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3479
mem_region_show(seq, memory[meminfo.avail[i].idx],
drivers/net/ethernet/sfc/ef10.c
3454
tests->memory = efx_ef10_run_bist(efx, MC_CMD_MC_MEM_BIST) ? -1 : 1;
drivers/net/ethernet/sfc/ethtool_common.c
370
efx_fill_test(n++, strings, data, &tests->memory,
drivers/net/ethernet/sfc/falcon/ethtool.c
323
ef4_fill_test(n++, strings, data, &tests->memory,
drivers/net/ethernet/sfc/falcon/selftest.c
751
if ((tests->memory < 0 || tests->registers < 0) && !rc_test)
drivers/net/ethernet/sfc/falcon/selftest.h
38
int memory;
drivers/net/ethernet/sfc/selftest.c
744
if ((tests->memory < 0 || tests->registers < 0) && !rc_test)
drivers/net/ethernet/sfc/selftest.h
38
int memory;
drivers/net/ethernet/sfc/siena/ethtool_common.c
322
efx_fill_test(n++, strings, data, &tests->memory,
drivers/net/ethernet/sfc/siena/selftest.c
749
if ((tests->memory < 0 || tests->registers < 0) && !rc_test)
drivers/net/ethernet/sfc/siena/selftest.h
38
int memory;
drivers/net/ethernet/smsc/smc9194.c
1007
if_string, memory);
drivers/net/ethernet/smsc/smc9194.c
847
int i, memory, retval;
drivers/net/ethernet/smsc/smc9194.c
938
memory = ( memory_cfg_register >> 9 ) & 0x7; /* multiplier */
drivers/net/ethernet/smsc/smc9194.c
939
memory *= 256 * ( memory_info_register & 0xFF );
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3924
static DEVICE_ATTR_RW(memory);
drivers/of/fdt.c
52
int memory;
drivers/of/fdt.c
57
memory = fdt_path_offset(initial_boot_params, "/memory");
drivers/of/fdt.c
58
if (memory > 0) {
drivers/of/fdt.c
59
val = fdt_getprop(initial_boot_params, memory, "reg", &len);
drivers/of/fdt.c
63
fdt_setprop(initial_boot_params, memory, "reg", val,
drivers/of/of_reserved_mem.c
141
if (memblock_overlaps_region(&memblock.memory, base, size) &&
drivers/perf/fujitsu_uncore_pmu.c
334
MAC_EVENT_ATTR(memory-read-count, 0x40),
drivers/perf/fujitsu_uncore_pmu.c
335
MAC_EVENT_ATTR(memory-write-count, 0x50),
drivers/perf/fujitsu_uncore_pmu.c
336
MAC_EVENT_ATTR(memory-pwrite-count, 0x60),
drivers/perf/fujitsu_uncore_pmu.c
338
MAC_EVENT_ATTR(ea-memory, 0x90),
drivers/perf/fujitsu_uncore_pmu.c
339
MAC_EVENT_ATTR(ea-memory-mac-write, 0x92),
drivers/scsi/53c700.c
298
__u8 *memory;
drivers/scsi/53c700.c
307
memory = dma_alloc_coherent(dev, TOTAL_MEM_SIZE, &pScript, GFP_KERNEL);
drivers/scsi/53c700.c
308
if (!memory) {
drivers/scsi/53c700.c
310
memory = dma_alloc_noncoherent(dev, TOTAL_MEM_SIZE, &pScript,
drivers/scsi/53c700.c
313
if (!memory) {
drivers/scsi/53c700.c
318
script = (__u32 *)memory;
drivers/scsi/53c700.c
319
hostdata->msgin = memory + MSGIN_OFFSET;
drivers/scsi/53c700.c
320
hostdata->msgout = memory + MSGOUT_OFFSET;
drivers/scsi/53c700.c
321
hostdata->status = memory + STATUS_OFFSET;
drivers/scsi/53c700.c
322
hostdata->slots = (struct NCR_700_command_slot *)(memory + SLOTS_OFFSET);
drivers/soc/fsl/qbman/qman_ccsr.c
344
static int qm_set_memory(enum qm_memory memory, u64 ba, u32 size)
drivers/soc/fsl/qbman/qman_ccsr.c
347
u32 offset = (memory == qm_memory_fqd) ? REG_FQD_BARE : REG_PFDR_BARE;
drivers/soc/qcom/ocmem.c
146
if (addr < ocmem->memory->start || addr >= ocmem->memory->end)
drivers/soc/qcom/ocmem.c
149
return addr - ocmem->memory->start;
drivers/soc/qcom/ocmem.c
326
ocmem->memory = platform_get_resource_byname(pdev, IORESOURCE_MEM,
drivers/soc/qcom/ocmem.c
328
if (!ocmem->memory) {
drivers/soc/qcom/ocmem.c
57
struct resource *memory;
drivers/spi/spi-wpcm-fiu.c
409
if (!fiu->memory || offs >= fiu->memory_size)
drivers/spi/spi-wpcm-fiu.c
413
memcpy_fromio(buf, fiu->memory + offs, len);
drivers/spi/spi-wpcm-fiu.c
461
fiu->memory = devm_ioremap_resource(dev, res);
drivers/spi/spi-wpcm-fiu.c
462
if (IS_ERR(fiu->memory))
drivers/spi/spi-wpcm-fiu.c
463
return dev_err_probe(dev, PTR_ERR(fiu->memory),
drivers/spi/spi-wpcm-fiu.c
61
void __iomem *memory;
drivers/staging/media/atomisp/pci/sh_css.c
5012
bool memory, continuous;
drivers/staging/media/atomisp/pci/sh_css.c
5033
memory = pipe->stream->config.mode == IA_CSS_INPUT_MODE_MEMORY;
drivers/staging/media/atomisp/pci/sh_css.c
5246
need_isp_copy_binary = !online && !continuous && !memory;
drivers/staging/octeon/ethernet-mem.c
100
fpa = (char *)(((unsigned long)memory + 256) & ~0x7fUL);
drivers/staging/octeon/ethernet-mem.c
101
*((char **)fpa - 1) = memory;
drivers/staging/octeon/ethernet-mem.c
116
char *memory;
drivers/staging/octeon/ethernet-mem.c
124
memory = *((char **)fpa - 1);
drivers/staging/octeon/ethernet-mem.c
125
kfree(memory);
drivers/staging/octeon/ethernet-mem.c
49
char *memory;
drivers/staging/octeon/ethernet-mem.c
52
memory = cvmx_fpa_alloc(pool);
drivers/staging/octeon/ethernet-mem.c
53
if (memory) {
drivers/staging/octeon/ethernet-mem.c
55
*(struct sk_buff **)(memory - sizeof(void *));
drivers/staging/octeon/ethernet-mem.c
59
} while (memory);
drivers/staging/octeon/ethernet-mem.c
79
char *memory;
drivers/staging/octeon/ethernet-mem.c
94
memory = kmalloc(size + 256, GFP_ATOMIC);
drivers/staging/octeon/ethernet-mem.c
95
if (unlikely(!memory)) {
drivers/usb/gadget/udc/pxa25x_udc.c
1982
static struct pxa25x_udc memory = {
drivers/usb/gadget/udc/pxa25x_udc.c
1985
.ep0 = &memory.ep[0].ep,
drivers/usb/gadget/udc/pxa25x_udc.c
2002
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2016
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2031
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2048
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2063
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2078
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2095
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2110
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2126
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2141
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2156
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2173
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2188
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2204
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2219
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2234
.dev = &memory,
drivers/usb/gadget/udc/pxa25x_udc.c
2283
struct pxa25x_udc *dev = &memory;
drivers/usb/gadget/udc/pxa27x_udc.c
2292
static struct pxa_udc memory = {
drivers/usb/gadget/udc/pxa27x_udc.c
2295
.ep0 = &memory.udc_usb_ep[0].usb_ep,
drivers/usb/gadget/udc/pxa27x_udc.c
2356
struct pxa_udc *udc = &memory;
drivers/usb/gadget/udc/pxa27x_udc.h
239
.dev = &memory \
drivers/usb/gadget/udc/pxa27x_udc.h
265
.dev = &memory, \
drivers/video/fbdev/matrox/matroxfb_DAC1064.c
770
if (minfo->values.memory.ddr && (!minfo->values.memory.emrswen || !minfo->values.memory.dll)) {
drivers/video/fbdev/matrox/matroxfb_base.h
491
} memory;
drivers/video/fbdev/matrox/matroxfb_misc.c
670
minfo->values.memory.ddr = (bd->pins[114] & 0x60) == 0x20;
drivers/video/fbdev/matrox/matroxfb_misc.c
671
minfo->values.memory.dll = (bd->pins[115] & 0x02) != 0;
drivers/video/fbdev/matrox/matroxfb_misc.c
672
minfo->values.memory.emrswen = (bd->pins[115] & 0x01) != 0;
drivers/video/fbdev/matrox/matroxfb_misc.c
673
minfo->values.reg.maccess = minfo->values.memory.emrswen ? 0x00004000 : 0x00000000;
drivers/video/fbdev/matrox/matroxfb_misc.c
707
minfo->values.memory.ddr = 1;
drivers/video/fbdev/matrox/matroxfb_misc.c
708
minfo->values.memory.dll = 1;
drivers/video/fbdev/matrox/matroxfb_misc.c
709
minfo->values.memory.emrswen = 1;
drivers/video/fbdev/simplefb.c
133
struct resource memory;
drivers/video/fbdev/simplefb.c
179
ret = of_reserved_mem_region_to_resource(np, 0, ¶ms->memory);
drivers/video/fbdev/simplefb.c
184
memset(¶ms->memory, 0, sizeof(params->memory));
drivers/video/fbdev/simplefb.c
214
memset(¶ms->memory, 0, sizeof(params->memory));
drivers/video/fbdev/simplefb.c
552
if (params.memory.start == 0 && params.memory.end == 0) {
drivers/video/fbdev/simplefb.c
559
res = ¶ms.memory;
drivers/video/fbdev/sis/sis_main.c
1722
ivideo->sisfb_infoblock.memory = ivideo->video_size / 1024;
drivers/w1/slaves/w1_ds2433.c
102
w1_read_block(sl->master, &data->memory[off], W1_PAGE_SIZE);
drivers/w1/slaves/w1_ds2433.c
105
if (crc16(CRC16_INIT, &data->memory[off], W1_PAGE_SIZE) == CRC16_VALID)
drivers/w1/slaves/w1_ds2433.c
140
memcpy(buf, &data->memory[off], count);
drivers/w1/slaves/w1_ds2433.c
342
data->memory = kzalloc(data->cfg->eeprom_size, GFP_KERNEL);
drivers/w1/slaves/w1_ds2433.c
343
if (!data->memory) {
drivers/w1/slaves/w1_ds2433.c
359
kfree(data->memory);
drivers/w1/slaves/w1_ds2433.c
62
u8 *memory;
drivers/w1/slaves/w1_ds28e04.c
137
memcpy(buf, &data->memory[off], count);
drivers/w1/slaves/w1_ds28e04.c
52
u8 memory[W1_EEPROM_SIZE];
drivers/w1/slaves/w1_ds28e04.c
89
w1_read_block(sl->master, &data->memory[off], W1_PAGE_SIZE);
drivers/w1/slaves/w1_ds28e04.c
92
if (crc16(CRC16_INIT, &data->memory[off], W1_PAGE_SIZE) == CRC16_VALID)
fs/btrfs/tests/extent-io-tests.c
669
static void dump_eb_and_memory_contents(struct extent_buffer *eb, void *memory,
fs/btrfs/tests/extent-io-tests.c
676
if (memcmp(addr, memory + i, 1) != 0) {
fs/btrfs/tests/extent-io-tests.c
679
i, *(u8 *)addr, *(u8 *)(memory + i));
fs/btrfs/tests/extent-io-tests.c
685
static int verify_eb_and_memory(struct extent_buffer *eb, void *memory,
fs/btrfs/tests/extent-io-tests.c
691
if (memcmp(memory + (i << PAGE_SHIFT), eb_addr, PAGE_SIZE) != 0) {
fs/btrfs/tests/extent-io-tests.c
692
dump_eb_and_memory_contents(eb, memory, test_name);
fs/btrfs/tests/extent-io-tests.c
703
static void init_eb_and_memory(struct extent_buffer *eb, void *memory)
fs/btrfs/tests/extent-io-tests.c
705
get_random_bytes(memory, eb->len);
fs/btrfs/tests/extent-io-tests.c
706
write_extent_buffer(eb, memory, 0, eb->len);
fs/btrfs/tests/extent-io-tests.c
713
void *memory = NULL;
fs/btrfs/tests/extent-io-tests.c
724
memory = kvzalloc(nodesize, GFP_KERNEL);
fs/btrfs/tests/extent-io-tests.c
725
if (!memory) {
fs/btrfs/tests/extent-io-tests.c
738
init_eb_and_memory(eb, memory);
fs/btrfs/tests/extent-io-tests.c
739
ret = verify_eb_and_memory(eb, memory, "full eb write");
fs/btrfs/tests/extent-io-tests.c
743
memcpy(memory, memory + 16, 16);
fs/btrfs/tests/extent-io-tests.c
745
ret = verify_eb_and_memory(eb, memory, "same page non-overlapping memcpy 1");
fs/btrfs/tests/extent-io-tests.c
749
memcpy(memory, memory + 2048, 16);
fs/btrfs/tests/extent-io-tests.c
751
ret = verify_eb_and_memory(eb, memory, "same page non-overlapping memcpy 2");
fs/btrfs/tests/extent-io-tests.c
754
memcpy(memory, memory + 2048, 2048);
fs/btrfs/tests/extent-io-tests.c
756
ret = verify_eb_and_memory(eb, memory, "same page non-overlapping memcpy 3");
fs/btrfs/tests/extent-io-tests.c
760
memmove(memory + 512, memory + 256, 512);
fs/btrfs/tests/extent-io-tests.c
762
ret = verify_eb_and_memory(eb, memory, "same page overlapping memcpy 1");
fs/btrfs/tests/extent-io-tests.c
766
memmove(memory + 2048, memory + 512, 2048);
fs/btrfs/tests/extent-io-tests.c
768
ret = verify_eb_and_memory(eb, memory, "same page overlapping memcpy 2");
fs/btrfs/tests/extent-io-tests.c
771
memmove(memory + 512, memory + 2048, 2048);
fs/btrfs/tests/extent-io-tests.c
773
ret = verify_eb_and_memory(eb, memory, "same page overlapping memcpy 3");
fs/btrfs/tests/extent-io-tests.c
778
memcpy(memory, memory + 4096 - 128, 256);
fs/btrfs/tests/extent-io-tests.c
780
ret = verify_eb_and_memory(eb, memory, "cross page non-overlapping memcpy 1");
fs/btrfs/tests/extent-io-tests.c
784
memcpy(memory + 4096 - 128, memory + 4096 + 128, 256);
fs/btrfs/tests/extent-io-tests.c
786
ret = verify_eb_and_memory(eb, memory, "cross page non-overlapping memcpy 2");
fs/btrfs/tests/extent-io-tests.c
790
memmove(memory + 4096 - 128, memory + 4096 - 64, 256);
fs/btrfs/tests/extent-io-tests.c
792
ret = verify_eb_and_memory(eb, memory, "cross page overlapping memcpy 1");
fs/btrfs/tests/extent-io-tests.c
796
memmove(memory + 4096 - 64, memory + 4096 - 128, 256);
fs/btrfs/tests/extent-io-tests.c
798
ret = verify_eb_and_memory(eb, memory, "cross page overlapping memcpy 2");
fs/btrfs/tests/extent-io-tests.c
804
kvfree(memory);
include/acpi/acpiosxf.h
179
void acpi_os_free(void *memory);
include/acpi/actbl2.h
2366
#define ACPI_NFIT_BUILD_DEVICE_HANDLE(dimm, channel, memory, socket, node) \
include/acpi/actbl2.h
2369
((memory) << ACPI_NFIT_MEMORY_ID_OFFSET) | \
include/acpi/platform/aclinuxex.h
62
static inline void acpi_os_free(void *memory)
include/acpi/platform/aclinuxex.h
64
kfree(memory);
include/linux/cgroup_subsys.h
29
SUBSYS(memory)
include/linux/memblock.h
108
struct memblock_type memory;
include/linux/memblock.h
249
__for_each_mem_range(i, &memblock.memory, NULL, NUMA_NO_NODE, \
include/linux/memblock.h
261
__for_each_mem_range_rev(i, &memblock.memory, NULL, NUMA_NO_NODE, \
include/linux/memblock.h
341
__for_each_mem_range(i, &memblock.memory, &memblock.reserved, \
include/linux/memblock.h
358
__for_each_mem_range_rev(i, &memblock.memory, &memblock.reserved, \
include/linux/memblock.h
570
for (region = memblock.memory.regions; \
include/linux/memblock.h
571
region < (memblock.memory.regions + memblock.memory.cnt); \
include/linux/memcontrol.h
197
struct page_counter memory; /* Both v1 & v2 */
include/linux/memcontrol.h
567
*usage = page_counter_read(&memcg->memory);
include/linux/memcontrol.h
604
*min = READ_ONCE(memcg->memory.emin);
include/linux/memcontrol.h
605
*low = READ_ONCE(memcg->memory.elow);
include/linux/memcontrol.h
629
return READ_ONCE(memcg->memory.elow) >=
include/linux/memcontrol.h
630
page_counter_read(&memcg->memory);
include/linux/memcontrol.h
639
return READ_ONCE(memcg->memory.emin) >=
include/linux/memcontrol.h
640
page_counter_read(&memcg->memory);
include/linux/mlx5/device.h
691
} __packed memory;
include/media/v4l2-ioctl.h
781
__u32 memory;
include/media/videobuf2-core.h
1319
enum vb2_memory memory, unsigned int type);
include/media/videobuf2-core.h
256
unsigned int memory;
include/media/videobuf2-core.h
637
unsigned int memory;
include/media/videobuf2-core.h
688
return q->allow_cache_hints && q->memory == VB2_MEMORY_MMAP;
include/media/videobuf2-core.h
809
int vb2_core_reqbufs(struct vb2_queue *q, enum vb2_memory memory,
include/media/videobuf2-core.h
835
int vb2_core_create_bufs(struct vb2_queue *q, enum vb2_memory memory,
include/sound/gus.h
293
unsigned int memory; /* GUS's DRAM size in bytes */
include/trace/events/memory-failure.h
4
#define TRACE_INCLUDE_FILE memory-failure
include/uapi/linux/videodev2.h
1090
__u32 memory; /* enum v4l2_memory */
include/uapi/linux/videodev2.h
1193
__u32 memory;
include/uapi/linux/videodev2.h
2704
__u32 memory;
include/uapi/video/sisfb.h
84
__u32 memory; /* total video memory in KB */
kernel/liveupdate/kexec_handover.c
1436
__for_each_mem_range(i, &memblock.memory, NULL, NUMA_NO_NODE,
mm/bpf_memcontrol.c
115
return page_counter_read(&memcg->memory) * PAGE_SIZE;
mm/memblock.c
1029
return memblock_setclr_flag(&memblock.memory, base, size, 1, MEMBLOCK_HOTPLUG);
mm/memblock.c
1041
return memblock_setclr_flag(&memblock.memory, base, size, 0, MEMBLOCK_HOTPLUG);
mm/memblock.c
1058
return memblock_setclr_flag(&memblock.memory, base, size, 1, MEMBLOCK_MIRROR);
mm/memblock.c
1078
return memblock_setclr_flag(&memblock.memory, base, size, 1, MEMBLOCK_NOMAP);
mm/memblock.c
1090
return memblock_setclr_flag(&memblock.memory, base, size, 0, MEMBLOCK_NOMAP);
mm/memblock.c
1130
return memblock_setclr_flag(&memblock.memory, base, size, 1,
mm/memblock.c
1144
return memblock_setclr_flag(&memblock.memory, base, size, 0,
mm/memblock.c
130
.memory.regions = memblock_memory_init_regions,
mm/memblock.c
131
.memory.max = INIT_MEMBLOCK_MEMORY_REGIONS,
mm/memblock.c
132
.memory.name = "memory",
mm/memblock.c
1391
struct memblock_type *type = &memblock.memory;
mm/memblock.c
156
static __refdata struct memblock_type *memblock_memory = &memblock.memory;
mm/memblock.c
1786
return memblock.memory.total_size;
mm/memblock.c
1837
return memblock.memory.regions[0].base;
mm/memblock.c
1842
int idx = memblock.memory.cnt - 1;
mm/memblock.c
1844
return (memblock.memory.regions[idx].base + memblock.memory.regions[idx].size);
mm/memblock.c
1882
memblock_remove_range(&memblock.memory, max_addr,
mm/memblock.c
1901
ret = memblock_isolate_range(&memblock.memory, base, size,
mm/memblock.c
1907
for (i = memblock.memory.cnt - 1; i >= end_rgn; i--)
mm/memblock.c
1908
if (!memblock_is_nomap(&memblock.memory.regions[i]))
mm/memblock.c
1909
memblock_remove_region(&memblock.memory, i);
mm/memblock.c
1912
if (!memblock_is_nomap(&memblock.memory.regions[i]))
mm/memblock.c
1913
memblock_remove_region(&memblock.memory, i);
mm/memblock.c
1962
return memblock_search(&memblock.memory, addr) != -1;
mm/memblock.c
1967
int i = memblock_search(&memblock.memory, addr);
mm/memblock.c
1971
return !memblock_is_nomap(&memblock.memory.regions[i]);
mm/memblock.c
1977
struct memblock_type *type = &memblock.memory;
mm/memblock.c
2001
int idx = memblock_search(&memblock.memory, base);
mm/memblock.c
2006
return (memblock.memory.regions[idx].base +
mm/memblock.c
2007
memblock.memory.regions[idx].size) >= end;
mm/memblock.c
2044
memblock_remove_region(&memblock.memory,
mm/memblock.c
2045
r - memblock.memory.regions);
mm/memblock.c
2091
&memblock.memory.total_size,
mm/memblock.c
2094
memblock_dump(&memblock.memory);
mm/memblock.c
2756
&memblock.memory, &memblock_debug_fops);
mm/memblock.c
399
if (memblock.memory.regions != memblock_memory_init_regions) {
mm/memblock.c
400
addr = __pa(memblock.memory.regions);
mm/memblock.c
402
memblock.memory.max);
mm/memblock.c
404
kfree(memblock.memory.regions);
mm/memblock.c
455
if (type == &memblock.memory)
mm/memblock.c
736
return memblock_add_range(&memblock.memory, base, size, nid, flags);
mm/memblock.c
757
return memblock_add_range(&memblock.memory, base, size, MAX_NUMNODES, 0);
mm/memblock.c
893
return memblock_remove_range(&memblock.memory, base, size);
mm/memblock.c
978
__for_each_mem_range(i, &memblock.memory, NULL, NUMA_NO_NODE,
mm/memcontrol-v1.c
1468
struct page_counter *counter = memsw ? &memcg->memsw : &memcg->memory;
mm/memcontrol-v1.c
1481
limits_invariant = memsw ? max >= READ_ONCE(memcg->memory.max) :
mm/memcontrol-v1.c
1530
while (nr_retries && page_counter_read(&memcg->memory)) {
mm/memcontrol-v1.c
1580
counter = &memcg->memory;
mm/memcontrol-v1.c
1597
if (counter == &memcg->memory)
mm/memcontrol-v1.c
169
unsigned long nr_pages = page_counter_read(&memcg->memory);
mm/memcontrol-v1.c
1727
counter = &memcg->memory;
mm/memcontrol-v1.c
1892
unsigned long memory, memsw;
mm/memcontrol-v1.c
1917
memory = memsw = PAGE_COUNTER_MAX;
mm/memcontrol-v1.c
1919
memory = min(memory, READ_ONCE(mi->memory.max));
mm/memcontrol-v1.c
1923
(u64)memory * PAGE_SIZE);
mm/memcontrol-v1.c
445
val = page_counter_read(&memcg->memory);
mm/memcontrol-v1.c
651
page_counter_uncharge(&memcg->memory, nr_entries);
mm/memcontrol.c
1331
count = page_counter_read(&memcg->memory);
mm/memcontrol.c
1332
limit = READ_ONCE(memcg->memory.max);
mm/memcontrol.c
1591
memory_failcnt = memcg->memory.failcnt;
mm/memcontrol.c
1594
K((u64)page_counter_read(&memcg->memory)),
mm/memcontrol.c
1595
K((u64)READ_ONCE(memcg->memory.max)), memory_failcnt);
mm/memcontrol.c
1625
unsigned long max = READ_ONCE(memcg->memory.max);
mm/memcontrol.c
1873
page_counter_uncharge(&memcg->memory, nr_pages);
mm/memcontrol.c
2097
if (page_counter_read(&memcg->memory) <=
mm/memcontrol.c
2098
READ_ONCE(memcg->memory.high))
mm/memcontrol.c
2199
overage = calculate_overage(page_counter_read(&memcg->memory),
mm/memcontrol.c
2200
READ_ONCE(memcg->memory.high));
mm/memcontrol.c
2380
if (page_counter_try_charge(&memcg->memory, batch, &counter))
mm/memcontrol.c
2384
mem_over_limit = mem_cgroup_from_counter(counter, memory);
mm/memcontrol.c
2484
page_counter_charge(&memcg->memory, nr_pages);
mm/memcontrol.c
2506
mem_high = page_counter_read(&memcg->memory) >
mm/memcontrol.c
2507
READ_ONCE(memcg->memory.high);
mm/memcontrol.c
3430
unsigned long ceiling = min(READ_ONCE(memcg->memory.max),
mm/memcontrol.c
3431
READ_ONCE(memcg->memory.high));
mm/memcontrol.c
3432
unsigned long used = page_counter_read(&memcg->memory);
mm/memcontrol.c
3835
page_counter_set_high(&memcg->memory, PAGE_COUNTER_MAX);
mm/memcontrol.c
3845
page_counter_init(&memcg->memory, &parent->memory, memcg_on_dfl);
mm/memcontrol.c
3848
memcg->memory.track_failcnt = !memcg_on_dfl;
mm/memcontrol.c
3856
page_counter_init(&memcg->memory, NULL, true);
mm/memcontrol.c
3925
page_counter_set_min(&memcg->memory, 0);
mm/memcontrol.c
3926
page_counter_set_low(&memcg->memory, 0);
mm/memcontrol.c
3991
page_counter_set_max(&memcg->memory, PAGE_COUNTER_MAX);
mm/memcontrol.c
3997
page_counter_set_min(&memcg->memory, 0);
mm/memcontrol.c
3998
page_counter_set_low(&memcg->memory, 0);
mm/memcontrol.c
3999
page_counter_set_high(&memcg->memory, PAGE_COUNTER_MAX);
mm/memcontrol.c
4250
return (u64)page_counter_read(&memcg->memory) * PAGE_SIZE;
mm/memcontrol.c
4274
return peak_show(sf, v, &memcg->memory);
mm/memcontrol.c
4332
return peak_write(of, buf, nbytes, off, &memcg->memory,
mm/memcontrol.c
4341
READ_ONCE(mem_cgroup_from_seq(m)->memory.min));
mm/memcontrol.c
4356
page_counter_set_min(&memcg->memory, min);
mm/memcontrol.c
4364
READ_ONCE(mem_cgroup_from_seq(m)->memory.low));
mm/memcontrol.c
4379
page_counter_set_low(&memcg->memory, low);
mm/memcontrol.c
4387
READ_ONCE(mem_cgroup_from_seq(m)->memory.high));
mm/memcontrol.c
4404
page_counter_set_high(&memcg->memory, high);
mm/memcontrol.c
4410
unsigned long nr_pages = page_counter_read(&memcg->memory);
mm/memcontrol.c
4439
READ_ONCE(mem_cgroup_from_seq(m)->memory.max));
mm/memcontrol.c
4456
xchg(&memcg->memory.max, max);
mm/memcontrol.c
4462
unsigned long nr_pages = page_counter_read(&memcg->memory);
mm/memcontrol.c
4736
page_counter_calculate_protection(&root->memory, &memcg->memory, recursive_protection);
mm/memcontrol.c
4985
page_counter_charge(&memcg->memory, nr_pages);
mm/memcontrol.c
5677
K(atomic_long_read(&memcg->memory.children_min_usage)),
mm/memcontrol.c
5678
K(atomic_long_read(&memcg->memory.children_low_usage)));
mm/numa_memblks.c
248
if (!memblock_overlaps_region(&memblock.memory,
mm/numa_memblks.c
412
&memblock.memory, mb->nid);
mm/numa_memblks.c
455
WARN_ON(memblock_set_node(0, max_addr, &memblock.memory, NUMA_NO_NODE));
samples/vfio-mdev/mbochs.c
1314
static DEVICE_ATTR_RO(memory);
sound/isa/gus/gus_main.c
233
gus->gf1.memory = idx << 18;
sound/isa/gus/gus_main.c
235
gus->gf1.memory = 256 * 1024;
sound/isa/gus/gus_main.c
237
for (l = 0, local = gus->gf1.memory; l < 4; l++, local -= 256 * 1024) {
sound/isa/gus/gus_main.c
243
gus->gf1.mem_alloc.banks_8[0].size = gus->gf1.memory;
sound/isa/gus/gus_mem.c
217
if (!gus->gf1.memory)
sound/isa/gus/gus_pcm.c
115
begin = pcmp->memory + voice * (pcmp->dma_size / runtime->channels);
sound/isa/gus/gus_pcm.c
201
end = pcmp->memory + (((pcmp->bpos + 1) * pcmp->block_size) / runtime->channels);
sound/isa/gus/gus_pcm.c
234
snd_gf1_pcm_block_change(pcmp->substream, end, pcmp->memory + (end / 2), pcmp->block_size / 2);
sound/isa/gus/gus_pcm.c
235
snd_gf1_pcm_block_change(pcmp->substream, end + (pcmp->block_size / 2), pcmp->memory + (pcmp->dma_size / 2) + (end / 2), pcmp->block_size / 2);
sound/isa/gus/gus_pcm.c
237
snd_gf1_pcm_block_change(pcmp->substream, end, pcmp->memory + end, pcmp->block_size);
sound/isa/gus/gus_pcm.c
341
pcmp->memory + bpos, len);
sound/isa/gus/gus_pcm.c
346
pcmp->memory + bpos, len, w16, invert);
sound/isa/gus/gus_pcm.c
37
unsigned int memory;
sound/isa/gus/gus_pcm.c
392
if (pcmp->memory > 0) {
sound/isa/gus/gus_pcm.c
393
snd_gf1_mem_free(&gus->gf1.mem_alloc, pcmp->memory);
sound/isa/gus/gus_pcm.c
394
pcmp->memory = 0;
sound/isa/gus/gus_pcm.c
403
pcmp->memory = block->ptr;
sound/isa/gus/gus_pcm.c
445
if (pcmp->memory > 0) {
sound/isa/gus/gus_pcm.c
446
snd_gf1_mem_free(&pcmp->gus->gf1.mem_alloc, pcmp->memory);
sound/isa/gus/gus_pcm.c
447
pcmp->memory = 0;
sound/isa/gus/gus_pcm.c
503
pos = (snd_gf1_read_addr(gus, SNDRV_GF1_VA_CURRENT, voice_ctrl & 4) >> 4) - pcmp->memory;
sound/isa/gus/gus_reset.c
148
snd_gf1_write8(gus, SNDRV_GF1_VB_MODE, gus->gf1.memory ? 0x02 : 0x82); /* Deactivate voice */
sound/isa/gus/gus_reset.c
295
gus->gf1.default_voice_address = gus->gf1.memory > 0 ? 0 : 512 - 8;
sound/isa/gus/gus_reset.c
297
if (gus->gf1.enh_mode && gus->gf1.memory) {
sound/isa/gus/gus_reset.c
304
if (gus->gf1.memory > 0)
sound/isa/gus/interwave.c
422
gus->gf1.memory = pages;
sound/pci/cs46xx/cs46xx_lib.c
495
} memory[BA1_MEMORY_COUNT];
sound/pci/cs46xx/cs46xx_lib.c
525
size += chip->ba1->memory[i].size;
sound/pci/cs46xx/cs46xx_lib.c
543
ba1->memory[idx].offset,
sound/pci/cs46xx/cs46xx_lib.c
544
ba1->memory[idx].size);
sound/pci/cs46xx/cs46xx_lib.c
547
offset += ba1->memory[idx].size >> 2;
sound/pci/es1968.c
1116
init_capture_apu(chip, es, 0, es->memory->buf.addr, size,
sound/pci/es1968.c
1126
es->memory->buf.addr + size*2, size,
sound/pci/es1968.c
1414
if (chan->memory) {
sound/pci/es1968.c
1415
if (chan->memory->buf.bytes >= size) {
sound/pci/es1968.c
1419
snd_es1968_free_memory(chip, chan->memory);
sound/pci/es1968.c
1421
chan->memory = snd_es1968_new_memory(chip, size);
sound/pci/es1968.c
1422
if (chan->memory == NULL) {
sound/pci/es1968.c
1427
snd_pcm_set_runtime_buffer(substream, &chan->memory->buf);
sound/pci/es1968.c
1441
if (chan->memory) {
sound/pci/es1968.c
1442
snd_es1968_free_memory(chip, chan->memory);
sound/pci/es1968.c
1443
chan->memory = NULL;
sound/pci/es1968.c
1641
struct esm_memory *memory;
sound/pci/es1968.c
1654
memory = snd_es1968_new_memory(chip, CLOCK_MEASURE_BUFSIZE);
sound/pci/es1968.c
1655
if (!memory) {
sound/pci/es1968.c
1663
memset(memory->buf.area, 0, CLOCK_MEASURE_BUFSIZE);
sound/pci/es1968.c
1665
wave_set_register(chip, apu << 3, (memory->buf.addr - 0x10) & 0xfff8);
sound/pci/es1968.c
1667
pa = (unsigned int)((memory->buf.addr - chip->dma.addr) >> 1);
sound/pci/es1968.c
1725
snd_es1968_free_memory(chip, memory);
sound/pci/es1968.c
454
struct esm_memory *memory;
sound/pci/es1968.c
962
snd_es1968_program_wavecache(chip, es, channel, es->memory->buf.addr, 0);
sound/pci/es1968.c
965
pa = es->memory->buf.addr;
sound/soc/sof/imx/imx-common.c
288
for (i = 0; chip_info->memory[i].name; i++) {
sound/soc/sof/imx/imx-common.c
289
blk_type = imx_region_name_to_blk_type(chip_info->memory[i].name);
sound/soc/sof/imx/imx-common.c
293
chip_info->memory[i].name);
sound/soc/sof/imx/imx-common.c
295
if (!chip_info->memory[i].reserved) {
sound/soc/sof/imx/imx-common.c
297
chip_info->memory[i].name);
sound/soc/sof/imx/imx-common.c
301
chip_info->memory[i].name);
sound/soc/sof/imx/imx-common.c
305
chip_info->memory[i].name,
sound/soc/sof/imx/imx-common.c
310
chip_info->memory[i].name);
sound/soc/sof/imx/imx-common.c
319
chip_info->memory[i].name);
sound/soc/sof/imx/imx-common.h
102
struct imx_memory_info *memory;
sound/soc/sof/imx/imx8.c
341
.memory = imx8_memory_regions,
sound/soc/sof/imx/imx8.c
353
.memory = imx8_memory_regions,
sound/soc/sof/imx/imx8.c
365
.memory = imx8m_memory_regions,
sound/soc/sof/imx/imx8.c
378
.memory = imx8ulp_memory_regions,
sound/soc/sof/imx/imx9.c
76
.memory = imx95_memory_regions,
tools/perf/ui/browsers/hists.c
1289
__HPP_COLOR_MEM_STAT_FN(memory, MEMORY)
tools/perf/ui/hist.c
657
HPP_MEM_STAT_FNS(memory, MEMORY)
tools/perf/ui/hist.c
765
HPP__MEM_STAT_PRINT_FNS("Memory", memory, MEMORY),
tools/testing/memblock/tests/alloc_exact_nid_api.c
144
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
201
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
202
struct memblock_region *node2 = &memblock.memory.regions[6];
tools/testing/memblock/tests/alloc_exact_nid_api.c
259
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
260
struct memblock_region *min_node = &memblock.memory.regions[2];
tools/testing/memblock/tests/alloc_exact_nid_api.c
261
struct memblock_region *max_node = &memblock.memory.regions[5];
tools/testing/memblock/tests/alloc_exact_nid_api.c
300
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
31
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
352
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
415
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
472
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
473
struct memblock_region *node2 = &memblock.memory.regions[6];
tools/testing/memblock/tests/alloc_exact_nid_api.c
530
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
531
struct memblock_region *min_node = &memblock.memory.regions[2];
tools/testing/memblock/tests/alloc_exact_nid_api.c
532
struct memblock_region *max_node = &memblock.memory.regions[5];
tools/testing/memblock/tests/alloc_exact_nid_api.c
580
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
621
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
664
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
712
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
754
struct memblock_region *min_node = &memblock.memory.regions[2];
tools/testing/memblock/tests/alloc_exact_nid_api.c
755
struct memblock_region *max_node = &memblock.memory.regions[5];
tools/testing/memblock/tests/alloc_exact_nid_api.c
83
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
842
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_exact_nid_api.c
843
struct memblock_region *next_node = &memblock.memory.regions[nid_next];
tools/testing/memblock/tests/alloc_exact_nid_api.c
915
struct memblock_region *next_node = &memblock.memory.regions[7];
tools/testing/memblock/tests/alloc_nid_api.c
1247
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1299
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1300
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
1352
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1353
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
1404
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1463
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1464
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
1525
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1582
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
1638
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1639
struct memblock_region *node2 = &memblock.memory.regions[6];
tools/testing/memblock/tests/alloc_nid_api.c
1695
struct memblock_region *min_node = &memblock.memory.regions[2];
tools/testing/memblock/tests/alloc_nid_api.c
1696
struct memblock_region *max_node = &memblock.memory.regions[5];
tools/testing/memblock/tests/alloc_nid_api.c
1751
struct memblock_region *min_node = &memblock.memory.regions[2];
tools/testing/memblock/tests/alloc_nid_api.c
1752
struct memblock_region *max_node = &memblock.memory.regions[5];
tools/testing/memblock/tests/alloc_nid_api.c
1791
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1843
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1844
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
1897
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
1898
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
1949
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
2010
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
2011
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
2072
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
2129
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
2130
struct memblock_region *exp_node = &memblock.memory.regions[nid_exp];
tools/testing/memblock/tests/alloc_nid_api.c
2186
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
2187
struct memblock_region *node2 = &memblock.memory.regions[6];
tools/testing/memblock/tests/alloc_nid_api.c
2243
struct memblock_region *min_node = &memblock.memory.regions[2];
tools/testing/memblock/tests/alloc_nid_api.c
2244
struct memblock_region *max_node = &memblock.memory.regions[5];
tools/testing/memblock/tests/alloc_nid_api.c
2299
struct memblock_region *min_node = &memblock.memory.regions[2];
tools/testing/memblock/tests/alloc_nid_api.c
2300
struct memblock_region *max_node = &memblock.memory.regions[5];
tools/testing/memblock/tests/alloc_nid_api.c
2393
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/alloc_nid_api.c
2394
struct memblock_region *next_node = &memblock.memory.regions[nid_next];
tools/testing/memblock/tests/alloc_nid_api.c
2465
struct memblock_region *next_node = &memblock.memory.regions[7];
tools/testing/memblock/tests/alloc_nid_api.c
2507
struct memblock_region *req_node = &memblock.memory.regions[nid_req];
tools/testing/memblock/tests/basic_api.c
121
rgn1 = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
122
rgn2 = &memblock.memory.regions[1];
tools/testing/memblock/tests/basic_api.c
1284
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
1305
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
1306
ASSERT_EQ(memblock.memory.total_size, r2.size);
tools/testing/memblock/tests/basic_api.c
1333
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
1353
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
1354
ASSERT_EQ(memblock.memory.total_size, r1.size);
tools/testing/memblock/tests/basic_api.c
1385
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
1409
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
1410
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
1439
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
145
ASSERT_EQ(memblock.memory.cnt, 2);
tools/testing/memblock/tests/basic_api.c
146
ASSERT_EQ(memblock.memory.total_size, r1.size + r2.size);
tools/testing/memblock/tests/basic_api.c
1461
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
1462
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
1493
rgn1 = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
1494
rgn2 = &memblock.memory.regions[1];
tools/testing/memblock/tests/basic_api.c
1521
ASSERT_EQ(memblock.memory.cnt, 2);
tools/testing/memblock/tests/basic_api.c
1522
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
1540
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
1556
ASSERT_EQ(memblock.memory.cnt, 0);
tools/testing/memblock/tests/basic_api.c
1557
ASSERT_EQ(memblock.memory.total_size, 0);
tools/testing/memblock/tests/basic_api.c
1584
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
1607
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
1608
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
1635
rgn1 = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
1636
rgn2 = &memblock.memory.regions[1];
tools/testing/memblock/tests/basic_api.c
1670
ASSERT_EQ(memblock.memory.cnt, 2);
tools/testing/memblock/tests/basic_api.c
1671
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
17
ASSERT_NE(memblock.memory.regions, NULL);
tools/testing/memblock/tests/basic_api.c
175
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
18
ASSERT_EQ(memblock.memory.cnt, 0);
tools/testing/memblock/tests/basic_api.c
19
ASSERT_EQ(memblock.memory.max, EXPECTED_MEMBLOCK_REGIONS);
tools/testing/memblock/tests/basic_api.c
197
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
198
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
20
ASSERT_EQ(strcmp(memblock.memory.name, "memory"), 0);
tools/testing/memblock/tests/basic_api.c
2179
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
2195
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
2225
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
2246
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
227
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
2279
rgn1 = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
2280
rgn2 = &memblock.memory.regions[1];
tools/testing/memblock/tests/basic_api.c
2307
ASSERT_EQ(memblock.memory.cnt, 2);
tools/testing/memblock/tests/basic_api.c
2340
rgn1 = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
2341
rgn2 = &memblock.memory.regions[1];
tools/testing/memblock/tests/basic_api.c
2367
ASSERT_EQ(memblock.memory.cnt, 2);
tools/testing/memblock/tests/basic_api.c
24
ASSERT_EQ(memblock.memory.max, EXPECTED_MEMBLOCK_REGIONS);
tools/testing/memblock/tests/basic_api.c
2403
ASSERT_FALSE(memblock_overlaps_region(&memblock.memory, SZ_1M, SZ_1M));
tools/testing/memblock/tests/basic_api.c
2404
ASSERT_FALSE(memblock_overlaps_region(&memblock.memory, SZ_2G, SZ_1M));
tools/testing/memblock/tests/basic_api.c
2407
ASSERT_FALSE(memblock_overlaps_region(&memblock.memory, SZ_1G - SZ_1M, SZ_1M));
tools/testing/memblock/tests/basic_api.c
2408
ASSERT_FALSE(memblock_overlaps_region(&memblock.memory, SZ_1G + SZ_4M, SZ_1M));
tools/testing/memblock/tests/basic_api.c
2411
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G - SZ_1M, SZ_2M));
tools/testing/memblock/tests/basic_api.c
2412
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G + SZ_2M, SZ_2M));
tools/testing/memblock/tests/basic_api.c
2415
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G, SZ_4M));
tools/testing/memblock/tests/basic_api.c
2416
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G - SZ_2M, SZ_8M));
tools/testing/memblock/tests/basic_api.c
2417
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G + SZ_1M, SZ_1M));
tools/testing/memblock/tests/basic_api.c
2455
memblock_phys_mem_size() / 2, &memblock.memory, 0);
tools/testing/memblock/tests/basic_api.c
2457
memblock_phys_mem_size() / 2, &memblock.memory, 1);
tools/testing/memblock/tests/basic_api.c
2459
ASSERT_EQ(memblock.memory.cnt, 2);
tools/testing/memblock/tests/basic_api.c
2460
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
2464
rgn = &memblock.memory.regions[1];
tools/testing/memblock/tests/basic_api.c
249
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
250
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
276
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
296
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
297
ASSERT_EQ(memblock.memory.total_size, r1.size);
tools/testing/memblock/tests/basic_api.c
322
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
323
ASSERT_EQ(memblock.memory.total_size, r.size);
tools/testing/memblock/tests/basic_api.c
347
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
374
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
375
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
401
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
418
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
419
ASSERT_EQ(memblock.memory.total_size, total_size);
tools/testing/memblock/tests/basic_api.c
45
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
463
orig_region = memblock.memory.regions;
tools/testing/memblock/tests/basic_api.c
473
ASSERT_EQ(memblock.memory.cnt, i + 2);
tools/testing/memblock/tests/basic_api.c
474
ASSERT_EQ(memblock.memory.total_size, new_memory_regions_size +
tools/testing/memblock/tests/basic_api.c
482
ASSERT_EQ(memblock.memory.max, INIT_MEMBLOCK_REGIONS * 2);
tools/testing/memblock/tests/basic_api.c
493
ASSERT_EQ(memblock.memory.regions[0].base, r.base);
tools/testing/memblock/tests/basic_api.c
494
ASSERT_EQ(memblock.memory.regions[0].size, r.size);
tools/testing/memblock/tests/basic_api.c
496
ASSERT_EQ(memblock.memory.cnt, INIT_MEMBLOCK_REGIONS + 2);
tools/testing/memblock/tests/basic_api.c
497
ASSERT_EQ(memblock.memory.total_size, INIT_MEMBLOCK_REGIONS * size +
tools/testing/memblock/tests/basic_api.c
500
ASSERT_EQ(memblock.memory.max, INIT_MEMBLOCK_REGIONS * 2);
tools/testing/memblock/tests/basic_api.c
510
memblock.memory.regions = orig_region;
tools/testing/memblock/tests/basic_api.c
511
memblock.memory.cnt = INIT_MEMBLOCK_REGIONS;
tools/testing/memblock/tests/basic_api.c
60
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
61
ASSERT_EQ(memblock.memory.total_size, r.size);
tools/testing/memblock/tests/basic_api.c
78
rgn = &memblock.memory.regions[0];
tools/testing/memblock/tests/basic_api.c
97
ASSERT_EQ(memblock.memory.cnt, 1);
tools/testing/memblock/tests/basic_api.c
98
ASSERT_EQ(memblock.memory.total_size, r.size);
tools/testing/memblock/tests/common.c
41
memset(memblock.memory.regions, 0,
tools/testing/memblock/tests/common.c
42
memblock.memory.cnt * sizeof(struct memblock_region));
tools/testing/memblock/tests/common.c
43
memblock.memory.cnt = 0;
tools/testing/memblock/tests/common.c
44
memblock.memory.max = INIT_MEMBLOCK_REGIONS;
tools/testing/memblock/tests/common.c
45
memblock.memory.total_size = 0;
tools/testing/memblock/tests/common.c
56
memblock.memory.name = "memory";